Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
11,163
13,957,694,060
IssuesEvent
2020-10-24 08:11:08
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
IT: Missing resources in the Geoportal
Geoportal Harvesting process IT - Italy
Collected from the Geoportal Workshop online survey answers: As regards the resource documented at the link http://inspire-geoportal.ec.europa.eu/resources/INSPIREc22038a7-4e03-11e8-a459-52540023a883_20181112-123130/services/1/PullResults/14701-14750/services/8/ , an error is reported concerning the resource locator (Service Temporarily Unavailable) in the INSPIRE Geoportal. FileID r_friuli:m10713-cc-i10968 http://inspire-geoportal.ec.europa.eu/proxybrowser/#q=%22r_friuli%3Am10713-cc-i10968%22 The service metadata available at the link http://geodati.gov.it/geoportalRNDTPA/rest/document?id=r_friuli%3Am10713-cc-i10968 includes a correct link to the coupled resource (http://irdat.regione.fvg.it/CSW/metadati/INSPIRE?request=GetRecordById&service=CSW&version=2.0.2&OUTPUTFORMAT=application/xml&OUTPUTSCHEMA=http://www.isotc211.org/2005/gmd&ELEMENTSETNAME=full&id=r_friuli:m10713-cc-i11021) that allow to retrieve the unique resource identifier. In the data set metadata a resource locator (http://irdat.regione.fvg.it/Distributore/download?idDset=11021&idFmt=383&type=wfs&path=PPR:v_ghiacciai) is provided that allow the download of the dataset. Although the data-service linking doesn't fully comply with the solution of the INSPIRE geoportal, the user can get the resource in any case. The same situation for the following examples: - http://geodati.gov.it/geoportalRNDTPA/rest/document?id=r_friuli:m10678-cc-i10871 - http://geodati.gov.it/geoportalRNDTPA/rest/document?id=r_friuli:m10677-cc-i10870 - http://geodati.gov.it/geoportalRNDTPA/rest/document?id=r_friuli:m10715-cc-i10970
1.0
IT: Missing resources in the Geoportal - Collected from the Geoportal Workshop online survey answers: As regards the resource documented at the link http://inspire-geoportal.ec.europa.eu/resources/INSPIREc22038a7-4e03-11e8-a459-52540023a883_20181112-123130/services/1/PullResults/14701-14750/services/8/ , an error is reported concerning the resource locator (Service Temporarily Unavailable) in the INSPIRE Geoportal. FileID r_friuli:m10713-cc-i10968 http://inspire-geoportal.ec.europa.eu/proxybrowser/#q=%22r_friuli%3Am10713-cc-i10968%22 The service metadata available at the link http://geodati.gov.it/geoportalRNDTPA/rest/document?id=r_friuli%3Am10713-cc-i10968 includes a correct link to the coupled resource (http://irdat.regione.fvg.it/CSW/metadati/INSPIRE?request=GetRecordById&service=CSW&version=2.0.2&OUTPUTFORMAT=application/xml&OUTPUTSCHEMA=http://www.isotc211.org/2005/gmd&ELEMENTSETNAME=full&id=r_friuli:m10713-cc-i11021) that allow to retrieve the unique resource identifier. In the data set metadata a resource locator (http://irdat.regione.fvg.it/Distributore/download?idDset=11021&idFmt=383&type=wfs&path=PPR:v_ghiacciai) is provided that allow the download of the dataset. Although the data-service linking doesn't fully comply with the solution of the INSPIRE geoportal, the user can get the resource in any case. The same situation for the following examples: - http://geodati.gov.it/geoportalRNDTPA/rest/document?id=r_friuli:m10678-cc-i10871 - http://geodati.gov.it/geoportalRNDTPA/rest/document?id=r_friuli:m10677-cc-i10870 - http://geodati.gov.it/geoportalRNDTPA/rest/document?id=r_friuli:m10715-cc-i10970
process
it missing resources in the geoportal collected from the geoportal workshop online survey answers as regards the resource documented at the link an error is reported concerning the resource locator service temporarily unavailable in the inspire geoportal fileid r friuli cc the service metadata available at the link includes a correct link to the coupled resource that allow to retrieve the unique resource identifier in the data set metadata a resource locator is provided that allow the download of the dataset although the data service linking doesn t fully comply with the solution of the inspire geoportal the user can get the resource in any case the same situation for the following examples
1
7,602
10,719,369,106
IssuesEvent
2019-10-26 09:17:55
tueit/it_management
https://api.github.com/repos/tueit/it_management
closed
minor changes in IT Service Report
Please check & close enhancement process
Needs field Project and Task (coming from IT Ticket) -> need to be passed on to Timesheet. Add field type status which is also played back to IT Ticket.
1.0
minor changes in IT Service Report - Needs field Project and Task (coming from IT Ticket) -> need to be passed on to Timesheet. Add field type status which is also played back to IT Ticket.
process
minor changes in it service report needs field project and task coming from it ticket need to be passed on to timesheet add field type status which is also played back to it ticket
1
151,665
12,054,695,640
IssuesEvent
2020-04-15 11:39:19
scylladb/scylla
https://api.github.com/repos/scylladb/scylla
closed
Run Alternator tests from test.py
Alternator Backport 4.0 Backport candidate test
`test.py` currently runs several different types of tests, but does not run Alternator's tests. We have a separate script (commit pending) `alternator-tests/run` which runs Scylla in a temporary directory and runs all the Alternator functional tests against it (at the moment of this writing, 342 tests), but since this existing Jenkins jobs run `test.py`, not the new script, we don't have these tests running periodically. The easiest way to do this is to run `alternator-tests/run` from `test.py`. The entire Alternator test suite currently takes just 50 seconds to complete (and more than half of that is, unfortunately, the time to boot a single-node Scylla!), so at this point it's probably acceptable to have all of these tests together as one single `test.py` test instead of trying to let `test.py` know of the 342 separate Alternator tests. Beyond adding to `test.py` code to run `alternator-tests/run`, we also need to consider this scripts prerequisites. `test.py` needs to run in an environment defined by `install-dependencies.sh`, and we may need to add to it more things (such as boto3) but I don't know how. We also need cqlsh, but it may be better to instead require a Cassandra driver in Python - and rewrite the tiny thing we need from cqlsh in Python. A Cassandra driver in Python will also be helpful for other things, probably (such as the idea to add functional tests for CQL running against Scylla instead of cqlrepl).
1.0
Run Alternator tests from test.py - `test.py` currently runs several different types of tests, but does not run Alternator's tests. We have a separate script (commit pending) `alternator-tests/run` which runs Scylla in a temporary directory and runs all the Alternator functional tests against it (at the moment of this writing, 342 tests), but since this existing Jenkins jobs run `test.py`, not the new script, we don't have these tests running periodically. The easiest way to do this is to run `alternator-tests/run` from `test.py`. The entire Alternator test suite currently takes just 50 seconds to complete (and more than half of that is, unfortunately, the time to boot a single-node Scylla!), so at this point it's probably acceptable to have all of these tests together as one single `test.py` test instead of trying to let `test.py` know of the 342 separate Alternator tests. Beyond adding to `test.py` code to run `alternator-tests/run`, we also need to consider this scripts prerequisites. `test.py` needs to run in an environment defined by `install-dependencies.sh`, and we may need to add to it more things (such as boto3) but I don't know how. We also need cqlsh, but it may be better to instead require a Cassandra driver in Python - and rewrite the tiny thing we need from cqlsh in Python. A Cassandra driver in Python will also be helpful for other things, probably (such as the idea to add functional tests for CQL running against Scylla instead of cqlrepl).
non_process
run alternator tests from test py test py currently runs several different types of tests but does not run alternator s tests we have a separate script commit pending alternator tests run which runs scylla in a temporary directory and runs all the alternator functional tests against it at the moment of this writing tests but since this existing jenkins jobs run test py not the new script we don t have these tests running periodically the easiest way to do this is to run alternator tests run from test py the entire alternator test suite currently takes just seconds to complete and more than half of that is unfortunately the time to boot a single node scylla so at this point it s probably acceptable to have all of these tests together as one single test py test instead of trying to let test py know of the separate alternator tests beyond adding to test py code to run alternator tests run we also need to consider this scripts prerequisites test py needs to run in an environment defined by install dependencies sh and we may need to add to it more things such as but i don t know how we also need cqlsh but it may be better to instead require a cassandra driver in python and rewrite the tiny thing we need from cqlsh in python a cassandra driver in python will also be helpful for other things probably such as the idea to add functional tests for cql running against scylla instead of cqlrepl
0
5,210
7,987,481,323
IssuesEvent
2018-07-19 07:57:04
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Proposal: use release tagging
process: release stage: proposal
We recently discussed as a team to add "Releases" in github. I'm against adding the binary as releases but for using tags so that it's easy for users to see the source code that makes up a specific version of Cypress. I think this would require a minor change to our build process (that could be further optimized). My only request is that we go back through and tag previous releases. There should always be a "Released X.X.x" commit for every version so we'd only have to add a git tag for those.
1.0
Proposal: use release tagging - We recently discussed as a team to add "Releases" in github. I'm against adding the binary as releases but for using tags so that it's easy for users to see the source code that makes up a specific version of Cypress. I think this would require a minor change to our build process (that could be further optimized). My only request is that we go back through and tag previous releases. There should always be a "Released X.X.x" commit for every version so we'd only have to add a git tag for those.
process
proposal use release tagging we recently discussed as a team to add releases in github i m against adding the binary as releases but for using tags so that it s easy for users to see the source code that makes up a specific version of cypress i think this would require a minor change to our build process that could be further optimized my only request is that we go back through and tag previous releases there should always be a released x x x commit for every version so we d only have to add a git tag for those
1
5,180
7,964,094,173
IssuesEvent
2018-07-13 20:03:42
CCALI/caw
https://api.github.com/repos/CCALI/caw
closed
Question Bank link in rh menu needs to move up to Quizzes section
in process ready
The Question Bank should be part of the QuizWright Toolbox, not CALI Teaching Resources.
1.0
Question Bank link in rh menu needs to move up to Quizzes section - The Question Bank should be part of the QuizWright Toolbox, not CALI Teaching Resources.
process
question bank link in rh menu needs to move up to quizzes section the question bank should be part of the quizwright toolbox not cali teaching resources
1
524,313
15,210,866,174
IssuesEvent
2021-02-17 08:13:21
wp-media/wp-rocket
https://api.github.com/repos/wp-media/wp-rocket
opened
Bump minimum support versions to WordPress 5.3
effort: [XS] priority: high type: enhancement
**Describe the solution you'd like** We would like to bump the minimum required WordPress version to 5.3. We have today 97.6% of our users using WordPress 5.3 or + **What do we need to change?** In `wp-rocket.php`, update `WP_ROCKET_WP_VERSION` to `5.3`
1.0
Bump minimum support versions to WordPress 5.3 - **Describe the solution you'd like** We would like to bump the minimum required WordPress version to 5.3. We have today 97.6% of our users using WordPress 5.3 or + **What do we need to change?** In `wp-rocket.php`, update `WP_ROCKET_WP_VERSION` to `5.3`
non_process
bump minimum support versions to wordpress describe the solution you d like we would like to bump the minimum required wordpress version to we have today of our users using wordpress or what do we need to change in wp rocket php update wp rocket wp version to
0
323,779
27,751,820,875
IssuesEvent
2023-03-15 21:25:27
david-thrower/cerebros-core-algorithm-alpha
https://api.github.com/repos/david-thrower/cerebros-core-algorithm-alpha
closed
Add a training template with an EfficientNet embedding
kind/enhancement status/ready-pending-tests triage/high-priority triage/required kind/validation kind/increase-accuracy kind/tuning-configuration kind/usability audience/technical kind/performance
Kind of issue: Feature-request-or-enhancement Expected behavior A clear description of what you expected to happen. We want to augment the performance of the image classification and other image intelligence models. Execution environment: all Suggested Labels (If you don't know, that's ok): kind/enhancement
1.0
Add a training template with an EfficientNet embedding - Kind of issue: Feature-request-or-enhancement Expected behavior A clear description of what you expected to happen. We want to augment the performance of the image classification and other image intelligence models. Execution environment: all Suggested Labels (If you don't know, that's ok): kind/enhancement
non_process
add a training template with an efficientnet embedding kind of issue feature request or enhancement expected behavior a clear description of what you expected to happen we want to augment the performance of the image classification and other image intelligence models execution environment all suggested labels if you don t know that s ok kind enhancement
0
21,279
28,442,550,469
IssuesEvent
2023-04-16 04:03:02
cse442-at-ub/project_s23-team-infinity
https://api.github.com/repos/cse442-at-ub/project_s23-team-infinity
closed
Create an account recovery system backend and integrate it with the frontend.
Processing Task Sprint 3
**Task tests** *Test 1* 1) Go to the account recovery page : https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442ad/recovery 2) Enter email and click submit 3) Check your email for the link and verification code. 4) Click on the link to navigate to the verification page. 5) Enter the code from the email and click verify. 6) You will now be redirected to the reset password page. 7) Enter your new password and confirm your new password then click reset. 8) Verify that your password has been reset by going to the login page: https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442ad/ 9) Enter your username or email and the new password then click login. 10) Check that you've been redirected to the homepage. *Test 2* 1) Go to the account recovery page : https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442ad/recovery 2) Enter email and click submit 3) Check your email for the link and verification code. 4) Click on the link to navigate to the verification page. 5) Enter the code from the email and click verify. 6) You will now be redirected to the reset password page. 7) Enter your new password and confirm your new password then click reset. 8) Verify that your password has been reset by going to the login page: https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442ad/ 9) Enter your username or email and the old password then click login. 10) Check that an error saying that you've entered the wrong information shows up and you did not get redirected to the homepage.
1.0
Create an account recovery system backend and integrate it with the frontend. - **Task tests** *Test 1* 1) Go to the account recovery page : https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442ad/recovery 2) Enter email and click submit 3) Check your email for the link and verification code. 4) Click on the link to navigate to the verification page. 5) Enter the code from the email and click verify. 6) You will now be redirected to the reset password page. 7) Enter your new password and confirm your new password then click reset. 8) Verify that your password has been reset by going to the login page: https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442ad/ 9) Enter your username or email and the new password then click login. 10) Check that you've been redirected to the homepage. *Test 2* 1) Go to the account recovery page : https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442ad/recovery 2) Enter email and click submit 3) Check your email for the link and verification code. 4) Click on the link to navigate to the verification page. 5) Enter the code from the email and click verify. 6) You will now be redirected to the reset password page. 7) Enter your new password and confirm your new password then click reset. 8) Verify that your password has been reset by going to the login page: https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442ad/ 9) Enter your username or email and the old password then click login. 10) Check that an error saying that you've entered the wrong information shows up and you did not get redirected to the homepage.
process
create an account recovery system backend and integrate it with the frontend task tests test go to the account recovery page enter email and click submit check your email for the link and verification code click on the link to navigate to the verification page enter the code from the email and click verify you will now be redirected to the reset password page enter your new password and confirm your new password then click reset verify that your password has been reset by going to the login page enter your username or email and the new password then click login check that you ve been redirected to the homepage test go to the account recovery page enter email and click submit check your email for the link and verification code click on the link to navigate to the verification page enter the code from the email and click verify you will now be redirected to the reset password page enter your new password and confirm your new password then click reset verify that your password has been reset by going to the login page enter your username or email and the old password then click login check that an error saying that you ve entered the wrong information shows up and you did not get redirected to the homepage
1
5,788
8,634,618,591
IssuesEvent
2018-11-22 17:33:22
ddediu/AdhereR
https://api.github.com/repos/ddediu/AdhereR
closed
Devise procdure for testing before release to CRAN
process
We need to find a very efficient way for distributing a source package to the team members for testing *before* the release to `CRAN`.
1.0
Devise procdure for testing before release to CRAN - We need to find a very efficient way for distributing a source package to the team members for testing *before* the release to `CRAN`.
process
devise procdure for testing before release to cran we need to find a very efficient way for distributing a source package to the team members for testing before the release to cran
1
119,698
25,561,641,147
IssuesEvent
2022-11-30 11:13:25
FerretDB/FerretDB
https://api.github.com/repos/FerretDB/FerretDB
closed
Refactor `pg`'s `MsgListDatabases` and `pgdb`
code/chore
`MsgListDatabases` calls `pg_total_relation_size` and `pg_database_size` directly. Code wrapping those functions should be incapsulated in `pgdb` package with proper transaction handling.
1.0
Refactor `pg`'s `MsgListDatabases` and `pgdb` - `MsgListDatabases` calls `pg_total_relation_size` and `pg_database_size` directly. Code wrapping those functions should be incapsulated in `pgdb` package with proper transaction handling.
non_process
refactor pg s msglistdatabases and pgdb msglistdatabases calls pg total relation size and pg database size directly code wrapping those functions should be incapsulated in pgdb package with proper transaction handling
0
20,930
27,776,652,809
IssuesEvent
2023-03-16 17:41:36
googleapis/java-bigqueryconnection
https://api.github.com/repos/googleapis/java-bigqueryconnection
closed
Dependency Dashboard
type: process api: bigqueryconnection
This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more. ## Edited/Blocked These updates have been manually edited so Renovate will no longer make changes. To discard all commits and start over, click on a checkbox. - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-shared-config-1.x -->[build(deps): update dependency com.google.cloud:google-cloud-shared-config to v1.5.4](../pull/799) - [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-26.x -->[chore(deps): update dependency com.google.cloud:libraries-bom to v26.1.4](../pull/805) - [ ] <!-- rebase-branch=renovate/google.cloud.shared-dependencies.version -->[deps: update dependency com.google.cloud:google-cloud-shared-dependencies to v3.0.6](../pull/800) - [ ] <!-- rebase-branch=renovate/com.google.protobuf-protobuf-java-util-3.x -->[deps: update dependency com.google.protobuf:protobuf-java-util to v3.21.9](../pull/797) - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-bigqueryconnection-2.x -->[chore(deps): update dependency com.google.cloud:google-cloud-bigqueryconnection to v2.7.0](../pull/791) - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-bigqueryconnection-parent-2.x -->[chore(deps): update dependency com.google.cloud:google-cloud-bigqueryconnection-parent to v2.7.0](../pull/792) - [ ] <!-- rebase-branch=renovate/com.google.api.grpc-grpc-google-cloud-bigqueryconnection-v1-2.x -->[deps: update dependency com.google.api.grpc:grpc-google-cloud-bigqueryconnection-v1 to v2.7.0](../pull/793) - [ ] <!-- rebase-branch=renovate/com.google.api.grpc-grpc-google-cloud-bigqueryconnection-v1beta1-0.x -->[deps: update dependency com.google.api.grpc:grpc-google-cloud-bigqueryconnection-v1beta1 to v0.15.0](../pull/794) - [ ] <!-- rebase-branch=renovate/com.google.api.grpc-proto-google-cloud-bigqueryconnection-v1-2.x -->[deps: update dependency com.google.api.grpc:proto-google-cloud-bigqueryconnection-v1 to v2.7.0](../pull/795) - [ ] <!-- rebase-branch=renovate/com.google.api.grpc-proto-google-cloud-bigqueryconnection-v1beta1-0.x -->[deps: update dependency com.google.api.grpc:proto-google-cloud-bigqueryconnection-v1beta1 to v0.15.0](../pull/796) ## Detected dependencies <details><summary>github-actions</summary> <blockquote> <details><summary>.github/workflows/approve-readme.yaml</summary> - `actions/github-script v6` </details> <details><summary>.github/workflows/auto-release.yaml</summary> - `actions/github-script v6` </details> <details><summary>.github/workflows/ci.yaml</summary> - `actions/checkout v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` </details> <details><summary>.github/workflows/samples.yaml</summary> - `actions/checkout v3` - `actions/setup-java v3` </details> </blockquote> </details> <details><summary>maven</summary> <blockquote> <details><summary>google-cloud-bigqueryconnection-bom/pom.xml</summary> - `com.google.cloud:google-cloud-shared-config 1.5.3` - `com.google.cloud:google-cloud-bigqueryconnection 2.5.7-SNAPSHOT` - `com.google.api.grpc:grpc-google-cloud-bigqueryconnection-v1 2.5.7-SNAPSHOT` - `com.google.api.grpc:grpc-google-cloud-bigqueryconnection-v1beta1 0.13.7-SNAPSHOT` - `com.google.api.grpc:proto-google-cloud-bigqueryconnection-v1 2.5.7-SNAPSHOT` - `com.google.api.grpc:proto-google-cloud-bigqueryconnection-v1beta1 0.13.7-SNAPSHOT` - `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13` </details> <details><summary>google-cloud-bigqueryconnection/pom.xml</summary> - `com.google.cloud:google-cloud-bigqueryconnection-parent 2.5.7-SNAPSHOT` </details> <details><summary>grpc-google-cloud-bigqueryconnection-v1/pom.xml</summary> - `com.google.cloud:google-cloud-bigqueryconnection-parent 2.5.7-SNAPSHOT` </details> <details><summary>grpc-google-cloud-bigqueryconnection-v1beta1/pom.xml</summary> - `com.google.cloud:google-cloud-bigqueryconnection-parent 2.5.7-SNAPSHOT` </details> <details><summary>pom.xml</summary> - `com.google.cloud:google-cloud-shared-config 1.5.3` - `com.google.cloud:google-cloud-bigqueryconnection 2.5.7-SNAPSHOT` - `com.google.api.grpc:proto-google-cloud-bigqueryconnection-v1 2.5.7-SNAPSHOT` - `com.google.api.grpc:grpc-google-cloud-bigqueryconnection-v1 2.5.7-SNAPSHOT` - `com.google.api.grpc:proto-google-cloud-bigqueryconnection-v1beta1 0.13.7-SNAPSHOT` - `com.google.api.grpc:grpc-google-cloud-bigqueryconnection-v1beta1 0.13.7-SNAPSHOT` - `com.google.cloud:google-cloud-shared-dependencies 3.0.4` - `junit:junit 4.13.2` - `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13` - `org.apache.maven.plugins:maven-project-info-reports-plugin 3.4.1` - `org.apache.maven.plugins:maven-javadoc-plugin 3.4.1` </details> <details><summary>proto-google-cloud-bigqueryconnection-v1/pom.xml</summary> - `com.google.cloud:google-cloud-bigqueryconnection-parent 2.5.7-SNAPSHOT` </details> <details><summary>proto-google-cloud-bigqueryconnection-v1beta1/pom.xml</summary> - `com.google.cloud:google-cloud-bigqueryconnection-parent 2.5.7-SNAPSHOT` </details> <details><summary>samples/install-without-bom/pom.xml</summary> - `com.google.cloud.samples:shared-configuration 1.2.0` - `com.google.cloud:google-cloud-bigqueryconnection 2.5.6` - `com.google.protobuf:protobuf-java-util 3.21.7` - `junit:junit 4.13.2` - `com.google.truth:truth 1.1.3` - `org.codehaus.mojo:build-helper-maven-plugin 3.3.0` </details> <details><summary>samples/pom.xml</summary> - `com.google.cloud.samples:shared-configuration 1.2.0` - `org.apache.maven.plugins:maven-deploy-plugin 3.0.0` - `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13` </details> <details><summary>samples/snapshot/pom.xml</summary> - `com.google.cloud.samples:shared-configuration 1.2.0` - `com.google.cloud:google-cloud-bigqueryconnection 2.5.7-SNAPSHOT` - `com.google.protobuf:protobuf-java-util 3.21.7` - `junit:junit 4.13.2` - `com.google.truth:truth 1.1.3` - `org.codehaus.mojo:build-helper-maven-plugin 3.3.0` </details> <details><summary>samples/snippets/pom.xml</summary> - `com.google.cloud.samples:shared-configuration 1.2.0` - `com.google.cloud:libraries-bom 26.1.3` - `com.google.cloud:google-cloud-bigqueryconnection 2.5.6` - `com.google.protobuf:protobuf-java-util 3.21.7` - `junit:junit 4.13.2` - `com.google.truth:truth 1.1.3` </details> </blockquote> </details> --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more. ## Edited/Blocked These updates have been manually edited so Renovate will no longer make changes. To discard all commits and start over, click on a checkbox. - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-shared-config-1.x -->[build(deps): update dependency com.google.cloud:google-cloud-shared-config to v1.5.4](../pull/799) - [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-26.x -->[chore(deps): update dependency com.google.cloud:libraries-bom to v26.1.4](../pull/805) - [ ] <!-- rebase-branch=renovate/google.cloud.shared-dependencies.version -->[deps: update dependency com.google.cloud:google-cloud-shared-dependencies to v3.0.6](../pull/800) - [ ] <!-- rebase-branch=renovate/com.google.protobuf-protobuf-java-util-3.x -->[deps: update dependency com.google.protobuf:protobuf-java-util to v3.21.9](../pull/797) - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-bigqueryconnection-2.x -->[chore(deps): update dependency com.google.cloud:google-cloud-bigqueryconnection to v2.7.0](../pull/791) - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-bigqueryconnection-parent-2.x -->[chore(deps): update dependency com.google.cloud:google-cloud-bigqueryconnection-parent to v2.7.0](../pull/792) - [ ] <!-- rebase-branch=renovate/com.google.api.grpc-grpc-google-cloud-bigqueryconnection-v1-2.x -->[deps: update dependency com.google.api.grpc:grpc-google-cloud-bigqueryconnection-v1 to v2.7.0](../pull/793) - [ ] <!-- rebase-branch=renovate/com.google.api.grpc-grpc-google-cloud-bigqueryconnection-v1beta1-0.x -->[deps: update dependency com.google.api.grpc:grpc-google-cloud-bigqueryconnection-v1beta1 to v0.15.0](../pull/794) - [ ] <!-- rebase-branch=renovate/com.google.api.grpc-proto-google-cloud-bigqueryconnection-v1-2.x -->[deps: update dependency com.google.api.grpc:proto-google-cloud-bigqueryconnection-v1 to v2.7.0](../pull/795) - [ ] <!-- rebase-branch=renovate/com.google.api.grpc-proto-google-cloud-bigqueryconnection-v1beta1-0.x -->[deps: update dependency com.google.api.grpc:proto-google-cloud-bigqueryconnection-v1beta1 to v0.15.0](../pull/796) ## Detected dependencies <details><summary>github-actions</summary> <blockquote> <details><summary>.github/workflows/approve-readme.yaml</summary> - `actions/github-script v6` </details> <details><summary>.github/workflows/auto-release.yaml</summary> - `actions/github-script v6` </details> <details><summary>.github/workflows/ci.yaml</summary> - `actions/checkout v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` - `actions/checkout v3` - `actions/setup-java v3` </details> <details><summary>.github/workflows/samples.yaml</summary> - `actions/checkout v3` - `actions/setup-java v3` </details> </blockquote> </details> <details><summary>maven</summary> <blockquote> <details><summary>google-cloud-bigqueryconnection-bom/pom.xml</summary> - `com.google.cloud:google-cloud-shared-config 1.5.3` - `com.google.cloud:google-cloud-bigqueryconnection 2.5.7-SNAPSHOT` - `com.google.api.grpc:grpc-google-cloud-bigqueryconnection-v1 2.5.7-SNAPSHOT` - `com.google.api.grpc:grpc-google-cloud-bigqueryconnection-v1beta1 0.13.7-SNAPSHOT` - `com.google.api.grpc:proto-google-cloud-bigqueryconnection-v1 2.5.7-SNAPSHOT` - `com.google.api.grpc:proto-google-cloud-bigqueryconnection-v1beta1 0.13.7-SNAPSHOT` - `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13` </details> <details><summary>google-cloud-bigqueryconnection/pom.xml</summary> - `com.google.cloud:google-cloud-bigqueryconnection-parent 2.5.7-SNAPSHOT` </details> <details><summary>grpc-google-cloud-bigqueryconnection-v1/pom.xml</summary> - `com.google.cloud:google-cloud-bigqueryconnection-parent 2.5.7-SNAPSHOT` </details> <details><summary>grpc-google-cloud-bigqueryconnection-v1beta1/pom.xml</summary> - `com.google.cloud:google-cloud-bigqueryconnection-parent 2.5.7-SNAPSHOT` </details> <details><summary>pom.xml</summary> - `com.google.cloud:google-cloud-shared-config 1.5.3` - `com.google.cloud:google-cloud-bigqueryconnection 2.5.7-SNAPSHOT` - `com.google.api.grpc:proto-google-cloud-bigqueryconnection-v1 2.5.7-SNAPSHOT` - `com.google.api.grpc:grpc-google-cloud-bigqueryconnection-v1 2.5.7-SNAPSHOT` - `com.google.api.grpc:proto-google-cloud-bigqueryconnection-v1beta1 0.13.7-SNAPSHOT` - `com.google.api.grpc:grpc-google-cloud-bigqueryconnection-v1beta1 0.13.7-SNAPSHOT` - `com.google.cloud:google-cloud-shared-dependencies 3.0.4` - `junit:junit 4.13.2` - `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13` - `org.apache.maven.plugins:maven-project-info-reports-plugin 3.4.1` - `org.apache.maven.plugins:maven-javadoc-plugin 3.4.1` </details> <details><summary>proto-google-cloud-bigqueryconnection-v1/pom.xml</summary> - `com.google.cloud:google-cloud-bigqueryconnection-parent 2.5.7-SNAPSHOT` </details> <details><summary>proto-google-cloud-bigqueryconnection-v1beta1/pom.xml</summary> - `com.google.cloud:google-cloud-bigqueryconnection-parent 2.5.7-SNAPSHOT` </details> <details><summary>samples/install-without-bom/pom.xml</summary> - `com.google.cloud.samples:shared-configuration 1.2.0` - `com.google.cloud:google-cloud-bigqueryconnection 2.5.6` - `com.google.protobuf:protobuf-java-util 3.21.7` - `junit:junit 4.13.2` - `com.google.truth:truth 1.1.3` - `org.codehaus.mojo:build-helper-maven-plugin 3.3.0` </details> <details><summary>samples/pom.xml</summary> - `com.google.cloud.samples:shared-configuration 1.2.0` - `org.apache.maven.plugins:maven-deploy-plugin 3.0.0` - `org.sonatype.plugins:nexus-staging-maven-plugin 1.6.13` </details> <details><summary>samples/snapshot/pom.xml</summary> - `com.google.cloud.samples:shared-configuration 1.2.0` - `com.google.cloud:google-cloud-bigqueryconnection 2.5.7-SNAPSHOT` - `com.google.protobuf:protobuf-java-util 3.21.7` - `junit:junit 4.13.2` - `com.google.truth:truth 1.1.3` - `org.codehaus.mojo:build-helper-maven-plugin 3.3.0` </details> <details><summary>samples/snippets/pom.xml</summary> - `com.google.cloud.samples:shared-configuration 1.2.0` - `com.google.cloud:libraries-bom 26.1.3` - `com.google.cloud:google-cloud-bigqueryconnection 2.5.6` - `com.google.protobuf:protobuf-java-util 3.21.7` - `junit:junit 4.13.2` - `com.google.truth:truth 1.1.3` </details> </blockquote> </details> --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue lists renovate updates and detected dependencies read the docs to learn more edited blocked these updates have been manually edited so renovate will no longer make changes to discard all commits and start over click on a checkbox pull pull pull pull pull pull pull pull pull pull detected dependencies github actions github workflows approve readme yaml actions github script github workflows auto release yaml actions github script github workflows ci yaml actions checkout actions setup java actions checkout actions setup java actions checkout actions setup java actions checkout actions setup java actions checkout actions setup java github workflows samples yaml actions checkout actions setup java maven google cloud bigqueryconnection bom pom xml com google cloud google cloud shared config com google cloud google cloud bigqueryconnection snapshot com google api grpc grpc google cloud bigqueryconnection snapshot com google api grpc grpc google cloud bigqueryconnection snapshot com google api grpc proto google cloud bigqueryconnection snapshot com google api grpc proto google cloud bigqueryconnection snapshot org sonatype plugins nexus staging maven plugin google cloud bigqueryconnection pom xml com google cloud google cloud bigqueryconnection parent snapshot grpc google cloud bigqueryconnection pom xml com google cloud google cloud bigqueryconnection parent snapshot grpc google cloud bigqueryconnection pom xml com google cloud google cloud bigqueryconnection parent snapshot pom xml com google cloud google cloud shared config com google cloud google cloud bigqueryconnection snapshot com google api grpc proto google cloud bigqueryconnection snapshot com google api grpc grpc google cloud bigqueryconnection snapshot com google api grpc proto google cloud bigqueryconnection snapshot com google api grpc grpc google cloud bigqueryconnection snapshot com google cloud google cloud shared dependencies junit junit org sonatype plugins nexus staging maven plugin org apache maven plugins maven project info reports plugin org apache maven plugins maven javadoc plugin proto google cloud bigqueryconnection pom xml com google cloud google cloud bigqueryconnection parent snapshot proto google cloud bigqueryconnection pom xml com google cloud google cloud bigqueryconnection parent snapshot samples install without bom pom xml com google cloud samples shared configuration com google cloud google cloud bigqueryconnection com google protobuf protobuf java util junit junit com google truth truth org codehaus mojo build helper maven plugin samples pom xml com google cloud samples shared configuration org apache maven plugins maven deploy plugin org sonatype plugins nexus staging maven plugin samples snapshot pom xml com google cloud samples shared configuration com google cloud google cloud bigqueryconnection snapshot com google protobuf protobuf java util junit junit com google truth truth org codehaus mojo build helper maven plugin samples snippets pom xml com google cloud samples shared configuration com google cloud libraries bom com google cloud google cloud bigqueryconnection com google protobuf protobuf java util junit junit com google truth truth check this box to trigger a request for renovate to run again on this repository
1
195,437
22,339,631,304
IssuesEvent
2022-06-14 22:33:22
vincenzodistasio97/events-manager-io
https://api.github.com/repos/vincenzodistasio97/events-manager-io
closed
CVE-2021-32804 (High) detected in tar-2.2.1.tgz - autoclosed
security vulnerability
## CVE-2021-32804 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-2.2.1.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p> <p> Dependency Hierarchy: - babel-cli-6.26.0.tgz (Root Library) - chokidar-1.7.0.tgz - fsevents-1.1.3.tgz - node-pre-gyp-0.6.39.tgz - :x: **tar-2.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vincenzodistasio97/events-manager-io/commit/34b4ee7777ec330308085b59cefc667c68e51123">34b4ee7777ec330308085b59cefc667c68e51123</a></p> <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar. <p>Publish Date: 2021-08-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804>CVE-2021-32804</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9">https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9</a></p> <p>Release Date: 2021-08-03</p> <p>Fix Resolution (tar): 3.2.2</p> <p>Direct dependency fix Resolution (babel-cli): 7.0.0-alpha.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-32804 (High) detected in tar-2.2.1.tgz - autoclosed - ## CVE-2021-32804 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-2.2.1.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p> <p> Dependency Hierarchy: - babel-cli-6.26.0.tgz (Root Library) - chokidar-1.7.0.tgz - fsevents-1.1.3.tgz - node-pre-gyp-0.6.39.tgz - :x: **tar-2.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vincenzodistasio97/events-manager-io/commit/34b4ee7777ec330308085b59cefc667c68e51123">34b4ee7777ec330308085b59cefc667c68e51123</a></p> <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar. <p>Publish Date: 2021-08-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804>CVE-2021-32804</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9">https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9</a></p> <p>Release Date: 2021-08-03</p> <p>Fix Resolution (tar): 3.2.2</p> <p>Direct dependency fix Resolution (babel-cli): 7.0.0-alpha.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in tar tgz autoclosed cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href dependency hierarchy babel cli tgz root library chokidar tgz fsevents tgz node pre gyp tgz x tar tgz vulnerable library found in head commit a href found in base branch dev vulnerability details the npm package tar aka node tar before versions and has a arbitrary file creation overwrite vulnerability due to insufficient absolute path sanitization node tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the preservepaths flag is not set to true this is achieved by stripping the absolute path root from any absolute file paths contained in a tar file for example home user bashrc would turn into home user bashrc this logic was insufficient when file paths contained repeated path roots such as home user bashrc node tar would only strip a single path root from such paths when given an absolute file path with repeating path roots the resulting path e g home user bashrc would still resolve to an absolute path thus allowing arbitrary file creation and overwrite this issue was addressed in releases and users may work around this vulnerability without upgrading by creating a custom onentry method which sanitizes the entry path or a filter method which removes entries with absolute paths see referenced github advisory for details be aware of cve which fixes a similar bug in later versions of tar publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar direct dependency fix resolution babel cli alpha step up your open source security game with mend
0
4,163
7,107,918,897
IssuesEvent
2018-01-16 21:45:52
18F/product-guide
https://api.github.com/repos/18F/product-guide
closed
NEW SECTION - Hiring & Interviewing
process change waiting for content
Consulting built [this handy doc](https://docs.google.com/document/d/15zFqquxrj37O3s0LoHsJLVUrxQzmCsRM1_Rfu8r6FUI/edit#heading=h.zcoelxl3g5tv) to help guide the hiring and interviewing process. Let's turn this into another section of the guide for ALL product folk to use! Precedence is set on publicly sharing this info by Engineering, who does it [here](https://pages.18f.gov/joining-18f/interview-process/interviews/engineering-team/). They're planning on adding even more detail. We briefly outline the interview process [here](https://pages.18f.gov/joining-18f/interview-process/interviews/product-team/) for Product, and should point to this link, but also probably add content and potentially point to the guide from that page. **Resources** [mural.ly board team product built around interview questions](https://app.mural.ly/t/gsa6/m/gsa6/1440169660195) - red cards are elements of overall position description, purple are the bullets in the description, blue are the refined topic that we wanted to probe as a starting point for that part of the interview, and orange are the โ€œopening gambitโ€ questions to explore that topic. never got to the point of filling in โ€œwhat good/bad looks likeโ€ for each of the orange cards. yellow cards are all the brainstorming that happened between each of the other colors [slack convo for background](https://18f.slack.com/archives/product/p1453392885000160)
1.0
NEW SECTION - Hiring & Interviewing - Consulting built [this handy doc](https://docs.google.com/document/d/15zFqquxrj37O3s0LoHsJLVUrxQzmCsRM1_Rfu8r6FUI/edit#heading=h.zcoelxl3g5tv) to help guide the hiring and interviewing process. Let's turn this into another section of the guide for ALL product folk to use! Precedence is set on publicly sharing this info by Engineering, who does it [here](https://pages.18f.gov/joining-18f/interview-process/interviews/engineering-team/). They're planning on adding even more detail. We briefly outline the interview process [here](https://pages.18f.gov/joining-18f/interview-process/interviews/product-team/) for Product, and should point to this link, but also probably add content and potentially point to the guide from that page. **Resources** [mural.ly board team product built around interview questions](https://app.mural.ly/t/gsa6/m/gsa6/1440169660195) - red cards are elements of overall position description, purple are the bullets in the description, blue are the refined topic that we wanted to probe as a starting point for that part of the interview, and orange are the โ€œopening gambitโ€ questions to explore that topic. never got to the point of filling in โ€œwhat good/bad looks likeโ€ for each of the orange cards. yellow cards are all the brainstorming that happened between each of the other colors [slack convo for background](https://18f.slack.com/archives/product/p1453392885000160)
process
new section hiring interviewing consulting built to help guide the hiring and interviewing process let s turn this into another section of the guide for all product folk to use precedence is set on publicly sharing this info by engineering who does it they re planning on adding even more detail we briefly outline the interview process for product and should point to this link but also probably add content and potentially point to the guide from that page resources red cards are elements of overall position description purple are the bullets in the description blue are the refined topic that we wanted to probe as a starting point for that part of the interview and orange are the โ€œopening gambitโ€ questions to explore that topic never got to the point of filling in โ€œwhat good bad looks likeโ€ for each of the orange cards yellow cards are all the brainstorming that happened between each of the other colors
1
15,105
18,954,633,576
IssuesEvent
2021-11-18 18:43:52
WayofTime/BloodMagic
https://api.github.com/repos/WayofTime/BloodMagic
closed
Ritual Of The Martyr [woot] Loop Crash
invalid compatibility 1.16
#### Issue Description: Curiously searching for something that is not present. #### What happens: When activating Ritual Of The Martyr [woot] above a Tier V (or possibly all tier's, not sure) Farm that does not contain the appropriate upgrades in it causes a dedicated server to go into a crash loop. #### What you expected to happen: The ritual to active but not provide LP as the woot farm was not filled, however it was completed. #### Steps to reproduce: 1. Build T4 Altar below T5 Woot farm and do not place upgrades in it, then place ritual above 2. Activate Ritual Of The Martyr [woot] 3. Watch the magic happen #### Affected Versions (Do *not* use "latest"): - BloodMagic-1.16.4-3.1.5-25.jar - Minecraft: 1.16.5 - Forge: 36.2.8
True
Ritual Of The Martyr [woot] Loop Crash - #### Issue Description: Curiously searching for something that is not present. #### What happens: When activating Ritual Of The Martyr [woot] above a Tier V (or possibly all tier's, not sure) Farm that does not contain the appropriate upgrades in it causes a dedicated server to go into a crash loop. #### What you expected to happen: The ritual to active but not provide LP as the woot farm was not filled, however it was completed. #### Steps to reproduce: 1. Build T4 Altar below T5 Woot farm and do not place upgrades in it, then place ritual above 2. Activate Ritual Of The Martyr [woot] 3. Watch the magic happen #### Affected Versions (Do *not* use "latest"): - BloodMagic-1.16.4-3.1.5-25.jar - Minecraft: 1.16.5 - Forge: 36.2.8
non_process
ritual of the martyr loop crash issue description curiously searching for something that is not present what happens when activating ritual of the martyr above a tier v or possibly all tier s not sure farm that does not contain the appropriate upgrades in it causes a dedicated server to go into a crash loop what you expected to happen the ritual to active but not provide lp as the woot farm was not filled however it was completed steps to reproduce build altar below woot farm and do not place upgrades in it then place ritual above activate ritual of the martyr watch the magic happen affected versions do not use latest bloodmagic jar minecraft forge
0
59,572
17,023,165,415
IssuesEvent
2021-07-03 00:40:05
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Potlatch continues pretending to work after a login session expires
Component: potlatch (flash editor) Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 5.10pm, Monday, 21st May 2007]** If your login cookie expires then Potlatch contines allowing you to draw ways, but fails to upload them. This happens even if you use another tab in the browser to login again - only after the editor page has been reloaded does it start uploading again.
1.0
Potlatch continues pretending to work after a login session expires - **[Submitted to the original trac issue database at 5.10pm, Monday, 21st May 2007]** If your login cookie expires then Potlatch contines allowing you to draw ways, but fails to upload them. This happens even if you use another tab in the browser to login again - only after the editor page has been reloaded does it start uploading again.
non_process
potlatch continues pretending to work after a login session expires if your login cookie expires then potlatch contines allowing you to draw ways but fails to upload them this happens even if you use another tab in the browser to login again only after the editor page has been reloaded does it start uploading again
0
88,534
11,101,605,761
IssuesEvent
2019-12-16 21:48:50
18F/cms-hitech-apd
https://api.github.com/repos/18F/cms-hitech-apd
closed
Convert Goals and Objectives Section to align with Objectives and Key Results
blocked design
Per leadership request (and it just makes sense!), content should do a review of the language in the Goals and Objective Section and revise it to align with CMS' move towards utilizing Objectives and Key Results. Primer on OKR: https://felipecastro.com/en/okr/what-is-okr/ ### This task is done when... - [ ] @NAretakis has provided some feedback to team about what is expected from OKRs - [ ] Content has had a chance to review current state and see what needs to be done - [ ] Design (user interaction) and Dev are consulted as needed if changes outside of content are required - [ ] Appropriate issues/pull requests are added to appropriate sections
1.0
Convert Goals and Objectives Section to align with Objectives and Key Results - Per leadership request (and it just makes sense!), content should do a review of the language in the Goals and Objective Section and revise it to align with CMS' move towards utilizing Objectives and Key Results. Primer on OKR: https://felipecastro.com/en/okr/what-is-okr/ ### This task is done when... - [ ] @NAretakis has provided some feedback to team about what is expected from OKRs - [ ] Content has had a chance to review current state and see what needs to be done - [ ] Design (user interaction) and Dev are consulted as needed if changes outside of content are required - [ ] Appropriate issues/pull requests are added to appropriate sections
non_process
convert goals and objectives section to align with objectives and key results per leadership request and it just makes sense content should do a review of the language in the goals and objective section and revise it to align with cms move towards utilizing objectives and key results primer on okr this task is done when naretakis has provided some feedback to team about what is expected from okrs content has had a chance to review current state and see what needs to be done design user interaction and dev are consulted as needed if changes outside of content are required appropriate issues pull requests are added to appropriate sections
0
9,485
12,478,034,668
IssuesEvent
2020-05-29 15:55:17
kubeflow/pipelines
https://api.github.com/repos/kubeflow/pipelines
closed
[FR] Implement optional component level usage tracking
area/engprod kind/process platform/gcp priority/p1 status/triaged
As #3578 checked in, we have an example of how to set beacon to collect critical usage information to help us make the right business decision, without compromising user privacy. One natural extension of that work is to instrument telemetry of our first-party components shipped with KFP. To be concrete, the proposed scope includes components under [1]. We can consider whether/how to include user-defined component later on. Still my proposed approach will be facilitated by pod labels append to each ContainerOp. 1. We establish a whitelisting dictionary, only including first-party components; I am considering using the component url as signature. 2. In [2], when loading a component, we'll append telemetry pod labels, which will then surfaced in GKE. 3. We provide toggle to allow users to opt out this collection, at both component loading, and compiler. [1] https://github.com/kubeflow/pipelines/tree/master/components [2] https://github.com/kubeflow/pipelines/blob/master/sdk/python/kfp/components/_components.py Any input will be very helpful :) @Ark-kun @rmgogogo
1.0
[FR] Implement optional component level usage tracking - As #3578 checked in, we have an example of how to set beacon to collect critical usage information to help us make the right business decision, without compromising user privacy. One natural extension of that work is to instrument telemetry of our first-party components shipped with KFP. To be concrete, the proposed scope includes components under [1]. We can consider whether/how to include user-defined component later on. Still my proposed approach will be facilitated by pod labels append to each ContainerOp. 1. We establish a whitelisting dictionary, only including first-party components; I am considering using the component url as signature. 2. In [2], when loading a component, we'll append telemetry pod labels, which will then surfaced in GKE. 3. We provide toggle to allow users to opt out this collection, at both component loading, and compiler. [1] https://github.com/kubeflow/pipelines/tree/master/components [2] https://github.com/kubeflow/pipelines/blob/master/sdk/python/kfp/components/_components.py Any input will be very helpful :) @Ark-kun @rmgogogo
process
implement optional component level usage tracking as checked in we have an example of how to set beacon to collect critical usage information to help us make the right business decision without compromising user privacy one natural extension of that work is to instrument telemetry of our first party components shipped with kfp to be concrete the proposed scope includes components under we can consider whether how to include user defined component later on still my proposed approach will be facilitated by pod labels append to each containerop we establish a whitelisting dictionary only including first party components i am considering using the component url as signature in when loading a component we ll append telemetry pod labels which will then surfaced in gke we provide toggle to allow users to opt out this collection at both component loading and compiler any input will be very helpful ark kun rmgogogo
1
20,205
26,783,198,965
IssuesEvent
2023-01-31 23:18:34
googleapis/java-storage
https://api.github.com/repos/googleapis/java-storage
closed
test: add `storage.notification.*` conformance tests
type: process api: storage
Now that we have notifications integrated into the api, we should ensure our retries are following the prescribed best practice. * [ ] Add setup steps to RpcMethodMapping * [ ] Add mappings into RpcMethodMappings * [ ] Update StorageRetryStrategy and associated classes to account for notification operations * [ ] Use Retrying#run instead of runWithRetries
1.0
test: add `storage.notification.*` conformance tests - Now that we have notifications integrated into the api, we should ensure our retries are following the prescribed best practice. * [ ] Add setup steps to RpcMethodMapping * [ ] Add mappings into RpcMethodMappings * [ ] Update StorageRetryStrategy and associated classes to account for notification operations * [ ] Use Retrying#run instead of runWithRetries
process
test add storage notification conformance tests now that we have notifications integrated into the api we should ensure our retries are following the prescribed best practice add setup steps to rpcmethodmapping add mappings into rpcmethodmappings update storageretrystrategy and associated classes to account for notification operations use retrying run instead of runwithretries
1
13,331
15,790,053,100
IssuesEvent
2021-04-02 00:23:19
SaviorD7/applicant-service
https://api.github.com/repos/SaviorD7/applicant-service
closed
Combine data
ALARM data-processing
Calculate the average score of students. Eliminate unnecessary data. Make new column in students data with avg score. Combine data with "ะะพะผะตั€ ะ›ะ”" key. **Time spent: XX hours**
1.0
Combine data - Calculate the average score of students. Eliminate unnecessary data. Make new column in students data with avg score. Combine data with "ะะพะผะตั€ ะ›ะ”" key. **Time spent: XX hours**
process
combine data calculate the average score of students eliminate unnecessary data make new column in students data with avg score combine data with ะฝะพะผะตั€ ะปะด key time spent xx hours
1
7,942
11,137,519,126
IssuesEvent
2019-12-20 19:34:55
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Add "We're pulling data" modal after Next Steps
Apply Process Approved Requirements Ready State Dept.
Who: Applicants What: Provide modal to let them know we're pulling over USAJOBS data Why: in order to let them know there may be a time lag and also make it clear we're pulling over the data Acceptance: When a user clicks "Continue" on next steps, display a modal that lets them know we are pulling data over from USAJOBS. Title: Importing your information from USAJOBS Message: "Please wait while we import your education, work experience, references, and languages from your USAJOBS profile." No buttons and now "X" since you won't be able to stop this process
1.0
Add "We're pulling data" modal after Next Steps - Who: Applicants What: Provide modal to let them know we're pulling over USAJOBS data Why: in order to let them know there may be a time lag and also make it clear we're pulling over the data Acceptance: When a user clicks "Continue" on next steps, display a modal that lets them know we are pulling data over from USAJOBS. Title: Importing your information from USAJOBS Message: "Please wait while we import your education, work experience, references, and languages from your USAJOBS profile." No buttons and now "X" since you won't be able to stop this process
process
add we re pulling data modal after next steps who applicants what provide modal to let them know we re pulling over usajobs data why in order to let them know there may be a time lag and also make it clear we re pulling over the data acceptance when a user clicks continue on next steps display a modal that lets them know we are pulling data over from usajobs title importing your information from usajobs message please wait while we import your education work experience references and languages from your usajobs profile no buttons and now x since you won t be able to stop this process
1
12,370
14,895,927,871
IssuesEvent
2021-01-21 09:44:52
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
closed
Ambiguous cache retrieval in regrid preprocessor
bear at a dinner party bug preprocessor
**Describe the bug** In #907 I came across the following test failure: ``` self = <tests.unit.preprocessor._regrid.test_regrid.Test testMethod=test_regrid__cell_specification> def test_regrid__cell_specification(self): specs = ['1x1', '2x2', '3x3', '4x4', '5x5'] scheme = 'linear' for spec in specs: result = regrid(self.src_cube, spec, scheme) self.assertEqual(result, self.regridded_cube) > self._check(spec, scheme, spec=True) tests/unit/preprocessor/_regrid/test_regrid.py:107: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ tests/unit/preprocessor/_regrid/test_regrid.py:23: in _check self.assertEqual(_CACHE[spec], self.tgt_grid) E AssertionError: <iris 'Cube' of unknown / (unknown) (latitude: 180; longitude: 360)> != <Mock spec='Cube' id='140424322565072'> ``` I could not reproduce the error on my machine, and it occasionally passed on CircleCI. I tracked it down to being related to the `_CACHE`: https://github.com/ESMValGroup/ESMValCore/blob/de112e449c2aa8fb1d132816c8c5cd7a3186b821/esmvalcore/preprocessor/_regrid.py#L42 The cache stores the results of the regridding as a simpley key, `1x1`, `2x2`, etc. The problem is that the key assignment is ambiguous. This means after multiple calls, there _will_ be a name collision, so that the wrong cached value is retrieved. This may lead to data corruption, and in my case, random test failures. For #907, I believe that the outcome of the test depends on whether the regridder goes first (pass) or last (second). I happened to catch this, because I implement additional tests which also call the regridder at some point. I have worked around the issue the issue by clearing `_CACHE` in #907 , but a less naive implementation of the cache is necessary.
1.0
Ambiguous cache retrieval in regrid preprocessor - **Describe the bug** In #907 I came across the following test failure: ``` self = <tests.unit.preprocessor._regrid.test_regrid.Test testMethod=test_regrid__cell_specification> def test_regrid__cell_specification(self): specs = ['1x1', '2x2', '3x3', '4x4', '5x5'] scheme = 'linear' for spec in specs: result = regrid(self.src_cube, spec, scheme) self.assertEqual(result, self.regridded_cube) > self._check(spec, scheme, spec=True) tests/unit/preprocessor/_regrid/test_regrid.py:107: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ tests/unit/preprocessor/_regrid/test_regrid.py:23: in _check self.assertEqual(_CACHE[spec], self.tgt_grid) E AssertionError: <iris 'Cube' of unknown / (unknown) (latitude: 180; longitude: 360)> != <Mock spec='Cube' id='140424322565072'> ``` I could not reproduce the error on my machine, and it occasionally passed on CircleCI. I tracked it down to being related to the `_CACHE`: https://github.com/ESMValGroup/ESMValCore/blob/de112e449c2aa8fb1d132816c8c5cd7a3186b821/esmvalcore/preprocessor/_regrid.py#L42 The cache stores the results of the regridding as a simpley key, `1x1`, `2x2`, etc. The problem is that the key assignment is ambiguous. This means after multiple calls, there _will_ be a name collision, so that the wrong cached value is retrieved. This may lead to data corruption, and in my case, random test failures. For #907, I believe that the outcome of the test depends on whether the regridder goes first (pass) or last (second). I happened to catch this, because I implement additional tests which also call the regridder at some point. I have worked around the issue the issue by clearing `_CACHE` in #907 , but a less naive implementation of the cache is necessary.
process
ambiguous cache retrieval in regrid preprocessor describe the bug in i came across the following test failure self def test regrid cell specification self specs scheme linear for spec in specs result regrid self src cube spec scheme self assertequal result self regridded cube self check spec scheme spec true tests unit preprocessor regrid test regrid py tests unit preprocessor regrid test regrid py in check self assertequal cache self tgt grid e assertionerror i could not reproduce the error on my machine and it occasionally passed on circleci i tracked it down to being related to the cache the cache stores the results of the regridding as a simpley key etc the problem is that the key assignment is ambiguous this means after multiple calls there will be a name collision so that the wrong cached value is retrieved this may lead to data corruption and in my case random test failures for i believe that the outcome of the test depends on whether the regridder goes first pass or last second i happened to catch this because i implement additional tests which also call the regridder at some point i have worked around the issue the issue by clearing cache in but a less naive implementation of the cache is necessary
1
4,612
7,453,766,536
IssuesEvent
2018-03-29 13:07:15
contentmonkey/contentmonkey
https://api.github.com/repos/contentmonkey/contentmonkey
opened
Admin interface rework
enhancement important in-process
The admin interface should be fully reworked, in order to provide a future save, modern dashboard. Current dashboard is using an old version of the bootstrap framework, the new dashboard should be written in **TypeScript** and **React**.
1.0
Admin interface rework - The admin interface should be fully reworked, in order to provide a future save, modern dashboard. Current dashboard is using an old version of the bootstrap framework, the new dashboard should be written in **TypeScript** and **React**.
process
admin interface rework the admin interface should be fully reworked in order to provide a future save modern dashboard current dashboard is using an old version of the bootstrap framework the new dashboard should be written in typescript and react
1
317,089
9,660,760,309
IssuesEvent
2019-05-20 16:12:00
coding-eval-platform/evaluations-service
https://api.github.com/repos/coding-eval-platform/evaluations-service
closed
Remove this
Chore Priority: High
https://github.com/coding-eval-platform/evaluations-service/blob/ad45f4d0a81b49d348301e64c41bd0695526ebd6/evaluations-service-core/evaluations-service-ports/evaluations-service-services/src/main/java/ar/edu/itba/cep/evaluations_service/services/ExamService.java#L336 Ralated to #22 and #62, and in #60 (in which another port is defined for this)
1.0
Remove this - https://github.com/coding-eval-platform/evaluations-service/blob/ad45f4d0a81b49d348301e64c41bd0695526ebd6/evaluations-service-core/evaluations-service-ports/evaluations-service-services/src/main/java/ar/edu/itba/cep/evaluations_service/services/ExamService.java#L336 Ralated to #22 and #62, and in #60 (in which another port is defined for this)
non_process
remove this ralated to and and in in which another port is defined for this
0
9,227
12,259,316,407
IssuesEvent
2020-05-06 16:24:21
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
chore: put Kokoro build badge in the README
priority: p2 type: process
Currently README.md has a stale build badge pointing to Travis build. We should use Kokoro build badges.
1.0
chore: put Kokoro build badge in the README - Currently README.md has a stale build badge pointing to Travis build. We should use Kokoro build badges.
process
chore put kokoro build badge in the readme currently readme md has a stale build badge pointing to travis build we should use kokoro build badges
1
452,581
13,056,077,026
IssuesEvent
2020-07-30 03:35:14
kubesphere/console
https://api.github.com/repos/kubesphere/console
opened
wrong custom ws role
area/console area/iam kind/bug priority/high
**Describe the bug** `ws-tester8` is the uer with the custom ws role `Workspace Settings`, but log in with the user, it is totally wrong as you can see from the following screenshots. ![Screen Shot 2020-07-30 at 11 32 31 AM](https://user-images.githubusercontent.com/28859385/88877659-9fe8a380-d258-11ea-879f-15a354052992.png) ![Screen Shot 2020-07-30 at 11 31 38 AM](https://user-images.githubusercontent.com/28859385/88877664-a6771b00-d258-11ea-87fa-df0ed1698d42.png) **Versions used(KubeSphere/Kubernetes)** KubeSphere: 3.0.0-dev
1.0
wrong custom ws role - **Describe the bug** `ws-tester8` is the uer with the custom ws role `Workspace Settings`, but log in with the user, it is totally wrong as you can see from the following screenshots. ![Screen Shot 2020-07-30 at 11 32 31 AM](https://user-images.githubusercontent.com/28859385/88877659-9fe8a380-d258-11ea-879f-15a354052992.png) ![Screen Shot 2020-07-30 at 11 31 38 AM](https://user-images.githubusercontent.com/28859385/88877664-a6771b00-d258-11ea-87fa-df0ed1698d42.png) **Versions used(KubeSphere/Kubernetes)** KubeSphere: 3.0.0-dev
non_process
wrong custom ws role describe the bug ws is the uer with the custom ws role workspace settings but log in with the user it is totally wrong as you can see from the following screenshots versions used kubesphere kubernetes kubesphere dev
0
360,460
25,290,831,074
IssuesEvent
2022-11-17 00:00:12
jupyterlab/jupyterlab
https://api.github.com/repos/jupyterlab/jupyterlab
opened
Document how to set text on CodeEditor in JupyterLab 4.0
enhancement documentation tag:Documentation
In JupyterLab 3.x we extension authors could do `widget.content.model.value.text = 'something'`. In JupyterLab 4.0 this no longer works but can be achieved with: ```js widget.content.model.sharedModel.setSource('foo'); // or // widget.content.model.sharedModel.updateSource(0, 0, 'foo'); ``` This should be added to [migration guide](https://jupyterlab.readthedocs.io/en/latest/extension/extension_migration.html) for easy reference. Currently `sharedModel` does not appear in the migration guide at all.
2.0
Document how to set text on CodeEditor in JupyterLab 4.0 - In JupyterLab 3.x we extension authors could do `widget.content.model.value.text = 'something'`. In JupyterLab 4.0 this no longer works but can be achieved with: ```js widget.content.model.sharedModel.setSource('foo'); // or // widget.content.model.sharedModel.updateSource(0, 0, 'foo'); ``` This should be added to [migration guide](https://jupyterlab.readthedocs.io/en/latest/extension/extension_migration.html) for easy reference. Currently `sharedModel` does not appear in the migration guide at all.
non_process
document how to set text on codeeditor in jupyterlab in jupyterlab x we extension authors could do widget content model value text something in jupyterlab this no longer works but can be achieved with js widget content model sharedmodel setsource foo or widget content model sharedmodel updatesource foo this should be added to for easy reference currently sharedmodel does not appear in the migration guide at all
0
389,476
26,814,618,071
IssuesEvent
2023-02-02 02:42:45
Supergrammer/supergrammer.github.io
https://api.github.com/repos/Supergrammer/supergrammer.github.io
closed
Vuetify ์ถ”๊ฐ€
๋ฌธ์„œ (Documentation) ๊ธฐ๋Šฅ ์ถ”๊ฐ€ (Enhancement) ์ž˜ ๋ด๋ผ (Good first issue)
*** ### ๐Ÿ’ฃ ์ด์Šˆ (Issue) : > Vuetify ๋ฐ Vuetify ์ปดํฌ๋„ŒํŠธ ์ถ”๊ฐ€ ### ๐Ÿ“‹ ์—…๋ฌด ์ฒดํฌ๋ฆฌ์ŠคํŠธ (Task Checklist) : > ์—ฌ๊ธฐ์— ์—…๋ฌด ์ฒดํฌ๋ฆฌ์ŠคํŠธ๋ฅผ ์ž‘์„ฑํ•˜์„ธ์š”. > > - [x] Vuetify ์ถ”๊ฐ€ > - [x] Vuetify submodule (icon ๋“ฑ) ์ถ”๊ฐ€ ### ๐Ÿ“œ ์ƒ์„ธ ์„ค๋ช… (Description) : > Vue project ์— Vuetify ๋ฐ Vuetify ์˜ submodule ๋ฅผ ์ถ”๊ฐ€ํ•˜๋Š” ์ž‘์—… ### ๐Ÿ”— ๊ด€๋ จ ์ด์Šˆ (Related Issues) : > ์—ฌ๊ธฐ์— ๊ด€๋ จ ์ด์Šˆ์— ๋Œ€ํ•ด ๊ธฐ์ˆ ํ•˜๊ฑฐ๋‚˜ ๋งํฌ๋ฅผ ์ฒจ๋ถ€ํ•˜์„ธ์š”. ### ๐Ÿ“š ์ฐธ๊ณ  (Reference) : > https://github.com/Supergrammer/JIB-SEE/issues/1 *** ### โœ๏ธ ์ฒ˜๋ฆฌ ๋‚ด์šฉ > ํ•ด๋‹น ์ด์Šˆ ์ฒ˜๋ฆฌ ํ›„, ์ฒ˜๋ฆฌ ๋‚ด์šฉ์— ๋Œ€ํ•ด ๊ธฐ์ˆ ํ•˜๋Š” ๋ถ€๋ถ„์ž…๋‹ˆ๋‹ค. > ์ด์Šˆ ์ฒ˜๋ฆฌ ํ›„ ์ž‘์„ฑํ•ด ์ฃผ์„ธ์š” #### Vuetify ์ถ”๊ฐ€ ```bash % npm i --save vuetify@^3.0.6 ``` - plugins/vuetify.js ```javascript import { createVuetify } from "vuetify"; import * as components from "vuetify/components"; import * as directives from "vuetify/directives"; import { fa } from "vuetify/iconsets/fa" import "vuetify/styles"; const vuetify = createVuetify({ components, directives, icons: { defaultSet: "fa", sets: { fa }, } }); export default vuetify; ``` - main.js ```javascript import { createApp } from "vue"; import { createPinia } from "pinia"; import App from "@/App.vue"; import router from "@/router"; import vuetify from "@/plugins/vuetify"; const app = createApp(App); const pinia = createPinia(); app.use(router); app.use(pinia); app.use(vuetify); app.mount("#app"); ``` #### Vuetify Icons module ์ถ”๊ฐ€ ```bash # Material Design Icon ์ถ”๊ฐ€ % npm i --save @mdi/font # Fontawesome Icon ์ถ”๊ฐ€ % npm i --save @fortawesome/fontawesome-free ``` - main.js ```javascript ... import "@mdi/font/css/materialdesignicons.css" import "@fortawesome/fontawesome-free/css/all.css" ... ``` ***
1.0
Vuetify ์ถ”๊ฐ€ - *** ### ๐Ÿ’ฃ ์ด์Šˆ (Issue) : > Vuetify ๋ฐ Vuetify ์ปดํฌ๋„ŒํŠธ ์ถ”๊ฐ€ ### ๐Ÿ“‹ ์—…๋ฌด ์ฒดํฌ๋ฆฌ์ŠคํŠธ (Task Checklist) : > ์—ฌ๊ธฐ์— ์—…๋ฌด ์ฒดํฌ๋ฆฌ์ŠคํŠธ๋ฅผ ์ž‘์„ฑํ•˜์„ธ์š”. > > - [x] Vuetify ์ถ”๊ฐ€ > - [x] Vuetify submodule (icon ๋“ฑ) ์ถ”๊ฐ€ ### ๐Ÿ“œ ์ƒ์„ธ ์„ค๋ช… (Description) : > Vue project ์— Vuetify ๋ฐ Vuetify ์˜ submodule ๋ฅผ ์ถ”๊ฐ€ํ•˜๋Š” ์ž‘์—… ### ๐Ÿ”— ๊ด€๋ จ ์ด์Šˆ (Related Issues) : > ์—ฌ๊ธฐ์— ๊ด€๋ จ ์ด์Šˆ์— ๋Œ€ํ•ด ๊ธฐ์ˆ ํ•˜๊ฑฐ๋‚˜ ๋งํฌ๋ฅผ ์ฒจ๋ถ€ํ•˜์„ธ์š”. ### ๐Ÿ“š ์ฐธ๊ณ  (Reference) : > https://github.com/Supergrammer/JIB-SEE/issues/1 *** ### โœ๏ธ ์ฒ˜๋ฆฌ ๋‚ด์šฉ > ํ•ด๋‹น ์ด์Šˆ ์ฒ˜๋ฆฌ ํ›„, ์ฒ˜๋ฆฌ ๋‚ด์šฉ์— ๋Œ€ํ•ด ๊ธฐ์ˆ ํ•˜๋Š” ๋ถ€๋ถ„์ž…๋‹ˆ๋‹ค. > ์ด์Šˆ ์ฒ˜๋ฆฌ ํ›„ ์ž‘์„ฑํ•ด ์ฃผ์„ธ์š” #### Vuetify ์ถ”๊ฐ€ ```bash % npm i --save vuetify@^3.0.6 ``` - plugins/vuetify.js ```javascript import { createVuetify } from "vuetify"; import * as components from "vuetify/components"; import * as directives from "vuetify/directives"; import { fa } from "vuetify/iconsets/fa" import "vuetify/styles"; const vuetify = createVuetify({ components, directives, icons: { defaultSet: "fa", sets: { fa }, } }); export default vuetify; ``` - main.js ```javascript import { createApp } from "vue"; import { createPinia } from "pinia"; import App from "@/App.vue"; import router from "@/router"; import vuetify from "@/plugins/vuetify"; const app = createApp(App); const pinia = createPinia(); app.use(router); app.use(pinia); app.use(vuetify); app.mount("#app"); ``` #### Vuetify Icons module ์ถ”๊ฐ€ ```bash # Material Design Icon ์ถ”๊ฐ€ % npm i --save @mdi/font # Fontawesome Icon ์ถ”๊ฐ€ % npm i --save @fortawesome/fontawesome-free ``` - main.js ```javascript ... import "@mdi/font/css/materialdesignicons.css" import "@fortawesome/fontawesome-free/css/all.css" ... ``` ***
non_process
vuetify ์ถ”๊ฐ€ ๐Ÿ’ฃ ์ด์Šˆ issue vuetify ๋ฐ vuetify ์ปดํฌ๋„ŒํŠธ ์ถ”๊ฐ€ ๐Ÿ“‹ ์—…๋ฌด ์ฒดํฌ๋ฆฌ์ŠคํŠธ task checklist ์—ฌ๊ธฐ์— ์—…๋ฌด ์ฒดํฌ๋ฆฌ์ŠคํŠธ๋ฅผ ์ž‘์„ฑํ•˜์„ธ์š” vuetify ์ถ”๊ฐ€ vuetify submodule icon ๋“ฑ ์ถ”๊ฐ€ ๐Ÿ“œ ์ƒ์„ธ ์„ค๋ช… description vue project ์— vuetify ๋ฐ vuetify ์˜ submodule ๋ฅผ ์ถ”๊ฐ€ํ•˜๋Š” ์ž‘์—… ๐Ÿ”— ๊ด€๋ จ ์ด์Šˆ related issues ์—ฌ๊ธฐ์— ๊ด€๋ จ ์ด์Šˆ์— ๋Œ€ํ•ด ๊ธฐ์ˆ ํ•˜๊ฑฐ๋‚˜ ๋งํฌ๋ฅผ ์ฒจ๋ถ€ํ•˜์„ธ์š” ๐Ÿ“š ์ฐธ๊ณ  reference โœ๏ธ ์ฒ˜๋ฆฌ ๋‚ด์šฉ ํ•ด๋‹น ์ด์Šˆ ์ฒ˜๋ฆฌ ํ›„ ์ฒ˜๋ฆฌ ๋‚ด์šฉ์— ๋Œ€ํ•ด ๊ธฐ์ˆ ํ•˜๋Š” ๋ถ€๋ถ„์ž…๋‹ˆ๋‹ค ์ด์Šˆ ์ฒ˜๋ฆฌ ํ›„ ์ž‘์„ฑํ•ด ์ฃผ์„ธ์š” vuetify ์ถ”๊ฐ€ bash npm i save vuetify plugins vuetify js javascript import createvuetify from vuetify import as components from vuetify components import as directives from vuetify directives import fa from vuetify iconsets fa import vuetify styles const vuetify createvuetify components directives icons defaultset fa sets fa export default vuetify main js javascript import createapp from vue import createpinia from pinia import app from app vue import router from router import vuetify from plugins vuetify const app createapp app const pinia createpinia app use router app use pinia app use vuetify app mount app vuetify icons module ์ถ”๊ฐ€ bash material design icon ์ถ”๊ฐ€ npm i save mdi font fontawesome icon ์ถ”๊ฐ€ npm i save fortawesome fontawesome free main js javascript import mdi font css materialdesignicons css import fortawesome fontawesome free css all css
0
194,583
14,681,759,241
IssuesEvent
2020-12-31 14:16:17
seannich/RatVenture_ETI_Assignment
https://api.github.com/repos/seannich/RatVenture_ETI_Assignment
closed
Town menu functions test โ€” View Character
test
- shows the playerโ€™s stats properly. - if player has 10hp it should be reflected there
1.0
Town menu functions test โ€” View Character - - shows the playerโ€™s stats properly. - if player has 10hp it should be reflected there
non_process
town menu functions test โ€” view character shows the playerโ€™s stats properly if player has it should be reflected there
0
1,578
4,173,252,406
IssuesEvent
2016-06-21 09:50:21
ProgrammingLife2016/PL2-2016
https://api.github.com/repos/ProgrammingLife2016/PL2-2016
closed
Exact base positions within a gene cannot be easily located
data processing development enhancement wontfix
The client specifically requires indexing within genes. An example is to jump to the 100th base / nucleotide in a gene. This way, the client would want to look for mutations that happen inside that gene.
1.0
Exact base positions within a gene cannot be easily located - The client specifically requires indexing within genes. An example is to jump to the 100th base / nucleotide in a gene. This way, the client would want to look for mutations that happen inside that gene.
process
exact base positions within a gene cannot be easily located the client specifically requires indexing within genes an example is to jump to the base nucleotide in a gene this way the client would want to look for mutations that happen inside that gene
1
208,440
23,605,493,081
IssuesEvent
2022-08-24 07:54:36
ioana-nicolae/second
https://api.github.com/repos/ioana-nicolae/second
closed
CVE-2019-12814 (Medium) detected in jackson-databind-2.7.9.jar - autoclosed
security vulnerability
## CVE-2019-12814 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.9.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.9/a4c0b14c7dd85bdf4d25da074e90a10fa4b9b88b/jackson-databind-2.7.9.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.7.9.jar** (Vulnerable Library) <p>Found in base branch: <b>branch3</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.x through 2.9.9. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has JDOM 1.x or 2.x jar in the classpath, an attacker can send a specifically crafted JSON message that allows them to read arbitrary local files on the server. <p>Publish Date: 2019-06-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12814>CVE-2019-12814</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2020-10-20</p> <p>Fix Resolution: 2.7.9.6</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
True
CVE-2019-12814 (Medium) detected in jackson-databind-2.7.9.jar - autoclosed - ## CVE-2019-12814 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.9.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.9/a4c0b14c7dd85bdf4d25da074e90a10fa4b9b88b/jackson-databind-2.7.9.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.7.9.jar** (Vulnerable Library) <p>Found in base branch: <b>branch3</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.x through 2.9.9. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has JDOM 1.x or 2.x jar in the classpath, an attacker can send a specifically crafted JSON message that allows them to read arbitrary local files on the server. <p>Publish Date: 2019-06-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12814>CVE-2019-12814</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2020-10-20</p> <p>Fix Resolution: 2.7.9.6</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
non_process
cve medium detected in jackson databind jar autoclosed cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file build gradle path to vulnerable library gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in base branch vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind x through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has jdom x or x jar in the classpath an attacker can send a specifically crafted json message that allows them to read arbitrary local files on the server publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution rescue worker helmet automatic remediation is available for this issue
0
22,120
30,656,188,582
IssuesEvent
2023-07-25 12:19:18
AvaloniaUI/Avalonia
https://api.github.com/repos/AvaloniaUI/Avalonia
closed
Cannot select text in multi-line SelectableTextBlock easily
bug area-textprocessing
**Describe the bug** I dont really know how to explain the bug but it is difficult to select text manually in a `SelectableTextBlock` that has multiple lines in it. See the video below. **To Reproduce** Steps to reproduce the behavior: 1. Add a `SelectableTextBlock`. 2. Add some texts with multiple line. 3. Try to select the text as shown in the video. **Expected behavior** Can easily select the text inside the control as shown with Notepad in the latter part of the video below. **Desktop (please complete the following information):** - OS: Windows - Version 11.0.999-cibuild0032750-beta **Additional context** [Sample project](https://github.com/AvaloniaUI/Avalonia/files/11156598/App.zip) https://user-images.githubusercontent.com/12814796/230022153-e57ba4b2-832b-4f1c-ae6e-7e0ddd2533eb.mp4
1.0
Cannot select text in multi-line SelectableTextBlock easily - **Describe the bug** I dont really know how to explain the bug but it is difficult to select text manually in a `SelectableTextBlock` that has multiple lines in it. See the video below. **To Reproduce** Steps to reproduce the behavior: 1. Add a `SelectableTextBlock`. 2. Add some texts with multiple line. 3. Try to select the text as shown in the video. **Expected behavior** Can easily select the text inside the control as shown with Notepad in the latter part of the video below. **Desktop (please complete the following information):** - OS: Windows - Version 11.0.999-cibuild0032750-beta **Additional context** [Sample project](https://github.com/AvaloniaUI/Avalonia/files/11156598/App.zip) https://user-images.githubusercontent.com/12814796/230022153-e57ba4b2-832b-4f1c-ae6e-7e0ddd2533eb.mp4
process
cannot select text in multi line selectabletextblock easily describe the bug i dont really know how to explain the bug but it is difficult to select text manually in a selectabletextblock that has multiple lines in it see the video below to reproduce steps to reproduce the behavior add a selectabletextblock add some texts with multiple line try to select the text as shown in the video expected behavior can easily select the text inside the control as shown with notepad in the latter part of the video below desktop please complete the following information os windows version beta additional context
1
12,009
14,738,365,386
IssuesEvent
2021-01-07 04:32:57
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Notification: Is your merchantโ€™s payment platform ready to go?
anc-process anp-3 ant-support
In GitLab by @kdjstudios on May 15, 2018, 10:25 HD: http://www.servicedesk.answernet.com/profiles/ticket/2018-05-15-82213/conversation Notification email from Authorize: http://app.payment.authorize.net/e/es.aspx?s=986383348&e=1580378&elqTrackId=b307147cf4ef4925bd108180234867d4&elq=b87ae2edc3a44e3b9e80b9066c2426f5&elqaid=715&elqat=1
1.0
Notification: Is your merchantโ€™s payment platform ready to go? - In GitLab by @kdjstudios on May 15, 2018, 10:25 HD: http://www.servicedesk.answernet.com/profiles/ticket/2018-05-15-82213/conversation Notification email from Authorize: http://app.payment.authorize.net/e/es.aspx?s=986383348&e=1580378&elqTrackId=b307147cf4ef4925bd108180234867d4&elq=b87ae2edc3a44e3b9e80b9066c2426f5&elqaid=715&elqat=1
process
notification is your merchantโ€™s payment platform ready to go in gitlab by kdjstudios on may hd notification email from authorize
1
194,098
14,669,476,346
IssuesEvent
2020-12-30 00:52:27
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
pingcap/tidb-operator: pkg/backup/backupschedule/backup_schedule_manager_test.go; 12 LoC
fresh small test
Found a possible issue in [pingcap/tidb-operator](https://www.github.com/pingcap/tidb-operator) at [pkg/backup/backupschedule/backup_schedule_manager_test.go](https://github.com/pingcap/tidb-operator/blob/7236eaba7d5924f47ae0cd3b0bf928886ca8cfc3/pkg/backup/backupschedule/backup_schedule_manager_test.go#L97-L108) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call at line 105 passes reference to bk to third-party code [Click here to see the code in its original context.](https://github.com/pingcap/tidb-operator/blob/7236eaba7d5924f47ae0cd3b0bf928886ca8cfc3/pkg/backup/backupschedule/backup_schedule_manager_test.go#L97-L108) <details> <summary>Click here to show the 12 line(s) of Go which triggered the analyzer.</summary> ```go for _, bk := range bks.Items { changed := v1alpha1.UpdateBackupCondition(&bk.Status, &v1alpha1.BackupCondition{ Type: v1alpha1.BackupComplete, Status: v1.ConditionTrue, }) if changed { bk.CreationTimestamp = metav1.Time{Time: m.now()} t.Log("complete backup: ", bk.Name) helper.updateBackup(&bk) } g.Expect(err).Should(BeNil()) } ``` </details> <details> <summary>Click here to show extra information the analyzer produced.</summary> ``` No path was found through the callgraph that could lead to a function which writes a pointer argument. The following graphviz dot graph describes paths through the callgraph that could lead to a function which passes a pointer to third-party code: digraph G { "(Update, 1)" -> {} "(Scale, 3)" -> {"(ScaleIn, 3)";"(scaleOne, 2)";"(ScaleOut, 3)";} "(syncTiDBService, 1)" -> {"(CreateService, 2)";"(UpdateService, 2)";} "(syncTiKVConfigMap, 2)" -> {"(CreateOrUpdateConfigMap, 2)";} "(gracefulUpgrade, 3)" -> {"(upgradePDPod, 3)";} "(syncTiDBStatefulSetForTidbCluster, 1)" -> {"(UpdateStatefulSet, 4)";"(syncTiDBConfigMap, 2)";"(Upgrade, 3)";"(CreateStatefulSet, 2)";"(syncTidbClusterStatus, 2)";"(Failover, 1)";} "(setStoreLabelsForTiKV, 1)" -> {} "(syncRestoreJob, 1)" -> {"(CreateJob, 2)";"(ensureRestorePVCExist, 1)";"(Update, 2)";} "(Upgrade, 3)" -> {"(gracefulUpgrade, 3)";} "(resetReplicas, 2)" -> {} "(updateBackup, 1)" -> {"(Sync, 1)";} "(CreateService, 2)" -> {"(Create, 1)";} "(ScaleIn, 3)" -> {"(setReplicasAndDeleteSlots, 3)";"(resetReplicas, 2)";"(scaleOne, 2)";} "(syncConfigMap, 2)" -> {"(CreateOrUpdateConfigMap, 2)";} "(ensureRestorePVCExist, 1)" -> {"(CreatePVC, 2)";} "(syncServiceForTidbCluster, 2)" -> {"(CreateService, 2)";"(UpdateService, 2)";} "(syncBackupJob, 1)" -> {"(CreateJob, 2)";"(ensureBackupPVCExist, 1)";"(Update, 2)";} "(CreateJob, 2)" -> {"(Create, 1)";} "(syncTiDBConfigMap, 2)" -> {"(CreateOrUpdateConfigMap, 2)";} "(Update, 2)" -> {"(Update, 1)";} "(syncPDHeadlessServiceForTidbCluster, 1)" -> {"(CreateService, 2)";"(UpdateService, 2)";} "(syncTiCDCStatus, 2)" -> {} "(setReplicasAndDeleteSlots, 3)" -> {} "(Sync, 1)" -> {"(syncCDCHeadlessService, 1)";"(syncTiDBHeadlessServiceForTidbCluster, 1)";"(syncPDServiceForTidbCluster, 1)";"(syncStatefulSet, 1)";"(UpdateMetaInfo, 2)";"(syncTiDBStatefulSetForTidbCluster, 1)";"(syncPDHeadlessServiceForTidbCluster, 1)";"(syncTiDBService, 1)";"(syncStatefulSetForTidbCluster, 1)";"(syncTiDBInitJob, 1)";"(syncPDStatefulSetForTidbCluster, 1)";"(syncRestoreJob, 1)";"(UpdateMetaInfo, 3)";"(syncTiDBInitConfigMap, 1)";"(syncServiceForTidbCluster, 2)";"(syncHeadlessService, 1)";"(syncPumpStatefulSetForTidbCluster, 1)";"(Clean, 1)";"(syncBackupJob, 1)";} "(Failover, 1)" -> {"(isPDInQuorum, 1)";"(tryToDeleteAFailureMember, 1)";"(tryToMarkAPeerAsFailure, 1)";} "(syncCDCHeadlessService, 1)" -> {"(CreateService, 2)";"(UpdateService, 2)";} "(syncHeadlessService, 1)" -> {"(UpdateService, 2)";"(CreateService, 2)";} "(syncStatefulSetForTidbCluster, 1)" -> {"(UpdateStatefulSet, 4)";"(syncTiKVConfigMap, 2)";"(Scale, 3)";"(Upgrade, 3)";"(CreateStatefulSet, 2)";"(syncTidbClusterStatus, 2)";"(Failover, 1)";"(setStoreLabelsForTiKV, 1)";} "(syncTiDBInitJob, 1)" -> {} "(syncPDStatefulSetForTidbCluster, 1)" -> {"(Scale, 3)";"(Upgrade, 3)";"(CreateStatefulSet, 2)";"(syncTidbClusterStatus, 2)";"(Failover, 1)";"(UpdateStatefulSet, 4)";"(syncPDConfigMap, 2)";} "(syncTidbClusterStatus, 2)" -> {} "(CreatePVC, 2)" -> {} "(isPDInQuorum, 1)" -> {} "(ensureBackupPVCExist, 1)" -> {"(CreatePVC, 2)";} "(Clean, 1)" -> {"(CreateJob, 2)";"(Update, 2)";} "(Create, 1)" -> {} "(ScaleOut, 3)" -> {"(setReplicasAndDeleteSlots, 3)";"(resetReplicas, 2)";"(scaleOne, 2)";} "(UpdateStatefulSet, 4)" -> {} "(UpdateMetaInfo, 2)" -> {"(Update, 1)";} "(tryToMarkAPeerAsFailure, 1)" -> {} "(syncPDConfigMap, 2)" -> {"(CreateOrUpdateConfigMap, 2)";} "(UpdateService, 2)" -> {"(Update, 1)";} "(CreateOrUpdateConfigMap, 2)" -> {} "(syncStatefulSet, 1)" -> {"(CreateStatefulSet, 2)";"(syncTidbClusterStatus, 2)";"(Failover, 1)";"(UpdateStatefulSet, 4)";"(syncTiCDCStatus, 2)";"(syncConfigMap, 2)";"(Scale, 3)";"(Upgrade, 3)";} "(syncPumpStatefulSetForTidbCluster, 1)" -> {"(CreateStatefulSet, 2)";"(UpdateStatefulSet, 4)";"(syncConfigMap, 2)";} "(tryToDeleteAFailureMember, 1)" -> {} "(syncTiDBInitConfigMap, 1)" -> {} "(syncPDServiceForTidbCluster, 1)" -> {"(CreateService, 2)";"(UpdateService, 2)";} "(UpdateMetaInfo, 3)" -> {"(Update, 1)";} "(scaleOne, 2)" -> {} "(upgradePDPod, 3)" -> {} "(CreateStatefulSet, 2)" -> {"(Create, 1)";} "(syncTiDBHeadlessServiceForTidbCluster, 1)" -> {"(CreateService, 2)";"(UpdateService, 2)";} } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 7236eaba7d5924f47ae0cd3b0bf928886ca8cfc3
1.0
pingcap/tidb-operator: pkg/backup/backupschedule/backup_schedule_manager_test.go; 12 LoC - Found a possible issue in [pingcap/tidb-operator](https://www.github.com/pingcap/tidb-operator) at [pkg/backup/backupschedule/backup_schedule_manager_test.go](https://github.com/pingcap/tidb-operator/blob/7236eaba7d5924f47ae0cd3b0bf928886ca8cfc3/pkg/backup/backupschedule/backup_schedule_manager_test.go#L97-L108) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call at line 105 passes reference to bk to third-party code [Click here to see the code in its original context.](https://github.com/pingcap/tidb-operator/blob/7236eaba7d5924f47ae0cd3b0bf928886ca8cfc3/pkg/backup/backupschedule/backup_schedule_manager_test.go#L97-L108) <details> <summary>Click here to show the 12 line(s) of Go which triggered the analyzer.</summary> ```go for _, bk := range bks.Items { changed := v1alpha1.UpdateBackupCondition(&bk.Status, &v1alpha1.BackupCondition{ Type: v1alpha1.BackupComplete, Status: v1.ConditionTrue, }) if changed { bk.CreationTimestamp = metav1.Time{Time: m.now()} t.Log("complete backup: ", bk.Name) helper.updateBackup(&bk) } g.Expect(err).Should(BeNil()) } ``` </details> <details> <summary>Click here to show extra information the analyzer produced.</summary> ``` No path was found through the callgraph that could lead to a function which writes a pointer argument. The following graphviz dot graph describes paths through the callgraph that could lead to a function which passes a pointer to third-party code: digraph G { "(Update, 1)" -> {} "(Scale, 3)" -> {"(ScaleIn, 3)";"(scaleOne, 2)";"(ScaleOut, 3)";} "(syncTiDBService, 1)" -> {"(CreateService, 2)";"(UpdateService, 2)";} "(syncTiKVConfigMap, 2)" -> {"(CreateOrUpdateConfigMap, 2)";} "(gracefulUpgrade, 3)" -> {"(upgradePDPod, 3)";} "(syncTiDBStatefulSetForTidbCluster, 1)" -> {"(UpdateStatefulSet, 4)";"(syncTiDBConfigMap, 2)";"(Upgrade, 3)";"(CreateStatefulSet, 2)";"(syncTidbClusterStatus, 2)";"(Failover, 1)";} "(setStoreLabelsForTiKV, 1)" -> {} "(syncRestoreJob, 1)" -> {"(CreateJob, 2)";"(ensureRestorePVCExist, 1)";"(Update, 2)";} "(Upgrade, 3)" -> {"(gracefulUpgrade, 3)";} "(resetReplicas, 2)" -> {} "(updateBackup, 1)" -> {"(Sync, 1)";} "(CreateService, 2)" -> {"(Create, 1)";} "(ScaleIn, 3)" -> {"(setReplicasAndDeleteSlots, 3)";"(resetReplicas, 2)";"(scaleOne, 2)";} "(syncConfigMap, 2)" -> {"(CreateOrUpdateConfigMap, 2)";} "(ensureRestorePVCExist, 1)" -> {"(CreatePVC, 2)";} "(syncServiceForTidbCluster, 2)" -> {"(CreateService, 2)";"(UpdateService, 2)";} "(syncBackupJob, 1)" -> {"(CreateJob, 2)";"(ensureBackupPVCExist, 1)";"(Update, 2)";} "(CreateJob, 2)" -> {"(Create, 1)";} "(syncTiDBConfigMap, 2)" -> {"(CreateOrUpdateConfigMap, 2)";} "(Update, 2)" -> {"(Update, 1)";} "(syncPDHeadlessServiceForTidbCluster, 1)" -> {"(CreateService, 2)";"(UpdateService, 2)";} "(syncTiCDCStatus, 2)" -> {} "(setReplicasAndDeleteSlots, 3)" -> {} "(Sync, 1)" -> {"(syncCDCHeadlessService, 1)";"(syncTiDBHeadlessServiceForTidbCluster, 1)";"(syncPDServiceForTidbCluster, 1)";"(syncStatefulSet, 1)";"(UpdateMetaInfo, 2)";"(syncTiDBStatefulSetForTidbCluster, 1)";"(syncPDHeadlessServiceForTidbCluster, 1)";"(syncTiDBService, 1)";"(syncStatefulSetForTidbCluster, 1)";"(syncTiDBInitJob, 1)";"(syncPDStatefulSetForTidbCluster, 1)";"(syncRestoreJob, 1)";"(UpdateMetaInfo, 3)";"(syncTiDBInitConfigMap, 1)";"(syncServiceForTidbCluster, 2)";"(syncHeadlessService, 1)";"(syncPumpStatefulSetForTidbCluster, 1)";"(Clean, 1)";"(syncBackupJob, 1)";} "(Failover, 1)" -> {"(isPDInQuorum, 1)";"(tryToDeleteAFailureMember, 1)";"(tryToMarkAPeerAsFailure, 1)";} "(syncCDCHeadlessService, 1)" -> {"(CreateService, 2)";"(UpdateService, 2)";} "(syncHeadlessService, 1)" -> {"(UpdateService, 2)";"(CreateService, 2)";} "(syncStatefulSetForTidbCluster, 1)" -> {"(UpdateStatefulSet, 4)";"(syncTiKVConfigMap, 2)";"(Scale, 3)";"(Upgrade, 3)";"(CreateStatefulSet, 2)";"(syncTidbClusterStatus, 2)";"(Failover, 1)";"(setStoreLabelsForTiKV, 1)";} "(syncTiDBInitJob, 1)" -> {} "(syncPDStatefulSetForTidbCluster, 1)" -> {"(Scale, 3)";"(Upgrade, 3)";"(CreateStatefulSet, 2)";"(syncTidbClusterStatus, 2)";"(Failover, 1)";"(UpdateStatefulSet, 4)";"(syncPDConfigMap, 2)";} "(syncTidbClusterStatus, 2)" -> {} "(CreatePVC, 2)" -> {} "(isPDInQuorum, 1)" -> {} "(ensureBackupPVCExist, 1)" -> {"(CreatePVC, 2)";} "(Clean, 1)" -> {"(CreateJob, 2)";"(Update, 2)";} "(Create, 1)" -> {} "(ScaleOut, 3)" -> {"(setReplicasAndDeleteSlots, 3)";"(resetReplicas, 2)";"(scaleOne, 2)";} "(UpdateStatefulSet, 4)" -> {} "(UpdateMetaInfo, 2)" -> {"(Update, 1)";} "(tryToMarkAPeerAsFailure, 1)" -> {} "(syncPDConfigMap, 2)" -> {"(CreateOrUpdateConfigMap, 2)";} "(UpdateService, 2)" -> {"(Update, 1)";} "(CreateOrUpdateConfigMap, 2)" -> {} "(syncStatefulSet, 1)" -> {"(CreateStatefulSet, 2)";"(syncTidbClusterStatus, 2)";"(Failover, 1)";"(UpdateStatefulSet, 4)";"(syncTiCDCStatus, 2)";"(syncConfigMap, 2)";"(Scale, 3)";"(Upgrade, 3)";} "(syncPumpStatefulSetForTidbCluster, 1)" -> {"(CreateStatefulSet, 2)";"(UpdateStatefulSet, 4)";"(syncConfigMap, 2)";} "(tryToDeleteAFailureMember, 1)" -> {} "(syncTiDBInitConfigMap, 1)" -> {} "(syncPDServiceForTidbCluster, 1)" -> {"(CreateService, 2)";"(UpdateService, 2)";} "(UpdateMetaInfo, 3)" -> {"(Update, 1)";} "(scaleOne, 2)" -> {} "(upgradePDPod, 3)" -> {} "(CreateStatefulSet, 2)" -> {"(Create, 1)";} "(syncTiDBHeadlessServiceForTidbCluster, 1)" -> {"(CreateService, 2)";"(UpdateService, 2)";} } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 7236eaba7d5924f47ae0cd3b0bf928886ca8cfc3
non_process
pingcap tidb operator pkg backup backupschedule backup schedule manager test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call at line passes reference to bk to third party code click here to show the line s of go which triggered the analyzer go for bk range bks items changed updatebackupcondition bk status backupcondition type backupcomplete status conditiontrue if changed bk creationtimestamp time time m now t log complete backup bk name helper updatebackup bk g expect err should benil click here to show extra information the analyzer produced no path was found through the callgraph that could lead to a function which writes a pointer argument the following graphviz dot graph describes paths through the callgraph that could lead to a function which passes a pointer to third party code digraph g update scale scalein scaleone scaleout synctidbservice createservice updateservice synctikvconfigmap createorupdateconfigmap gracefulupgrade upgradepdpod synctidbstatefulsetfortidbcluster updatestatefulset synctidbconfigmap upgrade createstatefulset synctidbclusterstatus failover setstorelabelsfortikv syncrestorejob createjob ensurerestorepvcexist update upgrade gracefulupgrade resetreplicas updatebackup sync createservice create scalein setreplicasanddeleteslots resetreplicas scaleone syncconfigmap createorupdateconfigmap ensurerestorepvcexist createpvc syncservicefortidbcluster createservice updateservice syncbackupjob createjob ensurebackuppvcexist update createjob create synctidbconfigmap createorupdateconfigmap update update syncpdheadlessservicefortidbcluster createservice updateservice syncticdcstatus setreplicasanddeleteslots sync synccdcheadlessservice synctidbheadlessservicefortidbcluster syncpdservicefortidbcluster syncstatefulset updatemetainfo synctidbstatefulsetfortidbcluster syncpdheadlessservicefortidbcluster synctidbservice syncstatefulsetfortidbcluster synctidbinitjob syncpdstatefulsetfortidbcluster syncrestorejob updatemetainfo synctidbinitconfigmap syncservicefortidbcluster syncheadlessservice syncpumpstatefulsetfortidbcluster clean syncbackupjob failover ispdinquorum trytodeleteafailuremember trytomarkapeerasfailure synccdcheadlessservice createservice updateservice syncheadlessservice updateservice createservice syncstatefulsetfortidbcluster updatestatefulset synctikvconfigmap scale upgrade createstatefulset synctidbclusterstatus failover setstorelabelsfortikv synctidbinitjob syncpdstatefulsetfortidbcluster scale upgrade createstatefulset synctidbclusterstatus failover updatestatefulset syncpdconfigmap synctidbclusterstatus createpvc ispdinquorum ensurebackuppvcexist createpvc clean createjob update create scaleout setreplicasanddeleteslots resetreplicas scaleone updatestatefulset updatemetainfo update trytomarkapeerasfailure syncpdconfigmap createorupdateconfigmap updateservice update createorupdateconfigmap syncstatefulset createstatefulset synctidbclusterstatus failover updatestatefulset syncticdcstatus syncconfigmap scale upgrade syncpumpstatefulsetfortidbcluster createstatefulset updatestatefulset syncconfigmap trytodeleteafailuremember synctidbinitconfigmap syncpdservicefortidbcluster createservice updateservice updatemetainfo update scaleone upgradepdpod createstatefulset create synctidbheadlessservicefortidbcluster createservice updateservice leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
369,758
10,917,335,745
IssuesEvent
2019-11-21 14:59:31
tensorwerk/hangar-py
https://api.github.com/repos/tensorwerk/hangar-py
closed
[BUG REPORT] `repo.checkout` accepts `commit` argument when `write=True`
Bug: Priority 2 PR In Progress
**Describe the bug** It's a UI issue on our `checkout` API. `commit` and `write=True` must be mutually exclusive since we are not returning write checkout on a particular commit. If we don't raise an error or warning when the user specifies a commit, user expectation might not be met **Severity** <!--- fill in the space between `[ ]` with and `x` (ie. `[x]`) ---> Select an option: - [ ] Data Corruption / Loss of Any Kind - [x] Unexpected Behavior, Exceptions or Error Thrown - [ ] Performance Bottleneck **To Reproduce** ```python co = repo.checkout(write=True, commit='an_old_commit') assert co.commit_hash == `an_old_commit` ``` **Expected behavior** An exception could be raised when `commit` is passed as an argument while `write=True`
1.0
[BUG REPORT] `repo.checkout` accepts `commit` argument when `write=True` - **Describe the bug** It's a UI issue on our `checkout` API. `commit` and `write=True` must be mutually exclusive since we are not returning write checkout on a particular commit. If we don't raise an error or warning when the user specifies a commit, user expectation might not be met **Severity** <!--- fill in the space between `[ ]` with and `x` (ie. `[x]`) ---> Select an option: - [ ] Data Corruption / Loss of Any Kind - [x] Unexpected Behavior, Exceptions or Error Thrown - [ ] Performance Bottleneck **To Reproduce** ```python co = repo.checkout(write=True, commit='an_old_commit') assert co.commit_hash == `an_old_commit` ``` **Expected behavior** An exception could be raised when `commit` is passed as an argument while `write=True`
non_process
repo checkout accepts commit argument when write true describe the bug it s a ui issue on our checkout api commit and write true must be mutually exclusive since we are not returning write checkout on a particular commit if we don t raise an error or warning when the user specifies a commit user expectation might not be met severity select an option data corruption loss of any kind unexpected behavior exceptions or error thrown performance bottleneck to reproduce python co repo checkout write true commit an old commit assert co commit hash an old commit expected behavior an exception could be raised when commit is passed as an argument while write true
0
6,109
8,536,825,629
IssuesEvent
2018-11-05 16:20:26
MichaelClerx/myokit
https://api.github.com/repos/MichaelClerx/myokit
opened
CVODE detection fails on Appveyor, but simulation does work
bug compatibility
Output from Appveyor for Python2.7: https://ci.appveyor.com/project/MichaelClerx/myokit/builds/20057698/job/kxknixceof7qmfrk ``` %PYTHON%\python.exe -m myokit system WARNING:myokit:Using development version of Myokit. This may contain untested features and bugs. Please see http://myokit.org for the latest stable release. WARNING:myokit._config:Unable to auto-detect Sundials version. WARNING:myokit._config:Sundials version not set in myokit.ini and version auto-detection failed. == System information == Myokit: 1.27.8.dev Python: 2.7.15 (v2.7.15:ca079a3ea3, Apr 30 2018, 16:30:26) [MSC v.1500 64 bit (AMD64)] OS: Windows (win32, nt) == Python requirements == NumPy: 1.15.4 SciPy: 1.1.0 Matplotlib: 2.2.3 ConfigParser: OK Setuptools: 40.5.0 == Python extras == SymPy: OK CMAES: Not found MoviePy: Not found == GUI == PyQt5: Not found PyQt4: Not found PySide: Not found PySide2: Not found == Development tools == Sphinx: Not found Flake8: Not found == Simulation tools == Compiler: Microsoft Visual Studio 9.0 (2008) - 150030729 Sundials: Not found OpenCL: No OpenCL support detected. ```
True
CVODE detection fails on Appveyor, but simulation does work - Output from Appveyor for Python2.7: https://ci.appveyor.com/project/MichaelClerx/myokit/builds/20057698/job/kxknixceof7qmfrk ``` %PYTHON%\python.exe -m myokit system WARNING:myokit:Using development version of Myokit. This may contain untested features and bugs. Please see http://myokit.org for the latest stable release. WARNING:myokit._config:Unable to auto-detect Sundials version. WARNING:myokit._config:Sundials version not set in myokit.ini and version auto-detection failed. == System information == Myokit: 1.27.8.dev Python: 2.7.15 (v2.7.15:ca079a3ea3, Apr 30 2018, 16:30:26) [MSC v.1500 64 bit (AMD64)] OS: Windows (win32, nt) == Python requirements == NumPy: 1.15.4 SciPy: 1.1.0 Matplotlib: 2.2.3 ConfigParser: OK Setuptools: 40.5.0 == Python extras == SymPy: OK CMAES: Not found MoviePy: Not found == GUI == PyQt5: Not found PyQt4: Not found PySide: Not found PySide2: Not found == Development tools == Sphinx: Not found Flake8: Not found == Simulation tools == Compiler: Microsoft Visual Studio 9.0 (2008) - 150030729 Sundials: Not found OpenCL: No OpenCL support detected. ```
non_process
cvode detection fails on appveyor but simulation does work output from appveyor for python python exe m myokit system warning myokit using development version of myokit this may contain untested features and bugs please see for the latest stable release warning myokit config unable to auto detect sundials version warning myokit config sundials version not set in myokit ini and version auto detection failed system information myokit dev python apr os windows nt python requirements numpy scipy matplotlib configparser ok setuptools python extras sympy ok cmaes not found moviepy not found gui not found not found pyside not found not found development tools sphinx not found not found simulation tools compiler microsoft visual studio sundials not found opencl no opencl support detected
0
11,674
14,532,909,745
IssuesEvent
2020-12-14 23:20:49
pacificclimate/quail
https://api.github.com/repos/pacificclimate/quail
closed
Monthly Minimum of Daily Minimum Temperature
process
## Description This function takes a climdexInput object as input and computes the monthly or annual minimum of daily minimum temperature. ## Function to wrap [`climdex.tnn`](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L919)
1.0
Monthly Minimum of Daily Minimum Temperature - ## Description This function takes a climdexInput object as input and computes the monthly or annual minimum of daily minimum temperature. ## Function to wrap [`climdex.tnn`](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L919)
process
monthly minimum of daily minimum temperature description this function takes a climdexinput object as input and computes the monthly or annual minimum of daily minimum temperature function to wrap
1
6,845
9,986,767,858
IssuesEvent
2019-07-10 19:58:46
leighmurdick/CRT_S-Actuarial
https://api.github.com/repos/leighmurdick/CRT_S-Actuarial
opened
Complete migration of surveillance tasks from Daragh to Analytics/Dev team
Analytics/Dev Team Data/Processes QBR/Quarterly Surveillance
Goal: no reliance on Daragh (other than questions) for Q3 surveillance run
1.0
Complete migration of surveillance tasks from Daragh to Analytics/Dev team - Goal: no reliance on Daragh (other than questions) for Q3 surveillance run
process
complete migration of surveillance tasks from daragh to analytics dev team goal no reliance on daragh other than questions for surveillance run
1
7,669
10,758,646,410
IssuesEvent
2019-10-31 15:19:00
linnovate/root
https://api.github.com/repos/linnovate/root
closed
setting cannot delete document
2.0.8 Internal Test Process bug
go to setting open new folder or offices upload a file try to delete result: the file not delete
1.0
setting cannot delete document - go to setting open new folder or offices upload a file try to delete result: the file not delete
process
setting cannot delete document go to setting open new folder or offices upload a file try to delete result the file not delete
1
26,439
2,684,495,037
IssuesEvent
2015-03-29 01:39:01
gtcasl/gpuocelot
https://api.github.com/repos/gtcasl/gpuocelot
closed
Error during make install
bug imported Priority-Medium wontfix
_From [kecelif...@gmail.com](https://code.google.com/u/110071228144016332643/) on March 08, 2011 13:11:59_ What steps will reproduce the problem? 1. Download the ocelot-1.3.967 package from https://code.google.com/p/gpuocelot/downloads/detail?name=ocelot-1.3.967.tar.bz2&can=2&q= 2. Run ./configure; make; sudo make install What is the expected output? What do you see instead? Below is the error I get. Looks like TestLLVMKernels.h file is included in the list. /usr/bin/install -c -m 644 ocelot/executive/test/TestGPUKernel.h ocelot/executive/test/TestLLVMKernels.h ocelot/executive/test/TestEmulator.h ocelot/executive/test/TestLLVMKernels.h ocelot/executive/test/sequence.ptx ocelot/executive/test/kernels.ptx '/usr/local/include/ocelot/executive/test' /usr/bin/install: will not overwrite just-created `/usr/local/include/ocelot/executive/test/TestLLVMKernels.h' with `ocelot/executive/test/TestLLVMKernels.h' make[2]: *** [install-nobase_includeHEADERS] Error 1 make[2]: Leaving directory `/home/animus/Work/simulators/ocelot-1.3.967' make[1]: *** [install-am] Error 2 make[1]: Leaving directory `/home/animus/Work/simulators/ocelot-1.3.967' make: *** [install] Error 2 What version of the product are you using? On what operating system? I am using ocelot-1.3.967 on Ubuntu 10.04. Make version is 3.81. Install version is 8.5. Please provide any additional information below. This doesn't seem to be fatal error for NVIDIA GPU emulation. The libraries are installed and they work. I just am not sure if this error has any effect on the working of the simulator. _Original issue: http://code.google.com/p/gpuocelot/issues/detail?id=48_
1.0
Error during make install - _From [kecelif...@gmail.com](https://code.google.com/u/110071228144016332643/) on March 08, 2011 13:11:59_ What steps will reproduce the problem? 1. Download the ocelot-1.3.967 package from https://code.google.com/p/gpuocelot/downloads/detail?name=ocelot-1.3.967.tar.bz2&can=2&q= 2. Run ./configure; make; sudo make install What is the expected output? What do you see instead? Below is the error I get. Looks like TestLLVMKernels.h file is included in the list. /usr/bin/install -c -m 644 ocelot/executive/test/TestGPUKernel.h ocelot/executive/test/TestLLVMKernels.h ocelot/executive/test/TestEmulator.h ocelot/executive/test/TestLLVMKernels.h ocelot/executive/test/sequence.ptx ocelot/executive/test/kernels.ptx '/usr/local/include/ocelot/executive/test' /usr/bin/install: will not overwrite just-created `/usr/local/include/ocelot/executive/test/TestLLVMKernels.h' with `ocelot/executive/test/TestLLVMKernels.h' make[2]: *** [install-nobase_includeHEADERS] Error 1 make[2]: Leaving directory `/home/animus/Work/simulators/ocelot-1.3.967' make[1]: *** [install-am] Error 2 make[1]: Leaving directory `/home/animus/Work/simulators/ocelot-1.3.967' make: *** [install] Error 2 What version of the product are you using? On what operating system? I am using ocelot-1.3.967 on Ubuntu 10.04. Make version is 3.81. Install version is 8.5. Please provide any additional information below. This doesn't seem to be fatal error for NVIDIA GPU emulation. The libraries are installed and they work. I just am not sure if this error has any effect on the working of the simulator. _Original issue: http://code.google.com/p/gpuocelot/issues/detail?id=48_
non_process
error during make install from on march what steps will reproduce the problem download the ocelot package from run configure make sudo make install what is the expected output what do you see instead below is the error i get looks like testllvmkernels h file is included in the list usr bin install c m ocelot executive test testgpukernel h ocelot executive test testllvmkernels h ocelot executive test testemulator h ocelot executive test testllvmkernels h ocelot executive test sequence ptx ocelot executive test kernels ptx usr local include ocelot executive test usr bin install will not overwrite just created usr local include ocelot executive test testllvmkernels h with ocelot executive test testllvmkernels h make error make leaving directory home animus work simulators ocelot make error make leaving directory home animus work simulators ocelot make error what version of the product are you using on what operating system i am using ocelot on ubuntu make version is install version is please provide any additional information below this doesn t seem to be fatal error for nvidia gpu emulation the libraries are installed and they work i just am not sure if this error has any effect on the working of the simulator original issue
0
343,244
10,326,881,488
IssuesEvent
2019-09-02 04:28:55
cuappdev/acai-ios
https://api.github.com/repos/cuappdev/acai-ios
opened
Fix bug in TextTableViewCell
Priority: Medium Refactor
Labeled: `// TODO: fix bug -- if set backgroundColor to clear and switch between signup and login, then label keeps generating on top of itself (text gets thicker)` in code base
1.0
Fix bug in TextTableViewCell - Labeled: `// TODO: fix bug -- if set backgroundColor to clear and switch between signup and login, then label keeps generating on top of itself (text gets thicker)` in code base
non_process
fix bug in texttableviewcell labeled todo fix bug if set backgroundcolor to clear and switch between signup and login then label keeps generating on top of itself text gets thicker in code base
0
15,021
18,735,733,457
IssuesEvent
2021-11-04 07:12:14
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
Analyze table failed due to tikv server timeout error during lighting import
type/bug sig/coprocessor severity/critical
## Bug Report <!-- Thanks for your bug report! Don't worry if you can't fill out all the sections. --> ### What version of TiKV are you using? / # ./tikv-server -V TiKV Release Version: 5.2.0 Edition: Community Git Commit Hash: 556783c314a9bfca36c818256182eeef364120d7 Git Commit Branch: heads/refs/tags/v5.2.0 UTC Build Time: 2021-08-26 05:47:23 Rust Version: rustc 1.56.0-nightly (2faabf579 2021-07-27) Enable Features: jemalloc mem-profiling portable sse protobuf-codec test-engines-rocksdb cloud-aws cloud-gcp Profile: dist_release ### What operating system and CPU are you using? TiKV: 8U, 61GB ### Steps to reproduce <!-- If possible, provide a recipe for reproducing the error. A complete runnable program is good. --> 1, Create DBaaS Cluster in Dev Env - T5.xlarge, TiKV x 3, 8U 61GB 2, Create CSV Import job, import data from: s3://data-platform-uw2/csv_export_default_10GiB ### What did you expect? Lightning import succeed ### What did happened? Lightning import failed. ย  | 2021-09-09 05:59:06 | [2021/09/09 05:59:06.942 +00:00] [ERROR] [restore.go:154] [-] [table=`aaa`.`bbb`] [status=analyzed] [error="analyze table failed: Error 9002: TiKV server timeout"] ย  | 2021-09-09 05:59:06 | [2021/09/09 05:59:06.942 +00:00] [ERROR] [restore.go:152] ["tables failed to be imported"] [count=1] ย  | 2021-09-09 05:59:06 | [2021/09/09 05:59:06.942 +00:00] [ERROR] [restore.go:431] ["the whole procedure failed"] [takeTime=1h52m11.9212808s] [error="analyze table failed: Error 9002: TiKV server timeout"] ย  | 2021-09-09 05:59:06 | [2021/09/09 05:59:06.942 +00:00] [ERROR] [restore.go:420] ["run failed"] [step=3] [error="analyze table failed: Error 9002: TiKV server timeout"] ย  | 2021-09-09 05:59:06 | Error: analyze table failed: Error 9002: TiKV server timeout ย  | 2021-09-09 05:59:06 | [2021/09/09 05:59:06.651 +00:00] [INFO] [restore.go:1038] ["switch to normal mode"] Log: https://grafana.oauth.dev.shared.aws.tidbcloud.com/explore?orgId=1&left=%5B%22now-3h%22,%22now%22,%22loki%22,%7B%22expr%22:%22%7Bcontainer%3D%5C%22tidb-lightning%5C%22,namespace%3D%5C%22tidb120007%5C%22%7D%22%7D%5D
1.0
Analyze table failed due to tikv server timeout error during lighting import - ## Bug Report <!-- Thanks for your bug report! Don't worry if you can't fill out all the sections. --> ### What version of TiKV are you using? / # ./tikv-server -V TiKV Release Version: 5.2.0 Edition: Community Git Commit Hash: 556783c314a9bfca36c818256182eeef364120d7 Git Commit Branch: heads/refs/tags/v5.2.0 UTC Build Time: 2021-08-26 05:47:23 Rust Version: rustc 1.56.0-nightly (2faabf579 2021-07-27) Enable Features: jemalloc mem-profiling portable sse protobuf-codec test-engines-rocksdb cloud-aws cloud-gcp Profile: dist_release ### What operating system and CPU are you using? TiKV: 8U, 61GB ### Steps to reproduce <!-- If possible, provide a recipe for reproducing the error. A complete runnable program is good. --> 1, Create DBaaS Cluster in Dev Env - T5.xlarge, TiKV x 3, 8U 61GB 2, Create CSV Import job, import data from: s3://data-platform-uw2/csv_export_default_10GiB ### What did you expect? Lightning import succeed ### What did happened? Lightning import failed. ย  | 2021-09-09 05:59:06 | [2021/09/09 05:59:06.942 +00:00] [ERROR] [restore.go:154] [-] [table=`aaa`.`bbb`] [status=analyzed] [error="analyze table failed: Error 9002: TiKV server timeout"] ย  | 2021-09-09 05:59:06 | [2021/09/09 05:59:06.942 +00:00] [ERROR] [restore.go:152] ["tables failed to be imported"] [count=1] ย  | 2021-09-09 05:59:06 | [2021/09/09 05:59:06.942 +00:00] [ERROR] [restore.go:431] ["the whole procedure failed"] [takeTime=1h52m11.9212808s] [error="analyze table failed: Error 9002: TiKV server timeout"] ย  | 2021-09-09 05:59:06 | [2021/09/09 05:59:06.942 +00:00] [ERROR] [restore.go:420] ["run failed"] [step=3] [error="analyze table failed: Error 9002: TiKV server timeout"] ย  | 2021-09-09 05:59:06 | Error: analyze table failed: Error 9002: TiKV server timeout ย  | 2021-09-09 05:59:06 | [2021/09/09 05:59:06.651 +00:00] [INFO] [restore.go:1038] ["switch to normal mode"] Log: https://grafana.oauth.dev.shared.aws.tidbcloud.com/explore?orgId=1&left=%5B%22now-3h%22,%22now%22,%22loki%22,%7B%22expr%22:%22%7Bcontainer%3D%5C%22tidb-lightning%5C%22,namespace%3D%5C%22tidb120007%5C%22%7D%22%7D%5D
process
analyze table failed due to tikv server timeout error during lighting import bug report what version of tikv are you using tikv server v tikv release version edition community git commit hash git commit branch heads refs tags utc build time rust version rustc nightly enable features jemalloc mem profiling portable sse protobuf codec test engines rocksdb cloud aws cloud gcp profile dist release what operating system and cpu are you using tikv steps to reproduce create dbaas cluster in dev env xlarge tikv x create csv import job import data from data platform csv export default what did you expect lightning import succeed what did happened lightning import failed ย  ย  ย  ย  ย  error analyze table failed error tikv server timeout ย  log
1
19,199
25,331,963,284
IssuesEvent
2022-11-18 13:57:50
Ultimate-Hosts-Blacklist/whitelist
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
opened
[FALSE-POSITIVE?]
whitelisting process
**Domains or links** <!-- Please list below any domains and links listed here which you believe are a false positive. --> 1. example.com 2. bkrenergy.ca 172.64.80.1 **More Information** <!-- How did you discover your web site or domain was listed here? --> 1. This must be a new addition, I stopped being able to access my thermostat through my app 2. I then found the IP of the company, found it in your list, tested by disabling the list and trying again and it worked **Have you requested removal from other sources?** <!-- Please include all relevant links to your existing removals / whitelistings. --> ... **Additional context** <!-- Add any other context about the problem here. --> Not sure how this was flagged originally, but just know it is a canadian energy company that helps manage heat pump vs furnace usage in homes. With this IP blocked the user can't use the phone app, at the very least, I'm not sure if it's stopping other comms in the background as well <!-- โ— We understand being listed on a list like this can be frustrating and embarrassing for many web site owners. The first step is to remain calm. The second step is to rest assured one of our maintainers will address your issue as soon as possible. Please make sure you have provided as much information as possible to help speed up the process. -->
1.0
[FALSE-POSITIVE?] - **Domains or links** <!-- Please list below any domains and links listed here which you believe are a false positive. --> 1. example.com 2. bkrenergy.ca 172.64.80.1 **More Information** <!-- How did you discover your web site or domain was listed here? --> 1. This must be a new addition, I stopped being able to access my thermostat through my app 2. I then found the IP of the company, found it in your list, tested by disabling the list and trying again and it worked **Have you requested removal from other sources?** <!-- Please include all relevant links to your existing removals / whitelistings. --> ... **Additional context** <!-- Add any other context about the problem here. --> Not sure how this was flagged originally, but just know it is a canadian energy company that helps manage heat pump vs furnace usage in homes. With this IP blocked the user can't use the phone app, at the very least, I'm not sure if it's stopping other comms in the background as well <!-- โ— We understand being listed on a list like this can be frustrating and embarrassing for many web site owners. The first step is to remain calm. The second step is to rest assured one of our maintainers will address your issue as soon as possible. Please make sure you have provided as much information as possible to help speed up the process. -->
process
domains or links example com bkrenergy ca more information this must be a new addition i stopped being able to access my thermostat through my app i then found the ip of the company found it in your list tested by disabling the list and trying again and it worked have you requested removal from other sources additional context not sure how this was flagged originally but just know it is a canadian energy company that helps manage heat pump vs furnace usage in homes with this ip blocked the user can t use the phone app at the very least i m not sure if it s stopping other comms in the background as well โ— we understand being listed on a list like this can be frustrating and embarrassing for many web site owners the first step is to remain calm the second step is to rest assured one of our maintainers will address your issue as soon as possible please make sure you have provided as much information as possible to help speed up the process
1
269,709
23,460,429,845
IssuesEvent
2022-08-16 12:41:52
wazuh/wazuh
https://api.github.com/repos/wazuh/wazuh
opened
Release 4.3.7 - Release Candidate 1 - WPK upgrade tests
module/upgrade/wpk type/test/manual
The following issue aims to run `upgrade WPK tests` for the current release candidate, report the results, and open new issues for any encountered errors. ## WPK upgrade tests information |Main RC issue|Version|Release candidate|Tag|Previous issue| |---|---|---|---|---| |https://github.com/wazuh/wazuh/issues/14562|4.3.7|RC1|[v4.3.7-rc1](https://github.com/wazuh/wazuh/tree/v4.3.7-rc1)|https://github.com/wazuh/wazuh/issues/14267| ## Test report procedure All individual test checks must be marked as: | | | |---------------------------------|--------------------------------------------| | Pass | The test ran successfully. | | Xfail | The test was expected to fail and it failed. It must be properly justified and reported in an issue. | | Skip | The test was not run. It must be properly justified and reported in an issue. | | Fail | The test failed. A new issue must be opened to evaluate and address the problem. | All test results must have one the following statuses: | | | |---------------------------------|--------------------------------------------| | :green_circle: | All checks passed. | | :red_circle: | There is at least one failed check. | | :yellow_circle: | There is at least one expected fail or skipped test and no failures. | Any failing test must be properly addressed with a new issue, detailing the error and the possible cause. It must be included in the `Fixes` section of the current release candidate main issue. Any expected fail or skipped test must have an issue justifying the reason. All auditors must validate the justification for an expected fail or skipped test. An extended report of the test results must be attached as a zip or txt. This report can be used by the auditors to dig deeper into any possible failures and details. ## Tests To evaluate this feature, it is necessary to test upgrading the agent and also the case when the upgrade fails (rollback). The `tree` command will be used to compare, before and after the upgrade/rollback process, and check that the presence, ownership and permissions of the files and directories are expected. Wazuh versions to test (Upgrade to the current agent version): ### Linux |OS|Version|Status|Upgrade fail|Upgrade OK| |----|-----|------|---------------|------------------| |CentOS 6|3.6|:white_circle: skipped | | | | |CentOS 6|3.7|:white_circle: skipped | | | | |CentOS 6|3.13.3|:white_circle: skipped | | | | |CentOS 6|4.0.4| :white_circle: skipped| | | | |CentOS 6|4.1.5| :white_circle: skipped| | | | |CentOS 6|4.2.7|:yellow_circle: pending | | | | |CentOS 6|4.3.6|:yellow_circle: pending | | | | |CentOS 8|3.6| :white_circle: skipped| | | | |CentOS 8|3.7| :white_circle: skipped| | | | |CentOS 8|3.13.3| :white_circle: skipped| | | | |CentOS 8|4.0.4| :white_circle: skipped| | | | |CentOS 8|4.1.5| :white_circle: skipped| | | | |CentOS 8|4.2.7|:yellow_circle: pending | | | | |CentOS 8|4.3.6|:yellow_circle: pending | | | | |RHEL 9|3.6| :white_circle: skipped| | | | |RHEL 9|3.7| :white_circle: skipped| | | | |RHEL 9|3.13.3| :white_circle: skipped| | | | |RHEL 9|4.0.4| :white_circle: skipped| | | | |RHEL 9|4.1.5| :white_circle: skipped| | | | |RHEL 9|4.2.7|:yellow_circle: pending | | | | |RHEL 9|4.3.6|:yellow_circle: pending | | | | |Ubuntu 20|3.6|:white_circle: skipped | | | | |Ubuntu 20|3.7| :white_circle: skipped| | | | |Ubuntu 20|3.13.3|:white_circle: skipped| | | | |Ubuntu 20|4.0.4|:white_circle: skipped| | | | |Ubuntu 20|4.1.5| :white_circle: skipped ||| | |Ubuntu 20|4.2.7| :white_circle: skipped||| | |Ubuntu 20|4.3 |:white_circle: skipped| | | |openSUSE Tumbleweed|3.6| :white_circle: skipped| | | | |openSUSE Tumbleweed|3.7|:white_circle: skipped | | | | |openSUSE Tumbleweed|3.13.3|:white_circle: skipped| || | |openSUSE Tumbleweed|4.0.4|:white_circle: skipped| | || |openSUSE Tumbleweed|4.1.5|:white_circle: skipped| | || |openSUSE Tumbleweed|4.2.7|:white_circle: skipped| || | |openSUSE Tumbleweed|4.3|:white_circle: skipped | | |Amazon Linux 2|3.6|:white_circle: skipped | | | | |Amazon Linux 2|3.7|:white_circle: skipped | | | | |Amazon Linux 2|3.13.3|:white_circle: skipped| || | |Amazon Linux 2|4.0.4|:white_circle: skipped| | || |Amazon Linux 2|4.1.5|:white_circle: skipped| | || |Amazon Linux 2|4.2.7|:white_circle: skipped| || | |Amazon Linux 2|4.3|:white_circle: skipped | | ### Windows |OS|Version|Status|Upgrade fail|Upgrade OK| |----|-----|------|---------------|------------------| |Server 2008|3.6|:white_circle: skipped | | | | |Server 2008|3.7|:white_circle: skipped | | | | |Server 2008|3.13.3|:white_circle: skipped| || | |Server 2008|4.0.4|:white_circle: skipped| | || |Server 2008|4.1.5|:white_circle: skipped| | || |Server 2008|4.2.7|:white_circle: skipped| || | |Server 2008|4.3| :white_circle: skipped| | |Server 2012 R2|3.6| :white_circle: skipped| | | | |Server 2012 R2|3.7|:white_circle: skipped | | | | |Server 2012 R2|3.13.3|:white_circle: skipped| || | |Server 2012 R2|4.0.4|:white_circle: skipped| | || |Server 2012 R2|4.1.5|:white_circle: skipped| | || |Server 2012 R2|4.2.7|:white_circle: skipped| || | |Server 2012 R2|4.3|:white_circle: skipped | | |Server 2016|3.6|:white_circle: skipped | | | | |Server 2016|3.7| :white_circle: skipped| | | | |Server 2016|3.13.3|:white_circle: skipped| || | |Server 2016|4.0.4|:white_circle: skipped| | || |Server 2016|4.1.5|:white_circle: skipped| | || |Server 2016|4.2.7|:white_circle: skipped| || | |Server 2016|4.3| :white_circle: skipped| | |Server 2019|3.6|:white_circle: skipped | | | | |Server 2019|3.7|:white_circle: skipped | | | | |Server 2019|3.13.3|:white_circle: skipped| || | |Server 2019|4.0.4|:white_circle: skipped| | || |Server 2019|4.1.5|:white_circle: skipped| | || |Server 2019|4.2.7|:white_circle: skipped| || | |Server 2019|4.3| :white_circle: skipped| | |Windows 10|3.6| :white_circle: skipped| | | | |Windows 10|3.7| :white_circle: skipped| | | | |Windows 10|3.13.3|:white_circle: skipped| || | |Windows 10|4.0.4|:white_circle: skipped| | || |Windows 10|4.1.5|:white_circle: skipped| | || |Windows 10|4.2.7|:white_circle: skipped| || | |Windows 10|4.3| :white_circle: skipped| | |Server 2022|3.6| :white_circle: skipped| | | | |Server 2022|3.7| :white_circle: skipped| | | | |Server 2022|3.13.3|:white_circle: skipped| || | |Server 2022|4.0.4|:white_circle: skipped| | || |Server 2022|4.1.5|:white_circle: skipped| | || |Server 2022|4.2.7|:white_circle: skipped| || | |Server 2022|4.3|:white_circle: skipped | | ### macOS |OS|Version|Status|Upgrade fail|Upgrade OK| |----|-----------|--------|-------|------------------| | Sierra |4.3.0| :white_circle: skipped|| | | | | Sierra |4.3.4|:white_circle: skipped || | | | | Catalina |4.3.0| :white_circle: skipped|| | | | | Catalina |4.3.4|:white_circle: skipped || | | | | Big Sur |4.3.0| :white_circle: skipped|| | | | | Big Sur |4.3.4| :white_circle: skipped|| | | | | Monterey |4.3.0| :white_circle: skipped|| | | | | Monterey |4.3.4| :white_circle: skipped|| | | | <!-- For each operating system and version, check the following points and add a comment for each OS tested. ## Linux: ### UPGRADE FAIL - [ ] The wazuh home backup is restored correctly (no traces of the installation, but only the `.tar.gz` backup and the logs). - [ ] The permissions and owners of the following directories did NOT change: - `/` - `/var` - `/usr`, `/usr/lib/`, `/usr/lib/systemd/`, `/usr/lib/systemd/system/` - `/etc`, `/etc/systemd/`, `/etc/systemd/system/`, `/etc/rc.d`, `/etc/initd.d/`, `/etc/initd.d/rc.d/` - [ ] Wazuh service runs wazuh-control (`systemctl cat wazuh-agent.service`) - [ ] Wazuh service runs ossec-control (`systemctl cat wazuh-agent.service`) - [ ] The service was enabled (`systemctl is-enabled wazuh-agent.service`) - [ ] Init file runs wazuh-control (`cat /etc/rc.d/init.d/wazuh-agent`) - [ ] Init file runs ossec-control (`cat /etc/rc.d/init.d/wazuh-agent`) - [ ] Wazuh as service is enabled `chkconfig --list` - [ ] Wazuh starts and connects when the backup is restored (`cat /var/ossec/var/run/ossec-agentd.state`) - [ ] Wazuh starts and connects automatically when the system is rebooted. - [ ] Restore SELinux policies (`semodule -l | grep -i wazuh`) (DISABLED) ### UPGRADE OK - [ ] Upgrade is performed successfully (agent connects to the manager after upgrading) - [ ] Service starts automatically after rebooting - [ ] Agent connects to the manager after rebooting ## Windows: ### UPGRADE FAIL - [ ] Wazuh-Agent folder tree: No files are lost after the rollback. The logs of the failed upgrade (`ossec.log`) are kept. - [ ] After the rollback the agent connects to the manager - [ ] After reboot, the Wazuh-Agent starts and connects to the manager. - [ ] The correct Wazuh-Agent version is shown in the list of Windows' `programs and features`. - [ ] A new version of Wazuh-Agent can be manually installed via MSI after the rollback process. ### UPGRADE OK - [ ] Message `Upgrade finished successfully.` is shown in `upgrade.log` file. - [ ] Wazuh service is started and the agent is connected to the manager. - [ ] The version shown in the control panel is 4.3 ## MacOS: ### UPGRADE FAIL - [ ] Wazuh-Agent folder tree: No files are lost after the rollback. The logs of the failed upgrade (`ossec.log`) are kept. - [ ] After the rollback the agent connects to the manager - [ ] After reboot, the Wazuh-Agent starts and connects to the manager. ### UPGRADE OK - [ ] Message `Upgrade finished successfully.` is shown in `upgrade.log` file. - [ ] Wazuh service is started and the agent is connected to the manager. --> ## Auditors validation The definition of done for this one is the validation of the conclusions and the test results from all auditors. All checks from below must be accepted in order to close this issue. - [ ] @vikman90
1.0
Release 4.3.7 - Release Candidate 1 - WPK upgrade tests - The following issue aims to run `upgrade WPK tests` for the current release candidate, report the results, and open new issues for any encountered errors. ## WPK upgrade tests information |Main RC issue|Version|Release candidate|Tag|Previous issue| |---|---|---|---|---| |https://github.com/wazuh/wazuh/issues/14562|4.3.7|RC1|[v4.3.7-rc1](https://github.com/wazuh/wazuh/tree/v4.3.7-rc1)|https://github.com/wazuh/wazuh/issues/14267| ## Test report procedure All individual test checks must be marked as: | | | |---------------------------------|--------------------------------------------| | Pass | The test ran successfully. | | Xfail | The test was expected to fail and it failed. It must be properly justified and reported in an issue. | | Skip | The test was not run. It must be properly justified and reported in an issue. | | Fail | The test failed. A new issue must be opened to evaluate and address the problem. | All test results must have one the following statuses: | | | |---------------------------------|--------------------------------------------| | :green_circle: | All checks passed. | | :red_circle: | There is at least one failed check. | | :yellow_circle: | There is at least one expected fail or skipped test and no failures. | Any failing test must be properly addressed with a new issue, detailing the error and the possible cause. It must be included in the `Fixes` section of the current release candidate main issue. Any expected fail or skipped test must have an issue justifying the reason. All auditors must validate the justification for an expected fail or skipped test. An extended report of the test results must be attached as a zip or txt. This report can be used by the auditors to dig deeper into any possible failures and details. ## Tests To evaluate this feature, it is necessary to test upgrading the agent and also the case when the upgrade fails (rollback). The `tree` command will be used to compare, before and after the upgrade/rollback process, and check that the presence, ownership and permissions of the files and directories are expected. Wazuh versions to test (Upgrade to the current agent version): ### Linux |OS|Version|Status|Upgrade fail|Upgrade OK| |----|-----|------|---------------|------------------| |CentOS 6|3.6|:white_circle: skipped | | | | |CentOS 6|3.7|:white_circle: skipped | | | | |CentOS 6|3.13.3|:white_circle: skipped | | | | |CentOS 6|4.0.4| :white_circle: skipped| | | | |CentOS 6|4.1.5| :white_circle: skipped| | | | |CentOS 6|4.2.7|:yellow_circle: pending | | | | |CentOS 6|4.3.6|:yellow_circle: pending | | | | |CentOS 8|3.6| :white_circle: skipped| | | | |CentOS 8|3.7| :white_circle: skipped| | | | |CentOS 8|3.13.3| :white_circle: skipped| | | | |CentOS 8|4.0.4| :white_circle: skipped| | | | |CentOS 8|4.1.5| :white_circle: skipped| | | | |CentOS 8|4.2.7|:yellow_circle: pending | | | | |CentOS 8|4.3.6|:yellow_circle: pending | | | | |RHEL 9|3.6| :white_circle: skipped| | | | |RHEL 9|3.7| :white_circle: skipped| | | | |RHEL 9|3.13.3| :white_circle: skipped| | | | |RHEL 9|4.0.4| :white_circle: skipped| | | | |RHEL 9|4.1.5| :white_circle: skipped| | | | |RHEL 9|4.2.7|:yellow_circle: pending | | | | |RHEL 9|4.3.6|:yellow_circle: pending | | | | |Ubuntu 20|3.6|:white_circle: skipped | | | | |Ubuntu 20|3.7| :white_circle: skipped| | | | |Ubuntu 20|3.13.3|:white_circle: skipped| | | | |Ubuntu 20|4.0.4|:white_circle: skipped| | | | |Ubuntu 20|4.1.5| :white_circle: skipped ||| | |Ubuntu 20|4.2.7| :white_circle: skipped||| | |Ubuntu 20|4.3 |:white_circle: skipped| | | |openSUSE Tumbleweed|3.6| :white_circle: skipped| | | | |openSUSE Tumbleweed|3.7|:white_circle: skipped | | | | |openSUSE Tumbleweed|3.13.3|:white_circle: skipped| || | |openSUSE Tumbleweed|4.0.4|:white_circle: skipped| | || |openSUSE Tumbleweed|4.1.5|:white_circle: skipped| | || |openSUSE Tumbleweed|4.2.7|:white_circle: skipped| || | |openSUSE Tumbleweed|4.3|:white_circle: skipped | | |Amazon Linux 2|3.6|:white_circle: skipped | | | | |Amazon Linux 2|3.7|:white_circle: skipped | | | | |Amazon Linux 2|3.13.3|:white_circle: skipped| || | |Amazon Linux 2|4.0.4|:white_circle: skipped| | || |Amazon Linux 2|4.1.5|:white_circle: skipped| | || |Amazon Linux 2|4.2.7|:white_circle: skipped| || | |Amazon Linux 2|4.3|:white_circle: skipped | | ### Windows |OS|Version|Status|Upgrade fail|Upgrade OK| |----|-----|------|---------------|------------------| |Server 2008|3.6|:white_circle: skipped | | | | |Server 2008|3.7|:white_circle: skipped | | | | |Server 2008|3.13.3|:white_circle: skipped| || | |Server 2008|4.0.4|:white_circle: skipped| | || |Server 2008|4.1.5|:white_circle: skipped| | || |Server 2008|4.2.7|:white_circle: skipped| || | |Server 2008|4.3| :white_circle: skipped| | |Server 2012 R2|3.6| :white_circle: skipped| | | | |Server 2012 R2|3.7|:white_circle: skipped | | | | |Server 2012 R2|3.13.3|:white_circle: skipped| || | |Server 2012 R2|4.0.4|:white_circle: skipped| | || |Server 2012 R2|4.1.5|:white_circle: skipped| | || |Server 2012 R2|4.2.7|:white_circle: skipped| || | |Server 2012 R2|4.3|:white_circle: skipped | | |Server 2016|3.6|:white_circle: skipped | | | | |Server 2016|3.7| :white_circle: skipped| | | | |Server 2016|3.13.3|:white_circle: skipped| || | |Server 2016|4.0.4|:white_circle: skipped| | || |Server 2016|4.1.5|:white_circle: skipped| | || |Server 2016|4.2.7|:white_circle: skipped| || | |Server 2016|4.3| :white_circle: skipped| | |Server 2019|3.6|:white_circle: skipped | | | | |Server 2019|3.7|:white_circle: skipped | | | | |Server 2019|3.13.3|:white_circle: skipped| || | |Server 2019|4.0.4|:white_circle: skipped| | || |Server 2019|4.1.5|:white_circle: skipped| | || |Server 2019|4.2.7|:white_circle: skipped| || | |Server 2019|4.3| :white_circle: skipped| | |Windows 10|3.6| :white_circle: skipped| | | | |Windows 10|3.7| :white_circle: skipped| | | | |Windows 10|3.13.3|:white_circle: skipped| || | |Windows 10|4.0.4|:white_circle: skipped| | || |Windows 10|4.1.5|:white_circle: skipped| | || |Windows 10|4.2.7|:white_circle: skipped| || | |Windows 10|4.3| :white_circle: skipped| | |Server 2022|3.6| :white_circle: skipped| | | | |Server 2022|3.7| :white_circle: skipped| | | | |Server 2022|3.13.3|:white_circle: skipped| || | |Server 2022|4.0.4|:white_circle: skipped| | || |Server 2022|4.1.5|:white_circle: skipped| | || |Server 2022|4.2.7|:white_circle: skipped| || | |Server 2022|4.3|:white_circle: skipped | | ### macOS |OS|Version|Status|Upgrade fail|Upgrade OK| |----|-----------|--------|-------|------------------| | Sierra |4.3.0| :white_circle: skipped|| | | | | Sierra |4.3.4|:white_circle: skipped || | | | | Catalina |4.3.0| :white_circle: skipped|| | | | | Catalina |4.3.4|:white_circle: skipped || | | | | Big Sur |4.3.0| :white_circle: skipped|| | | | | Big Sur |4.3.4| :white_circle: skipped|| | | | | Monterey |4.3.0| :white_circle: skipped|| | | | | Monterey |4.3.4| :white_circle: skipped|| | | | <!-- For each operating system and version, check the following points and add a comment for each OS tested. ## Linux: ### UPGRADE FAIL - [ ] The wazuh home backup is restored correctly (no traces of the installation, but only the `.tar.gz` backup and the logs). - [ ] The permissions and owners of the following directories did NOT change: - `/` - `/var` - `/usr`, `/usr/lib/`, `/usr/lib/systemd/`, `/usr/lib/systemd/system/` - `/etc`, `/etc/systemd/`, `/etc/systemd/system/`, `/etc/rc.d`, `/etc/initd.d/`, `/etc/initd.d/rc.d/` - [ ] Wazuh service runs wazuh-control (`systemctl cat wazuh-agent.service`) - [ ] Wazuh service runs ossec-control (`systemctl cat wazuh-agent.service`) - [ ] The service was enabled (`systemctl is-enabled wazuh-agent.service`) - [ ] Init file runs wazuh-control (`cat /etc/rc.d/init.d/wazuh-agent`) - [ ] Init file runs ossec-control (`cat /etc/rc.d/init.d/wazuh-agent`) - [ ] Wazuh as service is enabled `chkconfig --list` - [ ] Wazuh starts and connects when the backup is restored (`cat /var/ossec/var/run/ossec-agentd.state`) - [ ] Wazuh starts and connects automatically when the system is rebooted. - [ ] Restore SELinux policies (`semodule -l | grep -i wazuh`) (DISABLED) ### UPGRADE OK - [ ] Upgrade is performed successfully (agent connects to the manager after upgrading) - [ ] Service starts automatically after rebooting - [ ] Agent connects to the manager after rebooting ## Windows: ### UPGRADE FAIL - [ ] Wazuh-Agent folder tree: No files are lost after the rollback. The logs of the failed upgrade (`ossec.log`) are kept. - [ ] After the rollback the agent connects to the manager - [ ] After reboot, the Wazuh-Agent starts and connects to the manager. - [ ] The correct Wazuh-Agent version is shown in the list of Windows' `programs and features`. - [ ] A new version of Wazuh-Agent can be manually installed via MSI after the rollback process. ### UPGRADE OK - [ ] Message `Upgrade finished successfully.` is shown in `upgrade.log` file. - [ ] Wazuh service is started and the agent is connected to the manager. - [ ] The version shown in the control panel is 4.3 ## MacOS: ### UPGRADE FAIL - [ ] Wazuh-Agent folder tree: No files are lost after the rollback. The logs of the failed upgrade (`ossec.log`) are kept. - [ ] After the rollback the agent connects to the manager - [ ] After reboot, the Wazuh-Agent starts and connects to the manager. ### UPGRADE OK - [ ] Message `Upgrade finished successfully.` is shown in `upgrade.log` file. - [ ] Wazuh service is started and the agent is connected to the manager. --> ## Auditors validation The definition of done for this one is the validation of the conclusions and the test results from all auditors. All checks from below must be accepted in order to close this issue. - [ ] @vikman90
non_process
release release candidate wpk upgrade tests the following issue aims to run upgrade wpk tests for the current release candidate report the results and open new issues for any encountered errors wpk upgrade tests information main rc issue version release candidate tag previous issue test report procedure all individual test checks must be marked as pass the test ran successfully xfail the test was expected to fail and it failed it must be properly justified and reported in an issue skip the test was not run it must be properly justified and reported in an issue fail the test failed a new issue must be opened to evaluate and address the problem all test results must have one the following statuses green circle all checks passed red circle there is at least one failed check yellow circle there is at least one expected fail or skipped test and no failures any failing test must be properly addressed with a new issue detailing the error and the possible cause it must be included in the fixes section of the current release candidate main issue any expected fail or skipped test must have an issue justifying the reason all auditors must validate the justification for an expected fail or skipped test an extended report of the test results must be attached as a zip or txt this report can be used by the auditors to dig deeper into any possible failures and details tests to evaluate this feature it is necessary to test upgrading the agent and also the case when the upgrade fails rollback the tree command will be used to compare before and after the upgrade rollback process and check that the presence ownership and permissions of the files and directories are expected wazuh versions to test upgrade to the current agent version linux os version status upgrade fail upgrade ok centos white circle skipped centos white circle skipped centos white circle skipped centos white circle skipped centos white circle skipped centos yellow circle pending centos yellow circle pending centos white circle skipped centos white circle skipped centos white circle skipped centos white circle skipped centos white circle skipped centos yellow circle pending centos yellow circle pending rhel white circle skipped rhel white circle skipped rhel white circle skipped rhel white circle skipped rhel white circle skipped rhel yellow circle pending rhel yellow circle pending ubuntu white circle skipped ubuntu white circle skipped ubuntu white circle skipped ubuntu white circle skipped ubuntu white circle skipped ubuntu white circle skipped ubuntu white circle skipped opensuse tumbleweed white circle skipped opensuse tumbleweed white circle skipped opensuse tumbleweed white circle skipped opensuse tumbleweed white circle skipped opensuse tumbleweed white circle skipped opensuse tumbleweed white circle skipped opensuse tumbleweed white circle skipped amazon linux white circle skipped amazon linux white circle skipped amazon linux white circle skipped amazon linux white circle skipped amazon linux white circle skipped amazon linux white circle skipped amazon linux white circle skipped windows os version status upgrade fail upgrade ok server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped windows white circle skipped windows white circle skipped windows white circle skipped windows white circle skipped windows white circle skipped windows white circle skipped windows white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped server white circle skipped macos os version status upgrade fail upgrade ok sierra white circle skipped sierra white circle skipped catalina white circle skipped catalina white circle skipped big sur white circle skipped big sur white circle skipped monterey white circle skipped monterey white circle skipped for each operating system and version check the following points and add a comment for each os tested linux upgrade fail the wazuh home backup is restored correctly no traces of the installation but only the tar gz backup and the logs the permissions and owners of the following directories did not change var usr usr lib usr lib systemd usr lib systemd system etc etc systemd etc systemd system etc rc d etc initd d etc initd d rc d wazuh service runs wazuh control systemctl cat wazuh agent service wazuh service runs ossec control systemctl cat wazuh agent service the service was enabled systemctl is enabled wazuh agent service init file runs wazuh control cat etc rc d init d wazuh agent init file runs ossec control cat etc rc d init d wazuh agent wazuh as service is enabled chkconfig list wazuh starts and connects when the backup is restored cat var ossec var run ossec agentd state wazuh starts and connects automatically when the system is rebooted restore selinux policies semodule l grep i wazuh disabled upgrade ok upgrade is performed successfully agent connects to the manager after upgrading service starts automatically after rebooting agent connects to the manager after rebooting windows upgrade fail wazuh agent folder tree no files are lost after the rollback the logs of the failed upgrade ossec log are kept after the rollback the agent connects to the manager after reboot the wazuh agent starts and connects to the manager the correct wazuh agent version is shown in the list of windows programs and features a new version of wazuh agent can be manually installed via msi after the rollback process upgrade ok message upgrade finished successfully is shown in upgrade log file wazuh service is started and the agent is connected to the manager the version shown in the control panel is macos upgrade fail wazuh agent folder tree no files are lost after the rollback the logs of the failed upgrade ossec log are kept after the rollback the agent connects to the manager after reboot the wazuh agent starts and connects to the manager upgrade ok message upgrade finished successfully is shown in upgrade log file wazuh service is started and the agent is connected to the manager auditors validation the definition of done for this one is the validation of the conclusions and the test results from all auditors all checks from below must be accepted in order to close this issue
0
220,337
7,359,155,200
IssuesEvent
2018-03-10 02:41:56
eustasy/puff-members
https://api.github.com/repos/eustasy/puff-members
opened
Use Dropbox's password strength estimator "zxcvbn"
Language: HTML Language: JavaScript Language: PHP Priority: Medium Status: Confirmed Type: Enhancement
It's better than just requiring long passwords. 1. https://github.com/dropbox/zxcvbn 2. https://github.com/bjeavons/zxcvbn-php
1.0
Use Dropbox's password strength estimator "zxcvbn" - It's better than just requiring long passwords. 1. https://github.com/dropbox/zxcvbn 2. https://github.com/bjeavons/zxcvbn-php
non_process
use dropbox s password strength estimator zxcvbn it s better than just requiring long passwords
0
1,401
3,967,722,018
IssuesEvent
2016-05-03 17:10:57
opentrials/opentrials
https://api.github.com/repos/opentrials/opentrials
opened
Strategy of records->trial merging?
Processors
It could be: - some algorithm of merge - only primary is using (current)
1.0
Strategy of records->trial merging? - It could be: - some algorithm of merge - only primary is using (current)
process
strategy of records trial merging it could be some algorithm of merge only primary is using current
1
17,098
22,613,471,355
IssuesEvent
2022-06-29 19:25:56
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
opened
[processor/transform] Add ability to delete entries from a map.
priority:p2 comp: transformprocessor
**Is your feature request related to a problem? Please describe.** The transform processor has no way to delete entries from a map. If an attribute needs to be delete the only solution right now is `keep_keys`, which is unrealistic. **Describe the solution you'd like** Add a `delete_key` and `delete_all_matching_keys` functions. **Additional context** [Deleting capability is outlined in the processing doc.](https://github.com/open-telemetry/opentelemetry-collector/blob/main/docs/processing.md#examples)
1.0
[processor/transform] Add ability to delete entries from a map. - **Is your feature request related to a problem? Please describe.** The transform processor has no way to delete entries from a map. If an attribute needs to be delete the only solution right now is `keep_keys`, which is unrealistic. **Describe the solution you'd like** Add a `delete_key` and `delete_all_matching_keys` functions. **Additional context** [Deleting capability is outlined in the processing doc.](https://github.com/open-telemetry/opentelemetry-collector/blob/main/docs/processing.md#examples)
process
add ability to delete entries from a map is your feature request related to a problem please describe the transform processor has no way to delete entries from a map if an attribute needs to be delete the only solution right now is keep keys which is unrealistic describe the solution you d like add a delete key and delete all matching keys functions additional context
1
257,342
8,136,123,431
IssuesEvent
2018-08-20 07:15:15
OpenSourceEconomics/respy
https://api.github.com/repos/OpenSourceEconomics/respy
opened
FutureWarning
enhancement priority-low size-S
/home/peisenha/external_storage/ownCloud/office/workspace/research/papers/OptionValueHumanCapital/respy/respy/estimate.py:41: FutureWarning: Method .as_matrix will be removed in a future version. Use .values instead. data_array = data_frame.as_matrix() @janosg , just in case it is still relevant in your branch as well, please fix.
1.0
FutureWarning - /home/peisenha/external_storage/ownCloud/office/workspace/research/papers/OptionValueHumanCapital/respy/respy/estimate.py:41: FutureWarning: Method .as_matrix will be removed in a future version. Use .values instead. data_array = data_frame.as_matrix() @janosg , just in case it is still relevant in your branch as well, please fix.
non_process
futurewarning home peisenha external storage owncloud office workspace research papers optionvaluehumancapital respy respy estimate py futurewarning method as matrix will be removed in a future version use values instead data array data frame as matrix janosg just in case it is still relevant in your branch as well please fix
0
20,702
27,389,122,787
IssuesEvent
2023-02-28 15:11:39
esdc-devcop/esdc-devcop.github.io
https://api.github.com/repos/esdc-devcop/esdc-devcop.github.io
closed
Add a new Guide: NSD Application catalogue
ESDC processes
Create a new guide in the docs/_guides folder, using the same format as **_template.md** Title: NSD Application catalogue Content: Add front matter to this guide, something like this: ``` --- title: NSD Application catalogue layout: default category: ESDC: How We Operate summary: Add a short description of the guide here. date: YYYY-MM-DD --- ``` Add further content here
1.0
Add a new Guide: NSD Application catalogue - Create a new guide in the docs/_guides folder, using the same format as **_template.md** Title: NSD Application catalogue Content: Add front matter to this guide, something like this: ``` --- title: NSD Application catalogue layout: default category: ESDC: How We Operate summary: Add a short description of the guide here. date: YYYY-MM-DD --- ``` Add further content here
process
add a new guide nsd application catalogue create a new guide in the docs guides folder using the same format as template md title nsd application catalogue content add front matter to this guide something like this title nsd application catalogue layout default category esdc how we operate summary add a short description of the guide here date yyyy mm dd add further content here
1
226,905
25,010,636,907
IssuesEvent
2022-11-03 15:01:16
ondat/discoblocks
https://api.github.com/repos/ondat/discoblocks
closed
Fix code scanning alert - Incorrect conversion between integer types
bug security
<!-- Warning: The suggested title contains the alert rule name. This can expose security information. --> Tracking issue for: - [ ] https://github.com/ondat/discoblocks/security/code-scanning/15
True
Fix code scanning alert - Incorrect conversion between integer types - <!-- Warning: The suggested title contains the alert rule name. This can expose security information. --> Tracking issue for: - [ ] https://github.com/ondat/discoblocks/security/code-scanning/15
non_process
fix code scanning alert incorrect conversion between integer types tracking issue for
0
2,026
4,846,901,937
IssuesEvent
2016-11-10 13:24:51
Alfresco/alfresco-ng2-components
https://api.github.com/repos/Alfresco/alfresco-ng2-components
opened
After completing a task within the start process dialog this is not displayed in the completed tasks in tasklists
browser: all bug comp: activiti-processList comp: activiti-taskList
Firstly with the two being seperate componets should this occur? 1. Start a new process that includes a form to complete in its start event 2. Complete the form 3. Go to the task list component 4. Go to the completed tasks filter **Expected results** Task completed in step 1 is displayed **Actual results** Task completed in step 1 is not displayed
1.0
After completing a task within the start process dialog this is not displayed in the completed tasks in tasklists - Firstly with the two being seperate componets should this occur? 1. Start a new process that includes a form to complete in its start event 2. Complete the form 3. Go to the task list component 4. Go to the completed tasks filter **Expected results** Task completed in step 1 is displayed **Actual results** Task completed in step 1 is not displayed
process
after completing a task within the start process dialog this is not displayed in the completed tasks in tasklists firstly with the two being seperate componets should this occur start a new process that includes a form to complete in its start event complete the form go to the task list component go to the completed tasks filter expected results task completed in step is displayed actual results task completed in step is not displayed
1
18,664
24,582,495,888
IssuesEvent
2022-10-13 16:43:00
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Export mesh faces no longer works with tested E3T 2DM files - Wrong or missing parameter value: Dataset groups
Processing Regression Bug Mesh
### What is the bug or the crash? The "Export mesh faces" processing tool does not seem to recognize datasets in 2DM files and therefore won't run and yields the "Wrong or missing parameter value: Dataset groups" warning. ### Steps to reproduce the issue 1) Load sample 2DM file [test_simple_e3t_2dm_file.zip](https://github.com/qgis/QGIS/files/9435395/test_simple_e3t_2dm_file.zip) 2) Check Available Datasets (see attached screenshot) ![image](https://user-images.githubusercontent.com/13538726/186964985-eca96184-201f-44af-bb6d-1bd82406cc50.png) 3) Attempt to run the "Export mesh faces" processing tool. Note that no "Dataset groups" can be selected as none are available. ### Versions QGIS version 3.22.8-Biaล‚owieลผa QGIS code revision 8d5e9761df Qt version 5.15.3 Python version 3.9.5 GDAL/OGR version 3.5.0 PROJ version 9.0.1 EPSG Registry database version v10.064 (2022-05-19) GEOS version 3.10.3-CAPI-1.16.1 SQLite version 3.38.1 PDAL version 2.3.0 PostgreSQL client version 14.3 SpatiaLite version 5.0.1 QWT version 6.1.6 QScintilla2 version 2.13.1 OS version Windows 10 Version 2009 Active Python plugins basemesh 2.0.11 changeDataSource 3.1 DataPlotly 3.9 gmsh version 0.2.2 HCMGIS 22.8.8 ImportPhotos 3.0.3 LAStools 1.4 latlontools 3.6.4 leastcostpath 1.0 openlayers_plugin 2.0.0 processing_fusion 3.3 QPackage 1.5 QuickMultiAttributeEdit3 version 3.0.3 quick_map_services 0.19.29 splitmultipart 1.0.0 valuetool 3.0.15 db_manager 0.1.20 grassprovider 2.12.99 otbprovider 2.12.99 processing 2.12.99 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context _No response_
1.0
Export mesh faces no longer works with tested E3T 2DM files - Wrong or missing parameter value: Dataset groups - ### What is the bug or the crash? The "Export mesh faces" processing tool does not seem to recognize datasets in 2DM files and therefore won't run and yields the "Wrong or missing parameter value: Dataset groups" warning. ### Steps to reproduce the issue 1) Load sample 2DM file [test_simple_e3t_2dm_file.zip](https://github.com/qgis/QGIS/files/9435395/test_simple_e3t_2dm_file.zip) 2) Check Available Datasets (see attached screenshot) ![image](https://user-images.githubusercontent.com/13538726/186964985-eca96184-201f-44af-bb6d-1bd82406cc50.png) 3) Attempt to run the "Export mesh faces" processing tool. Note that no "Dataset groups" can be selected as none are available. ### Versions QGIS version 3.22.8-Biaล‚owieลผa QGIS code revision 8d5e9761df Qt version 5.15.3 Python version 3.9.5 GDAL/OGR version 3.5.0 PROJ version 9.0.1 EPSG Registry database version v10.064 (2022-05-19) GEOS version 3.10.3-CAPI-1.16.1 SQLite version 3.38.1 PDAL version 2.3.0 PostgreSQL client version 14.3 SpatiaLite version 5.0.1 QWT version 6.1.6 QScintilla2 version 2.13.1 OS version Windows 10 Version 2009 Active Python plugins basemesh 2.0.11 changeDataSource 3.1 DataPlotly 3.9 gmsh version 0.2.2 HCMGIS 22.8.8 ImportPhotos 3.0.3 LAStools 1.4 latlontools 3.6.4 leastcostpath 1.0 openlayers_plugin 2.0.0 processing_fusion 3.3 QPackage 1.5 QuickMultiAttributeEdit3 version 3.0.3 quick_map_services 0.19.29 splitmultipart 1.0.0 valuetool 3.0.15 db_manager 0.1.20 grassprovider 2.12.99 otbprovider 2.12.99 processing 2.12.99 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context _No response_
process
export mesh faces no longer works with tested files wrong or missing parameter value dataset groups what is the bug or the crash the export mesh faces processing tool does not seem to recognize datasets in files and therefore won t run and yields the wrong or missing parameter value dataset groups warning steps to reproduce the issue load sample file check available datasets see attached screenshot attempt to run the export mesh faces processing tool note that no dataset groups can be selected as none are available versions qgis version biaล‚owieลผa qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version spatialite version qwt version version os version windows version active python plugins basemesh changedatasource dataplotly gmsh version hcmgis importphotos lastools latlontools leastcostpath openlayers plugin processing fusion qpackage version quick map services splitmultipart valuetool db manager grassprovider otbprovider processing supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
1
309,167
26,654,287,157
IssuesEvent
2023-01-25 15:46:02
nrwl/nx
https://api.github.com/repos/nrwl/nx
closed
Cypress E2E w/ Storybook baseUrl
type: bug scope: testing tools scope: storybook
### Current Behavior I am getting this error when running the e2e tests for my library with storybook configuration ### Expected Behavior To work ### Github Repo _No response_ ### Steps to Reproduce 1. Run e2e test on library e2e app with storybook configuration (react) ### Nx Report ```shell Node : 16.18.1 OS : darwin arm64 yarn : 3.3.1 nx : 15.5.1 @nrwl/angular : Not Found @nrwl/cypress : 15.5.1 @nrwl/detox : 15.5.1 @nrwl/devkit : 15.5.1 @nrwl/esbuild : Not Found @nrwl/eslint-plugin-nx : 15.5.1 @nrwl/expo : 15.5.1 @nrwl/express : Not Found @nrwl/jest : 15.5.1 @nrwl/js : 15.5.1 @nrwl/linter : 15.5.1 @nrwl/nest : 15.5.1 @nrwl/next : 15.5.1 @nrwl/node : 15.5.1 @nrwl/nx-cloud : 15.0.2 @nrwl/nx-plugin : Not Found @nrwl/react : 15.5.1 @nrwl/react-native : Not Found @nrwl/rollup : 15.5.1 @nrwl/schematics : Not Found @nrwl/storybook : 15.5.1 @nrwl/web : 15.5.1 @nrwl/webpack : 15.5.1 @nrwl/workspace : 15.5.1 @nrwl/vite : 15.5.1 typescript : 4.8.4 --------------------------------------- Local workspace plugins: --------------------------------------- Community plugins: ``` ### Failure Logs ```shell <s> [webpack.Progress] 99% done plugins <s> [webpack.Progress] 99% <s> [webpack.Progress] 99% cache store build dependencies <s> [webpack.Progress] 99% cache store build dependencies <s> [webpack.Progress] 99% cache begin idle <s> [webpack.Progress] 99% cache begin idle <s> [webpack.Progress] 100% โ•ญโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ•ฎ โ”‚ โ”‚ โ”‚ Storybook 6.5.15 for React started โ”‚ โ”‚ 5.75 s for preview โ”‚ โ”‚ โ”‚ โ”‚ Local: http://localhost:4400/ โ”‚ โ”‚ On your network: http://localhost:4400/ โ”‚ โ”‚ โ”‚ โ•ฐโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ•ฏ > NX Cannot read properties of undefined (reading 'baseUrl') TypeError: Cannot read properties of undefined (reading 'baseUrl') ``` ### Additional Information _No response_
1.0
Cypress E2E w/ Storybook baseUrl - ### Current Behavior I am getting this error when running the e2e tests for my library with storybook configuration ### Expected Behavior To work ### Github Repo _No response_ ### Steps to Reproduce 1. Run e2e test on library e2e app with storybook configuration (react) ### Nx Report ```shell Node : 16.18.1 OS : darwin arm64 yarn : 3.3.1 nx : 15.5.1 @nrwl/angular : Not Found @nrwl/cypress : 15.5.1 @nrwl/detox : 15.5.1 @nrwl/devkit : 15.5.1 @nrwl/esbuild : Not Found @nrwl/eslint-plugin-nx : 15.5.1 @nrwl/expo : 15.5.1 @nrwl/express : Not Found @nrwl/jest : 15.5.1 @nrwl/js : 15.5.1 @nrwl/linter : 15.5.1 @nrwl/nest : 15.5.1 @nrwl/next : 15.5.1 @nrwl/node : 15.5.1 @nrwl/nx-cloud : 15.0.2 @nrwl/nx-plugin : Not Found @nrwl/react : 15.5.1 @nrwl/react-native : Not Found @nrwl/rollup : 15.5.1 @nrwl/schematics : Not Found @nrwl/storybook : 15.5.1 @nrwl/web : 15.5.1 @nrwl/webpack : 15.5.1 @nrwl/workspace : 15.5.1 @nrwl/vite : 15.5.1 typescript : 4.8.4 --------------------------------------- Local workspace plugins: --------------------------------------- Community plugins: ``` ### Failure Logs ```shell <s> [webpack.Progress] 99% done plugins <s> [webpack.Progress] 99% <s> [webpack.Progress] 99% cache store build dependencies <s> [webpack.Progress] 99% cache store build dependencies <s> [webpack.Progress] 99% cache begin idle <s> [webpack.Progress] 99% cache begin idle <s> [webpack.Progress] 100% โ•ญโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ•ฎ โ”‚ โ”‚ โ”‚ Storybook 6.5.15 for React started โ”‚ โ”‚ 5.75 s for preview โ”‚ โ”‚ โ”‚ โ”‚ Local: http://localhost:4400/ โ”‚ โ”‚ On your network: http://localhost:4400/ โ”‚ โ”‚ โ”‚ โ•ฐโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ•ฏ > NX Cannot read properties of undefined (reading 'baseUrl') TypeError: Cannot read properties of undefined (reading 'baseUrl') ``` ### Additional Information _No response_
non_process
cypress w storybook baseurl current behavior i am getting this error when running the tests for my library with storybook configuration expected behavior to work github repo no response steps to reproduce run test on library app with storybook configuration react nx report shell node os darwin yarn nx nrwl angular not found nrwl cypress nrwl detox nrwl devkit nrwl esbuild not found nrwl eslint plugin nx nrwl expo nrwl express not found nrwl jest nrwl js nrwl linter nrwl nest nrwl next nrwl node nrwl nx cloud nrwl nx plugin not found nrwl react nrwl react native not found nrwl rollup nrwl schematics not found nrwl storybook nrwl web nrwl webpack nrwl workspace nrwl vite typescript local workspace plugins community plugins failure logs shell done plugins cache store build dependencies cache store build dependencies cache begin idle cache begin idle โ•ญโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ•ฎ โ”‚ โ”‚ โ”‚ storybook for react started โ”‚ โ”‚ s for preview โ”‚ โ”‚ โ”‚ โ”‚ local โ”‚ โ”‚ on your network โ”‚ โ”‚ โ”‚ โ•ฐโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ•ฏ nx cannot read properties of undefined reading baseurl typeerror cannot read properties of undefined reading baseurl additional information no response
0
9,424
6,889,443,205
IssuesEvent
2017-11-22 10:23:25
OctopusDeploy/Issues
https://api.github.com/repos/OctopusDeploy/Issues
closed
/api/useronboarding is slow when there are lots of deployments in the system
area/scale bug feature/performance
The `/api/useronboarding` call pulls in all the `DeploymentProcess` objects to check if any step required a package. This is causing memory issues once the system has a lot of deployments, or the deployments are large memory objects (lots of embedded scripts, lots of steps). # Root cause Every time the Octopus Web UI loads in the browser it makes an initial call to the `/api/useronboarding` endpoint. If you have a very large installation of Octopus this can be a very expensive call. Then scale that out across potentially hundreds of users coming in to work in the morning and loading Octopus. # Suggested solution Once you have completed a step of the user onboarding, like creating a project, we can assume you don't need our help to do that step again. We are planning to remember as soon as you check off each step in the onboarding process. This means we will only run the (potentially) expensive queries for the pieces of onboarding you haven't completed by that point in time. This means when people start the Octopus UI, it will still make a GET request to `/api/useronboarding` - but the resulting query will be very cheap, instead of blindly loading every deployment process in existence. :)
True
/api/useronboarding is slow when there are lots of deployments in the system - The `/api/useronboarding` call pulls in all the `DeploymentProcess` objects to check if any step required a package. This is causing memory issues once the system has a lot of deployments, or the deployments are large memory objects (lots of embedded scripts, lots of steps). # Root cause Every time the Octopus Web UI loads in the browser it makes an initial call to the `/api/useronboarding` endpoint. If you have a very large installation of Octopus this can be a very expensive call. Then scale that out across potentially hundreds of users coming in to work in the morning and loading Octopus. # Suggested solution Once you have completed a step of the user onboarding, like creating a project, we can assume you don't need our help to do that step again. We are planning to remember as soon as you check off each step in the onboarding process. This means we will only run the (potentially) expensive queries for the pieces of onboarding you haven't completed by that point in time. This means when people start the Octopus UI, it will still make a GET request to `/api/useronboarding` - but the resulting query will be very cheap, instead of blindly loading every deployment process in existence. :)
non_process
api useronboarding is slow when there are lots of deployments in the system the api useronboarding call pulls in all the deploymentprocess objects to check if any step required a package this is causing memory issues once the system has a lot of deployments or the deployments are large memory objects lots of embedded scripts lots of steps root cause every time the octopus web ui loads in the browser it makes an initial call to the api useronboarding endpoint if you have a very large installation of octopus this can be a very expensive call then scale that out across potentially hundreds of users coming in to work in the morning and loading octopus suggested solution once you have completed a step of the user onboarding like creating a project we can assume you don t need our help to do that step again we are planning to remember as soon as you check off each step in the onboarding process this means we will only run the potentially expensive queries for the pieces of onboarding you haven t completed by that point in time this means when people start the octopus ui it will still make a get request to api useronboarding but the resulting query will be very cheap instead of blindly loading every deployment process in existence
0
77,852
10,023,247,081
IssuesEvent
2019-07-16 18:42:39
spring-io/initializr
https://api.github.com/repos/spring-io/initializr
closed
Document how to generate a project structure
type: documentation
Now that we have a proper API (rather than templates to override and core classes to have to extend) we should have a section in the ref guide that describes how to do it. We've started [here](https://docs.spring.io/initializr/docs/current-SNAPSHOT/reference/html/#project-generation) but we need to expand this with a step by step guide.
1.0
Document how to generate a project structure - Now that we have a proper API (rather than templates to override and core classes to have to extend) we should have a section in the ref guide that describes how to do it. We've started [here](https://docs.spring.io/initializr/docs/current-SNAPSHOT/reference/html/#project-generation) but we need to expand this with a step by step guide.
non_process
document how to generate a project structure now that we have a proper api rather than templates to override and core classes to have to extend we should have a section in the ref guide that describes how to do it we ve started but we need to expand this with a step by step guide
0
6,713
9,819,520,741
IssuesEvent
2019-06-13 22:18:40
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Department of State OOPs Page
Apply Process Approved Requirements Ready State Dept.
Who: Student applicant What: OOPs Page Why: As a student I would like to know why there are no internship opportunities listed. A/C - The design should follow the design system and current OOPs pages - The header will be **Sorry, there are no internships available right now.** - The following text will appear under the header: The U.S. Department of State Student Internship Program is not accepting applications at this time. Our next available session will be [insert next internship session]. [Sign up for email updates to know when weโ€™re accepting new applications](link to: https://public.govdelivery.com/accounts/USSTATE/subscriber/new). Learn more about the [U.S. Department of State Student Internship Program] (link to: https://careers.state.gov/intern/student-internships/) - All links will open in a new window - The following sentence is conditional "Our next available session will be [insert next internship session]." - If the next cycle has not been created by the CM this sentence will not be visible.
1.0
Department of State OOPs Page - Who: Student applicant What: OOPs Page Why: As a student I would like to know why there are no internship opportunities listed. A/C - The design should follow the design system and current OOPs pages - The header will be **Sorry, there are no internships available right now.** - The following text will appear under the header: The U.S. Department of State Student Internship Program is not accepting applications at this time. Our next available session will be [insert next internship session]. [Sign up for email updates to know when weโ€™re accepting new applications](link to: https://public.govdelivery.com/accounts/USSTATE/subscriber/new). Learn more about the [U.S. Department of State Student Internship Program] (link to: https://careers.state.gov/intern/student-internships/) - All links will open in a new window - The following sentence is conditional "Our next available session will be [insert next internship session]." - If the next cycle has not been created by the CM this sentence will not be visible.
process
department of state oops page who student applicant what oops page why as a student i would like to know why there are no internship opportunities listed a c the design should follow the design system and current oops pages the header will be sorry there are no internships available right now the following text will appear under the header the u s department of state student internship program is not accepting applications at this time our next available session will be link to learn more about the link to all links will open in a new window the following sentence is conditional our next available session will be if the next cycle has not been created by the cm this sentence will not be visible
1
15,914
20,120,730,261
IssuesEvent
2022-02-08 01:49:47
hoprnet/hoprnet
https://api.github.com/repos/hoprnet/hoprnet
closed
sprint TLDR for community
processes stale
<!--- Please DO NOT remove the automatically added 'new issue' label --> <!--- Provide a general summary of the issue in the Title above --> Describe the process for representatives give a tldr to comm team for what was completed previous sprint and what's coming up
1.0
sprint TLDR for community - <!--- Please DO NOT remove the automatically added 'new issue' label --> <!--- Provide a general summary of the issue in the Title above --> Describe the process for representatives give a tldr to comm team for what was completed previous sprint and what's coming up
process
sprint tldr for community describe the process for representatives give a tldr to comm team for what was completed previous sprint and what s coming up
1
17,377
23,200,172,821
IssuesEvent
2022-08-01 20:34:10
Ultimate-Hosts-Blacklist/whitelist
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
closed
[FALSE-POSITIVE?] continuum.dds.microsoft.com
whitelisting process upstream issue list removal request
This domain is required for `Your Phone Companion` to work. Else it will be stuck on "connecting to PC" and "connecting to phone"
1.0
[FALSE-POSITIVE?] continuum.dds.microsoft.com - This domain is required for `Your Phone Companion` to work. Else it will be stuck on "connecting to PC" and "connecting to phone"
process
continuum dds microsoft com this domain is required for your phone companion to work else it will be stuck on connecting to pc and connecting to phone
1
7,332
10,469,052,814
IssuesEvent
2019-09-22 17:59:29
produvia/ai-platform
https://api.github.com/repos/produvia/ai-platform
closed
Language Modelling
natural-language-processing task wontfix
# Goal(s) - Predict the next word or character in a document # Input(s) - Document # Output(s) - Document # Objective Function(s) - TBD
1.0
Language Modelling - # Goal(s) - Predict the next word or character in a document # Input(s) - Document # Output(s) - Document # Objective Function(s) - TBD
process
language modelling goal s predict the next word or character in a document input s document output s document objective function s tbd
1
108,124
23,537,226,864
IssuesEvent
2022-08-19 22:53:23
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
closed
Use CodeMirror query input in Code Monitors form
team/search-product code-monitors
Due to an oversight of mine, the code monitor creation form still uses the Monaco query input. I created https://github.com/sourcegraph/sourcegraph/pull/39178 to fix this but ran into issues getting the tests to work. While I made good progress and I even was able to get the tests to pass in isolation, they don't seem pass when the whole file is run. We an either investigate more to try to fix them or disable them for the time being. /cc @benvenker @lguychard
1.0
Use CodeMirror query input in Code Monitors form - Due to an oversight of mine, the code monitor creation form still uses the Monaco query input. I created https://github.com/sourcegraph/sourcegraph/pull/39178 to fix this but ran into issues getting the tests to work. While I made good progress and I even was able to get the tests to pass in isolation, they don't seem pass when the whole file is run. We an either investigate more to try to fix them or disable them for the time being. /cc @benvenker @lguychard
non_process
use codemirror query input in code monitors form due to an oversight of mine the code monitor creation form still uses the monaco query input i created to fix this but ran into issues getting the tests to work while i made good progress and i even was able to get the tests to pass in isolation they don t seem pass when the whole file is run we an either investigate more to try to fix them or disable them for the time being cc benvenker lguychard
0
2,359
5,166,317,344
IssuesEvent
2017-01-17 15:57:09
openvstorage/framework
https://api.github.com/repos/openvstorage/framework
closed
Durandal activate triggers twice for all wizards
priority_normal process_duplicate type_bug
The 'activate' code in our wizards, which is supposed to only happen once on launch of a wizard, gets always triggered twice. This causes a bunch of APIs to be executed twice upon opening a wizard and thus causing unnecessary load and delay
1.0
Durandal activate triggers twice for all wizards - The 'activate' code in our wizards, which is supposed to only happen once on launch of a wizard, gets always triggered twice. This causes a bunch of APIs to be executed twice upon opening a wizard and thus causing unnecessary load and delay
process
durandal activate triggers twice for all wizards the activate code in our wizards which is supposed to only happen once on launch of a wizard gets always triggered twice this causes a bunch of apis to be executed twice upon opening a wizard and thus causing unnecessary load and delay
1
20,070
26,562,974,919
IssuesEvent
2023-01-20 17:24:06
spinalcordtoolbox/spinalcordtoolbox
https://api.github.com/repos/spinalcordtoolbox/spinalcordtoolbox
opened
Refactor the `aggregate_per_slice_or_level` function to make it easier to understand and modify
sct_process_segmentation refactoring API: aggregate_slicewise.py
[`aggregate_per_slice_or_level`](https://github.com/spinalcordtoolbox/spinalcordtoolbox/blob/377bfb88a9845bb8348e760aea5d70bb4f4bcb42/spinalcordtoolbox/aggregate_slicewise.py#L233-L366) is a lengthy function with many nested conditionals. This complexity makes it difficult to understand and make changes to. Even once a change has been proposed, reviewers then have to parse this complicated nest of conditionals to understand the impact that the change has on the rest of the function. This makes a relatively simple change like #4004 a headache to review. Additionally, `aggregate_per_slice_or_level` is a core component of `sct_extract_metric` and `sct_process_segmentation`, and these functions have been an active focus for quite a lot of past/present/future changes to SCT: - Adding new features/metrics to `sct_process_segmentation` - https://github.com/spinalcordtoolbox/spinalcordtoolbox/issues/3888 - https://github.com/spinalcordtoolbox/spinalcordtoolbox/pull/3946 - https://github.com/spinalcordtoolbox/spinalcordtoolbox/pull/3977 - `sct_process_segmentation`/`sct_analyze_lesion` refactor effort - https://github.com/spinalcordtoolbox/spinalcordtoolbox/issues/2677 - https://github.com/spinalcordtoolbox/spinalcordtoolbox/issues/3378 - https://github.com/spinalcordtoolbox/spinalcordtoolbox/issues/2707 Considering that this is a high-activity part of SCT, it's worth taking account the cost of [tech debt](https://www.maxcountryman.com/articles/a-framework-for-prioritizing-tech-debt): > Is this part of the system active or dormantโ€“are we building new features on it today or is it in maintenance mode? > > * If it's dormant, how much time is spent on maintenance each engineering cycle? > * If it's active, take note of that: you're building on a foundation that may not be stable. So, I think `aggregate_per_slice_or_level` would be a good candidate for refactoring to make it easier to work with.
1.0
Refactor the `aggregate_per_slice_or_level` function to make it easier to understand and modify - [`aggregate_per_slice_or_level`](https://github.com/spinalcordtoolbox/spinalcordtoolbox/blob/377bfb88a9845bb8348e760aea5d70bb4f4bcb42/spinalcordtoolbox/aggregate_slicewise.py#L233-L366) is a lengthy function with many nested conditionals. This complexity makes it difficult to understand and make changes to. Even once a change has been proposed, reviewers then have to parse this complicated nest of conditionals to understand the impact that the change has on the rest of the function. This makes a relatively simple change like #4004 a headache to review. Additionally, `aggregate_per_slice_or_level` is a core component of `sct_extract_metric` and `sct_process_segmentation`, and these functions have been an active focus for quite a lot of past/present/future changes to SCT: - Adding new features/metrics to `sct_process_segmentation` - https://github.com/spinalcordtoolbox/spinalcordtoolbox/issues/3888 - https://github.com/spinalcordtoolbox/spinalcordtoolbox/pull/3946 - https://github.com/spinalcordtoolbox/spinalcordtoolbox/pull/3977 - `sct_process_segmentation`/`sct_analyze_lesion` refactor effort - https://github.com/spinalcordtoolbox/spinalcordtoolbox/issues/2677 - https://github.com/spinalcordtoolbox/spinalcordtoolbox/issues/3378 - https://github.com/spinalcordtoolbox/spinalcordtoolbox/issues/2707 Considering that this is a high-activity part of SCT, it's worth taking account the cost of [tech debt](https://www.maxcountryman.com/articles/a-framework-for-prioritizing-tech-debt): > Is this part of the system active or dormantโ€“are we building new features on it today or is it in maintenance mode? > > * If it's dormant, how much time is spent on maintenance each engineering cycle? > * If it's active, take note of that: you're building on a foundation that may not be stable. So, I think `aggregate_per_slice_or_level` would be a good candidate for refactoring to make it easier to work with.
process
refactor the aggregate per slice or level function to make it easier to understand and modify is a lengthy function with many nested conditionals this complexity makes it difficult to understand and make changes to even once a change has been proposed reviewers then have to parse this complicated nest of conditionals to understand the impact that the change has on the rest of the function this makes a relatively simple change like a headache to review additionally aggregate per slice or level is a core component of sct extract metric and sct process segmentation and these functions have been an active focus for quite a lot of past present future changes to sct adding new features metrics to sct process segmentation sct process segmentation sct analyze lesion refactor effort considering that this is a high activity part of sct it s worth taking account the cost of is this part of the system active or dormantโ€“are we building new features on it today or is it in maintenance mode if it s dormant how much time is spent on maintenance each engineering cycle if it s active take note of that you re building on a foundation that may not be stable so i think aggregate per slice or level would be a good candidate for refactoring to make it easier to work with
1
67,756
28,039,947,170
IssuesEvent
2023-03-28 17:42:43
hashicorp/terraform-provider-aws
https://api.github.com/repos/hashicorp/terraform-provider-aws
closed
bug: Error creating appsync api key / functions with ConcurrentModificationException ,Schema is currently being altered
bug service/appsync stale
<!--- Please note the following potential times when an issue might be in Terraform core: * [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues * [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues * [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues * [Registry](https://registry.terraform.io/) issues * Spans resources across multiple providers If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead. ---> <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a ๐Ÿ‘ [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform CLI and Terraform AWS Provider Version terraform version : 0.13.5 AWS provider: 3.9.0 ### Affected Resource(s) aws_appsync_resolver aws_appsync_function aws_appsync_api_key ### Terraform Configuration Files The following link has the required terraform files https://github.com/terraform-aws-modules/terraform-aws-appsync/tree/master/examples/complete ### Expected Behavior app sync functions/api keys /resolvers are provisioned without any errors ### Actual Behavior Intermittent issues with following messages ``` Error: error creating Appsync API Key: ConcurrentModificationException: Schema is currently being altered, please wait until that is complete. ``` I have also observed cases when app sync functions are added ``` Error: Error creating AppSync Function: ConcurrentModificationException: Schema is currently being altered, please wait until that is complete. Error: Error creating AppSync Function: ConcurrentModificationException: Schema is currently being altered, please wait until that is complete. Error: Error creating AppSync Function: ConcurrentModificationException: Schema is currently being altered, please wait until that is complete. ``` The issue gets solved when we re trigger the job 30-60 secs later ### Steps to Reproduce 1. Clone the repo https://github.com/terraform-aws-modules/terraform-aws-appsync.git 2. cd examples/complete 3. terraform init 4. terraform apply As the issue is intermittent, job might need to be triggered several times before getting a reproduction of error. ### References https://github.com/aws/aws-cdk/pull/9737 https://docs.aws.amazon.com/appsync/latest/APIReference/API_UpdateGraphqlApi.htmlhttps://github.com/aws/aws-cdk/pull/9737 * #0000
1.0
bug: Error creating appsync api key / functions with ConcurrentModificationException ,Schema is currently being altered - <!--- Please note the following potential times when an issue might be in Terraform core: * [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues * [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues * [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues * [Registry](https://registry.terraform.io/) issues * Spans resources across multiple providers If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead. ---> <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a ๐Ÿ‘ [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform CLI and Terraform AWS Provider Version terraform version : 0.13.5 AWS provider: 3.9.0 ### Affected Resource(s) aws_appsync_resolver aws_appsync_function aws_appsync_api_key ### Terraform Configuration Files The following link has the required terraform files https://github.com/terraform-aws-modules/terraform-aws-appsync/tree/master/examples/complete ### Expected Behavior app sync functions/api keys /resolvers are provisioned without any errors ### Actual Behavior Intermittent issues with following messages ``` Error: error creating Appsync API Key: ConcurrentModificationException: Schema is currently being altered, please wait until that is complete. ``` I have also observed cases when app sync functions are added ``` Error: Error creating AppSync Function: ConcurrentModificationException: Schema is currently being altered, please wait until that is complete. Error: Error creating AppSync Function: ConcurrentModificationException: Schema is currently being altered, please wait until that is complete. Error: Error creating AppSync Function: ConcurrentModificationException: Schema is currently being altered, please wait until that is complete. ``` The issue gets solved when we re trigger the job 30-60 secs later ### Steps to Reproduce 1. Clone the repo https://github.com/terraform-aws-modules/terraform-aws-appsync.git 2. cd examples/complete 3. terraform init 4. terraform apply As the issue is intermittent, job might need to be triggered several times before getting a reproduction of error. ### References https://github.com/aws/aws-cdk/pull/9737 https://docs.aws.amazon.com/appsync/latest/APIReference/API_UpdateGraphqlApi.htmlhttps://github.com/aws/aws-cdk/pull/9737 * #0000
non_process
bug error creating appsync api key functions with concurrentmodificationexception schema is currently being altered please note the following potential times when an issue might be in terraform core or resource ordering issues and issues issues issues spans resources across multiple providers if you are running into one of these scenarios we recommend opening an issue in the instead community note please vote on this issue by adding a ๐Ÿ‘ to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform cli and terraform aws provider version terraform version aws provider affected resource s aws appsync resolver aws appsync function aws appsync api key terraform configuration files the following link has the required terraform files expected behavior app sync functions api keys resolvers are provisioned without any errors actual behavior intermittent issues with following messages error error creating appsync api key concurrentmodificationexception schema is currently being altered please wait until that is complete i have also observed cases when app sync functions are added error error creating appsync function concurrentmodificationexception schema is currently being altered please wait until that is complete error error creating appsync function concurrentmodificationexception schema is currently being altered please wait until that is complete error error creating appsync function concurrentmodificationexception schema is currently being altered please wait until that is complete the issue gets solved when we re trigger the job secs later steps to reproduce clone the repo cd examples complete terraform init terraform apply as the issue is intermittent job might need to be triggered several times before getting a reproduction of error references
0
20,893
27,723,789,486
IssuesEvent
2023-03-14 23:22:54
CodeForPhilly/paws-data-pipeline
https://api.github.com/repos/CodeForPhilly/paws-data-pipeline
closed
UI needs to check /api/execute response for status
UX Async processes
We shouldn't get timeouts on nginx if we crank the timeout high enough (ugh) but we should catch that or other errors. In `handleExecute()`, before parsing the return from /api/execute we should check `response.ok`. If False, we need to display to user that an error (of type `response.status`) has occurred.
1.0
UI needs to check /api/execute response for status - We shouldn't get timeouts on nginx if we crank the timeout high enough (ugh) but we should catch that or other errors. In `handleExecute()`, before parsing the return from /api/execute we should check `response.ok`. If False, we need to display to user that an error (of type `response.status`) has occurred.
process
ui needs to check api execute response for status we shouldn t get timeouts on nginx if we crank the timeout high enough ugh but we should catch that or other errors in handleexecute before parsing the return from api execute we should check response ok if false we need to display to user that an error of type response status has occurred
1
19,051
25,063,019,075
IssuesEvent
2022-11-07 04:54:34
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
attributes processor - regex replace action
Stale processor/attributes
**Is your feature request related to a problem? Please describe.** I have a need to remove only certain URL parameters from the `http.url` attribute. Some parameters such as email, user, etc can leak PII data. **Describe the solution you'd like** A way to rewrite the `http.url` attribute removing only the PII parameters. This could be accomplished with a "regex_replace" type of action that will allow the operator to specify a regex with capture groups, which are referenced in replace string. For example using this regex/replace regex: `^(.*)((&|\?)user\=.+?)(&(.*)$|$)` replace: `$1$4` We would get the following input / output sets Input1: `https://otel.io/path?foo=bar&user=hide_me&param2=value2` Output1: `https://otel.io/path?foo=bar&param2=value2` Input2: `https://otel.io/path?foo=bar&user=hide_me` Output2: `https://otel.io/path?foo=bar` Another potential solution is to have a new action that can "concatenate" multiple attributes. This could then be used in conjunction with the existing `extract` action. **Describe alternatives you've considered** I have tried to accomplish this using extract, which allowed me to extract any string before and after the `user` parameter, however, concatenating the "before" and "after" attributes back into a single attribute was not possible.
1.0
attributes processor - regex replace action - **Is your feature request related to a problem? Please describe.** I have a need to remove only certain URL parameters from the `http.url` attribute. Some parameters such as email, user, etc can leak PII data. **Describe the solution you'd like** A way to rewrite the `http.url` attribute removing only the PII parameters. This could be accomplished with a "regex_replace" type of action that will allow the operator to specify a regex with capture groups, which are referenced in replace string. For example using this regex/replace regex: `^(.*)((&|\?)user\=.+?)(&(.*)$|$)` replace: `$1$4` We would get the following input / output sets Input1: `https://otel.io/path?foo=bar&user=hide_me&param2=value2` Output1: `https://otel.io/path?foo=bar&param2=value2` Input2: `https://otel.io/path?foo=bar&user=hide_me` Output2: `https://otel.io/path?foo=bar` Another potential solution is to have a new action that can "concatenate" multiple attributes. This could then be used in conjunction with the existing `extract` action. **Describe alternatives you've considered** I have tried to accomplish this using extract, which allowed me to extract any string before and after the `user` parameter, however, concatenating the "before" and "after" attributes back into a single attribute was not possible.
process
attributes processor regex replace action is your feature request related to a problem please describe i have a need to remove only certain url parameters from the http url attribute some parameters such as email user etc can leak pii data describe the solution you d like a way to rewrite the http url attribute removing only the pii parameters this could be accomplished with a regex replace type of action that will allow the operator to specify a regex with capture groups which are referenced in replace string for example using this regex replace regex user replace we would get the following input output sets another potential solution is to have a new action that can concatenate multiple attributes this could then be used in conjunction with the existing extract action describe alternatives you ve considered i have tried to accomplish this using extract which allowed me to extract any string before and after the user parameter however concatenating the before and after attributes back into a single attribute was not possible
1
81,322
30,798,379,426
IssuesEvent
2023-07-31 22:01:15
openziti/zrok
https://api.github.com/repos/openziti/zrok
opened
'zrok status', 'zrok config', etc. Broken For Uninitialized Environment
defect
Running `zrok status` or `zrok config`, or `zrok enable` on an environment that has no `~/.zrok` directory will result in panics.
1.0
'zrok status', 'zrok config', etc. Broken For Uninitialized Environment - Running `zrok status` or `zrok config`, or `zrok enable` on an environment that has no `~/.zrok` directory will result in panics.
non_process
zrok status zrok config etc broken for uninitialized environment running zrok status or zrok config or zrok enable on an environment that has no zrok directory will result in panics
0
13
2,496,237,777
IssuesEvent
2015-01-06 18:02:44
vivo-isf/vivo-isf-ontology
https://api.github.com/repos/vivo-isf/vivo-isf-ontology
closed
sex differentiation
biological_process imported
_From [vasil...@ohsu.edu](https://code.google.com/u/108803237899917466626/) on November 12, 2012 16:48:33_ GO:0007548 Parent: developmental process \<a href="http://purl.obolibrary.org/obo/GO_0032502" rel="nofollow">http://purl.obolibrary.org/obo/GO_0032502</a>&#13; &#13; \<a href="http://ohsu.eagle-i.net/i/00000130-6c5b-d11f-0c8d-bb9080000000" rel="nofollow">http://ohsu.eagle-i.net/i/00000130-6c5b-d11f-0c8d-bb9080000000</a>&#13; _Original issue: http://code.google.com/p/eagle-i/issues/detail?id=164_
1.0
sex differentiation - _From [vasil...@ohsu.edu](https://code.google.com/u/108803237899917466626/) on November 12, 2012 16:48:33_ GO:0007548 Parent: developmental process \<a href="http://purl.obolibrary.org/obo/GO_0032502" rel="nofollow">http://purl.obolibrary.org/obo/GO_0032502</a>&#13; &#13; \<a href="http://ohsu.eagle-i.net/i/00000130-6c5b-d11f-0c8d-bb9080000000" rel="nofollow">http://ohsu.eagle-i.net/i/00000130-6c5b-d11f-0c8d-bb9080000000</a>&#13; _Original issue: http://code.google.com/p/eagle-i/issues/detail?id=164_
process
sex differentiation from on november go parent developmental process original issue
1
3,907
6,825,118,216
IssuesEvent
2017-11-08 09:23:10
DynareTeam/dynare
https://api.github.com/repos/DynareTeam/dynare
closed
Pre-processor bug: plot_shock_decomposition
preprocessor
The type option is not properly handled, since it always provides qoq field `options_.plot_shock_decomp.type = 'qoq'; ` whatever this choice `qoq | yoy | aoa`
1.0
Pre-processor bug: plot_shock_decomposition - The type option is not properly handled, since it always provides qoq field `options_.plot_shock_decomp.type = 'qoq'; ` whatever this choice `qoq | yoy | aoa`
process
pre processor bug plot shock decomposition the type option is not properly handled since it always provides qoq field options plot shock decomp type qoq whatever this choice qoq yoy aoa
1
216,438
16,659,424,563
IssuesEvent
2021-06-06 04:35:44
zhangferry/iOSWeeklyLearning
https://api.github.com/repos/zhangferry/iOSWeeklyLearning
closed
ๆ–‡็ซ ๅˆ†ไบซใ€Šๆญ็ง˜่‹นๆžœๅบ”็”จๅฎกๆ ธๅ›ข้˜Ÿ๏ผˆๅฒไธŠๆœ€ๅ…จ็‰ˆ๏ผ‰ใ€‹
documentation
## ๆ–‡็ซ ๆ ‡้ข˜ใ€่‘—ๅๅ‡บๅค„ๅ’Œไฝœ่€… - ๆ–‡็ซ ๆ ‡้ข˜๏ผšๆญ็ง˜่‹นๆžœๅบ”็”จๅฎกๆ ธๅ›ข้˜Ÿ๏ผˆๅฒไธŠๆœ€ๅ…จ็‰ˆ๏ผ‰ - ๅ‡บๅค„๏ผšๆŽ˜้‡‘ - ไฝœ่€…๏ผš37ๆ‰‹ๆธธiOSๆŠ€ๆœฏ่ฟ่ฅๅ›ข้˜Ÿ ## ๆ–‡็ซ ้“พๆŽฅ https://juejin.cn/post/6970363897668698148/ ## ๆŽจ่็†็”ฑ๏ผˆๅฆ‚ๆžœๆฒกๆœ‰ๅฏไปฅๆ‘˜ๆŠ„ๆ–‡็ซ ๅ†…ไธ€ๆฎต้‡่ฆ็š„่ฏ๏ผ‰ ๅฒไธŠๆœ€ๅ…จ็‰ˆ๏ผšๆญ็ง˜่‹นๆžœๅบ”็”จๅฎกๆ ธๅ›ข้˜Ÿ๏ผŒๅ‘Š่ฏ‰ไฝ  App Store ็š„็”ฑๆฅ๏ผŒๆ˜ฏๆ€Žๆ ทๅ‘ๅฑ•ๅˆฐไปŠๅคฉ่ฟ™ๆ ท่พ‰็…Œ๏ผŒๅฆ‚ไฝ•่‡ชๅŠจๅŒ–๏ผšๆœบๅ™จๅฎกๆ ธ+ไบบๅทฅๅฎกๆ ธ๏ผŸๅฎกๆ ธ้€Ÿๅบฆๆ˜ฏๆ€ŽไนˆไปŽ7-10ๅคฉ๏ผŒๆๅ‡ๅˆฐๅฆ‚ไปŠๅช้œ€่ฆ 48 ๅฐๆ—ถ็š„๏ผŸ
1.0
ๆ–‡็ซ ๅˆ†ไบซใ€Šๆญ็ง˜่‹นๆžœๅบ”็”จๅฎกๆ ธๅ›ข้˜Ÿ๏ผˆๅฒไธŠๆœ€ๅ…จ็‰ˆ๏ผ‰ใ€‹ - ## ๆ–‡็ซ ๆ ‡้ข˜ใ€่‘—ๅๅ‡บๅค„ๅ’Œไฝœ่€… - ๆ–‡็ซ ๆ ‡้ข˜๏ผšๆญ็ง˜่‹นๆžœๅบ”็”จๅฎกๆ ธๅ›ข้˜Ÿ๏ผˆๅฒไธŠๆœ€ๅ…จ็‰ˆ๏ผ‰ - ๅ‡บๅค„๏ผšๆŽ˜้‡‘ - ไฝœ่€…๏ผš37ๆ‰‹ๆธธiOSๆŠ€ๆœฏ่ฟ่ฅๅ›ข้˜Ÿ ## ๆ–‡็ซ ้“พๆŽฅ https://juejin.cn/post/6970363897668698148/ ## ๆŽจ่็†็”ฑ๏ผˆๅฆ‚ๆžœๆฒกๆœ‰ๅฏไปฅๆ‘˜ๆŠ„ๆ–‡็ซ ๅ†…ไธ€ๆฎต้‡่ฆ็š„่ฏ๏ผ‰ ๅฒไธŠๆœ€ๅ…จ็‰ˆ๏ผšๆญ็ง˜่‹นๆžœๅบ”็”จๅฎกๆ ธๅ›ข้˜Ÿ๏ผŒๅ‘Š่ฏ‰ไฝ  App Store ็š„็”ฑๆฅ๏ผŒๆ˜ฏๆ€Žๆ ทๅ‘ๅฑ•ๅˆฐไปŠๅคฉ่ฟ™ๆ ท่พ‰็…Œ๏ผŒๅฆ‚ไฝ•่‡ชๅŠจๅŒ–๏ผšๆœบๅ™จๅฎกๆ ธ+ไบบๅทฅๅฎกๆ ธ๏ผŸๅฎกๆ ธ้€Ÿๅบฆๆ˜ฏๆ€ŽไนˆไปŽ7-10ๅคฉ๏ผŒๆๅ‡ๅˆฐๅฆ‚ไปŠๅช้œ€่ฆ 48 ๅฐๆ—ถ็š„๏ผŸ
non_process
ๆ–‡็ซ ๅˆ†ไบซใ€Šๆญ็ง˜่‹นๆžœๅบ”็”จๅฎกๆ ธๅ›ข้˜Ÿ๏ผˆๅฒไธŠๆœ€ๅ…จ็‰ˆ๏ผ‰ใ€‹ ๆ–‡็ซ ๆ ‡้ข˜ใ€่‘—ๅๅ‡บๅค„ๅ’Œไฝœ่€… ๆ–‡็ซ ๆ ‡้ข˜๏ผšๆญ็ง˜่‹นๆžœๅบ”็”จๅฎกๆ ธๅ›ข้˜Ÿ๏ผˆๅฒไธŠๆœ€ๅ…จ็‰ˆ๏ผ‰ ๅ‡บๅค„๏ผšๆŽ˜้‡‘ ไฝœ่€…๏ผš ๆ–‡็ซ ้“พๆŽฅ ๆŽจ่็†็”ฑ๏ผˆๅฆ‚ๆžœๆฒกๆœ‰ๅฏไปฅๆ‘˜ๆŠ„ๆ–‡็ซ ๅ†…ไธ€ๆฎต้‡่ฆ็š„่ฏ๏ผ‰ ๅฒไธŠๆœ€ๅ…จ็‰ˆ๏ผšๆญ็ง˜่‹นๆžœๅบ”็”จๅฎกๆ ธๅ›ข้˜Ÿ๏ผŒๅ‘Š่ฏ‰ไฝ  app store ็š„็”ฑๆฅ๏ผŒๆ˜ฏๆ€Žๆ ทๅ‘ๅฑ•ๅˆฐไปŠๅคฉ่ฟ™ๆ ท่พ‰็…Œ๏ผŒๅฆ‚ไฝ•่‡ชๅŠจๅŒ–๏ผšๆœบๅ™จๅฎกๆ ธ ไบบๅทฅๅฎกๆ ธ๏ผŸ ๏ผŒๆๅ‡ๅˆฐๅฆ‚ไปŠๅช้œ€่ฆ ๅฐๆ—ถ็š„๏ผŸ
0
8,912
12,016,759,356
IssuesEvent
2020-04-10 16:49:25
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Error: You must first sign the relevant contracts online.
process: release stage: needs investigating
### Current behavior: We're getting an error thrown from `cypress/node_modules/electron-notarize` during building of the binary in `develop` branch here: https://circleci.com/gh/cypress-io/cypress/297860 ``` 1 package(s) were not uploaded because they had problems: /var/folders/1b/gl7yt7ds26vcyr1pkgld6l040000gn/T/393FBFE1-5903-4AF1-893C-3545527D7A99/com.electron.cypress.itmsp - Error Messages: You must first sign the relevant contracts online. (1048) 2020-04-07 10:06:18.588 altool[1215:8304] *** Error: You must first sign the relevant contracts online. (1048) at /Users/distiller/cypress/node_modules/electron-notarize/src/index.ts:83:13 at Generator.next (<anonymous>) at fulfilled (/Users/distiller/cypress/node_modules/electron-notarize/lib/index.js:4:58) at processTicksAndRejections (internal/process/task_queues.js:93:5) ๐Ÿ”ฅ deploy error ``` ### Versions `develop` branch - failed checks from this commit specifically https://github.com/cypress-io/cypress/commit/f47f3d64648f09b5b922eb423375580ff8d4902c
1.0
Error: You must first sign the relevant contracts online. - ### Current behavior: We're getting an error thrown from `cypress/node_modules/electron-notarize` during building of the binary in `develop` branch here: https://circleci.com/gh/cypress-io/cypress/297860 ``` 1 package(s) were not uploaded because they had problems: /var/folders/1b/gl7yt7ds26vcyr1pkgld6l040000gn/T/393FBFE1-5903-4AF1-893C-3545527D7A99/com.electron.cypress.itmsp - Error Messages: You must first sign the relevant contracts online. (1048) 2020-04-07 10:06:18.588 altool[1215:8304] *** Error: You must first sign the relevant contracts online. (1048) at /Users/distiller/cypress/node_modules/electron-notarize/src/index.ts:83:13 at Generator.next (<anonymous>) at fulfilled (/Users/distiller/cypress/node_modules/electron-notarize/lib/index.js:4:58) at processTicksAndRejections (internal/process/task_queues.js:93:5) ๐Ÿ”ฅ deploy error ``` ### Versions `develop` branch - failed checks from this commit specifically https://github.com/cypress-io/cypress/commit/f47f3d64648f09b5b922eb423375580ff8d4902c
process
error you must first sign the relevant contracts online current behavior we re getting an error thrown from cypress node modules electron notarize during building of the binary in develop branch here package s were not uploaded because they had problems var folders t com electron cypress itmsp error messages you must first sign the relevant contracts online altool error you must first sign the relevant contracts online at users distiller cypress node modules electron notarize src index ts at generator next at fulfilled users distiller cypress node modules electron notarize lib index js at processticksandrejections internal process task queues js ๐Ÿ”ฅ deploy error versions develop branch failed checks from this commit specifically
1
20,134
26,678,187,513
IssuesEvent
2023-01-26 15:43:35
apache/arrow-datafusion
https://api.github.com/repos/apache/arrow-datafusion
closed
Add `release-crates` bash script
enhancement good first issue help wanted development-process
**Is your feature request related to a problem or challenge? Please describe what you are trying to do.** One part of the release process that is time-consuming is running the individual `cargo publish` commands. The `cargo publish` process has improved recently and the command will not complete until the crate has been propagated to the crates.io index so this makes it more compelling to automate this process. **Describe the solution you'd like** Write a script that runs the following commands. The script should check that the command is not being run from a git repository (we need to run from the downloaded tarball). The documentation at `dev/release/README.md` should also be updated. ```shell (cd datafusion/common && cargo publish) (cd datafusion/expr && cargo publish) (cd datafusion/sql && cargo publish) (cd datafusion/jit && cargo publish) (cd datafusion/row && cargo publish) (cd datafusion/physical-expr && cargo publish) (cd datafusion/optimizer && cargo publish) (cd datafusion/core && cargo publish) (cd datafusion/proto && cargo publish) (cd datafusion-cli && cargo publish --no-verify) ``` **Describe alternatives you've considered** Keep doing this manually **Additional context** None
1.0
Add `release-crates` bash script - **Is your feature request related to a problem or challenge? Please describe what you are trying to do.** One part of the release process that is time-consuming is running the individual `cargo publish` commands. The `cargo publish` process has improved recently and the command will not complete until the crate has been propagated to the crates.io index so this makes it more compelling to automate this process. **Describe the solution you'd like** Write a script that runs the following commands. The script should check that the command is not being run from a git repository (we need to run from the downloaded tarball). The documentation at `dev/release/README.md` should also be updated. ```shell (cd datafusion/common && cargo publish) (cd datafusion/expr && cargo publish) (cd datafusion/sql && cargo publish) (cd datafusion/jit && cargo publish) (cd datafusion/row && cargo publish) (cd datafusion/physical-expr && cargo publish) (cd datafusion/optimizer && cargo publish) (cd datafusion/core && cargo publish) (cd datafusion/proto && cargo publish) (cd datafusion-cli && cargo publish --no-verify) ``` **Describe alternatives you've considered** Keep doing this manually **Additional context** None
process
add release crates bash script is your feature request related to a problem or challenge please describe what you are trying to do one part of the release process that is time consuming is running the individual cargo publish commands the cargo publish process has improved recently and the command will not complete until the crate has been propagated to the crates io index so this makes it more compelling to automate this process describe the solution you d like write a script that runs the following commands the script should check that the command is not being run from a git repository we need to run from the downloaded tarball the documentation at dev release readme md should also be updated shell cd datafusion common cargo publish cd datafusion expr cargo publish cd datafusion sql cargo publish cd datafusion jit cargo publish cd datafusion row cargo publish cd datafusion physical expr cargo publish cd datafusion optimizer cargo publish cd datafusion core cargo publish cd datafusion proto cargo publish cd datafusion cli cargo publish no verify describe alternatives you ve considered keep doing this manually additional context none
1
538,275
15,765,676,695
IssuesEvent
2021-03-31 14:22:06
snowplow/snowplow-javascript-tracker
https://api.github.com/repos/snowplow/snowplow-javascript-tracker
closed
Add linting
category:setup priority:medium status:in_progress type:internal
I think this would help a bit, maybe a combination of these in a pre-commit hook: - https://github.com/prettier/prettier - https://eslint.org/
1.0
Add linting - I think this would help a bit, maybe a combination of these in a pre-commit hook: - https://github.com/prettier/prettier - https://eslint.org/
non_process
add linting i think this would help a bit maybe a combination of these in a pre commit hook
0
606
3,075,332,786
IssuesEvent
2015-08-20 13:09:24
maraujop/django-crispy-forms
https://api.github.com/repos/maraujop/django-crispy-forms
closed
Consider testing against Django versions 1.4, 1.4.1, 1.4.2, 1.4.3
Testing/Process
In current test suite there are 2 places with conditions related to these versions. [Here](https://github.com/maraujop/django-crispy-forms/blob/dev/crispy_forms/tests/test_layout.py#L283) and [here](https://github.com/maraujop/django-crispy-forms/blob/dev/crispy_forms/tests/test_layout.py#L348) Also in travis config there are only latest versions of 1.4, 1.7, 1.8 and master branches. My question is - Should we drop these conditions or add old Django version to config? Personally, I prefer to drop.
1.0
Consider testing against Django versions 1.4, 1.4.1, 1.4.2, 1.4.3 - In current test suite there are 2 places with conditions related to these versions. [Here](https://github.com/maraujop/django-crispy-forms/blob/dev/crispy_forms/tests/test_layout.py#L283) and [here](https://github.com/maraujop/django-crispy-forms/blob/dev/crispy_forms/tests/test_layout.py#L348) Also in travis config there are only latest versions of 1.4, 1.7, 1.8 and master branches. My question is - Should we drop these conditions or add old Django version to config? Personally, I prefer to drop.
process
consider testing against django versions in current test suite there are places with conditions related to these versions and also in travis config there are only latest versions of and master branches my question is should we drop these conditions or add old django version to config personally i prefer to drop
1
18,124
25,040,844,971
IssuesEvent
2022-11-04 20:34:43
dotnet/docs
https://api.github.com/repos/dotnet/docs
closed
[Breaking change]: Updating CreationTime[Utc], LastAccessTime[Utc] and LastWriteTime[Utc] on a symbolic link no longer affects the target
doc-idea breaking-change Pri1 source incompatible :checkered_flag: Release: .NET 7 in-pr
### Description Updating those fields on a symbolic link had the effect of not change the information in the symlink itself but instead it was affecting the target. ```cs string filename = "file"; string linkname = "link"; File.Create(filename).Dispose(); File.CreateSymbolicLink(linkname, filename); Console.WriteLine("Before update:"); PrintMetadata(filename); PrintMetadata(linkname); UpdateMetadata(linkname); Console.WriteLine("\nAfter update:"); PrintMetadata(filename); PrintMetadata(linkname); static void UpdateMetadata(string filename) { DateTime tomorrow = DateTime.Now.AddDays(1); File.SetCreationTime(filename, tomorrow); File.SetLastAccessTime(filename, tomorrow); File.SetLastWriteTime(filename, tomorrow); File.SetAttributes(filename, File.GetAttributes(filename) | FileAttributes.Offline); } static void PrintMetadata(string filename) { Console.WriteLine($"---{filename}---"); Console.WriteLine("Creation:\t" + File.GetCreationTime(filename)); Console.WriteLine("Last access:\t" + File.GetLastAccessTime(filename)); Console.WriteLine("Last write:\t" + File.GetLastWriteTime(filename)); Console.WriteLine("Attributes:\t" + File.GetAttributes(filename)); } ``` Version: .NET 7 preview 1 ### Version Other (please put exact version in description textbox) ### Previous behavior Updating any of the aforementioned fields on a symlink would affect the fields of the symlink's target. Output form above program: ``` Before update: ---file--- Creation: 9/29/2022 10:35:40 AM Last access: 9/29/2022 10:35:40 AM Last write: 9/29/2022 10:35:40 AM Attributes: Archive ---link--- Creation: 9/29/2022 10:35:40 AM Last access: 9/29/2022 10:35:40 AM Last write: 9/29/2022 10:35:40 AM Attributes: Archive, ReparsePoint After update: ---file--- Creation: 9/30/2022 10:35:40 AM Last access: 9/30/2022 10:35:40 AM Last write: 9/30/2022 10:35:40 AM Attributes: Archive ---link--- Creation: 9/29/2022 10:35:40 AM Last access: 9/29/2022 10:35:40 AM Last write: 9/29/2022 10:35:40 AM Attributes: Archive, ReparsePoint, Offline ``` ### New behavior Updating any of the aforementioned fields on a symlink would affect the fields of the symlink itself. Output: ``` Before update: ---file--- Creation: 9/29/2022 10:33:39 AM Last access: 9/29/2022 10:33:39 AM Last write: 9/29/2022 10:33:39 AM Attributes: Archive ---link--- Creation: 9/29/2022 10:33:39 AM Last access: 9/29/2022 10:33:39 AM Last write: 9/29/2022 10:33:39 AM Attributes: Archive, ReparsePoint After update: ---file--- Creation: 9/29/2022 10:33:39 AM Last access: 9/29/2022 10:33:39 AM Last write: 9/29/2022 10:33:39 AM Attributes: Archive ---link--- Creation: 9/30/2022 10:33:39 AM Last access: 9/30/2022 10:33:39 AM Last write: 9/30/2022 10:33:39 AM Attributes: Archive, ReparsePoint, Offline ``` ### Type of breaking change - [ ] **Binary incompatible**: Existing binaries may encounter a breaking change in behavior, such as failure to load/execute or different run-time behavior. - [X] **Source incompatible**: Source code may encounter a breaking change in behavior when targeting the new runtime/component/SDK, such as compile errors or different run-time behavior. ### Reason for change Behavior was determined to be unexpected and undesirable in some cases, inconsistent with the behavior of the properties/methods that `get` the same fields, and it was impossible to actually update said fields in the symlink itself thorught .NET APIs, so we switched it for consistency. ### Recommended action If you were relying in this behavior, be advised that setting one of the `*Time` fields in a symlink won't affect the target anymore. You can use the new symbolic link APIs to obtain the target of a symlink and proceed to update that file system object instaed. ```cs FileSystemInfo? targetInfo = linkInfo.ResolveLinkTarget(returnFinalTarget: true); if (targetInfo != null) { // update the properties accordingly. targetInfo.LastWriteTime = DateTime.Now; } ``` ### Feature area Core .NET libraries ### Affected APIs System.IO.File.SetCreationTime(String, DateTime) System.IO.File.SetCreationTimeUtc(String, DateTime) System.IO.File.SetLastAccessTime(String, DateTime) System.IO.File.SetLastAccessTimeUtc(String, DateTime) System.IO.File.SetLastWriteTime(String, DateTime) System.IO.File.SetLastWriteTimeUtc(String, DateTime) System.IO.Directory.SetCreationTime(String, DateTime) System.IO.Directory.SetCreationTimeUtc(String, DateTime) System.IO.Directory.SetLastAccessTime(String, DateTime) System.IO.Directory.SetLastAccessTimeUtc(String, DateTime) System.IO.Directory.SetLastWriteTime(String, DateTime) System.IO.Directory.SetLastWriteTimeUtc(String, DateTime) System.IO.FileSystemInfo.CreationTime System.IO.FileSystemInfo.CreationTimeUtc System.IO.FileSystemInfo.LastAccessTime System.IO.FileSystemInfo.LastAccessTimeUtc System.IO.FileSystemInfo.LastWriteTime System.IO.FileSystemInfo.LastWriteTimeUtc
True
[Breaking change]: Updating CreationTime[Utc], LastAccessTime[Utc] and LastWriteTime[Utc] on a symbolic link no longer affects the target - ### Description Updating those fields on a symbolic link had the effect of not change the information in the symlink itself but instead it was affecting the target. ```cs string filename = "file"; string linkname = "link"; File.Create(filename).Dispose(); File.CreateSymbolicLink(linkname, filename); Console.WriteLine("Before update:"); PrintMetadata(filename); PrintMetadata(linkname); UpdateMetadata(linkname); Console.WriteLine("\nAfter update:"); PrintMetadata(filename); PrintMetadata(linkname); static void UpdateMetadata(string filename) { DateTime tomorrow = DateTime.Now.AddDays(1); File.SetCreationTime(filename, tomorrow); File.SetLastAccessTime(filename, tomorrow); File.SetLastWriteTime(filename, tomorrow); File.SetAttributes(filename, File.GetAttributes(filename) | FileAttributes.Offline); } static void PrintMetadata(string filename) { Console.WriteLine($"---{filename}---"); Console.WriteLine("Creation:\t" + File.GetCreationTime(filename)); Console.WriteLine("Last access:\t" + File.GetLastAccessTime(filename)); Console.WriteLine("Last write:\t" + File.GetLastWriteTime(filename)); Console.WriteLine("Attributes:\t" + File.GetAttributes(filename)); } ``` Version: .NET 7 preview 1 ### Version Other (please put exact version in description textbox) ### Previous behavior Updating any of the aforementioned fields on a symlink would affect the fields of the symlink's target. Output form above program: ``` Before update: ---file--- Creation: 9/29/2022 10:35:40 AM Last access: 9/29/2022 10:35:40 AM Last write: 9/29/2022 10:35:40 AM Attributes: Archive ---link--- Creation: 9/29/2022 10:35:40 AM Last access: 9/29/2022 10:35:40 AM Last write: 9/29/2022 10:35:40 AM Attributes: Archive, ReparsePoint After update: ---file--- Creation: 9/30/2022 10:35:40 AM Last access: 9/30/2022 10:35:40 AM Last write: 9/30/2022 10:35:40 AM Attributes: Archive ---link--- Creation: 9/29/2022 10:35:40 AM Last access: 9/29/2022 10:35:40 AM Last write: 9/29/2022 10:35:40 AM Attributes: Archive, ReparsePoint, Offline ``` ### New behavior Updating any of the aforementioned fields on a symlink would affect the fields of the symlink itself. Output: ``` Before update: ---file--- Creation: 9/29/2022 10:33:39 AM Last access: 9/29/2022 10:33:39 AM Last write: 9/29/2022 10:33:39 AM Attributes: Archive ---link--- Creation: 9/29/2022 10:33:39 AM Last access: 9/29/2022 10:33:39 AM Last write: 9/29/2022 10:33:39 AM Attributes: Archive, ReparsePoint After update: ---file--- Creation: 9/29/2022 10:33:39 AM Last access: 9/29/2022 10:33:39 AM Last write: 9/29/2022 10:33:39 AM Attributes: Archive ---link--- Creation: 9/30/2022 10:33:39 AM Last access: 9/30/2022 10:33:39 AM Last write: 9/30/2022 10:33:39 AM Attributes: Archive, ReparsePoint, Offline ``` ### Type of breaking change - [ ] **Binary incompatible**: Existing binaries may encounter a breaking change in behavior, such as failure to load/execute or different run-time behavior. - [X] **Source incompatible**: Source code may encounter a breaking change in behavior when targeting the new runtime/component/SDK, such as compile errors or different run-time behavior. ### Reason for change Behavior was determined to be unexpected and undesirable in some cases, inconsistent with the behavior of the properties/methods that `get` the same fields, and it was impossible to actually update said fields in the symlink itself thorught .NET APIs, so we switched it for consistency. ### Recommended action If you were relying in this behavior, be advised that setting one of the `*Time` fields in a symlink won't affect the target anymore. You can use the new symbolic link APIs to obtain the target of a symlink and proceed to update that file system object instaed. ```cs FileSystemInfo? targetInfo = linkInfo.ResolveLinkTarget(returnFinalTarget: true); if (targetInfo != null) { // update the properties accordingly. targetInfo.LastWriteTime = DateTime.Now; } ``` ### Feature area Core .NET libraries ### Affected APIs System.IO.File.SetCreationTime(String, DateTime) System.IO.File.SetCreationTimeUtc(String, DateTime) System.IO.File.SetLastAccessTime(String, DateTime) System.IO.File.SetLastAccessTimeUtc(String, DateTime) System.IO.File.SetLastWriteTime(String, DateTime) System.IO.File.SetLastWriteTimeUtc(String, DateTime) System.IO.Directory.SetCreationTime(String, DateTime) System.IO.Directory.SetCreationTimeUtc(String, DateTime) System.IO.Directory.SetLastAccessTime(String, DateTime) System.IO.Directory.SetLastAccessTimeUtc(String, DateTime) System.IO.Directory.SetLastWriteTime(String, DateTime) System.IO.Directory.SetLastWriteTimeUtc(String, DateTime) System.IO.FileSystemInfo.CreationTime System.IO.FileSystemInfo.CreationTimeUtc System.IO.FileSystemInfo.LastAccessTime System.IO.FileSystemInfo.LastAccessTimeUtc System.IO.FileSystemInfo.LastWriteTime System.IO.FileSystemInfo.LastWriteTimeUtc
non_process
updating creationtime lastaccesstime and lastwritetime on a symbolic link no longer affects the target description updating those fields on a symbolic link had the effect of not change the information in the symlink itself but instead it was affecting the target cs string filename file string linkname link file create filename dispose file createsymboliclink linkname filename console writeline before update printmetadata filename printmetadata linkname updatemetadata linkname console writeline nafter update printmetadata filename printmetadata linkname static void updatemetadata string filename datetime tomorrow datetime now adddays file setcreationtime filename tomorrow file setlastaccesstime filename tomorrow file setlastwritetime filename tomorrow file setattributes filename file getattributes filename fileattributes offline static void printmetadata string filename console writeline filename console writeline creation t file getcreationtime filename console writeline last access t file getlastaccesstime filename console writeline last write t file getlastwritetime filename console writeline attributes t file getattributes filename version net preview version other please put exact version in description textbox previous behavior updating any of the aforementioned fields on a symlink would affect the fields of the symlink s target output form above program before update file creation am last access am last write am attributes archive link creation am last access am last write am attributes archive reparsepoint after update file creation am last access am last write am attributes archive link creation am last access am last write am attributes archive reparsepoint offline new behavior updating any of the aforementioned fields on a symlink would affect the fields of the symlink itself output before update file creation am last access am last write am attributes archive link creation am last access am last write am attributes archive reparsepoint after update file creation am last access am last write am attributes archive link creation am last access am last write am attributes archive reparsepoint offline type of breaking change binary incompatible existing binaries may encounter a breaking change in behavior such as failure to load execute or different run time behavior source incompatible source code may encounter a breaking change in behavior when targeting the new runtime component sdk such as compile errors or different run time behavior reason for change behavior was determined to be unexpected and undesirable in some cases inconsistent with the behavior of the properties methods that get the same fields and it was impossible to actually update said fields in the symlink itself thorught net apis so we switched it for consistency recommended action if you were relying in this behavior be advised that setting one of the time fields in a symlink won t affect the target anymore you can use the new symbolic link apis to obtain the target of a symlink and proceed to update that file system object instaed cs filesysteminfo targetinfo linkinfo resolvelinktarget returnfinaltarget true if targetinfo null update the properties accordingly targetinfo lastwritetime datetime now feature area core net libraries affected apis system io file setcreationtime string datetime system io file setcreationtimeutc string datetime system io file setlastaccesstime string datetime system io file setlastaccesstimeutc string datetime system io file setlastwritetime string datetime system io file setlastwritetimeutc string datetime system io directory setcreationtime string datetime system io directory setcreationtimeutc string datetime system io directory setlastaccesstime string datetime system io directory setlastaccesstimeutc string datetime system io directory setlastwritetime string datetime system io directory setlastwritetimeutc string datetime system io filesysteminfo creationtime system io filesysteminfo creationtimeutc system io filesysteminfo lastaccesstime system io filesysteminfo lastaccesstimeutc system io filesysteminfo lastwritetime system io filesysteminfo lastwritetimeutc
0
21,107
28,068,462,260
IssuesEvent
2023-03-29 17:10:14
ethereum/EIPs
https://api.github.com/repos/ethereum/EIPs
closed
Dynamic Required Approvals
w-stale enhancement r-ci r-process e-consensus
> My recommendation was to have changes to non-EIPs require a decreasing number of approvals based on how long they have been open. Something like: > > 1 week: 5 approvals > 2 weeks: 4 approvals > 4 weeks: 3 approvals > 8 weeks: 2 approvals > 16 weeks: 1 approval > and a single rejection would prevent auto-merging. > > The idea here is that if there is consensus among all editors then things would go through quickly and smoothly. If, however, there aren't enough actively engaging editors then system doesn't permanently freeze/halt, it just slows down. _Originally posted by @MicahZoltu in https://github.com/ethereum/EIPs/issues/5640#issuecomment-1245065644_ My suggestion is a little bit more general. If all editors approve, then a change gets instantly merged. If an editor approves, it starts a one-month countdown before it is automatically merged. If subsequent editors approve, it halves that countdown. A "request changes" review pauses this countdown.
1.0
Dynamic Required Approvals - > My recommendation was to have changes to non-EIPs require a decreasing number of approvals based on how long they have been open. Something like: > > 1 week: 5 approvals > 2 weeks: 4 approvals > 4 weeks: 3 approvals > 8 weeks: 2 approvals > 16 weeks: 1 approval > and a single rejection would prevent auto-merging. > > The idea here is that if there is consensus among all editors then things would go through quickly and smoothly. If, however, there aren't enough actively engaging editors then system doesn't permanently freeze/halt, it just slows down. _Originally posted by @MicahZoltu in https://github.com/ethereum/EIPs/issues/5640#issuecomment-1245065644_ My suggestion is a little bit more general. If all editors approve, then a change gets instantly merged. If an editor approves, it starts a one-month countdown before it is automatically merged. If subsequent editors approve, it halves that countdown. A "request changes" review pauses this countdown.
process
dynamic required approvals my recommendation was to have changes to non eips require a decreasing number of approvals based on how long they have been open something like week approvals weeks approvals weeks approvals weeks approvals weeks approval and a single rejection would prevent auto merging the idea here is that if there is consensus among all editors then things would go through quickly and smoothly if however there aren t enough actively engaging editors then system doesn t permanently freeze halt it just slows down originally posted by micahzoltu in my suggestion is a little bit more general if all editors approve then a change gets instantly merged if an editor approves it starts a one month countdown before it is automatically merged if subsequent editors approve it halves that countdown a request changes review pauses this countdown
1
10,809
13,609,288,466
IssuesEvent
2020-09-23 04:50:15
googleapis/java-conformance-tests
https://api.github.com/repos/googleapis/java-conformance-tests
closed
Dependency Dashboard
external type: process
This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/com.coveo-fmt-maven-plugin-2.x -->[build(deps): update dependency com.coveo:fmt-maven-plugin to v2.10](../pull/188) - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-bom-0.x -->[deps: update dependency com.google.cloud:google-cloud-bom to v0.135.0](../pull/189) - [ ] <!-- rebase-branch=renovate/com.google.googlejavaformat-google-java-format-1.x -->[deps: update dependency com.google.googlejavaformat:google-java-format to v1.9](../pull/190) - [ ] <!-- rebase-all-open-prs -->**Check this option to rebase all the above open PRs at once** --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/com.coveo-fmt-maven-plugin-2.x -->[build(deps): update dependency com.coveo:fmt-maven-plugin to v2.10](../pull/188) - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-bom-0.x -->[deps: update dependency com.google.cloud:google-cloud-bom to v0.135.0](../pull/189) - [ ] <!-- rebase-branch=renovate/com.google.googlejavaformat-google-java-format-1.x -->[deps: update dependency com.google.googlejavaformat:google-java-format to v1.9](../pull/190) - [ ] <!-- rebase-all-open-prs -->**Check this option to rebase all the above open PRs at once** --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any pull pull pull check this option to rebase all the above open prs at once check this box to trigger a request for renovate to run again on this repository
1
19,085
25,133,802,837
IssuesEvent
2022-11-09 16:58:45
paul-buerkner/brms
https://api.github.com/repos/paul-buerkner/brms
closed
Slowdown in emmeans since version 2.14.0
efficiency post-processing
It appears that calls to emmeans have substantially slowed since brms version 2.14.0. I've noticed this slowdown in a large model with an ar(1) structure but it could also be present more generally. Here is a quick example showing what I mean. I created a little dataset complicated enough to bring out the differences between brms versions for a call to an emmeans function (emtrends) - link to model [test_model_2022-11-07.rds - Google Drive](https://drive.google.com/file/d/1_8GrpQIZ1nTaeuDYfkTCxwCmAPFP8gLX/view?usp=sharing). The upshot is that brms v 2.18 takes about twice as long to do the emtrends call as brms v 2.14.0. In my actual model which is much larger than the example below, 2.14.0 is at least 10-20 times faster, making calls to emmeans in 2.18 almost unusable. A quick profile of the calls in each of versions suggested that it might have been due to differences in how the ref_grid is created. It seems that the ref_grid in 2.18 includes the time and grouping factor from the ar(1) structure while brms v 2.14.0 doesnโ€™t. Thanks! Original discussion: https://discourse.mc-stan.org/t/brms-and-emmeans-with-incl-autocor-argument/29371/4 ``` # Fake dataset with trend, grouping factors, largish x1 <- c(rep('A', 3000), rep('B', 3000)) x2 <- rnorm(6000, mean = 2, sd = 2) y_orig <- rnorm(6000, mean = 10, sd = 5) y_mod <- if_else(x1 == 'A', y_orig-x2, y_orig+x2) gr <- rep(1:60, each = 100) gr_mod <- rnorm(60, mean = 1, sd = .5) gr_mod2 <- rep(gr_mod, each = 100) t <- rep(1:100, 60) y_trend <- y_mod+y_mod*t/100 y_final <- y_trend + gr_mod2 df <- data.frame(y = y_final, x1 = x1, x2 = x2, gr = gr, t = t) fit <- brm(y ~ x1*x2 + (1 | gr) + ar(p = 1, time = t, gr = gr), data = df, chains = 4, cores = 4, file = 'test_model_2022-11-07') summary(fit) # system.time(emm1 <- emtrends(fit, ~ x1, var = 'x2')) emm1 <- emtrends(fit, ~ x1, var = 'x2') emm1 # Time for brms version 2.14.0: 0.861, 0.650, 0.850 (mean: 0.787) # Ref grid for brms version 2.14.0 # > ref_grid(fit, ~ x1, var = 'x2') # 'emmGrid' object with variables: # x1 = A, B # x2 = 1.9879 # Time for brms version 2.18: 1.674, 1.246, 1.363 (mean: 1.43; x1.82 longer) # Ref grid for brms version 2.18 # > ref_grid(fit, ~ x1, var = 'x2') # 'emmGrid' object with variables: # x1 = A, B # x2 = 1.9879 # t = 50.5 # gr = 30.5 ```
1.0
Slowdown in emmeans since version 2.14.0 - It appears that calls to emmeans have substantially slowed since brms version 2.14.0. I've noticed this slowdown in a large model with an ar(1) structure but it could also be present more generally. Here is a quick example showing what I mean. I created a little dataset complicated enough to bring out the differences between brms versions for a call to an emmeans function (emtrends) - link to model [test_model_2022-11-07.rds - Google Drive](https://drive.google.com/file/d/1_8GrpQIZ1nTaeuDYfkTCxwCmAPFP8gLX/view?usp=sharing). The upshot is that brms v 2.18 takes about twice as long to do the emtrends call as brms v 2.14.0. In my actual model which is much larger than the example below, 2.14.0 is at least 10-20 times faster, making calls to emmeans in 2.18 almost unusable. A quick profile of the calls in each of versions suggested that it might have been due to differences in how the ref_grid is created. It seems that the ref_grid in 2.18 includes the time and grouping factor from the ar(1) structure while brms v 2.14.0 doesnโ€™t. Thanks! Original discussion: https://discourse.mc-stan.org/t/brms-and-emmeans-with-incl-autocor-argument/29371/4 ``` # Fake dataset with trend, grouping factors, largish x1 <- c(rep('A', 3000), rep('B', 3000)) x2 <- rnorm(6000, mean = 2, sd = 2) y_orig <- rnorm(6000, mean = 10, sd = 5) y_mod <- if_else(x1 == 'A', y_orig-x2, y_orig+x2) gr <- rep(1:60, each = 100) gr_mod <- rnorm(60, mean = 1, sd = .5) gr_mod2 <- rep(gr_mod, each = 100) t <- rep(1:100, 60) y_trend <- y_mod+y_mod*t/100 y_final <- y_trend + gr_mod2 df <- data.frame(y = y_final, x1 = x1, x2 = x2, gr = gr, t = t) fit <- brm(y ~ x1*x2 + (1 | gr) + ar(p = 1, time = t, gr = gr), data = df, chains = 4, cores = 4, file = 'test_model_2022-11-07') summary(fit) # system.time(emm1 <- emtrends(fit, ~ x1, var = 'x2')) emm1 <- emtrends(fit, ~ x1, var = 'x2') emm1 # Time for brms version 2.14.0: 0.861, 0.650, 0.850 (mean: 0.787) # Ref grid for brms version 2.14.0 # > ref_grid(fit, ~ x1, var = 'x2') # 'emmGrid' object with variables: # x1 = A, B # x2 = 1.9879 # Time for brms version 2.18: 1.674, 1.246, 1.363 (mean: 1.43; x1.82 longer) # Ref grid for brms version 2.18 # > ref_grid(fit, ~ x1, var = 'x2') # 'emmGrid' object with variables: # x1 = A, B # x2 = 1.9879 # t = 50.5 # gr = 30.5 ```
process
slowdown in emmeans since version it appears that calls to emmeans have substantially slowed since brms version i ve noticed this slowdown in a large model with an ar structure but it could also be present more generally here is a quick example showing what i mean i created a little dataset complicated enough to bring out the differences between brms versions for a call to an emmeans function emtrends link to model the upshot is that brms v takes about twice as long to do the emtrends call as brms v in my actual model which is much larger than the example below is at least times faster making calls to emmeans in almost unusable a quick profile of the calls in each of versions suggested that it might have been due to differences in how the ref grid is created it seems that the ref grid in includes the time and grouping factor from the ar structure while brms v doesnโ€™t thanks original discussion fake dataset with trend grouping factors largish c rep a rep b rnorm mean sd y orig rnorm mean sd y mod if else a y orig y orig gr rep each gr mod rnorm mean sd gr rep gr mod each t rep y trend y mod y mod t y final y trend gr df data frame y y final gr gr t t fit brm y gr ar p time t gr gr data df chains cores file test model summary fit system time emtrends fit var emtrends fit var time for brms version mean ref grid for brms version ref grid fit var emmgrid object with variables a b time for brms version mean longer ref grid for brms version ref grid fit var emmgrid object with variables a b t gr
1
19,736
26,085,032,577
IssuesEvent
2022-12-26 00:53:53
ankidroid/Anki-Android
https://api.github.com/repos/ankidroid/Anki-Android
closed
Add lint for `Build.VERSION.SDK_INT` in test
Good First Issue! Stale Test process
I made a mistake that was not seen by the reviewers. Essentially `Build.VERSION.SDK_INT` is 0 unless we use `@Config(sdk = [21, 26])` or `targetApi`. So, I would love a linter that ensure that if either this constant or `CompatHelper` is used, then we configure the SDK.
1.0
Add lint for `Build.VERSION.SDK_INT` in test - I made a mistake that was not seen by the reviewers. Essentially `Build.VERSION.SDK_INT` is 0 unless we use `@Config(sdk = [21, 26])` or `targetApi`. So, I would love a linter that ensure that if either this constant or `CompatHelper` is used, then we configure the SDK.
process
add lint for build version sdk int in test i made a mistake that was not seen by the reviewers essentially build version sdk int is unless we use config sdk or targetapi so i would love a linter that ensure that if either this constant or compathelper is used then we configure the sdk
1
17,006
22,386,203,980
IssuesEvent
2022-06-17 00:50:17
figlesias221/ProyectoDevOps_Grupo3_IglesiasPerezMolinoloJuan
https://api.github.com/repos/figlesias221/ProyectoDevOps_Grupo3_IglesiasPerezMolinoloJuan
closed
Subir issues a github resultantes del testing exploratorio y deuda tรฉcnica
task process
Integrantes: Andrรฉs, Federico y Matรญas Esfuerzo en HS-P: - ESTIMADO: 0 (no hay estimaciones la primera iteraciรณn) - REAL: 0.5 (por integrante)
1.0
Subir issues a github resultantes del testing exploratorio y deuda tรฉcnica - Integrantes: Andrรฉs, Federico y Matรญas Esfuerzo en HS-P: - ESTIMADO: 0 (no hay estimaciones la primera iteraciรณn) - REAL: 0.5 (por integrante)
process
subir issues a github resultantes del testing exploratorio y deuda tรฉcnica integrantes andrรฉs federico y matรญas esfuerzo en hs p estimado no hay estimaciones la primera iteraciรณn real por integrante
1
13,007
15,366,217,287
IssuesEvent
2021-03-02 00:56:38
bridgetownrb/bridgetown
https://api.github.com/repos/bridgetownrb/bridgetown
opened
Research improved code patterns using the Adornable gem
process
I just came across this: https://github.com/kjleitz/adornable and I freakin' love that DSL. ๐Ÿ˜ There are a lot of code patterns which could be cleaned up/DRYed up using that. Will investigate.
1.0
Research improved code patterns using the Adornable gem - I just came across this: https://github.com/kjleitz/adornable and I freakin' love that DSL. ๐Ÿ˜ There are a lot of code patterns which could be cleaned up/DRYed up using that. Will investigate.
process
research improved code patterns using the adornable gem i just came across this and i freakin love that dsl ๐Ÿ˜ there are a lot of code patterns which could be cleaned up dryed up using that will investigate
1
139,425
11,267,981,006
IssuesEvent
2020-01-14 04:24:06
xqrzd/kudu-client-net
https://api.github.com/repos/xqrzd/kudu-client-net
opened
Tests sometimes stall
flaky test
Even the unit tests stall; complete deadlock. Maybe thread pool exhausted? Happens very rarely on CI build.
1.0
Tests sometimes stall - Even the unit tests stall; complete deadlock. Maybe thread pool exhausted? Happens very rarely on CI build.
non_process
tests sometimes stall even the unit tests stall complete deadlock maybe thread pool exhausted happens very rarely on ci build
0
7,750
10,865,574,407
IssuesEvent
2019-11-14 19:18:01
kubeflow/community
https://api.github.com/repos/kubeflow/community
closed
v0.7 blog post
area/docs kind/feature kind/process priority/p0
Opening this issue to track the blog post for 0.7. @jbottum has created a draft blog post. /assign @jbottum
1.0
v0.7 blog post - Opening this issue to track the blog post for 0.7. @jbottum has created a draft blog post. /assign @jbottum
process
blog post opening this issue to track the blog post for jbottum has created a draft blog post assign jbottum
1
21,798
11,387,029,999
IssuesEvent
2020-01-29 14:21:23
elastic/apm-agent-java
https://api.github.com/repos/elastic/apm-agent-java
closed
Instrument implementations of interfaces more efficiently
component: instrumentation performance
I have an idea to efficiently match all implementations of an interface, such as `org.asynchttpclient.AsyncHandler`. The problem is that those may be implemented as an anonymous inner class so we canโ€™t pre-filter on the name, like `endsWith("Handler")` which means we have to check the whole inheritance hierarchy of every class in order to determine whether a class implements AsyncHandler. Instead of eagerly searching for all implementations, we could just check in `DefaultAsyncHttpClient#executeRequest(Request, AsyncHandler)` whether we have already instrumented the provided `AsyncHandler`. If we did not, we can just retransform the class and apply the corresponding instrumentation to it. We currently donโ€™t have a facility to do that but should not be too complex to doโ„ข. The benefit is faster startups of the application.
True
Instrument implementations of interfaces more efficiently - I have an idea to efficiently match all implementations of an interface, such as `org.asynchttpclient.AsyncHandler`. The problem is that those may be implemented as an anonymous inner class so we canโ€™t pre-filter on the name, like `endsWith("Handler")` which means we have to check the whole inheritance hierarchy of every class in order to determine whether a class implements AsyncHandler. Instead of eagerly searching for all implementations, we could just check in `DefaultAsyncHttpClient#executeRequest(Request, AsyncHandler)` whether we have already instrumented the provided `AsyncHandler`. If we did not, we can just retransform the class and apply the corresponding instrumentation to it. We currently donโ€™t have a facility to do that but should not be too complex to doโ„ข. The benefit is faster startups of the application.
non_process
instrument implementations of interfaces more efficiently i have an idea to efficiently match all implementations of an interface such as org asynchttpclient asynchandler the problem is that those may be implemented as an anonymous inner class so we canโ€™t pre filter on the name like endswith handler which means we have to check the whole inheritance hierarchy of every class in order to determine whether a class implements asynchandler instead of eagerly searching for all implementations we could just check in defaultasynchttpclient executerequest request asynchandler whether we have already instrumented the provided asynchandler if we did not we can just retransform the class and apply the corresponding instrumentation to it we currently donโ€™t have a facility to do that but should not be too complex to doโ„ข the benefit is faster startups of the application
0
9,187
12,228,676,000
IssuesEvent
2020-05-03 20:26:57
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
[processing] SumLineLengths algorithm returns a wrong (!) length sum
Bug Processing
Sample GPKG data: [sumlinelenghts.zip](https://github.com/qgis/QGIS/files/4566765/sumlinelenghts.zip) + CRS: EPSG:3117 (projected CRS) + Contains a polygon and a line layer. Expected length sum: **51.168**,291572545866 [m] Obtained length sum: **1.243'634.138**,1295323 [m] !!!
1.0
[processing] SumLineLengths algorithm returns a wrong (!) length sum - Sample GPKG data: [sumlinelenghts.zip](https://github.com/qgis/QGIS/files/4566765/sumlinelenghts.zip) + CRS: EPSG:3117 (projected CRS) + Contains a polygon and a line layer. Expected length sum: **51.168**,291572545866 [m] Obtained length sum: **1.243'634.138**,1295323 [m] !!!
process
sumlinelengths algorithm returns a wrong length sum sample gpkg data crs epsg projected crs contains a polygon and a line layer expected length sum obtained length sum
1
138,183
20,366,453,200
IssuesEvent
2022-02-21 06:28:17
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
switch expression returns double on int, long, float condition
Area-Compilers Resolution-By Design
### Description We came across a behavior that we would classify as bug. If you use a switch expression (with condition on int, long, float) in a function that returns an object, all values have as return type double. As soon as another types (like string, decimal) are present, everything works as expected. I didn't found any information about that if this is an expected behavior or not, so if it's expected, it would be great if someone could provide met the link to the documentation. Thanks! ### Configuration Which version of .NET is the code running on? - .NET 5/6 What OS and version, and for Linux, what distro? - macOS Monterey 12.1 What is the architecture (x64, x86, ARM, ARM64)? - x64 ### Example ```csharp using System; using System.Collections.Generic; public class Program { public static void Main() { var test = new List<object>() { 1, (long)1, 1d, 1f, }; foreach (var item in test) { Console.WriteLine( $"ValueIsOfType: {item.GetType().Name,-6} | SwitchExpression: {SwitchExpression(item).GetType().Name} | SwitchExpressionWithString: {SwitchExpression_WithString(item).GetType().Name,-6} | SwitchCase: {SwitchCase(item).GetType().Name}" ); } } private static object SwitchExpression(object x) { return x switch { int v => v, long v => v, double v => v, float v => v, _ => throw new Exception("Unexpected type"), }; } private static object SwitchExpression_WithString(object x) { return x switch { int v => v, long v => v, double v => v, float v => v, string v => v, _ => throw new Exception("Unexpected type"), }; } private static object SwitchCase(object x) { switch (x) { case int v: return v; case long v: return v; case double v: return v; case float v: return v; default: throw new Exception("Unexpected type"); } } } ``` Result: ``` ValueIsOfType: Int32 | SwitchExpression: Double | SwitchExpressionWithString: Int32 | SwitchCase: Int32 ValueIsOfType: Int64 | SwitchExpression: Double | SwitchExpressionWithString: Int64 | SwitchCase: Int64 ValueIsOfType: Double | SwitchExpression: Double | SwitchExpressionWithString: Double | SwitchCase: Double ValueIsOfType: Single | SwitchExpression: Double | SwitchExpressionWithString: Single | SwitchCase: Single ```
1.0
switch expression returns double on int, long, float condition - ### Description We came across a behavior that we would classify as bug. If you use a switch expression (with condition on int, long, float) in a function that returns an object, all values have as return type double. As soon as another types (like string, decimal) are present, everything works as expected. I didn't found any information about that if this is an expected behavior or not, so if it's expected, it would be great if someone could provide met the link to the documentation. Thanks! ### Configuration Which version of .NET is the code running on? - .NET 5/6 What OS and version, and for Linux, what distro? - macOS Monterey 12.1 What is the architecture (x64, x86, ARM, ARM64)? - x64 ### Example ```csharp using System; using System.Collections.Generic; public class Program { public static void Main() { var test = new List<object>() { 1, (long)1, 1d, 1f, }; foreach (var item in test) { Console.WriteLine( $"ValueIsOfType: {item.GetType().Name,-6} | SwitchExpression: {SwitchExpression(item).GetType().Name} | SwitchExpressionWithString: {SwitchExpression_WithString(item).GetType().Name,-6} | SwitchCase: {SwitchCase(item).GetType().Name}" ); } } private static object SwitchExpression(object x) { return x switch { int v => v, long v => v, double v => v, float v => v, _ => throw new Exception("Unexpected type"), }; } private static object SwitchExpression_WithString(object x) { return x switch { int v => v, long v => v, double v => v, float v => v, string v => v, _ => throw new Exception("Unexpected type"), }; } private static object SwitchCase(object x) { switch (x) { case int v: return v; case long v: return v; case double v: return v; case float v: return v; default: throw new Exception("Unexpected type"); } } } ``` Result: ``` ValueIsOfType: Int32 | SwitchExpression: Double | SwitchExpressionWithString: Int32 | SwitchCase: Int32 ValueIsOfType: Int64 | SwitchExpression: Double | SwitchExpressionWithString: Int64 | SwitchCase: Int64 ValueIsOfType: Double | SwitchExpression: Double | SwitchExpressionWithString: Double | SwitchCase: Double ValueIsOfType: Single | SwitchExpression: Double | SwitchExpressionWithString: Single | SwitchCase: Single ```
non_process
switch expression returns double on int long float condition description we came across a behavior that we would classify as bug if you use a switch expression with condition on int long float in a function that returns an object all values have as return type double as soon as another types like string decimal are present everything works as expected i didn t found any information about that if this is an expected behavior or not so if it s expected it would be great if someone could provide met the link to the documentation thanks configuration which version of net is the code running on net what os and version and for linux what distro macos monterey what is the architecture arm example csharp using system using system collections generic public class program public static void main var test new list long foreach var item in test console writeline valueisoftype item gettype name switchexpression switchexpression item gettype name switchexpressionwithstring switchexpression withstring item gettype name switchcase switchcase item gettype name private static object switchexpression object x return x switch int v v long v v double v v float v v throw new exception unexpected type private static object switchexpression withstring object x return x switch int v v long v v double v v float v v string v v throw new exception unexpected type private static object switchcase object x switch x case int v return v case long v return v case double v return v case float v return v default throw new exception unexpected type result valueisoftype switchexpression double switchexpressionwithstring switchcase valueisoftype switchexpression double switchexpressionwithstring switchcase valueisoftype double switchexpression double switchexpressionwithstring double switchcase double valueisoftype single switchexpression double switchexpressionwithstring single switchcase single
0
153,443
24,127,914,134
IssuesEvent
2022-09-21 03:31:06
Winneyp/group-14
https://api.github.com/repos/Winneyp/group-14
closed
เธ‡เธฒเธ™เธ—เธตเนˆเน€เธซเธฅเธทเธญเธ‚เธญเธ‡design
21-09-2022 Design
เธ•เธญเธ™เธ™เธตเน‰เธ‡เธฒเธ™เธ”เธตเน„เธ‹เธ™เนŒเน€เธซเธฅเธทเธญ - [x] เนเธเน‰เธชเธตเธœเธกเธ•เธฑเธงเธฅเธฐเธ„เธฃเธซเธฅเธฑเธเธ•เธญเธ™เธ•เธฒเธข - [x] เธ›เธธเนˆเธกUIเธ•เธญเธ™เธ•เธฒเธข - [x] เธžเธทเน‰เธ™เธซเธฅเธฑเธ‡เธ›เธธเนˆเธกเน€เธฅเธทเธญเธเธ”เนˆเธฒเธ™ - [x] เธ•เธฑเธงเธฅเธฐเธ„เธฃเธเธฑเนˆเธ‡เธฃเน‰เธฒเธข - [x] เธซเธ™เน‰เธฒเนเธ™เธฐเธ™เธณเน„เธญเน€เธ—เธก - [x] เธชเธ•เธญเธฃเธตเนˆเธ•เธญเธ™เน€เธฃเธดเนˆเธกเน€เธเธก - [x] เธชเธ•เธญเธฃเธตเนˆเธ•เธญเธ™เธˆเธš/เธซเธ™เน‰เธฒเธŠเธ™เธฐ
1.0
เธ‡เธฒเธ™เธ—เธตเนˆเน€เธซเธฅเธทเธญเธ‚เธญเธ‡design - เธ•เธญเธ™เธ™เธตเน‰เธ‡เธฒเธ™เธ”เธตเน„เธ‹เธ™เนŒเน€เธซเธฅเธทเธญ - [x] เนเธเน‰เธชเธตเธœเธกเธ•เธฑเธงเธฅเธฐเธ„เธฃเธซเธฅเธฑเธเธ•เธญเธ™เธ•เธฒเธข - [x] เธ›เธธเนˆเธกUIเธ•เธญเธ™เธ•เธฒเธข - [x] เธžเธทเน‰เธ™เธซเธฅเธฑเธ‡เธ›เธธเนˆเธกเน€เธฅเธทเธญเธเธ”เนˆเธฒเธ™ - [x] เธ•เธฑเธงเธฅเธฐเธ„เธฃเธเธฑเนˆเธ‡เธฃเน‰เธฒเธข - [x] เธซเธ™เน‰เธฒเนเธ™เธฐเธ™เธณเน„เธญเน€เธ—เธก - [x] เธชเธ•เธญเธฃเธตเนˆเธ•เธญเธ™เน€เธฃเธดเนˆเธกเน€เธเธก - [x] เธชเธ•เธญเธฃเธตเนˆเธ•เธญเธ™เธˆเธš/เธซเธ™เน‰เธฒเธŠเธ™เธฐ
non_process
เธ‡เธฒเธ™เธ—เธตเนˆเน€เธซเธฅเธทเธญเธ‚เธญเธ‡design เธ•เธญเธ™เธ™เธตเน‰เธ‡เธฒเธ™เธ”เธตเน„เธ‹เธ™เนŒเน€เธซเธฅเธทเธญ เนเธเน‰เธชเธตเธœเธกเธ•เธฑเธงเธฅเธฐเธ„เธฃเธซเธฅเธฑเธเธ•เธญเธ™เธ•เธฒเธข เธ›เธธเนˆเธกuiเธ•เธญเธ™เธ•เธฒเธข เธžเธทเน‰เธ™เธซเธฅเธฑเธ‡เธ›เธธเนˆเธกเน€เธฅเธทเธญเธเธ”เนˆเธฒเธ™ เธ•เธฑเธงเธฅเธฐเธ„เธฃเธเธฑเนˆเธ‡เธฃเน‰เธฒเธข เธซเธ™เน‰เธฒเนเธ™เธฐเธ™เธณเน„เธญเน€เธ—เธก เธชเธ•เธญเธฃเธตเนˆเธ•เธญเธ™เน€เธฃเธดเนˆเธกเน€เธเธก เธชเธ•เธญเธฃเธตเนˆเธ•เธญเธ™เธˆเธš เธซเธ™เน‰เธฒเธŠเธ™เธฐ
0
15,257
26,794,833,371
IssuesEvent
2023-02-01 11:05:53
VildeStabell/master-thesis-game
https://api.github.com/repos/VildeStabell/master-thesis-game
closed
Objects should fall on the board
requirement high
The objects should fall in random places and with random size and shape. The frequency they fall in changes with times. Task: - [x] #13 - [x] #14
1.0
Objects should fall on the board - The objects should fall in random places and with random size and shape. The frequency they fall in changes with times. Task: - [x] #13 - [x] #14
non_process
objects should fall on the board the objects should fall in random places and with random size and shape the frequency they fall in changes with times task
0
45,751
7,199,484,181
IssuesEvent
2018-02-05 16:04:57
wedeploy/wedeploy.com
https://api.github.com/repos/wedeploy/wedeploy.com
closed
Document that the hosting service provides ETag and gzip OOTB
documentation
<a href="https://github.com/ipeychev"><img src="https://avatars0.githubusercontent.com/u/78014?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [ipeychev](https://github.com/ipeychev)** _Friday Jan 06, 2017 at 09:27 GMT_ _Originally opened as https://github.com/wedeploy/docs/issues/66_ ---- The hosting service provides OOTB some goodies for the developers. These include ETag and gzip compression. The fact that ETag is available would save some work of the developer, because he won't have to add versions to his CSS and JS files to force the browser to reload them when a new version is created.
1.0
Document that the hosting service provides ETag and gzip OOTB - <a href="https://github.com/ipeychev"><img src="https://avatars0.githubusercontent.com/u/78014?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [ipeychev](https://github.com/ipeychev)** _Friday Jan 06, 2017 at 09:27 GMT_ _Originally opened as https://github.com/wedeploy/docs/issues/66_ ---- The hosting service provides OOTB some goodies for the developers. These include ETag and gzip compression. The fact that ETag is available would save some work of the developer, because he won't have to add versions to his CSS and JS files to force the browser to reload them when a new version is created.
non_process
document that the hosting service provides etag and gzip ootb issue by friday jan at gmt originally opened as the hosting service provides ootb some goodies for the developers these include etag and gzip compression the fact that etag is available would save some work of the developer because he won t have to add versions to his css and js files to force the browser to reload them when a new version is created
0
1,847
4,647,464,566
IssuesEvent
2016-10-01 14:18:36
opentrials/opentrials
https://api.github.com/repos/opentrials/opentrials
opened
Remove 0-valued identifiers
3. In Development Processors
Some of our sources have invalid identifiers like `NCT00000000`, `U0000-0000-0000`, `ISRCTN00000000` and others. These end up messing with our deduplication process, because the records end up having a mix of valid and invalid identifiers like: ``` # Record 1 # https://www.clinicaltrialsregister.eu/ctr-search/trial/2005-001059-39/PT { "nct": "NCT00202878", "euctr": "EUCTR2005-001059-39", "isrctn": "ISRCTN00000000" } # Record 2 # https://www.clinicaltrialsregister.eu/ctr-search/trial/2014-000042-30/IT { "nct": "NCT02300558", "who": "U0000-0000-0000", "euctr": "EUCTR2014-000042-30", "isrctn": "ISRCTN00000000" } ``` Our deduplication process will consider these 2 records as being the same, because they share the identifier `ISRCTN00000000`. So we'll end up thinking that trials `NCT02300558` and `NCT00202878` are the same, which is wrong. This is an issue with the source data, so we should write a report and ask them to fix. However, in the meantime, we can do a simple check for the identifiers: if all their numbers are `0`, consider it as invalid and don't use it. From a brief research, I could only find this issue with trial on EUCTR and HRA research summaries (e.g. http://www.hra.nhs.uk/news/research-summaries/tacrolimus-and-glucose-metabolism-in-renal-transplantation-v-1-1/)
1.0
Remove 0-valued identifiers - Some of our sources have invalid identifiers like `NCT00000000`, `U0000-0000-0000`, `ISRCTN00000000` and others. These end up messing with our deduplication process, because the records end up having a mix of valid and invalid identifiers like: ``` # Record 1 # https://www.clinicaltrialsregister.eu/ctr-search/trial/2005-001059-39/PT { "nct": "NCT00202878", "euctr": "EUCTR2005-001059-39", "isrctn": "ISRCTN00000000" } # Record 2 # https://www.clinicaltrialsregister.eu/ctr-search/trial/2014-000042-30/IT { "nct": "NCT02300558", "who": "U0000-0000-0000", "euctr": "EUCTR2014-000042-30", "isrctn": "ISRCTN00000000" } ``` Our deduplication process will consider these 2 records as being the same, because they share the identifier `ISRCTN00000000`. So we'll end up thinking that trials `NCT02300558` and `NCT00202878` are the same, which is wrong. This is an issue with the source data, so we should write a report and ask them to fix. However, in the meantime, we can do a simple check for the identifiers: if all their numbers are `0`, consider it as invalid and don't use it. From a brief research, I could only find this issue with trial on EUCTR and HRA research summaries (e.g. http://www.hra.nhs.uk/news/research-summaries/tacrolimus-and-glucose-metabolism-in-renal-transplantation-v-1-1/)
process
remove valued identifiers some of our sources have invalid identifiers like and others these end up messing with our deduplication process because the records end up having a mix of valid and invalid identifiers like record nct euctr isrctn record nct who euctr isrctn our deduplication process will consider these records as being the same because they share the identifier so we ll end up thinking that trials and are the same which is wrong this is an issue with the source data so we should write a report and ask them to fix however in the meantime we can do a simple check for the identifiers if all their numbers are consider it as invalid and don t use it from a brief research i could only find this issue with trial on euctr and hra research summaries e g
1
324,412
23,998,152,968
IssuesEvent
2022-09-14 09:13:42
TechsophyOfficial/awgment-package
https://api.github.com/repos/TechsophyOfficial/awgment-package
closed
Link Missing
documentation
Link Missing for Troubleshooting under **"setup kind for awgment on linux"** sub-heading.
1.0
Link Missing - Link Missing for Troubleshooting under **"setup kind for awgment on linux"** sub-heading.
non_process
link missing link missing for troubleshooting under setup kind for awgment on linux sub heading
0
32,743
6,912,465,114
IssuesEvent
2017-11-28 12:05:06
mcnomad81/sshtunnel
https://api.github.com/repos/mcnomad81/sshtunnel
closed
Port 80
auto-migrated Priority-Medium Type-Defect
``` Hello my SSH server runs on port 80 ( not the usual 22) when I try to connect I get the error in the log file. I have tried multiple Android SSH client and they have the same problem: it never works on this particular port. I tested this port with putty on my windows and it connects without a problem, Can you please help me on this issue (I need this particular port) Best regards ``` Original issue reported on code.google.com by `edied...@gmail.com` on 30 Dec 2014 at 1:43 Attachments: - [logSShTunnel.txt](https://storage.googleapis.com/google-code-attachments/sshtunnel/issue-194/comment-0/logSShTunnel.txt)
1.0
Port 80 - ``` Hello my SSH server runs on port 80 ( not the usual 22) when I try to connect I get the error in the log file. I have tried multiple Android SSH client and they have the same problem: it never works on this particular port. I tested this port with putty on my windows and it connects without a problem, Can you please help me on this issue (I need this particular port) Best regards ``` Original issue reported on code.google.com by `edied...@gmail.com` on 30 Dec 2014 at 1:43 Attachments: - [logSShTunnel.txt](https://storage.googleapis.com/google-code-attachments/sshtunnel/issue-194/comment-0/logSShTunnel.txt)
non_process
port hello my ssh server runs on port not the usual when i try to connect i get the error in the log file i have tried multiple android ssh client and they have the same problem it never works on this particular port i tested this port with putty on my windows and it connects without a problem can you please help me on this issue i need this particular port best regards original issue reported on code google com by edied gmail com on dec at attachments
0
21,935
30,446,746,017
IssuesEvent
2023-07-15 19:20:35
okTurtles/group-income
https://api.github.com/repos/okTurtles/group-income
opened
Preserve author information when splitting out Chelonia
Note:Question App:Backend Note:Discussion Note:Research Kind:Process Kind:Core Kind:Community
### Problem When we split out Chelonia into a separate repo, by default whoever is creating that PR will get credit for all the code changes. ### Solution Figure out how to preserve authorship in some capacity. Any tips/tricks/suggestions welcome! Some random possibly related links I found: - https://community.atlassian.com/t5/Bitbucket-questions/How-to-preserve-the-pull-request-author-as-commit-author-in/qaq-p/892714#M32800 - https://www.git-tower.com/learn/git/faq/change-author-name-email
1.0
Preserve author information when splitting out Chelonia - ### Problem When we split out Chelonia into a separate repo, by default whoever is creating that PR will get credit for all the code changes. ### Solution Figure out how to preserve authorship in some capacity. Any tips/tricks/suggestions welcome! Some random possibly related links I found: - https://community.atlassian.com/t5/Bitbucket-questions/How-to-preserve-the-pull-request-author-as-commit-author-in/qaq-p/892714#M32800 - https://www.git-tower.com/learn/git/faq/change-author-name-email
process
preserve author information when splitting out chelonia problem when we split out chelonia into a separate repo by default whoever is creating that pr will get credit for all the code changes solution figure out how to preserve authorship in some capacity any tips tricks suggestions welcome some random possibly related links i found
1
233,067
17,841,664,652
IssuesEvent
2021-09-03 10:56:39
maple-labs/proxy-factory
https://api.github.com/repos/maple-labs/proxy-factory
opened
Logic: What is the use of the Proxied contract ?
documentation help wanted
Shouldn't the `Proxied` inherit the Proxy contract to support of full proxy https://github.com/maple-labs/proxy-factory/blob/b18909327abfb86857d81a382a2a9f75931a95ff/contracts/Proxied.sol#L9 otherwise I don't understand the use of the `Proxied` contract.
1.0
Logic: What is the use of the Proxied contract ? - Shouldn't the `Proxied` inherit the Proxy contract to support of full proxy https://github.com/maple-labs/proxy-factory/blob/b18909327abfb86857d81a382a2a9f75931a95ff/contracts/Proxied.sol#L9 otherwise I don't understand the use of the `Proxied` contract.
non_process
logic what is the use of the proxied contract shouldn t the proxied inherit the proxy contract to support of full proxy otherwise i don t understand the use of the proxied contract
0
42,582
22,692,450,389
IssuesEvent
2022-07-04 23:10:49
bopen/sarsen
https://api.github.com/repos/bopen/sarsen
closed
Radiometric terrain-correction algorithm is slow and memory hungry
performance
At the minimum: - [x] review what step can be simplified - [x] review in what step we may reduce memory pressure by moving computations to *dask* out-of-core
True
Radiometric terrain-correction algorithm is slow and memory hungry - At the minimum: - [x] review what step can be simplified - [x] review in what step we may reduce memory pressure by moving computations to *dask* out-of-core
non_process
radiometric terrain correction algorithm is slow and memory hungry at the minimum review what step can be simplified review in what step we may reduce memory pressure by moving computations to dask out of core
0
13,785
5,447,290,403
IssuesEvent
2017-03-07 13:10:53
dotnet/docs
https://api.github.com/repos/dotnet/docs
closed
Unit testing using xUnit with core: project.json to csproj
Goal - Update from project.json based projects to MSBuild based projects
./samples/core/getting-started/unit-testing-using-dotnet-test/global.json getting-started/unit-testing-using-dotnet-test/src/PrimeService/project.json getting-started/unit-testing-using-dotnet-test/test/PrimeService.Tests/project.json
1.0
Unit testing using xUnit with core: project.json to csproj - ./samples/core/getting-started/unit-testing-using-dotnet-test/global.json getting-started/unit-testing-using-dotnet-test/src/PrimeService/project.json getting-started/unit-testing-using-dotnet-test/test/PrimeService.Tests/project.json
non_process
unit testing using xunit with core project json to csproj samples core getting started unit testing using dotnet test global json getting started unit testing using dotnet test src primeservice project json getting started unit testing using dotnet test test primeservice tests project json
0
128,740
18,070,110,399
IssuesEvent
2021-09-21 01:12:26
Dima2021/NodeGoat
https://api.github.com/repos/Dima2021/NodeGoat
opened
CVE-2021-3807 (Medium) detected in multiple libraries
security vulnerability
## CVE-2021-3807 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ansi-regex-2.1.1.tgz</b>, <b>ansi-regex-2.0.0.tgz</b>, <b>ansi-regex-3.0.0.tgz</b></p></summary> <p> <details><summary><b>ansi-regex-2.1.1.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-2.1.1.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-2.1.1.tgz</a></p> <p> Dependency Hierarchy: - nodemon-1.19.1.tgz (Root Library) - chokidar-2.1.6.tgz - fsevents-1.2.9.tgz - node-pre-gyp-0.12.0.tgz - npmlog-4.1.2.tgz - gauge-2.7.4.tgz - strip-ansi-3.0.1.tgz - :x: **ansi-regex-2.1.1.tgz** (Vulnerable Library) </details> <details><summary><b>ansi-regex-2.0.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-2.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-2.0.0.tgz</a></p> <p>Path to dependency file: NodeGoat/package.json</p> <p>Path to vulnerable library: NodeGoat/node_modules/nyc/node_modules/ansi-regex/package.json,NodeGoat/node_modules/npm/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - :x: **ansi-regex-2.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>ansi-regex-3.0.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz</a></p> <p>Path to dependency file: NodeGoat/package.json</p> <p>Path to vulnerable library: NodeGoat/node_modules/widest-line/node_modules/ansi-regex/package.json,NodeGoat/node_modules/boxen/node_modules/ansi-regex/package.json,NodeGoat/node_modules/ansi-align/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - nodemon-1.19.1.tgz (Root Library) - update-notifier-2.5.0.tgz - boxen-1.3.0.tgz - ansi-align-2.0.0.tgz - string-width-2.1.1.tgz - strip-ansi-4.0.0.tgz - :x: **ansi-regex-3.0.0.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ansi-regex is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807>CVE-2021-3807</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution: ansi-regex - 5.0.1,6.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"ansi-regex","packageVersion":"2.1.1","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"nodemon:1.19.1;chokidar:2.1.6;fsevents:1.2.9;node-pre-gyp:0.12.0;npmlog:4.1.2;gauge:2.7.4;strip-ansi:3.0.1;ansi-regex:2.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ansi-regex - 5.0.1,6.0.1"},{"packageType":"javascript/Node.js","packageName":"ansi-regex","packageVersion":"2.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;ansi-regex:2.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ansi-regex - 5.0.1,6.0.1"},{"packageType":"javascript/Node.js","packageName":"ansi-regex","packageVersion":"3.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"nodemon:1.19.1;update-notifier:2.5.0;boxen:1.3.0;ansi-align:2.0.0;string-width:2.1.1;strip-ansi:4.0.0;ansi-regex:3.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ansi-regex - 5.0.1,6.0.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-3807","vulnerabilityDetails":"ansi-regex is vulnerable to Inefficient Regular Expression Complexity","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-3807 (Medium) detected in multiple libraries - ## CVE-2021-3807 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ansi-regex-2.1.1.tgz</b>, <b>ansi-regex-2.0.0.tgz</b>, <b>ansi-regex-3.0.0.tgz</b></p></summary> <p> <details><summary><b>ansi-regex-2.1.1.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-2.1.1.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-2.1.1.tgz</a></p> <p> Dependency Hierarchy: - nodemon-1.19.1.tgz (Root Library) - chokidar-2.1.6.tgz - fsevents-1.2.9.tgz - node-pre-gyp-0.12.0.tgz - npmlog-4.1.2.tgz - gauge-2.7.4.tgz - strip-ansi-3.0.1.tgz - :x: **ansi-regex-2.1.1.tgz** (Vulnerable Library) </details> <details><summary><b>ansi-regex-2.0.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-2.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-2.0.0.tgz</a></p> <p>Path to dependency file: NodeGoat/package.json</p> <p>Path to vulnerable library: NodeGoat/node_modules/nyc/node_modules/ansi-regex/package.json,NodeGoat/node_modules/npm/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - :x: **ansi-regex-2.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>ansi-regex-3.0.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz</a></p> <p>Path to dependency file: NodeGoat/package.json</p> <p>Path to vulnerable library: NodeGoat/node_modules/widest-line/node_modules/ansi-regex/package.json,NodeGoat/node_modules/boxen/node_modules/ansi-regex/package.json,NodeGoat/node_modules/ansi-align/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - nodemon-1.19.1.tgz (Root Library) - update-notifier-2.5.0.tgz - boxen-1.3.0.tgz - ansi-align-2.0.0.tgz - string-width-2.1.1.tgz - strip-ansi-4.0.0.tgz - :x: **ansi-regex-3.0.0.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ansi-regex is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807>CVE-2021-3807</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution: ansi-regex - 5.0.1,6.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"ansi-regex","packageVersion":"2.1.1","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"nodemon:1.19.1;chokidar:2.1.6;fsevents:1.2.9;node-pre-gyp:0.12.0;npmlog:4.1.2;gauge:2.7.4;strip-ansi:3.0.1;ansi-regex:2.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ansi-regex - 5.0.1,6.0.1"},{"packageType":"javascript/Node.js","packageName":"ansi-regex","packageVersion":"2.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;ansi-regex:2.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ansi-regex - 5.0.1,6.0.1"},{"packageType":"javascript/Node.js","packageName":"ansi-regex","packageVersion":"3.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"nodemon:1.19.1;update-notifier:2.5.0;boxen:1.3.0;ansi-align:2.0.0;string-width:2.1.1;strip-ansi:4.0.0;ansi-regex:3.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ansi-regex - 5.0.1,6.0.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-3807","vulnerabilityDetails":"ansi-regex is vulnerable to Inefficient Regular Expression Complexity","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries ansi regex tgz ansi regex tgz ansi regex tgz ansi regex tgz regular expression for matching ansi escape codes library home page a href dependency hierarchy nodemon tgz root library chokidar tgz fsevents tgz node pre gyp tgz npmlog tgz gauge tgz strip ansi tgz x ansi regex tgz vulnerable library ansi regex tgz regular expression for matching ansi escape codes library home page a href path to dependency file nodegoat package json path to vulnerable library nodegoat node modules nyc node modules ansi regex package json nodegoat node modules npm node modules ansi regex package json dependency hierarchy grunt npm install tgz root library npm tgz x ansi regex tgz vulnerable library ansi regex tgz regular expression for matching ansi escape codes library home page a href path to dependency file nodegoat package json path to vulnerable library nodegoat node modules widest line node modules ansi regex package json nodegoat node modules boxen node modules ansi regex package json nodegoat node modules ansi align node modules ansi regex package json dependency hierarchy nodemon tgz root library update notifier tgz boxen tgz ansi align tgz string width tgz strip ansi tgz x ansi regex tgz vulnerable library found in base branch master vulnerability details ansi regex is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ansi regex isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree nodemon chokidar fsevents node pre gyp npmlog gauge strip ansi ansi regex isminimumfixversionavailable true minimumfixversion ansi regex packagetype javascript node js packagename ansi regex packageversion packagefilepaths istransitivedependency true dependencytree grunt npm install npm ansi regex isminimumfixversionavailable true minimumfixversion ansi regex packagetype javascript node js packagename ansi regex packageversion packagefilepaths istransitivedependency true dependencytree nodemon update notifier boxen ansi align string width strip ansi ansi regex isminimumfixversionavailable true minimumfixversion ansi regex basebranches vulnerabilityidentifier cve vulnerabilitydetails ansi regex is vulnerable to inefficient regular expression complexity vulnerabilityurl
0
21,942
30,446,799,516
IssuesEvent
2023-07-15 19:28:43
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
pyutils 0.0.1b17 has 2 GuardDog issues
guarddog typosquatting silent-process-execution
https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1b17", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: python-utils, pytils", "silent-process-execution": [ { "location": "pyutils/exec_utils.py/pyutils/exec_utils.py:205", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpbqhta8b_/pyutils" } }```
1.0
pyutils 0.0.1b17 has 2 GuardDog issues - https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1b17", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: python-utils, pytils", "silent-process-execution": [ { "location": "pyutils/exec_utils.py/pyutils/exec_utils.py:205", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpbqhta8b_/pyutils" } }```
process
pyutils has guarddog issues dependency pyutils version result issues errors results typosquatting this package closely ressembles the following package names and might be a typosquatting attempt python utils pytils silent process execution location pyutils exec utils py pyutils exec utils py code subproc subprocess popen n args n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp pyutils
1
84,247
3,655,786,251
IssuesEvent
2016-02-17 17:26:15
aic-collections/aicdams-lakeshore
https://api.github.com/repos/aic-collections/aicdams-lakeshore
closed
Checksum-based duplicate validation
HIGH priority
When ingesting an asset, parse the uploaded file checksum and verify that a file with the same checksum does not already exist in the repository. If such file exists, the upload should fail with an explanatory error message.
1.0
Checksum-based duplicate validation - When ingesting an asset, parse the uploaded file checksum and verify that a file with the same checksum does not already exist in the repository. If such file exists, the upload should fail with an explanatory error message.
non_process
checksum based duplicate validation when ingesting an asset parse the uploaded file checksum and verify that a file with the same checksum does not already exist in the repository if such file exists the upload should fail with an explanatory error message
0
13,827
16,591,410,658
IssuesEvent
2021-06-01 08:11:22
symfony/symfony
https://api.github.com/repos/symfony/symfony
closed
\Symfony\Component\Process\Process::escapeArgument as a service class
Feature Process
**Description** Extract `\Symfony\Component\Process\Process::escapeArgument` to a separate service class, so that external classes can make use of the shell argument escaping. Commands like `\Symfony\Component\Process\Process::fromShellCommandline` accept raw input, but sometimes the client wants to make sure the arguments passed to that method are actually escaped. Or is argument binding the only preferred way to achieve this? **Example** ``` php <?php use Symfony\Component\Process\ArgumentEscaper; use Symfony\Component\Process\Process; function test_piped_command(string $param): void { $escaper = new ArgumentEscaper(); $command = []; $command[] = 'echo ' . $escaper->escapeArgument($param); $command[] = 'md5sum'; $shellCommand = implode('|', $command); $process = Process::fromShellCommandline($command); $process->run(); } ```
1.0
\Symfony\Component\Process\Process::escapeArgument as a service class - **Description** Extract `\Symfony\Component\Process\Process::escapeArgument` to a separate service class, so that external classes can make use of the shell argument escaping. Commands like `\Symfony\Component\Process\Process::fromShellCommandline` accept raw input, but sometimes the client wants to make sure the arguments passed to that method are actually escaped. Or is argument binding the only preferred way to achieve this? **Example** ``` php <?php use Symfony\Component\Process\ArgumentEscaper; use Symfony\Component\Process\Process; function test_piped_command(string $param): void { $escaper = new ArgumentEscaper(); $command = []; $command[] = 'echo ' . $escaper->escapeArgument($param); $command[] = 'md5sum'; $shellCommand = implode('|', $command); $process = Process::fromShellCommandline($command); $process->run(); } ```
process
symfony component process process escapeargument as a service class description extract symfony component process process escapeargument to a separate service class so that external classes can make use of the shell argument escaping commands like symfony component process process fromshellcommandline accept raw input but sometimes the client wants to make sure the arguments passed to that method are actually escaped or is argument binding the only preferred way to achieve this example php php use symfony component process argumentescaper use symfony component process process function test piped command string param void escaper new argumentescaper command command echo escaper escapeargument param command shellcommand implode command process process fromshellcommandline command process run
1
201,360
15,191,310,625
IssuesEvent
2021-02-15 19:38:40
aroberge/friendly-traceback
https://api.github.com/repos/aroberge/friendly-traceback
closed
SyntaxError when attempting to use keyword 'arguments' inside subscript
Valid contest entry
``` $ echo "foo[x=1]" > test.py && python -m friendly_traceback test.py โ•ญโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€ Python exception โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ•ฎ โ”‚ Traceback (most recent call last): โ”‚ โ”‚ File "test.py", line 1 โ”‚ โ”‚ foo[x=1] โ”‚ โ”‚ ^ โ”‚ โ”‚ SyntaxError: invalid syntax โ”‚ โ”‚ โ”‚ โ”‚ A SyntaxError occurs when Python cannot understand your code. โ”‚ โ”‚ โ”‚ โ”‚ Python could not understand the code in the file 'test.py' beyond the location indicated by --> and ^. โ”‚ โ”‚ โ”‚ โ”‚ -->1: foo[x=1] โ”‚ โ”‚ ^ โ”‚ โ”‚ โ”‚ โ”‚ The opening square bracket [ on line 1 is not closed. โ”‚ โ”‚ โ”‚ โ”‚ 1: foo[x=1] โ”‚ โ”‚ | โ”‚ โ•ฐโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ•ฏ ``` The hint doesn't seem correct in this case. There's a [draft PEP](https://www.python.org/dev/peps/pep-0637/) on this topic and a (somewhat unpopular) [question on SO](https://stackoverflow.com/questions/31916617/using-keyword-arguments-in-getitem-method-in-python/41999773).
1.0
SyntaxError when attempting to use keyword 'arguments' inside subscript - ``` $ echo "foo[x=1]" > test.py && python -m friendly_traceback test.py โ•ญโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€ Python exception โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ•ฎ โ”‚ Traceback (most recent call last): โ”‚ โ”‚ File "test.py", line 1 โ”‚ โ”‚ foo[x=1] โ”‚ โ”‚ ^ โ”‚ โ”‚ SyntaxError: invalid syntax โ”‚ โ”‚ โ”‚ โ”‚ A SyntaxError occurs when Python cannot understand your code. โ”‚ โ”‚ โ”‚ โ”‚ Python could not understand the code in the file 'test.py' beyond the location indicated by --> and ^. โ”‚ โ”‚ โ”‚ โ”‚ -->1: foo[x=1] โ”‚ โ”‚ ^ โ”‚ โ”‚ โ”‚ โ”‚ The opening square bracket [ on line 1 is not closed. โ”‚ โ”‚ โ”‚ โ”‚ 1: foo[x=1] โ”‚ โ”‚ | โ”‚ โ•ฐโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ•ฏ ``` The hint doesn't seem correct in this case. There's a [draft PEP](https://www.python.org/dev/peps/pep-0637/) on this topic and a (somewhat unpopular) [question on SO](https://stackoverflow.com/questions/31916617/using-keyword-arguments-in-getitem-method-in-python/41999773).
non_process
syntaxerror when attempting to use keyword arguments inside subscript echo foo test py python m friendly traceback test py โ•ญโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€ python exception โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ•ฎ โ”‚ traceback most recent call last โ”‚ โ”‚ file test py line โ”‚ โ”‚ foo โ”‚ โ”‚ โ”‚ โ”‚ syntaxerror invalid syntax โ”‚ โ”‚ โ”‚ โ”‚ a syntaxerror occurs when python cannot understand your code โ”‚ โ”‚ โ”‚ โ”‚ python could not understand the code in the file test py beyond the location indicated by and โ”‚ โ”‚ โ”‚ โ”‚ foo โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ โ”‚ the opening square bracket on line is not closed โ”‚ โ”‚ โ”‚ โ”‚ foo โ”‚ โ”‚ โ”‚ โ•ฐโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ•ฏ the hint doesn t seem correct in this case there s a on this topic and a somewhat unpopular
0
3,320
6,429,799,055
IssuesEvent
2017-08-10 03:10:22
nodejs/node
https://api.github.com/repos/nodejs/node
closed
process.cpuUsage accuracy
process question
<!-- Thank you for reporting an issue. Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify affected core module name If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able. --> - **v6.3.1**: - **Darwin Kernel Version 15.6.0 root:xnu-3248.60.11~1/RELEASE_X86_64 x86_64**: - **process**: <!-- Enter your issue details below this comment. --> I'm trying to find cpu utilization of the process as a percentage of the system as a whole - similar to how top, activity monitor, and similar tools work, but from within the program itself. It appears that process.cpuUsage is a relatively new feature, and I haven't been able to find a way to get the cpu percentage and have it accurately match values found in system monitors. I created a repository of some of the different methods I have tried: https://github.com/nickstanish/node-test-process-cpuusage It appears that there is [discussion](https://github.com/nodejs/node/pull/6157#issuecomment-208950650) on an approach to get the percentage, but uses the wrong precision and seems to not account for number of cpu cores. Is there a recommended approach in order to find this value? I'd also like to point out that the tests for this feature have the wrong precision as well which makes the possible value range huge. The process usage values are in microseconds and it multiplies by 1e6 https://github.com/nodejs/node/blob/master/test/pummel/test-process-cpuUsage.js#L27 Original Issue: https://github.com/nodejs/help/issues/283
1.0
process.cpuUsage accuracy - <!-- Thank you for reporting an issue. Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify affected core module name If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able. --> - **v6.3.1**: - **Darwin Kernel Version 15.6.0 root:xnu-3248.60.11~1/RELEASE_X86_64 x86_64**: - **process**: <!-- Enter your issue details below this comment. --> I'm trying to find cpu utilization of the process as a percentage of the system as a whole - similar to how top, activity monitor, and similar tools work, but from within the program itself. It appears that process.cpuUsage is a relatively new feature, and I haven't been able to find a way to get the cpu percentage and have it accurately match values found in system monitors. I created a repository of some of the different methods I have tried: https://github.com/nickstanish/node-test-process-cpuusage It appears that there is [discussion](https://github.com/nodejs/node/pull/6157#issuecomment-208950650) on an approach to get the percentage, but uses the wrong precision and seems to not account for number of cpu cores. Is there a recommended approach in order to find this value? I'd also like to point out that the tests for this feature have the wrong precision as well which makes the possible value range huge. The process usage values are in microseconds and it multiplies by 1e6 https://github.com/nodejs/node/blob/master/test/pummel/test-process-cpuUsage.js#L27 Original Issue: https://github.com/nodejs/help/issues/283
process
process cpuusage accuracy thank you for reporting an issue please fill in as much of the template below as you re able version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify affected core module name if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able darwin kernel version root xnu release process i m trying to find cpu utilization of the process as a percentage of the system as a whole similar to how top activity monitor and similar tools work but from within the program itself it appears that process cpuusage is a relatively new feature and i haven t been able to find a way to get the cpu percentage and have it accurately match values found in system monitors i created a repository of some of the different methods i have tried it appears that there is on an approach to get the percentage but uses the wrong precision and seems to not account for number of cpu cores is there a recommended approach in order to find this value i d also like to point out that the tests for this feature have the wrong precision as well which makes the possible value range huge the process usage values are in microseconds and it multiplies by original issue
1