Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 2
665
| labels
stringlengths 4
554
| body
stringlengths 3
235k
| index
stringclasses 6
values | text_combine
stringlengths 96
235k
| label
stringclasses 2
values | text
stringlengths 96
196k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
32,709
| 26,926,391,658
|
IssuesEvent
|
2023-02-07 14:05:36
|
opendatahub-io/odh-dashboard
|
https://api.github.com/repos/opendatahub-io/odh-dashboard
|
closed
|
[Feature Request]: Optional role to spawn pods from notebooks, used with Spark workloads
|
kind/enhancement kind/question infrastructure priority/low
|
### Feature description
The ServiceAccount used to run the notebooks does not have right to spawn Pods. This feature would be useful to directly launch Spark workloads from a notebook, which some teams are used to do in their current environments.
Details:
- Spark supports kubernetes to directly launch workloads. You can target your spark-submit command to a kubernetes cluster, which will schedule executor pods, do the computation, and remove the pods when finished. The Spark instance running in the notebook acts as the driver for the workload.
- This functionality requires rights to create/delete Pods and Services. However, the ServiceAccount under which the Notebook container is running does not have those rights.
- The requested feature would add another Role to the notebook ServiceAccount, enabling those capabilities.
- The drawback is that any notebook user would be able to create Pods in the namespace by invoking the kubernetes API, so this must be an admin-selectable feature.
Proposed implementation:
- Create a Role with needed capabilities. This can be used as a base as it's exactly what the Spark operator is doing: https://github.com/opendatahub-io-contrib/spark-on-openshift/blob/main/spark-operator/spark-rbac.yaml
- Create a toggle in admin to enable this role or not. As the drawback of the method is that people would be able to schedule Pods, admins must be able to allow this or not. The toggle would create/delete the associated RoleBinding. For DSP, this could be a per-project toggle.
### Describe alternatives you've considered
_No response_
### Anything else?
This would bring back some of the features we previously had in ODH with Spark notebooks, that customers are now lacking. Instead of spawning a pseudo-permanent Spark attached to a notebook, this would be more dynamic, with the Spark cluster being spawned when the command is submitted.
|
1.0
|
[Feature Request]: Optional role to spawn pods from notebooks, used with Spark workloads - ### Feature description
The ServiceAccount used to run the notebooks does not have right to spawn Pods. This feature would be useful to directly launch Spark workloads from a notebook, which some teams are used to do in their current environments.
Details:
- Spark supports kubernetes to directly launch workloads. You can target your spark-submit command to a kubernetes cluster, which will schedule executor pods, do the computation, and remove the pods when finished. The Spark instance running in the notebook acts as the driver for the workload.
- This functionality requires rights to create/delete Pods and Services. However, the ServiceAccount under which the Notebook container is running does not have those rights.
- The requested feature would add another Role to the notebook ServiceAccount, enabling those capabilities.
- The drawback is that any notebook user would be able to create Pods in the namespace by invoking the kubernetes API, so this must be an admin-selectable feature.
Proposed implementation:
- Create a Role with needed capabilities. This can be used as a base as it's exactly what the Spark operator is doing: https://github.com/opendatahub-io-contrib/spark-on-openshift/blob/main/spark-operator/spark-rbac.yaml
- Create a toggle in admin to enable this role or not. As the drawback of the method is that people would be able to schedule Pods, admins must be able to allow this or not. The toggle would create/delete the associated RoleBinding. For DSP, this could be a per-project toggle.
### Describe alternatives you've considered
_No response_
### Anything else?
This would bring back some of the features we previously had in ODH with Spark notebooks, that customers are now lacking. Instead of spawning a pseudo-permanent Spark attached to a notebook, this would be more dynamic, with the Spark cluster being spawned when the command is submitted.
|
infrastructure
|
optional role to spawn pods from notebooks used with spark workloads feature description the serviceaccount used to run the notebooks does not have right to spawn pods this feature would be useful to directly launch spark workloads from a notebook which some teams are used to do in their current environments details spark supports kubernetes to directly launch workloads you can target your spark submit command to a kubernetes cluster which will schedule executor pods do the computation and remove the pods when finished the spark instance running in the notebook acts as the driver for the workload this functionality requires rights to create delete pods and services however the serviceaccount under which the notebook container is running does not have those rights the requested feature would add another role to the notebook serviceaccount enabling those capabilities the drawback is that any notebook user would be able to create pods in the namespace by invoking the kubernetes api so this must be an admin selectable feature proposed implementation create a role with needed capabilities this can be used as a base as it s exactly what the spark operator is doing create a toggle in admin to enable this role or not as the drawback of the method is that people would be able to schedule pods admins must be able to allow this or not the toggle would create delete the associated rolebinding for dsp this could be a per project toggle describe alternatives you ve considered no response anything else this would bring back some of the features we previously had in odh with spark notebooks that customers are now lacking instead of spawning a pseudo permanent spark attached to a notebook this would be more dynamic with the spark cluster being spawned when the command is submitted
| 1
|
6,471
| 6,445,351,162
|
IssuesEvent
|
2017-08-13 03:03:38
|
timvideos/HDMI2USB-litex-firmware
|
https://api.github.com/repos/timvideos/HDMI2USB-litex-firmware
|
opened
|
Build qemu-litex as part of the Travis-CI
|
boards-all firmware-softcpu level-infrastructure status-help-wanted type-enhancement
|
It would be good to run the [`./scripts/build-qemu.sh`](./scripts/build-qemu.sh) script as part of the Travis-CI environment.
This will make sure we don't accidently break the script with other changes.
|
1.0
|
Build qemu-litex as part of the Travis-CI - It would be good to run the [`./scripts/build-qemu.sh`](./scripts/build-qemu.sh) script as part of the Travis-CI environment.
This will make sure we don't accidently break the script with other changes.
|
infrastructure
|
build qemu litex as part of the travis ci it would be good to run the scripts build qemu sh script as part of the travis ci environment this will make sure we don t accidently break the script with other changes
| 1
|
25,732
| 19,074,829,385
|
IssuesEvent
|
2021-11-27 15:31:05
|
Kavignon/FSSF-Mentorship
|
https://api.github.com/repos/Kavignon/FSSF-Mentorship
|
opened
|
Create the GitHub Action configuration for the project
|
Azure Docker Infrastructure
|
Everything related to the project will be handled in the repository, such as building, testing, package, and deploying the application to the cloud provider (Azure).
Let's start with something simple for an initial version and go from there.
- [ ] Restore the development dependencies
- [ ] Build the solution
- [ ] Test the solution
- [ ] Build the Docker image
- [ ] Deploy Docker image to Azure
- [ ] Add support for Windows and Linux machines
|
1.0
|
Create the GitHub Action configuration for the project - Everything related to the project will be handled in the repository, such as building, testing, package, and deploying the application to the cloud provider (Azure).
Let's start with something simple for an initial version and go from there.
- [ ] Restore the development dependencies
- [ ] Build the solution
- [ ] Test the solution
- [ ] Build the Docker image
- [ ] Deploy Docker image to Azure
- [ ] Add support for Windows and Linux machines
|
infrastructure
|
create the github action configuration for the project everything related to the project will be handled in the repository such as building testing package and deploying the application to the cloud provider azure let s start with something simple for an initial version and go from there restore the development dependencies build the solution test the solution build the docker image deploy docker image to azure add support for windows and linux machines
| 1
|
14,610
| 11,012,259,722
|
IssuesEvent
|
2019-12-04 17:52:59
|
celo-org/celo-monorepo
|
https://api.github.com/repos/celo-org/celo-monorepo
|
opened
|
Prom to SD for VM proxies
|
infrastructure
|
### Expected Behavior
Prom to SD works for tx-nodes, validators, and proxies
### Current Behavior
Prom to SD only works for tx-nodes and validators at the moment
|
1.0
|
Prom to SD for VM proxies - ### Expected Behavior
Prom to SD works for tx-nodes, validators, and proxies
### Current Behavior
Prom to SD only works for tx-nodes and validators at the moment
|
infrastructure
|
prom to sd for vm proxies expected behavior prom to sd works for tx nodes validators and proxies current behavior prom to sd only works for tx nodes and validators at the moment
| 1
|
20,021
| 13,631,975,017
|
IssuesEvent
|
2020-09-24 18:52:54
|
oci-labs/check-ins
|
https://api.github.com/repos/oci-labs/check-ins
|
closed
|
FE > Add storybook or similar (S)
|
infrastructure sprint 10
|
Add a component gallery like storybook
- [ ] Plan how storybook will be integrated into the project
- [ ] Analyze how React code will be contributed and leveraged outside of isolation
- [ ] Develop the code, run the npm scripts, and experiment with one component
- [ ] Implement the React code and demonstrate how a second team member would be able to contribute.
- [ ] Document any additional links that were helpful, and any notes or thoughts you found informative that could be shared in posterity.
|
1.0
|
FE > Add storybook or similar (S) - Add a component gallery like storybook
- [ ] Plan how storybook will be integrated into the project
- [ ] Analyze how React code will be contributed and leveraged outside of isolation
- [ ] Develop the code, run the npm scripts, and experiment with one component
- [ ] Implement the React code and demonstrate how a second team member would be able to contribute.
- [ ] Document any additional links that were helpful, and any notes or thoughts you found informative that could be shared in posterity.
|
infrastructure
|
fe add storybook or similar s add a component gallery like storybook plan how storybook will be integrated into the project analyze how react code will be contributed and leveraged outside of isolation develop the code run the npm scripts and experiment with one component implement the react code and demonstrate how a second team member would be able to contribute document any additional links that were helpful and any notes or thoughts you found informative that could be shared in posterity
| 1
|
15,741
| 11,693,535,419
|
IssuesEvent
|
2020-03-06 00:52:37
|
department-of-veterans-affairs/va.gov-cms
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
|
opened
|
Cannot find profiles directory ./profiles &&
|
Infrastructure
|
It appears the ./profiles folder went missing, likely due to a Core 8.8 Composer Scaffolding update which I think creates the folder usually. This is causing multiple errors on CI, which makes sure that folder exists.
Solution is to make the directory again with a .gitkeep OR figure out what changed with scaffolding upstream.
> PDOException: SQLSTATE[23000]: Integrity constraint violation: 1048 Column 'rid' cannot be null in /var/aegir/devmaster-1.x/includes/database/database.inc:2227
> Stack trace:
> #0 /var/aegir/devmaster-1.x/includes/database/database.inc(2227): PDOStatement->execute(Array)
> #1 /var/aegir/devmaster-1.x/includes/database/database.inc(697): DatabaseStatementBase->execute(Array, Array)
> #2 /var/aegir/devmaster-1.x/includes/database/mysql/query.inc(36): DatabaseConnection->query('INSERT INTO {ho...', Array, Array)
> #3 /var/aegir/devmaster-1.x/profiles/devmaster/modules/contrib/hosting/package/hosting_package.instance.inc(149): InsertQuery_mysql->execute()
> #4 /var/aegir/devmaster-1.x/profiles/devmaster/modules/contrib/hosting/package/hosting_package.instance.inc(257): hosting_package_instance_create(Object(stdClass))
> #5 /var/aegir/devmaster-1.x/profiles/devmaster/modules/contrib/hosting/package/hosting_package.instance.inc(83): hosting_package_instance_save(Object(stdClass))
> #6 /var/aegir/devmaster-1.x/profiles/devmaster/modules/contrib/hosting/platform/hosting_platform.drush.inc(83): hosting_package_instance_sync(NULL, NULL, Array, NULL, Array)
> #7 /var/aegir/devmaster-1.x/includes/module.inc(965): hosting_platform_post_hosting_verify_task(Object(stdClass), Array)
> #8 /var/aegir/devmaster-1.x/profiles/devmaster/modules/contrib/hosting/task.hosting.inc(201): module_invoke_all('post_hosting_ve...', Object(stdClass), Array)
> #9 phar:///usr/share/devshop/bin/drush/includes/command.inc(422): drush_hosting_post_hosting_task(Object(stdClass))
> #10 phar:///usr/share/devshop/bin/drush/includes/command.inc(231): _drush_invoke_hooks(Array, Array)
> #11 phar:///usr/share/devshop/bin/drush/includes/command.inc(199): drush_command('14864')
> #12 phar:///usr/share/devshop/bin/drush/lib/Drush/Boot/BaseBoot.php(67): drush_dispatch(Array)
> #13 phar:///usr/share/devshop/bin/drush/includes/preflight.inc(67): Drush\Boot\BaseBoot->bootstrap_and_dispatch()
> #14 phar:///usr/share/devshop/bin/drush/includes/startup.inc(465): drush_main()
> #15 phar:///usr/share/devshop/bin/drush/includes/startup.inc(369): drush_run_main(false, '/', 'Phar detected. ...')
> #16 phar:///usr/share/devshop/bin/drush/drush(114): drush_startup(Array)
> #17 /usr/share
|
1.0
|
Cannot find profiles directory ./profiles && - It appears the ./profiles folder went missing, likely due to a Core 8.8 Composer Scaffolding update which I think creates the folder usually. This is causing multiple errors on CI, which makes sure that folder exists.
Solution is to make the directory again with a .gitkeep OR figure out what changed with scaffolding upstream.
> PDOException: SQLSTATE[23000]: Integrity constraint violation: 1048 Column 'rid' cannot be null in /var/aegir/devmaster-1.x/includes/database/database.inc:2227
> Stack trace:
> #0 /var/aegir/devmaster-1.x/includes/database/database.inc(2227): PDOStatement->execute(Array)
> #1 /var/aegir/devmaster-1.x/includes/database/database.inc(697): DatabaseStatementBase->execute(Array, Array)
> #2 /var/aegir/devmaster-1.x/includes/database/mysql/query.inc(36): DatabaseConnection->query('INSERT INTO {ho...', Array, Array)
> #3 /var/aegir/devmaster-1.x/profiles/devmaster/modules/contrib/hosting/package/hosting_package.instance.inc(149): InsertQuery_mysql->execute()
> #4 /var/aegir/devmaster-1.x/profiles/devmaster/modules/contrib/hosting/package/hosting_package.instance.inc(257): hosting_package_instance_create(Object(stdClass))
> #5 /var/aegir/devmaster-1.x/profiles/devmaster/modules/contrib/hosting/package/hosting_package.instance.inc(83): hosting_package_instance_save(Object(stdClass))
> #6 /var/aegir/devmaster-1.x/profiles/devmaster/modules/contrib/hosting/platform/hosting_platform.drush.inc(83): hosting_package_instance_sync(NULL, NULL, Array, NULL, Array)
> #7 /var/aegir/devmaster-1.x/includes/module.inc(965): hosting_platform_post_hosting_verify_task(Object(stdClass), Array)
> #8 /var/aegir/devmaster-1.x/profiles/devmaster/modules/contrib/hosting/task.hosting.inc(201): module_invoke_all('post_hosting_ve...', Object(stdClass), Array)
> #9 phar:///usr/share/devshop/bin/drush/includes/command.inc(422): drush_hosting_post_hosting_task(Object(stdClass))
> #10 phar:///usr/share/devshop/bin/drush/includes/command.inc(231): _drush_invoke_hooks(Array, Array)
> #11 phar:///usr/share/devshop/bin/drush/includes/command.inc(199): drush_command('14864')
> #12 phar:///usr/share/devshop/bin/drush/lib/Drush/Boot/BaseBoot.php(67): drush_dispatch(Array)
> #13 phar:///usr/share/devshop/bin/drush/includes/preflight.inc(67): Drush\Boot\BaseBoot->bootstrap_and_dispatch()
> #14 phar:///usr/share/devshop/bin/drush/includes/startup.inc(465): drush_main()
> #15 phar:///usr/share/devshop/bin/drush/includes/startup.inc(369): drush_run_main(false, '/', 'Phar detected. ...')
> #16 phar:///usr/share/devshop/bin/drush/drush(114): drush_startup(Array)
> #17 /usr/share
|
infrastructure
|
cannot find profiles directory profiles it appears the profiles folder went missing likely due to a core composer scaffolding update which i think creates the folder usually this is causing multiple errors on ci which makes sure that folder exists solution is to make the directory again with a gitkeep or figure out what changed with scaffolding upstream pdoexception sqlstate integrity constraint violation column rid cannot be null in var aegir devmaster x includes database database inc stack trace var aegir devmaster x includes database database inc pdostatement execute array var aegir devmaster x includes database database inc databasestatementbase execute array array var aegir devmaster x includes database mysql query inc databaseconnection query insert into ho array array var aegir devmaster x profiles devmaster modules contrib hosting package hosting package instance inc insertquery mysql execute var aegir devmaster x profiles devmaster modules contrib hosting package hosting package instance inc hosting package instance create object stdclass var aegir devmaster x profiles devmaster modules contrib hosting package hosting package instance inc hosting package instance save object stdclass var aegir devmaster x profiles devmaster modules contrib hosting platform hosting platform drush inc hosting package instance sync null null array null array var aegir devmaster x includes module inc hosting platform post hosting verify task object stdclass array var aegir devmaster x profiles devmaster modules contrib hosting task hosting inc module invoke all post hosting ve object stdclass array phar usr share devshop bin drush includes command inc drush hosting post hosting task object stdclass phar usr share devshop bin drush includes command inc drush invoke hooks array array phar usr share devshop bin drush includes command inc drush command phar usr share devshop bin drush lib drush boot baseboot php drush dispatch array phar usr share devshop bin drush includes preflight inc drush boot baseboot bootstrap and dispatch phar usr share devshop bin drush includes startup inc drush main phar usr share devshop bin drush includes startup inc drush run main false phar detected phar usr share devshop bin drush drush drush startup array usr share
| 1
|
181,986
| 14,086,352,710
|
IssuesEvent
|
2020-11-05 03:28:02
|
CSU-Booking-Platform/application
|
https://api.github.com/repos/CSU-Booking-Platform/application
|
opened
|
Acceptance Tests for #39 Book a Room
|
acceptance-test
|
### User story
#39
### Acceptance criteria checklist
- [ ] Users can book rooms for specific time intervals
- [ ] Users cannot book a room if it is not available at that time
|
1.0
|
Acceptance Tests for #39 Book a Room - ### User story
#39
### Acceptance criteria checklist
- [ ] Users can book rooms for specific time intervals
- [ ] Users cannot book a room if it is not available at that time
|
non_infrastructure
|
acceptance tests for book a room user story acceptance criteria checklist users can book rooms for specific time intervals users cannot book a room if it is not available at that time
| 0
|
3,671
| 4,467,715,913
|
IssuesEvent
|
2016-08-25 06:31:48
|
Pinafore/qb
|
https://api.github.com/repos/Pinafore/qb
|
closed
|
Move non-compile targets to luigi
|
enhancement in progress: active infrastructure
|
Since wikification depends on the questions, this should not be in the makefile. If the questions change, this needs to be rerun.
# Tasks
- [x] Move glove target to luigi
- [x] Move wikipedia target to luigi
- [x] Move kenlm target to luigi
- [x] Move wikifier targets to luigi
- [x] Rename/move wikifier targets from `data/external` to `output`
- [x] Rename/move wikipedia target from `data/external` to `output`
|
1.0
|
Move non-compile targets to luigi - Since wikification depends on the questions, this should not be in the makefile. If the questions change, this needs to be rerun.
# Tasks
- [x] Move glove target to luigi
- [x] Move wikipedia target to luigi
- [x] Move kenlm target to luigi
- [x] Move wikifier targets to luigi
- [x] Rename/move wikifier targets from `data/external` to `output`
- [x] Rename/move wikipedia target from `data/external` to `output`
|
infrastructure
|
move non compile targets to luigi since wikification depends on the questions this should not be in the makefile if the questions change this needs to be rerun tasks move glove target to luigi move wikipedia target to luigi move kenlm target to luigi move wikifier targets to luigi rename move wikifier targets from data external to output rename move wikipedia target from data external to output
| 1
|
3,314
| 4,214,966,489
|
IssuesEvent
|
2016-06-30 00:57:28
|
azkaban/azkaban
|
https://api.github.com/repos/azkaban/azkaban
|
closed
|
Improve scalability through supporting multiple exec servers
|
enhancement infrastructure
|
## Current Challenge
Azkaban consists of two servers - web server and executor server. The current design is the web server assigns flow executions to a single executor server. Web server mainly provides the UI and the scheduling features, and executor server is responsible for running the jobs in the flow. Azkaban is running into a scalability issue due to the increased usage as well as user jobs are taking longer and longer to finish because of our large ever increasing data sets. Given the increased number of jobs the executor server has to execute concurrently, it is running into system memory capacity issue and therefore the server hosting Azkaban web server and executor is busy with performing swapping and the web server response is dramatically slow.
## Proposed Design
At the high level, Azkaban server scalability can be improvied by deploying the web server on its own box and deploying multiple executor servers on separate boxes . With this enhancement, the job executions load can spread across multiple machines and therefore Azkaban server can support more load and can tolerate server failure.
Since Azkaban web server controls the scheduler and there is only one web server, it can still guarantee that a single job is executed only once, even though there are multiple executor servers.
The implementation of this design should make sure it is still easy and possible to run Azkaban in single-server mode, which is very useful for development and debugging purposes.
## Enhancements to Web Server
Accept a list of one or more executor servers - i.e <server name>:<port>. This information should be backed by a database table so web server doesn't have to be bounced when this list is modified.
When executing a flow, web server needs to decide which executor to send a job execution request to. The logic for the load balancing algorithm should be pluggable. We will provide a default implementation out of the box.
The JMX page needs to be enhanced to show which flows are being executed on which executor server as well as their current states.
The web server may need to adjust its internal thread pool based # of running executors. This way, as we scale down from 5 to 1 executor servers, the web server doesn't overwhelm the one executor with flows to execute.
Need a new property to specify the implementation of the pluggable load balancing logic.
When it is time to figure out which executor to execute a flow, web server will delegate that responsibility to the load balancer.
## Enhancements to Executor Server
Executor server should support various states - INITIALIZING, ACCEPTING, NOT_ACCEPTING, SHUTTING_DOWN. May need to come up with better name for these states.
INITIALIZING - while server is coming up. It doesn't accept requests
ACCEPTING - up and running, accepting requests
NOT_ACCEPTING - up and running, but not accepting requests. Good for draining out the currently running flows as we are doing rolling upgrade.
SHUTTING_DOWN - coming down, not accepting requests. (not sure if we need this)
Need a new REST endpoint to change the state to NOT_ACCEPTING or ACCEPTING
While the server is INITIALIZING or SHUTTING_DOWN state, its state can't be changed
Need a new REST endpoint to return the current load information (how many flows are currently being executed)
## Flow Execution Load Balancer
The goal of this load balancer is to distribute the flow executions evenly across 1 or more executor servers
We can start with the basic logic by picking the executor server with the least amount of load
Only executor servers that are in the "ACCEPTING" state should be considered for a candidate for executing a flow
If an error while connecting to the selected executor server, the next best candidate is selected.
There should be a background thread to periodically ping the list of executors for their current status and update that information in memory.
|
1.0
|
Improve scalability through supporting multiple exec servers - ## Current Challenge
Azkaban consists of two servers - web server and executor server. The current design is the web server assigns flow executions to a single executor server. Web server mainly provides the UI and the scheduling features, and executor server is responsible for running the jobs in the flow. Azkaban is running into a scalability issue due to the increased usage as well as user jobs are taking longer and longer to finish because of our large ever increasing data sets. Given the increased number of jobs the executor server has to execute concurrently, it is running into system memory capacity issue and therefore the server hosting Azkaban web server and executor is busy with performing swapping and the web server response is dramatically slow.
## Proposed Design
At the high level, Azkaban server scalability can be improvied by deploying the web server on its own box and deploying multiple executor servers on separate boxes . With this enhancement, the job executions load can spread across multiple machines and therefore Azkaban server can support more load and can tolerate server failure.
Since Azkaban web server controls the scheduler and there is only one web server, it can still guarantee that a single job is executed only once, even though there are multiple executor servers.
The implementation of this design should make sure it is still easy and possible to run Azkaban in single-server mode, which is very useful for development and debugging purposes.
## Enhancements to Web Server
Accept a list of one or more executor servers - i.e <server name>:<port>. This information should be backed by a database table so web server doesn't have to be bounced when this list is modified.
When executing a flow, web server needs to decide which executor to send a job execution request to. The logic for the load balancing algorithm should be pluggable. We will provide a default implementation out of the box.
The JMX page needs to be enhanced to show which flows are being executed on which executor server as well as their current states.
The web server may need to adjust its internal thread pool based # of running executors. This way, as we scale down from 5 to 1 executor servers, the web server doesn't overwhelm the one executor with flows to execute.
Need a new property to specify the implementation of the pluggable load balancing logic.
When it is time to figure out which executor to execute a flow, web server will delegate that responsibility to the load balancer.
## Enhancements to Executor Server
Executor server should support various states - INITIALIZING, ACCEPTING, NOT_ACCEPTING, SHUTTING_DOWN. May need to come up with better name for these states.
INITIALIZING - while server is coming up. It doesn't accept requests
ACCEPTING - up and running, accepting requests
NOT_ACCEPTING - up and running, but not accepting requests. Good for draining out the currently running flows as we are doing rolling upgrade.
SHUTTING_DOWN - coming down, not accepting requests. (not sure if we need this)
Need a new REST endpoint to change the state to NOT_ACCEPTING or ACCEPTING
While the server is INITIALIZING or SHUTTING_DOWN state, its state can't be changed
Need a new REST endpoint to return the current load information (how many flows are currently being executed)
## Flow Execution Load Balancer
The goal of this load balancer is to distribute the flow executions evenly across 1 or more executor servers
We can start with the basic logic by picking the executor server with the least amount of load
Only executor servers that are in the "ACCEPTING" state should be considered for a candidate for executing a flow
If an error while connecting to the selected executor server, the next best candidate is selected.
There should be a background thread to periodically ping the list of executors for their current status and update that information in memory.
|
infrastructure
|
improve scalability through supporting multiple exec servers current challenge azkaban consists of two servers web server and executor server the current design is the web server assigns flow executions to a single executor server web server mainly provides the ui and the scheduling features and executor server is responsible for running the jobs in the flow azkaban is running into a scalability issue due to the increased usage as well as user jobs are taking longer and longer to finish because of our large ever increasing data sets given the increased number of jobs the executor server has to execute concurrently it is running into system memory capacity issue and therefore the server hosting azkaban web server and executor is busy with performing swapping and the web server response is dramatically slow proposed design at the high level azkaban server scalability can be improvied by deploying the web server on its own box and deploying multiple executor servers on separate boxes with this enhancement the job executions load can spread across multiple machines and therefore azkaban server can support more load and can tolerate server failure since azkaban web server controls the scheduler and there is only one web server it can still guarantee that a single job is executed only once even though there are multiple executor servers the implementation of this design should make sure it is still easy and possible to run azkaban in single server mode which is very useful for development and debugging purposes enhancements to web server accept a list of one or more executor servers i e this information should be backed by a database table so web server doesn t have to be bounced when this list is modified when executing a flow web server needs to decide which executor to send a job execution request to the logic for the load balancing algorithm should be pluggable we will provide a default implementation out of the box the jmx page needs to be enhanced to show which flows are being executed on which executor server as well as their current states the web server may need to adjust its internal thread pool based of running executors this way as we scale down from to executor servers the web server doesn t overwhelm the one executor with flows to execute need a new property to specify the implementation of the pluggable load balancing logic when it is time to figure out which executor to execute a flow web server will delegate that responsibility to the load balancer enhancements to executor server executor server should support various states initializing accepting not accepting shutting down may need to come up with better name for these states initializing while server is coming up it doesn t accept requests accepting up and running accepting requests not accepting up and running but not accepting requests good for draining out the currently running flows as we are doing rolling upgrade shutting down coming down not accepting requests not sure if we need this need a new rest endpoint to change the state to not accepting or accepting while the server is initializing or shutting down state its state can t be changed need a new rest endpoint to return the current load information how many flows are currently being executed flow execution load balancer the goal of this load balancer is to distribute the flow executions evenly across or more executor servers we can start with the basic logic by picking the executor server with the least amount of load only executor servers that are in the accepting state should be considered for a candidate for executing a flow if an error while connecting to the selected executor server the next best candidate is selected there should be a background thread to periodically ping the list of executors for their current status and update that information in memory
| 1
|
30,065
| 24,519,540,564
|
IssuesEvent
|
2022-10-11 08:28:14
|
Altinn/altinn-platform
|
https://api.github.com/repos/Altinn/altinn-platform
|
closed
|
Oppdater Terraform providere for Platform
|
ops/infrastructure
|
terraformversion: "1.3.1"
azuread_provider: "= 2.26.1" -> 2.29
azurerm_provider: "= 3.21.1" -> 2.26
helm_provider: "= 2.6.0" -> 2.7
kubernetes_provider: "= 2.12.1" -> 2.14
postgresql_provider: "= 1.16.0" -> 1.17.1
random_provider: "= 3.1.0" -> 3.4.3<
Utrulling
- [x] AT21
- [x] AT22
- [x] AT23
- [x] AT24
- [x] YT01
- [x] TT02
- [x] PROD
|
1.0
|
Oppdater Terraform providere for Platform - terraformversion: "1.3.1"
azuread_provider: "= 2.26.1" -> 2.29
azurerm_provider: "= 3.21.1" -> 2.26
helm_provider: "= 2.6.0" -> 2.7
kubernetes_provider: "= 2.12.1" -> 2.14
postgresql_provider: "= 1.16.0" -> 1.17.1
random_provider: "= 3.1.0" -> 3.4.3<
Utrulling
- [x] AT21
- [x] AT22
- [x] AT23
- [x] AT24
- [x] YT01
- [x] TT02
- [x] PROD
|
infrastructure
|
oppdater terraform providere for platform terraformversion azuread provider azurerm provider helm provider kubernetes provider postgresql provider random provider utrulling prod
| 1
|
5,495
| 5,707,411,437
|
IssuesEvent
|
2017-04-18 13:50:36
|
servicecatalog/development
|
https://api.github.com/repos/servicecatalog/development
|
closed
|
Fix travis
|
feature-branch/fb_glassfish4_upgrade infrastructure priority/P2
|
Fix travis builds for glassfish4 related branches. The build hangs after 34 minutes on the same project (oscm-intsvc-base-unittests). Find out how to exclude project from build or how to change VM on which travis executes the builds.
|
1.0
|
Fix travis - Fix travis builds for glassfish4 related branches. The build hangs after 34 minutes on the same project (oscm-intsvc-base-unittests). Find out how to exclude project from build or how to change VM on which travis executes the builds.
|
infrastructure
|
fix travis fix travis builds for related branches the build hangs after minutes on the same project oscm intsvc base unittests find out how to exclude project from build or how to change vm on which travis executes the builds
| 1
|
27,752
| 22,304,556,945
|
IssuesEvent
|
2022-06-13 11:53:12
|
flarum/framework
|
https://api.github.com/repos/flarum/framework
|
opened
|
Test with PHP 8.2
|
type/infrastructure
|
Need to:
- move workflow definitions from `.github` to the monorepo (here).
- start testing with PHP 8.2
|
1.0
|
Test with PHP 8.2 - Need to:
- move workflow definitions from `.github` to the monorepo (here).
- start testing with PHP 8.2
|
infrastructure
|
test with php need to move workflow definitions from github to the monorepo here start testing with php
| 1
|
150,042
| 13,308,259,426
|
IssuesEvent
|
2020-08-26 00:25:48
|
miurahr/picast
|
https://api.github.com/repos/miurahr/picast
|
closed
|
How to install after change in IP address but not as developer ?
|
documentation no-issue-activity question
|
How to install and run after modifying setting.ini from clone of you git repository?
|
1.0
|
How to install after change in IP address but not as developer ? - How to install and run after modifying setting.ini from clone of you git repository?
|
non_infrastructure
|
how to install after change in ip address but not as developer how to install and run after modifying setting ini from clone of you git repository
| 0
|
1,567
| 3,268,961,620
|
IssuesEvent
|
2015-10-23 14:18:26
|
dart-lang/sdk
|
https://api.github.com/repos/dart-lang/sdk
|
closed
|
apt warns about untrusted source on recent .deb update
|
Area-Infrastructure
|
Hi,
Recently I started to notice a trust issue both on Ubuntu and on my Debian servers. The official repository is generating warnings even though the official key was added.
```
root@build:~# aptitude safe-upgrade
The following packages will be upgraded:
dart
1 packages upgraded, 0 newly installed, 0 to remove and 0 not upgraded.
Need to get 14.7 MB of archives. After unpacking 1024 B will be used.
Do you want to continue? [Y/n/?] i dart
Package: dart
New: yes
State: installed; will be upgraded [1.12.1-1 -> 1.12.2-1]
Automatically installed: no
Version: 1.12.1-1
Priority: optional
Section: misc
Maintainer: Soren Gjesse <sgjesse@google.com>
Architecture: amd64
Uncompressed Size: 40.6 M
Depends: libc6 (>= 2.11), libgcc1 (>= 1:4.1.1), libstdc++6 (>= 4.4.0)
Description: Dart SDK
Press Return to continue.
WARNING: untrusted versions of the following packages will be installed!
Untrusted packages could compromise your system's security.
You should only proceed with the installation if you are certain that
this is what you want to do.
dart
Do you want to ignore this warning and proceed anyway?
To continue, enter "Yes"; to abort, enter "No": No
Abort.
```
Note that the linux signing key from Google is added to the system as described in [installing dart on linux](https://www.dartlang.org/downloads/linux.html)
```
pub 1024D/7FAC5991 2007-03-08
uid Google, Inc. Linux Package Signing Key <linux-packages-keymaster@google.com>
sub 2048g/C07CB649 2007-03-08
```
The last version that properly installed was:
```
root@build:~# dart --version
Dart VM version: 1.12.1 (Tue Sep 8 11:14:08 2015) on "linux_x64"
```
|
1.0
|
apt warns about untrusted source on recent .deb update - Hi,
Recently I started to notice a trust issue both on Ubuntu and on my Debian servers. The official repository is generating warnings even though the official key was added.
```
root@build:~# aptitude safe-upgrade
The following packages will be upgraded:
dart
1 packages upgraded, 0 newly installed, 0 to remove and 0 not upgraded.
Need to get 14.7 MB of archives. After unpacking 1024 B will be used.
Do you want to continue? [Y/n/?] i dart
Package: dart
New: yes
State: installed; will be upgraded [1.12.1-1 -> 1.12.2-1]
Automatically installed: no
Version: 1.12.1-1
Priority: optional
Section: misc
Maintainer: Soren Gjesse <sgjesse@google.com>
Architecture: amd64
Uncompressed Size: 40.6 M
Depends: libc6 (>= 2.11), libgcc1 (>= 1:4.1.1), libstdc++6 (>= 4.4.0)
Description: Dart SDK
Press Return to continue.
WARNING: untrusted versions of the following packages will be installed!
Untrusted packages could compromise your system's security.
You should only proceed with the installation if you are certain that
this is what you want to do.
dart
Do you want to ignore this warning and proceed anyway?
To continue, enter "Yes"; to abort, enter "No": No
Abort.
```
Note that the linux signing key from Google is added to the system as described in [installing dart on linux](https://www.dartlang.org/downloads/linux.html)
```
pub 1024D/7FAC5991 2007-03-08
uid Google, Inc. Linux Package Signing Key <linux-packages-keymaster@google.com>
sub 2048g/C07CB649 2007-03-08
```
The last version that properly installed was:
```
root@build:~# dart --version
Dart VM version: 1.12.1 (Tue Sep 8 11:14:08 2015) on "linux_x64"
```
|
infrastructure
|
apt warns about untrusted source on recent deb update hi recently i started to notice a trust issue both on ubuntu and on my debian servers the official repository is generating warnings even though the official key was added root build aptitude safe upgrade the following packages will be upgraded dart packages upgraded newly installed to remove and not upgraded need to get mb of archives after unpacking b will be used do you want to continue i dart package dart new yes state installed will be upgraded automatically installed no version priority optional section misc maintainer soren gjesse architecture uncompressed size m depends libstdc description dart sdk press return to continue warning untrusted versions of the following packages will be installed untrusted packages could compromise your system s security you should only proceed with the installation if you are certain that this is what you want to do dart do you want to ignore this warning and proceed anyway to continue enter yes to abort enter no no abort note that the linux signing key from google is added to the system as described in pub uid google inc linux package signing key sub the last version that properly installed was root build dart version dart vm version tue sep on linux
| 1
|
14,644
| 11,033,560,692
|
IssuesEvent
|
2019-12-06 23:28:15
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
coreclr build fails: patch-corefx.py exception
|
area-Infrastructure-coreclr untriaged
|
Built with:
```
src\coreclr\build.cmd x64 checked -priority=1
```
(presumably the `-priority=1` is not needed to repro the issue)
```
python "f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py" -clr_core_root "f:\gh\runtime\src\coreclr\..\..\artifacts\tests\coreclr\Windows_NT.x64.Checked\Tests\Core_Root" -fx_root "" -arch x64 -build_type Debug
[f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py]: Patching CoreFX binaries from local enlistment.
[f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py]: Configuration:
[f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py]: arch: x64
[f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py]: build_type: Debug
[f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py]: clr_core_root: f:\gh\runtime\artifacts\tests\coreclr\Windows_NT.x64.Checked\Tests\Core_Root
[f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py]: fx_root: .
Traceback (most recent call last):
File "f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py", line 263, in <module>
main(Args)
File "f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py", line 251, in main
raise Exception('CoreFX bin path %s does not exist.' % (fx_bin))
Exception: CoreFX bin path .\artifacts\bin\runtime\netcoreapp-Windows_NT-Debug-x64 does not exist.
```
@trylek Looks like your recent changes might have affected this?
@dotnet/runtime-infrastructure
|
1.0
|
coreclr build fails: patch-corefx.py exception - Built with:
```
src\coreclr\build.cmd x64 checked -priority=1
```
(presumably the `-priority=1` is not needed to repro the issue)
```
python "f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py" -clr_core_root "f:\gh\runtime\src\coreclr\..\..\artifacts\tests\coreclr\Windows_NT.x64.Checked\Tests\Core_Root" -fx_root "" -arch x64 -build_type Debug
[f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py]: Patching CoreFX binaries from local enlistment.
[f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py]: Configuration:
[f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py]: arch: x64
[f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py]: build_type: Debug
[f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py]: clr_core_root: f:\gh\runtime\artifacts\tests\coreclr\Windows_NT.x64.Checked\Tests\Core_Root
[f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py]: fx_root: .
Traceback (most recent call last):
File "f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py", line 263, in <module>
main(Args)
File "f:\gh\runtime\src\coreclr\tests\scripts\patch-corefx.py", line 251, in main
raise Exception('CoreFX bin path %s does not exist.' % (fx_bin))
Exception: CoreFX bin path .\artifacts\bin\runtime\netcoreapp-Windows_NT-Debug-x64 does not exist.
```
@trylek Looks like your recent changes might have affected this?
@dotnet/runtime-infrastructure
|
infrastructure
|
coreclr build fails patch corefx py exception built with src coreclr build cmd checked priority presumably the priority is not needed to repro the issue python f gh runtime src coreclr tests scripts patch corefx py clr core root f gh runtime src coreclr artifacts tests coreclr windows nt checked tests core root fx root arch build type debug patching corefx binaries from local enlistment configuration arch build type debug clr core root f gh runtime artifacts tests coreclr windows nt checked tests core root fx root traceback most recent call last file f gh runtime src coreclr tests scripts patch corefx py line in main args file f gh runtime src coreclr tests scripts patch corefx py line in main raise exception corefx bin path s does not exist fx bin exception corefx bin path artifacts bin runtime netcoreapp windows nt debug does not exist trylek looks like your recent changes might have affected this dotnet runtime infrastructure
| 1
|
30,070
| 6,010,288,679
|
IssuesEvent
|
2017-06-06 12:52:02
|
bridgedotnet/Bridge
|
https://api.github.com/repos/bridgedotnet/Bridge
|
closed
|
Metadata from dependant libraries not included in output
|
defect
|
(I'm using code from `master`, which I believe includes the recent reflection changes)
I have two Bridge libraries, TestFramework and Tests. Tests has a reference to TestFramework. When TestFramework builds, it outputs a `.meta.js` file. However, this file never makes it to the output of Tests:
- `combineScripts: true` - Tests.js ends up with the reflection metadata for the Tests project, but not for the TestRunner project.
- `combineScripts: false` then Tests.meta.js has the reflection metadata for the Tests project, but there is no TestRunner.meta.js
So it seems that no matter what I do, I am unable to get the reflection metadata for the referenced project.
Is this a bug, or is there something else I need to do?
|
1.0
|
Metadata from dependant libraries not included in output - (I'm using code from `master`, which I believe includes the recent reflection changes)
I have two Bridge libraries, TestFramework and Tests. Tests has a reference to TestFramework. When TestFramework builds, it outputs a `.meta.js` file. However, this file never makes it to the output of Tests:
- `combineScripts: true` - Tests.js ends up with the reflection metadata for the Tests project, but not for the TestRunner project.
- `combineScripts: false` then Tests.meta.js has the reflection metadata for the Tests project, but there is no TestRunner.meta.js
So it seems that no matter what I do, I am unable to get the reflection metadata for the referenced project.
Is this a bug, or is there something else I need to do?
|
non_infrastructure
|
metadata from dependant libraries not included in output i m using code from master which i believe includes the recent reflection changes i have two bridge libraries testframework and tests tests has a reference to testframework when testframework builds it outputs a meta js file however this file never makes it to the output of tests combinescripts true tests js ends up with the reflection metadata for the tests project but not for the testrunner project combinescripts false then tests meta js has the reflection metadata for the tests project but there is no testrunner meta js so it seems that no matter what i do i am unable to get the reflection metadata for the referenced project is this a bug or is there something else i need to do
| 0
|
96,226
| 12,103,759,155
|
IssuesEvent
|
2020-04-20 18:58:52
|
microsoft/BotFramework-Composer
|
https://api.github.com/repos/microsoft/BotFramework-Composer
|
closed
|
enable hide left panel and right panel; show scroll bar when can't show all
|
Type: Enhancement UX Design skill-reported
|
## Is your feature request related to a problem? Please describe.
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
I'm always frustrated when I am comparing two dialogs and I don't have a big monitor or two monitors:

It is very crowded and it should show scroll bars in this situation.
## Describe the solution you'd like
<!-- A clear and concise description of what you want to happen. -->
Enable hide and even auto-hide (only show when click an element)
## Describe alternatives you've considered
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
## Additional context
<!-- Add any other context or screenshots about the feature request here. -->
|
1.0
|
enable hide left panel and right panel; show scroll bar when can't show all - ## Is your feature request related to a problem? Please describe.
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
I'm always frustrated when I am comparing two dialogs and I don't have a big monitor or two monitors:

It is very crowded and it should show scroll bars in this situation.
## Describe the solution you'd like
<!-- A clear and concise description of what you want to happen. -->
Enable hide and even auto-hide (only show when click an element)
## Describe alternatives you've considered
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
## Additional context
<!-- Add any other context or screenshots about the feature request here. -->
|
non_infrastructure
|
enable hide left panel and right panel show scroll bar when can t show all is your feature request related to a problem please describe i m always frustrated when i am comparing two dialogs and i don t have a big monitor or two monitors it is very crowded and it should show scroll bars in this situation describe the solution you d like enable hide and even auto hide only show when click an element describe alternatives you ve considered additional context
| 0
|
18,750
| 13,091,767,790
|
IssuesEvent
|
2020-08-03 07:18:11
|
containous/maesh
|
https://api.github.com/repos/containous/maesh
|
opened
|
Upgrade Traefik dependency version to v2.3
|
area/infrastructure area/traefik kind/enhancement
|
## Feature Request
### Proposal
Upgrade Traefik dependency version to v2.3 when it is available. This version update is required because of the `smi-sdk-go` update to v0.4 which relies on k8s v0.18.x.
|
1.0
|
Upgrade Traefik dependency version to v2.3 - ## Feature Request
### Proposal
Upgrade Traefik dependency version to v2.3 when it is available. This version update is required because of the `smi-sdk-go` update to v0.4 which relies on k8s v0.18.x.
|
infrastructure
|
upgrade traefik dependency version to feature request proposal upgrade traefik dependency version to when it is available this version update is required because of the smi sdk go update to which relies on x
| 1
|
8,747
| 7,606,791,817
|
IssuesEvent
|
2018-04-30 14:30:26
|
servo/servo
|
https://api.github.com/repos/servo/servo
|
opened
|
File an issue when downstream WPT sync fails for some reason
|
A-infrastructure A-testing
|
When the syncing process breaks the process of running tests or updating the expected test results, we silently stop receiving PRs with the sync changes. It would be best to file a github issue with the log output so we don't need to go hunting for it when we notice several days later.
|
1.0
|
File an issue when downstream WPT sync fails for some reason - When the syncing process breaks the process of running tests or updating the expected test results, we silently stop receiving PRs with the sync changes. It would be best to file a github issue with the log output so we don't need to go hunting for it when we notice several days later.
|
infrastructure
|
file an issue when downstream wpt sync fails for some reason when the syncing process breaks the process of running tests or updating the expected test results we silently stop receiving prs with the sync changes it would be best to file a github issue with the log output so we don t need to go hunting for it when we notice several days later
| 1
|
135,715
| 19,654,089,517
|
IssuesEvent
|
2022-01-10 10:37:04
|
Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
|
https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
|
opened
|
First Death Knights Look
|
suggestion :question: priority critical :exclamation: character designing 😶
|
<!--
**DO NOT REMOVE PRE-EXISTING LINES**
------------------------------------------------------------------------------------------------------------
-->
**Describe your suggestion in full detail below:**
Teron'gor and Ragnok should look like orcs before turning into death knights and like humans after.
|
1.0
|
First Death Knights Look - <!--
**DO NOT REMOVE PRE-EXISTING LINES**
------------------------------------------------------------------------------------------------------------
-->
**Describe your suggestion in full detail below:**
Teron'gor and Ragnok should look like orcs before turning into death knights and like humans after.
|
non_infrastructure
|
first death knights look do not remove pre existing lines describe your suggestion in full detail below teron gor and ragnok should look like orcs before turning into death knights and like humans after
| 0
|
345,369
| 30,806,350,160
|
IssuesEvent
|
2023-08-01 07:27:01
|
codestates-seb/ShellWe
|
https://api.github.com/repos/codestates-seb/ShellWe
|
closed
|
도메인 테스트 코드
|
Backend feat refactor test
|
# 도메인 테스트 코드 작성
## 순서
- [ ] shell
- [ ] member
- [ ] trade
- [ ] cart
- [ ] picture
- [ ] category
- [ ] tag
|
1.0
|
도메인 테스트 코드 - # 도메인 테스트 코드 작성
## 순서
- [ ] shell
- [ ] member
- [ ] trade
- [ ] cart
- [ ] picture
- [ ] category
- [ ] tag
|
non_infrastructure
|
도메인 테스트 코드 도메인 테스트 코드 작성 순서 shell member trade cart picture category tag
| 0
|
107,418
| 11,542,824,366
|
IssuesEvent
|
2020-02-18 08:22:58
|
ampproject/amp.dev
|
https://api.github.com/repos/ampproject/amp.dev
|
closed
|
Component reference: replace sample row in overview table
|
Category: Documentation P2: Medium Status: WIP Type: Update
|
We should find a better way to integrate sample references:
<img width="802" alt="Screen Shot 2019-05-16 at 20 11 32" src="https://user-images.githubusercontent.com/380472/57879095-eef51500-781b-11e9-8251-3383beb495c2.png">
Many of them also still mention AMP by Example.
|
1.0
|
Component reference: replace sample row in overview table - We should find a better way to integrate sample references:
<img width="802" alt="Screen Shot 2019-05-16 at 20 11 32" src="https://user-images.githubusercontent.com/380472/57879095-eef51500-781b-11e9-8251-3383beb495c2.png">
Many of them also still mention AMP by Example.
|
non_infrastructure
|
component reference replace sample row in overview table we should find a better way to integrate sample references img width alt screen shot at src many of them also still mention amp by example
| 0
|
5,291
| 5,556,338,480
|
IssuesEvent
|
2017-03-24 08:52:15
|
hzi-braunschweig/SORMAS-Open
|
https://api.github.com/repos/hzi-braunschweig/SORMAS-Open
|
closed
|
Short captions for personas (user roles) [0]
|
accepted Infrastructure optional sormas-api sormas-app sormas-backend
|
Should be used for the text representation of a user (e.g. "Sanaa OBASANJO - SurvOff")
SurvOff, ContOff, CaseOff, RumOff, SurvSup, ContSup, CaseSup, LabMan
|
1.0
|
Short captions for personas (user roles) [0] - Should be used for the text representation of a user (e.g. "Sanaa OBASANJO - SurvOff")
SurvOff, ContOff, CaseOff, RumOff, SurvSup, ContSup, CaseSup, LabMan
|
infrastructure
|
short captions for personas user roles should be used for the text representation of a user e g sanaa obasanjo survoff survoff contoff caseoff rumoff survsup contsup casesup labman
| 1
|
21,145
| 14,399,939,796
|
IssuesEvent
|
2020-12-03 11:38:48
|
wellcomecollection/platform
|
https://api.github.com/repos/wellcomecollection/platform
|
opened
|
Logging failure: Add a throttle to fluentbit container
|
🌟Enhancement 🚧 Infrastructure
|
We’ve seen problems with logging.wellcomecollection.org when under high load - this can be caused by noisy services, or high volume events (reindexes, large scale operations on the storage service).
While we should be be aware these things might cause issues in log/metric recording the logging cluster is not resilient to them.
All fluentbit containers talk directly with Elasticsearch and do not have a throttle or circuit breaker.
We should throttle logs at the service level by adding a throttle filter to the base fluentbit config , see: https://fluentbit.io/documentation/0.13/filter/throttle.html
Base fluentbit image is defined here: https://github.com/wellcomecollection/platform-infrastructure/tree/master/images/fluentbit
|
1.0
|
Logging failure: Add a throttle to fluentbit container - We’ve seen problems with logging.wellcomecollection.org when under high load - this can be caused by noisy services, or high volume events (reindexes, large scale operations on the storage service).
While we should be be aware these things might cause issues in log/metric recording the logging cluster is not resilient to them.
All fluentbit containers talk directly with Elasticsearch and do not have a throttle or circuit breaker.
We should throttle logs at the service level by adding a throttle filter to the base fluentbit config , see: https://fluentbit.io/documentation/0.13/filter/throttle.html
Base fluentbit image is defined here: https://github.com/wellcomecollection/platform-infrastructure/tree/master/images/fluentbit
|
infrastructure
|
logging failure add a throttle to fluentbit container we’ve seen problems with logging wellcomecollection org when under high load this can be caused by noisy services or high volume events reindexes large scale operations on the storage service while we should be be aware these things might cause issues in log metric recording the logging cluster is not resilient to them all fluentbit containers talk directly with elasticsearch and do not have a throttle or circuit breaker we should throttle logs at the service level by adding a throttle filter to the base fluentbit config see base fluentbit image is defined here
| 1
|
33,527
| 27,551,698,447
|
IssuesEvent
|
2023-03-07 15:19:45
|
spdk/spdk
|
https://api.github.com/repos/spdk/spdk
|
closed
|
opal_revert_cleanup fails under WFP10
|
Infrastructure Intermittent Failure
|
# CI Intermittent Failure
https://ci.spdk.io/results/autotest-per-patch/builds/98893/archive/vfio-user-autotest/build.log
This node should not be enabled in the pool, got there by accident - all nvmes there are present behind the VMD (our autotest is not suite for that).
```
00:04:34.410 13:12:59 -- spdk/autotest.sh@135 -- # /var/jenkins/workspace/vfio-user-autotest/spdk/scripts/setup.sh
00:04:34.977 10001:02:00.0 (8086 0a54): Active devices: data@nvme1n1, so not binding PCI dev
00:04:35.236 0000:17:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:17:05.5
00:04:35.236 0000:5d:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:5d:05.5
00:04:35.236 0000:85:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:85:05.5
00:04:35.236 0000:ae:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:ae:05.5
00:04:35.236 0000:d7:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:d7:05.5
00:04:35.236 0000:00:04.7 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:00:04.6 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:00:04.5 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:00:04.4 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:00:04.3 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:00:04.2 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:00:04.1 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:00:04.0 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:80:04.7 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:80:04.6 (8086 2021): ioatdma -> vfio-pci
00:04:35.494 0000:80:04.5 (8086 2021): ioatdma -> vfio-pci
00:04:35.494 0000:80:04.4 (8086 2021): ioatdma -> vfio-pci
00:04:35.494 0000:80:04.3 (8086 2021): ioatdma -> vfio-pci
00:04:35.494 0000:80:04.2 (8086 2021): ioatdma -> vfio-pci
00:04:35.494 0000:80:04.1 (8086 2021): ioatdma -> vfio-pci
00:04:35.494 0000:80:04.0 (8086 2021): ioatdma -> vfio-pci
00:04:36.428 10001:01:00.0 (8086 0a54): nvme -> vfio-pci
00:04:36.428 10004:01:00.0 (8086 0a54): nvme -> vfio-pci
00:04:36.686 13:13:02 -- spdk/autotest.sh@136 -- # timing_exit afterboot
00:04:36.686 13:13:02 -- common/autotest_common.sh@690 -- # xtrace_disable
00:04:36.686 13:13:02 -- common/autotest_common.sh@19 -- # '[' no '!=' yes ']'
00:04:36.686 13:13:02 -- common/autotest_common.sh@20 -- # PREV_BASH_OPTS=ehxBET
00:04:36.686 13:13:02 -- common/autotest_common.sh@21 -- # [[ ehxBET == *\x* ]]
00:04:36.686 13:13:02 -- common/autotest_common.sh@22 -- # XTRACE_DISABLED=yes
00:04:36.686 13:13:02 -- common/autotest_common.sh@24 -- # set +x
00:04:36.687 13:13:02 -- common/autotest_common.sh@692 -- # xtrace_enable
00:04:36.687 13:13:02 -- spdk/autotest.sh@140 -- # opal_revert_cleanup
00:04:36.687 13:13:02 -- common/autotest_common.sh@1517 -- # mapfile -t bdfs
00:04:36.687 13:13:02 -- common/autotest_common.sh@1517 -- # get_nvme_bdfs_by_id 0x0a54
00:04:36.687 13:13:02 -- common/autotest_common.sh@1503 -- # bdfs=()
00:04:36.687 13:13:02 -- common/autotest_common.sh@1503 -- # local bdfs
00:04:36.687 13:13:02 -- common/autotest_common.sh@1505 -- # get_nvme_bdfs
00:04:36.687 13:13:02 -- common/autotest_common.sh@1443 -- # xtrace_disable
00:04:36.687 13:13:02 -- common/autotest_common.sh@19 -- # '[' no '!=' yes ']'
00:04:36.687 13:13:02 -- common/autotest_common.sh@20 -- # PREV_BASH_OPTS=hxBET
00:04:36.687 13:13:02 -- common/autotest_common.sh@21 -- # [[ hxBET == *\x* ]]
00:04:36.687 13:13:02 -- common/autotest_common.sh@22 -- # XTRACE_DISABLED=yes
00:04:36.687 13:13:02 -- common/autotest_common.sh@24 -- # set +x
00:04:36.687 13:13:02 -- common/autotest_common.sh@1450 -- # xtrace_enable
00:04:36.687 13:13:02 -- common/autotest_common.sh@1505 -- # for bdf in $(get_nvme_bdfs)
00:04:36.687 13:13:02 -- common/autotest_common.sh@1506 -- # cat /sys/bus/pci/devices/10001:01:00.0/device
00:04:36.687 13:13:02 -- common/autotest_common.sh@1506 -- # device=0x0a54
00:04:36.687 13:13:02 -- common/autotest_common.sh@1507 -- # [[ 0x0a54 == \0\x\0\a\5\4 ]]
00:04:36.687 13:13:02 -- common/autotest_common.sh@1508 -- # bdfs+=($bdf)
00:04:36.687 13:13:02 -- common/autotest_common.sh@1505 -- # for bdf in $(get_nvme_bdfs)
00:04:36.687 13:13:02 -- common/autotest_common.sh@1506 -- # cat /sys/bus/pci/devices/10004:01:00.0/device
00:04:36.687 13:13:02 -- common/autotest_common.sh@1506 -- # device=0x0a54
00:04:36.687 13:13:02 -- common/autotest_common.sh@1507 -- # [[ 0x0a54 == \0\x\0\a\5\4 ]]
00:04:36.687 13:13:02 -- common/autotest_common.sh@1508 -- # bdfs+=($bdf)
00:04:36.687 13:13:02 -- common/autotest_common.sh@1512 -- # printf '%s\n' 10001:01:00.0 10004:01:00.0
00:04:36.687 13:13:02 -- common/autotest_common.sh@1518 -- # [[ -z 10001:01:00.0 ]]
00:04:36.687 13:13:02 -- common/autotest_common.sh@1523 -- # spdk_tgt_pid=60923
00:04:36.687 13:13:02 -- common/autotest_common.sh@1522 -- # /var/jenkins/workspace/vfio-user-autotest/spdk/build/bin/spdk_tgt
00:04:36.687 13:13:02 -- common/autotest_common.sh@1524 -- # waitforlisten 60923
00:04:36.687 13:13:02 -- common/autotest_common.sh@793 -- # '[' -z 60923 ']'
00:04:36.687 13:13:02 -- common/autotest_common.sh@797 -- # local rpc_addr=/var/tmp/spdk.sock
00:04:36.687 13:13:02 -- common/autotest_common.sh@798 -- # local max_retries=100
00:04:36.687 13:13:02 -- common/autotest_common.sh@800 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...'
00:04:36.687 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...
00:04:36.687 13:13:02 -- common/autotest_common.sh@802 -- # xtrace_disable
00:04:36.687 13:13:02 -- common/autotest_common.sh@19 -- # '[' no '!=' yes ']'
00:04:36.687 13:13:02 -- common/autotest_common.sh@20 -- # PREV_BASH_OPTS=ehxBET
00:04:36.687 13:13:02 -- common/autotest_common.sh@21 -- # [[ ehxBET == *\x* ]]
00:04:36.687 13:13:02 -- common/autotest_common.sh@22 -- # XTRACE_DISABLED=yes
00:04:36.687 13:13:02 -- common/autotest_common.sh@24 -- # set +x
00:04:36.945 [2023-02-21 13:13:02.407453] Starting SPDK v23.05-pre git sha1 96a8515b2 / DPDK 22.11.1 initialization...
00:04:36.945 [2023-02-21 13:13:02.407567] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid60923 ]
00:04:36.945 EAL: No free 2048 kB hugepages reported on node 1
00:04:36.945 TELEMETRY: No legacy callbacks, legacy socket not created
00:04:37.204 [2023-02-21 13:13:02.557584] app.c: 712:spdk_app_start: *NOTICE*: Total cores available: 1
00:04:37.204 [2023-02-21 13:13:02.646089] reactor.c: 926:reactor_run: *NOTICE*: Reactor started on core 0
00:04:37.204 [2023-02-21 13:13:02.677720] accel_sw.c: 681:sw_accel_module_init: *NOTICE*: Accel framework software module initialized.
00:04:37.770 13:13:03 -- common/autotest_common.sh@821 -- # xtrace_enable
00:04:37.770 13:13:03 -- common/autotest_common.sh@822 -- # (( i == 0 ))
00:04:37.770 13:13:03 -- common/autotest_common.sh@826 -- # return 0
00:04:37.770 13:13:03 -- common/autotest_common.sh@1526 -- # bdf_id=0
00:04:37.770 13:13:03 -- common/autotest_common.sh@1527 -- # for bdf in "${bdfs[@]}"
00:04:37.770 13:13:03 -- common/autotest_common.sh@1528 -- # /var/jenkins/workspace/vfio-user-autotest/spdk/scripts/rpc.py bdev_nvme_attach_controller -b nvme0 -t pcie -a 10001:01:00.0
00:04:38.029 EAL: 10001:01:00.0 VFIO group is not viable! Not all devices in IOMMU group bound to VFIO or unbound
00:04:38.029 EAL: Driver cannot attach the device (10001:01:00.0)
00:04:38.029 EAL: Failed to attach device on primary process
00:04:38.029 [2023-02-21 13:13:03.503140] nvme.c: 815:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed
00:04:38.029 [2023-02-21 13:13:03.503174] bdev_nvme.c:5299:bdev_nvme_create: *ERROR*: No controller was found with provided trid (traddr: 10001:01:00.0)
00:04:38.029 request:
00:04:38.029 {
00:04:38.029 "name": "nvme0",
00:04:38.029 "trtype": "pcie",
00:04:38.029 "traddr": "10001:01:00.0",
00:04:38.029 "method": "bdev_nvme_attach_controller",
00:04:38.029 "req_id": 1
00:04:38.029 }
00:04:38.029 Got JSON-RPC error response
00:04:38.029 response:
00:04:38.029 {
00:04:38.029 "code": -19,
00:04:38.029 "message": "No such device"
00:04:38.029 }
00:04:38.029 13:13:03 -- common/autotest_common.sh@1528 -- # trap - ERR
00:04:38.029 13:13:03 -- common/autotest_common.sh@1528 -- # print_backtrace
00:04:38.029 13:13:03 -- common/autotest_common.sh@1100 -- # [[ ehxBET =~ e ]]
00:04:38.029 13:13:03 -- common/autotest_common.sh@1102 -- # args=('/var/jenkins/workspace/vfio-user-autotest/autorun-spdk.conf')
00:04:38.029 13:13:03 -- common/autotest_common.sh@1102 -- # local args
00:04:38.029 13:13:03 -- common/autotest_common.sh@1104 -- # xtrace_disable
00:04:38.029 13:13:03 -- common/autotest_common.sh@19 -- # '[' no '!=' yes ']'
00:04:38.029 13:13:03 -- common/autotest_common.sh@20 -- # PREV_BASH_OPTS=ehxBET
00:04:38.029 13:13:03 -- common/autotest_common.sh@21 -- # [[ ehxBET == *\x* ]]
00:04:38.029 13:13:03 -- common/autotest_common.sh@22 -- # XTRACE_DISABLED=yes
00:04:38.029 13:13:03 -- common/autotest_common.sh@24 -- # set +x
00:04:38.029 ========== Backtrace start: ==========
00:04:38.029
00:04:38.029 in /var/jenkins/workspace/vfio-user-autotest/spdk/test/common/autotest_common.sh:1528 -> opal_revert_cleanup([])
00:04:38.029 ...
00:04:38.029 1523 spdk_tgt_pid=$!
00:04:38.029 1524 waitforlisten $spdk_tgt_pid
00:04:38.029 1525
00:04:38.029 1526 bdf_id=0
00:04:38.029 1527 for bdf in "${bdfs[@]}"; do
00:04:38.029 1528 $rootdir/scripts/rpc.py bdev_nvme_attach_controller -b "nvme"${bdf_id} -t "pcie" -a ${bdf}
00:04:38.029 1529 # Ignore if this fails.
00:04:38.029 1530 $rootdir/scripts/rpc.py bdev_nvme_opal_revert -b "nvme"${bdf_id} -p test || true
00:04:38.029 1531 ((++bdf_id))
00:04:38.029 1532 done
00:04:38.029 1533
00:04:38.029 ...
00:04:38.029 in /var/jenkins/workspace/vfio-user-autotest/spdk/autotest.sh:140 -> main(["/var/jenkins/workspace/vfio-user-autotest/autorun-spdk.conf"])
00:04:38.029 ...
00:04:38.029 135 $rootdir/scripts/setup.sh
00:04:38.029 136 timing_exit afterboot
00:04:38.029 137
00:04:38.029 138 # Revert existing OPAL to factory settings that may have been left from earlier failed tests.
00:04:38.029 139 # This ensures we won't hit any unexpected failures due to NVMe SSDs being locked.
00:04:38.029 => 140 opal_revert_cleanup
00:04:38.029 141
00:04:38.029 142 #####################
00:04:38.029 143 # Unit Tests
00:04:38.029 144 #####################
00:04:38.029 145
00:04:38.029 ...
00:04:38.029
00:04:38.029 ========== Backtrace end ==========
```
|
1.0
|
opal_revert_cleanup fails under WFP10 - # CI Intermittent Failure
https://ci.spdk.io/results/autotest-per-patch/builds/98893/archive/vfio-user-autotest/build.log
This node should not be enabled in the pool, got there by accident - all nvmes there are present behind the VMD (our autotest is not suite for that).
```
00:04:34.410 13:12:59 -- spdk/autotest.sh@135 -- # /var/jenkins/workspace/vfio-user-autotest/spdk/scripts/setup.sh
00:04:34.977 10001:02:00.0 (8086 0a54): Active devices: data@nvme1n1, so not binding PCI dev
00:04:35.236 0000:17:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:17:05.5
00:04:35.236 0000:5d:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:5d:05.5
00:04:35.236 0000:85:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:85:05.5
00:04:35.236 0000:ae:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:ae:05.5
00:04:35.236 0000:d7:05.5 (8086 201d): Skipping not allowed VMD controller at 0000:d7:05.5
00:04:35.236 0000:00:04.7 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:00:04.6 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:00:04.5 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:00:04.4 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:00:04.3 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:00:04.2 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:00:04.1 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:00:04.0 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:80:04.7 (8086 2021): ioatdma -> vfio-pci
00:04:35.236 0000:80:04.6 (8086 2021): ioatdma -> vfio-pci
00:04:35.494 0000:80:04.5 (8086 2021): ioatdma -> vfio-pci
00:04:35.494 0000:80:04.4 (8086 2021): ioatdma -> vfio-pci
00:04:35.494 0000:80:04.3 (8086 2021): ioatdma -> vfio-pci
00:04:35.494 0000:80:04.2 (8086 2021): ioatdma -> vfio-pci
00:04:35.494 0000:80:04.1 (8086 2021): ioatdma -> vfio-pci
00:04:35.494 0000:80:04.0 (8086 2021): ioatdma -> vfio-pci
00:04:36.428 10001:01:00.0 (8086 0a54): nvme -> vfio-pci
00:04:36.428 10004:01:00.0 (8086 0a54): nvme -> vfio-pci
00:04:36.686 13:13:02 -- spdk/autotest.sh@136 -- # timing_exit afterboot
00:04:36.686 13:13:02 -- common/autotest_common.sh@690 -- # xtrace_disable
00:04:36.686 13:13:02 -- common/autotest_common.sh@19 -- # '[' no '!=' yes ']'
00:04:36.686 13:13:02 -- common/autotest_common.sh@20 -- # PREV_BASH_OPTS=ehxBET
00:04:36.686 13:13:02 -- common/autotest_common.sh@21 -- # [[ ehxBET == *\x* ]]
00:04:36.686 13:13:02 -- common/autotest_common.sh@22 -- # XTRACE_DISABLED=yes
00:04:36.686 13:13:02 -- common/autotest_common.sh@24 -- # set +x
00:04:36.687 13:13:02 -- common/autotest_common.sh@692 -- # xtrace_enable
00:04:36.687 13:13:02 -- spdk/autotest.sh@140 -- # opal_revert_cleanup
00:04:36.687 13:13:02 -- common/autotest_common.sh@1517 -- # mapfile -t bdfs
00:04:36.687 13:13:02 -- common/autotest_common.sh@1517 -- # get_nvme_bdfs_by_id 0x0a54
00:04:36.687 13:13:02 -- common/autotest_common.sh@1503 -- # bdfs=()
00:04:36.687 13:13:02 -- common/autotest_common.sh@1503 -- # local bdfs
00:04:36.687 13:13:02 -- common/autotest_common.sh@1505 -- # get_nvme_bdfs
00:04:36.687 13:13:02 -- common/autotest_common.sh@1443 -- # xtrace_disable
00:04:36.687 13:13:02 -- common/autotest_common.sh@19 -- # '[' no '!=' yes ']'
00:04:36.687 13:13:02 -- common/autotest_common.sh@20 -- # PREV_BASH_OPTS=hxBET
00:04:36.687 13:13:02 -- common/autotest_common.sh@21 -- # [[ hxBET == *\x* ]]
00:04:36.687 13:13:02 -- common/autotest_common.sh@22 -- # XTRACE_DISABLED=yes
00:04:36.687 13:13:02 -- common/autotest_common.sh@24 -- # set +x
00:04:36.687 13:13:02 -- common/autotest_common.sh@1450 -- # xtrace_enable
00:04:36.687 13:13:02 -- common/autotest_common.sh@1505 -- # for bdf in $(get_nvme_bdfs)
00:04:36.687 13:13:02 -- common/autotest_common.sh@1506 -- # cat /sys/bus/pci/devices/10001:01:00.0/device
00:04:36.687 13:13:02 -- common/autotest_common.sh@1506 -- # device=0x0a54
00:04:36.687 13:13:02 -- common/autotest_common.sh@1507 -- # [[ 0x0a54 == \0\x\0\a\5\4 ]]
00:04:36.687 13:13:02 -- common/autotest_common.sh@1508 -- # bdfs+=($bdf)
00:04:36.687 13:13:02 -- common/autotest_common.sh@1505 -- # for bdf in $(get_nvme_bdfs)
00:04:36.687 13:13:02 -- common/autotest_common.sh@1506 -- # cat /sys/bus/pci/devices/10004:01:00.0/device
00:04:36.687 13:13:02 -- common/autotest_common.sh@1506 -- # device=0x0a54
00:04:36.687 13:13:02 -- common/autotest_common.sh@1507 -- # [[ 0x0a54 == \0\x\0\a\5\4 ]]
00:04:36.687 13:13:02 -- common/autotest_common.sh@1508 -- # bdfs+=($bdf)
00:04:36.687 13:13:02 -- common/autotest_common.sh@1512 -- # printf '%s\n' 10001:01:00.0 10004:01:00.0
00:04:36.687 13:13:02 -- common/autotest_common.sh@1518 -- # [[ -z 10001:01:00.0 ]]
00:04:36.687 13:13:02 -- common/autotest_common.sh@1523 -- # spdk_tgt_pid=60923
00:04:36.687 13:13:02 -- common/autotest_common.sh@1522 -- # /var/jenkins/workspace/vfio-user-autotest/spdk/build/bin/spdk_tgt
00:04:36.687 13:13:02 -- common/autotest_common.sh@1524 -- # waitforlisten 60923
00:04:36.687 13:13:02 -- common/autotest_common.sh@793 -- # '[' -z 60923 ']'
00:04:36.687 13:13:02 -- common/autotest_common.sh@797 -- # local rpc_addr=/var/tmp/spdk.sock
00:04:36.687 13:13:02 -- common/autotest_common.sh@798 -- # local max_retries=100
00:04:36.687 13:13:02 -- common/autotest_common.sh@800 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...'
00:04:36.687 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...
00:04:36.687 13:13:02 -- common/autotest_common.sh@802 -- # xtrace_disable
00:04:36.687 13:13:02 -- common/autotest_common.sh@19 -- # '[' no '!=' yes ']'
00:04:36.687 13:13:02 -- common/autotest_common.sh@20 -- # PREV_BASH_OPTS=ehxBET
00:04:36.687 13:13:02 -- common/autotest_common.sh@21 -- # [[ ehxBET == *\x* ]]
00:04:36.687 13:13:02 -- common/autotest_common.sh@22 -- # XTRACE_DISABLED=yes
00:04:36.687 13:13:02 -- common/autotest_common.sh@24 -- # set +x
00:04:36.945 [2023-02-21 13:13:02.407453] Starting SPDK v23.05-pre git sha1 96a8515b2 / DPDK 22.11.1 initialization...
00:04:36.945 [2023-02-21 13:13:02.407567] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid60923 ]
00:04:36.945 EAL: No free 2048 kB hugepages reported on node 1
00:04:36.945 TELEMETRY: No legacy callbacks, legacy socket not created
00:04:37.204 [2023-02-21 13:13:02.557584] app.c: 712:spdk_app_start: *NOTICE*: Total cores available: 1
00:04:37.204 [2023-02-21 13:13:02.646089] reactor.c: 926:reactor_run: *NOTICE*: Reactor started on core 0
00:04:37.204 [2023-02-21 13:13:02.677720] accel_sw.c: 681:sw_accel_module_init: *NOTICE*: Accel framework software module initialized.
00:04:37.770 13:13:03 -- common/autotest_common.sh@821 -- # xtrace_enable
00:04:37.770 13:13:03 -- common/autotest_common.sh@822 -- # (( i == 0 ))
00:04:37.770 13:13:03 -- common/autotest_common.sh@826 -- # return 0
00:04:37.770 13:13:03 -- common/autotest_common.sh@1526 -- # bdf_id=0
00:04:37.770 13:13:03 -- common/autotest_common.sh@1527 -- # for bdf in "${bdfs[@]}"
00:04:37.770 13:13:03 -- common/autotest_common.sh@1528 -- # /var/jenkins/workspace/vfio-user-autotest/spdk/scripts/rpc.py bdev_nvme_attach_controller -b nvme0 -t pcie -a 10001:01:00.0
00:04:38.029 EAL: 10001:01:00.0 VFIO group is not viable! Not all devices in IOMMU group bound to VFIO or unbound
00:04:38.029 EAL: Driver cannot attach the device (10001:01:00.0)
00:04:38.029 EAL: Failed to attach device on primary process
00:04:38.029 [2023-02-21 13:13:03.503140] nvme.c: 815:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed
00:04:38.029 [2023-02-21 13:13:03.503174] bdev_nvme.c:5299:bdev_nvme_create: *ERROR*: No controller was found with provided trid (traddr: 10001:01:00.0)
00:04:38.029 request:
00:04:38.029 {
00:04:38.029 "name": "nvme0",
00:04:38.029 "trtype": "pcie",
00:04:38.029 "traddr": "10001:01:00.0",
00:04:38.029 "method": "bdev_nvme_attach_controller",
00:04:38.029 "req_id": 1
00:04:38.029 }
00:04:38.029 Got JSON-RPC error response
00:04:38.029 response:
00:04:38.029 {
00:04:38.029 "code": -19,
00:04:38.029 "message": "No such device"
00:04:38.029 }
00:04:38.029 13:13:03 -- common/autotest_common.sh@1528 -- # trap - ERR
00:04:38.029 13:13:03 -- common/autotest_common.sh@1528 -- # print_backtrace
00:04:38.029 13:13:03 -- common/autotest_common.sh@1100 -- # [[ ehxBET =~ e ]]
00:04:38.029 13:13:03 -- common/autotest_common.sh@1102 -- # args=('/var/jenkins/workspace/vfio-user-autotest/autorun-spdk.conf')
00:04:38.029 13:13:03 -- common/autotest_common.sh@1102 -- # local args
00:04:38.029 13:13:03 -- common/autotest_common.sh@1104 -- # xtrace_disable
00:04:38.029 13:13:03 -- common/autotest_common.sh@19 -- # '[' no '!=' yes ']'
00:04:38.029 13:13:03 -- common/autotest_common.sh@20 -- # PREV_BASH_OPTS=ehxBET
00:04:38.029 13:13:03 -- common/autotest_common.sh@21 -- # [[ ehxBET == *\x* ]]
00:04:38.029 13:13:03 -- common/autotest_common.sh@22 -- # XTRACE_DISABLED=yes
00:04:38.029 13:13:03 -- common/autotest_common.sh@24 -- # set +x
00:04:38.029 ========== Backtrace start: ==========
00:04:38.029
00:04:38.029 in /var/jenkins/workspace/vfio-user-autotest/spdk/test/common/autotest_common.sh:1528 -> opal_revert_cleanup([])
00:04:38.029 ...
00:04:38.029 1523 spdk_tgt_pid=$!
00:04:38.029 1524 waitforlisten $spdk_tgt_pid
00:04:38.029 1525
00:04:38.029 1526 bdf_id=0
00:04:38.029 1527 for bdf in "${bdfs[@]}"; do
00:04:38.029 1528 $rootdir/scripts/rpc.py bdev_nvme_attach_controller -b "nvme"${bdf_id} -t "pcie" -a ${bdf}
00:04:38.029 1529 # Ignore if this fails.
00:04:38.029 1530 $rootdir/scripts/rpc.py bdev_nvme_opal_revert -b "nvme"${bdf_id} -p test || true
00:04:38.029 1531 ((++bdf_id))
00:04:38.029 1532 done
00:04:38.029 1533
00:04:38.029 ...
00:04:38.029 in /var/jenkins/workspace/vfio-user-autotest/spdk/autotest.sh:140 -> main(["/var/jenkins/workspace/vfio-user-autotest/autorun-spdk.conf"])
00:04:38.029 ...
00:04:38.029 135 $rootdir/scripts/setup.sh
00:04:38.029 136 timing_exit afterboot
00:04:38.029 137
00:04:38.029 138 # Revert existing OPAL to factory settings that may have been left from earlier failed tests.
00:04:38.029 139 # This ensures we won't hit any unexpected failures due to NVMe SSDs being locked.
00:04:38.029 => 140 opal_revert_cleanup
00:04:38.029 141
00:04:38.029 142 #####################
00:04:38.029 143 # Unit Tests
00:04:38.029 144 #####################
00:04:38.029 145
00:04:38.029 ...
00:04:38.029
00:04:38.029 ========== Backtrace end ==========
```
|
infrastructure
|
opal revert cleanup fails under ci intermittent failure this node should not be enabled in the pool got there by accident all nvmes there are present behind the vmd our autotest is not suite for that spdk autotest sh var jenkins workspace vfio user autotest spdk scripts setup sh active devices data so not binding pci dev skipping not allowed vmd controller at skipping not allowed vmd controller at skipping not allowed vmd controller at ae skipping not allowed vmd controller at ae skipping not allowed vmd controller at ioatdma vfio pci ioatdma vfio pci ioatdma vfio pci ioatdma vfio pci ioatdma vfio pci ioatdma vfio pci ioatdma vfio pci ioatdma vfio pci ioatdma vfio pci ioatdma vfio pci ioatdma vfio pci ioatdma vfio pci ioatdma vfio pci ioatdma vfio pci ioatdma vfio pci ioatdma vfio pci nvme vfio pci nvme vfio pci spdk autotest sh timing exit afterboot common autotest common sh xtrace disable common autotest common sh common autotest common sh prev bash opts ehxbet common autotest common sh common autotest common sh xtrace disabled yes common autotest common sh set x common autotest common sh xtrace enable spdk autotest sh opal revert cleanup common autotest common sh mapfile t bdfs common autotest common sh get nvme bdfs by id common autotest common sh bdfs common autotest common sh local bdfs common autotest common sh get nvme bdfs common autotest common sh xtrace disable common autotest common sh common autotest common sh prev bash opts hxbet common autotest common sh common autotest common sh xtrace disabled yes common autotest common sh set x common autotest common sh xtrace enable common autotest common sh for bdf in get nvme bdfs common autotest common sh cat sys bus pci devices device common autotest common sh device common autotest common sh common autotest common sh bdfs bdf common autotest common sh for bdf in get nvme bdfs common autotest common sh cat sys bus pci devices device common autotest common sh device common autotest common sh common autotest common sh bdfs bdf common autotest common sh printf s n common autotest common sh common autotest common sh spdk tgt pid common autotest common sh var jenkins workspace vfio user autotest spdk build bin spdk tgt common autotest common sh waitforlisten common autotest common sh common autotest common sh local rpc addr var tmp spdk sock common autotest common sh local max retries common autotest common sh echo waiting for process to start up and listen on unix domain socket var tmp spdk sock waiting for process to start up and listen on unix domain socket var tmp spdk sock common autotest common sh xtrace disable common autotest common sh common autotest common sh prev bash opts ehxbet common autotest common sh common autotest common sh xtrace disabled yes common autotest common sh set x starting spdk pre git dpdk initialization eal no free kb hugepages reported on node telemetry no legacy callbacks legacy socket not created app c spdk app start notice total cores available reactor c reactor run notice reactor started on core accel sw c sw accel module init notice accel framework software module initialized common autotest common sh xtrace enable common autotest common sh i common autotest common sh return common autotest common sh bdf id common autotest common sh for bdf in bdfs common autotest common sh var jenkins workspace vfio user autotest spdk scripts rpc py bdev nvme attach controller b t pcie a eal vfio group is not viable not all devices in iommu group bound to vfio or unbound eal driver cannot attach the device eal failed to attach device on primary process nvme c nvme probe internal error nvme ctrlr scan failed bdev nvme c bdev nvme create error no controller was found with provided trid traddr request name trtype pcie traddr method bdev nvme attach controller req id got json rpc error response response code message no such device common autotest common sh trap err common autotest common sh print backtrace common autotest common sh common autotest common sh args var jenkins workspace vfio user autotest autorun spdk conf common autotest common sh local args common autotest common sh xtrace disable common autotest common sh common autotest common sh prev bash opts ehxbet common autotest common sh common autotest common sh xtrace disabled yes common autotest common sh set x backtrace start in var jenkins workspace vfio user autotest spdk test common autotest common sh opal revert cleanup spdk tgt pid waitforlisten spdk tgt pid bdf id for bdf in bdfs do rootdir scripts rpc py bdev nvme attach controller b nvme bdf id t pcie a bdf ignore if this fails rootdir scripts rpc py bdev nvme opal revert b nvme bdf id p test true bdf id done in var jenkins workspace vfio user autotest spdk autotest sh main rootdir scripts setup sh timing exit afterboot revert existing opal to factory settings that may have been left from earlier failed tests this ensures we won t hit any unexpected failures due to nvme ssds being locked opal revert cleanup unit tests backtrace end
| 1
|
359,814
| 10,681,187,207
|
IssuesEvent
|
2019-10-21 23:45:31
|
redcross/smoke-alarm-portal
|
https://api.github.com/repos/redcross/smoke-alarm-portal
|
closed
|
Include status field in CSV Export
|
high priority
|
Include the **status** field in CSV Exports.
<img width="1280" alt="getasmokealarm csv export" src="https://user-images.githubusercontent.com/4147733/39664344-761ea738-5047-11e8-8440-c26b6102142a.png">
|
1.0
|
Include status field in CSV Export - Include the **status** field in CSV Exports.
<img width="1280" alt="getasmokealarm csv export" src="https://user-images.githubusercontent.com/4147733/39664344-761ea738-5047-11e8-8440-c26b6102142a.png">
|
non_infrastructure
|
include status field in csv export include the status field in csv exports img width alt getasmokealarm csv export src
| 0
|
16,376
| 11,947,796,933
|
IssuesEvent
|
2020-04-03 10:35:11
|
cmu-lib/dhweb_app
|
https://api.github.com/repos/cmu-lib/dhweb_app
|
closed
|
Hook up to SMTP server
|
infrastructure
|
send registration&password change emails
- [x] request new inbox&email address from Rashid
- [x] https://docs.djangoproject.com/en/3.0/topics/email/#smtp-backend to register email backend
- [x] identify django libraries for password reset / first-login workflows
|
1.0
|
Hook up to SMTP server - send registration&password change emails
- [x] request new inbox&email address from Rashid
- [x] https://docs.djangoproject.com/en/3.0/topics/email/#smtp-backend to register email backend
- [x] identify django libraries for password reset / first-login workflows
|
infrastructure
|
hook up to smtp server send registration password change emails request new inbox email address from rashid to register email backend identify django libraries for password reset first login workflows
| 1
|
100,657
| 12,543,699,564
|
IssuesEvent
|
2020-06-05 15:59:02
|
NASA-IMPACT/covid-dashboard
|
https://api.github.com/repos/NASA-IMPACT/covid-dashboard
|
closed
|
Format NO2 dataset page
|
design
|
Quickly put up the content. Need to:
- improve the structure
- CTA linking back to global view
|
1.0
|
Format NO2 dataset page - Quickly put up the content. Need to:
- improve the structure
- CTA linking back to global view
|
non_infrastructure
|
format dataset page quickly put up the content need to improve the structure cta linking back to global view
| 0
|
514,557
| 14,940,904,684
|
IssuesEvent
|
2021-01-25 18:57:43
|
dmuhs/web3data-py
|
https://api.github.com/repos/dmuhs/web3data-py
|
closed
|
Check CSV format integration
|
Priority: High Type: Feature
|
Add integration for CSV format: https://docs.amberdata.io/reference#csv-format
The JSON parsing in the response routine will probably cause problems
|
1.0
|
Check CSV format integration - Add integration for CSV format: https://docs.amberdata.io/reference#csv-format
The JSON parsing in the response routine will probably cause problems
|
non_infrastructure
|
check csv format integration add integration for csv format the json parsing in the response routine will probably cause problems
| 0
|
280,289
| 21,216,497,632
|
IssuesEvent
|
2022-04-11 07:53:21
|
appsmithorg/appsmith
|
https://api.github.com/repos/appsmithorg/appsmith
|
opened
|
Docs for #1004 [Bug] No error is shown on navigation if page name is invalid
|
Documentation User Education Pod
|
Field | Details
-----|-----
**POD** | FE Coders Pod
**Parent Ticket** | #1004
Engineer |
Release Date |
Live Date |
First Draft |
Auto Assign |
Priority |
Environment |
|
1.0
|
Docs for #1004 [Bug] No error is shown on navigation if page name is invalid - Field | Details
-----|-----
**POD** | FE Coders Pod
**Parent Ticket** | #1004
Engineer |
Release Date |
Live Date |
First Draft |
Auto Assign |
Priority |
Environment |
|
non_infrastructure
|
docs for no error is shown on navigation if page name is invalid field details pod fe coders pod parent ticket engineer release date live date first draft auto assign priority environment
| 0
|
651,510
| 21,481,586,119
|
IssuesEvent
|
2022-04-26 18:17:44
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
closed
|
macos TIMEOUT: cmake/build/test_cpp_client_credentials_test --gtest_filter=CredentialsTest.TlsChannelCredentialsWithAsyncExternalVerifier
|
kind/bug priority/P2
|
```
[ RUN ] CredentialsTest.TlsChannelCredentialsWithAsyncExternalVerifier
D1223 17:42:00.480608000 4495672768 ev_posix.cc:172] Using polling engine: poll
D1223 17:42:00.506058000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "grpclb"
D1223 17:42:00.506094000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "priority_experimental"
D1223 17:42:00.506111000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "weighted_target_experimental"
D1223 17:42:00.506126000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "pick_first"
D1223 17:42:00.506139000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "round_robin"
D1223 17:42:00.506154000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "ring_hash_experimental"
D1223 17:42:00.506168000 4495672768 dns_resolver_ares.cc:624] Using ares dns resolver
D1223 17:42:00.528378000 4495672768 certificate_provider_registry.cc:33] registering certificate provider factory for "file_watcher"
D1223 17:42:00.528441000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "cds_experimental"
D1223 17:42:00.528473000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "xds_cluster_impl_experimental"
D1223 17:42:00.528493000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "xds_cluster_resolver_experimental"
D1223 17:42:00.528511000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "xds_cluster_manager_experimental"
D1223 17:42:00.528631000 4495672768 init.cc:179] grpc_shutdown starts clean-up now
2021-12-23 17:47:01,037 TIMEOUT: cmake/build/test_cpp_client_credentials_test --gtest_filter=CredentialsTest.TlsChannelCredentialsWithAsyncExternalVerifier GRPC_POLL_STRATEGY=poll [pid=45307, time=300.6sec]
```
https://source.cloud.google.com/results/invocations/edfbcb11-fb89-47b8-96fb-e29a5e8f4517/log
seen in https://github.com/grpc/grpc/pull/28415
|
1.0
|
macos TIMEOUT: cmake/build/test_cpp_client_credentials_test --gtest_filter=CredentialsTest.TlsChannelCredentialsWithAsyncExternalVerifier - ```
[ RUN ] CredentialsTest.TlsChannelCredentialsWithAsyncExternalVerifier
D1223 17:42:00.480608000 4495672768 ev_posix.cc:172] Using polling engine: poll
D1223 17:42:00.506058000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "grpclb"
D1223 17:42:00.506094000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "priority_experimental"
D1223 17:42:00.506111000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "weighted_target_experimental"
D1223 17:42:00.506126000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "pick_first"
D1223 17:42:00.506139000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "round_robin"
D1223 17:42:00.506154000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "ring_hash_experimental"
D1223 17:42:00.506168000 4495672768 dns_resolver_ares.cc:624] Using ares dns resolver
D1223 17:42:00.528378000 4495672768 certificate_provider_registry.cc:33] registering certificate provider factory for "file_watcher"
D1223 17:42:00.528441000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "cds_experimental"
D1223 17:42:00.528473000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "xds_cluster_impl_experimental"
D1223 17:42:00.528493000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "xds_cluster_resolver_experimental"
D1223 17:42:00.528511000 4495672768 lb_policy_registry.cc:42] registering LB policy factory for "xds_cluster_manager_experimental"
D1223 17:42:00.528631000 4495672768 init.cc:179] grpc_shutdown starts clean-up now
2021-12-23 17:47:01,037 TIMEOUT: cmake/build/test_cpp_client_credentials_test --gtest_filter=CredentialsTest.TlsChannelCredentialsWithAsyncExternalVerifier GRPC_POLL_STRATEGY=poll [pid=45307, time=300.6sec]
```
https://source.cloud.google.com/results/invocations/edfbcb11-fb89-47b8-96fb-e29a5e8f4517/log
seen in https://github.com/grpc/grpc/pull/28415
|
non_infrastructure
|
macos timeout cmake build test cpp client credentials test gtest filter credentialstest tlschannelcredentialswithasyncexternalverifier credentialstest tlschannelcredentialswithasyncexternalverifier ev posix cc using polling engine poll lb policy registry cc registering lb policy factory for grpclb lb policy registry cc registering lb policy factory for priority experimental lb policy registry cc registering lb policy factory for weighted target experimental lb policy registry cc registering lb policy factory for pick first lb policy registry cc registering lb policy factory for round robin lb policy registry cc registering lb policy factory for ring hash experimental dns resolver ares cc using ares dns resolver certificate provider registry cc registering certificate provider factory for file watcher lb policy registry cc registering lb policy factory for cds experimental lb policy registry cc registering lb policy factory for xds cluster impl experimental lb policy registry cc registering lb policy factory for xds cluster resolver experimental lb policy registry cc registering lb policy factory for xds cluster manager experimental init cc grpc shutdown starts clean up now timeout cmake build test cpp client credentials test gtest filter credentialstest tlschannelcredentialswithasyncexternalverifier grpc poll strategy poll seen in
| 0
|
18,165
| 12,824,614,386
|
IssuesEvent
|
2020-07-06 13:45:28
|
qutebrowser/qutebrowser
|
https://api.github.com/repos/qutebrowser/qutebrowser
|
closed
|
GitHub Actions: Install newer shellcheck?
|
component: infrastructure priority: 2 - low
|
We get v0.4.6 from March - might be a good idea to use their precompiled binary instead of the pre-installed one?
|
1.0
|
GitHub Actions: Install newer shellcheck? - We get v0.4.6 from March - might be a good idea to use their precompiled binary instead of the pre-installed one?
|
infrastructure
|
github actions install newer shellcheck we get from march might be a good idea to use their precompiled binary instead of the pre installed one
| 1
|
4,741
| 5,256,629,346
|
IssuesEvent
|
2017-02-02 18:25:36
|
dart-lang/sdk
|
https://api.github.com/repos/dart-lang/sdk
|
closed
|
Infrastructure support for sampling stacks on windows
|
area-infrastructure gardening os-windows
|
It would be nice to have support for sampling stacks on Windows for tests which time out to diagnose the state of the VM before we kill the timed-out test.
We have similar functionality already for Linux/MacOS [here](https://github.com/dart-lang/sdk/blob/master/tools/testing/dart/test_runner.dart#L1982).
|
1.0
|
Infrastructure support for sampling stacks on windows - It would be nice to have support for sampling stacks on Windows for tests which time out to diagnose the state of the VM before we kill the timed-out test.
We have similar functionality already for Linux/MacOS [here](https://github.com/dart-lang/sdk/blob/master/tools/testing/dart/test_runner.dart#L1982).
|
infrastructure
|
infrastructure support for sampling stacks on windows it would be nice to have support for sampling stacks on windows for tests which time out to diagnose the state of the vm before we kill the timed out test we have similar functionality already for linux macos
| 1
|
825,347
| 31,334,844,411
|
IssuesEvent
|
2023-08-24 04:44:27
|
googleapis/java-spanner-jdbc
|
https://api.github.com/repos/googleapis/java-spanner-jdbc
|
closed
|
The build failed
|
priority: p2 type: bug api: spanner flakybot: issue
|
This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 76fc16a7651120ac019ad13ca2c2e4bccd2bd7a1
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/246ce13a-77fb-4817-9759-2338c5b0690b), [Sponge](http://sponge2/246ce13a-77fb-4817-9759-2338c5b0690b)
status: failed
<details><summary>Test output</summary><br><pre>com.google.cloud.spanner.SpannerException: UNAVAILABLE: Failed to list the databases of projects/span-cloud-testing/instances/spanner-testing-east1 with pageToken <null>: UNAVAILABLE: com.google.api.gax.rpc.UnavailableException: io.grpc.StatusRuntimeException: UNAVAILABLE: Credentials failed to obtain metadata
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:291)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:297)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:61)
at com.google.cloud.spanner.DatabaseAdminClientImpl$5.getNextPage(DatabaseAdminClientImpl.java:469)
at com.google.cloud.spanner.SpannerImpl$PageFetcher.getNextPage(SpannerImpl.java:293)
at com.google.cloud.spanner.DatabaseAdminClientImpl.listDatabases(DatabaseAdminClientImpl.java:488)
at com.google.cloud.spanner.IntegrationTestEnv.cleanUpOldDatabases(IntegrationTestEnv.java:185)
at com.google.cloud.spanner.IntegrationTestEnv.before(IntegrationTestEnv.java:114)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:50)
at org.junit.rules.RunRules.evaluate(RunRules.java:20)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.junit.runners.Suite.runChild(Suite.java:128)
at org.junit.runners.Suite.runChild(Suite.java:27)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:119)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:87)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75)
at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:158)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:456)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:169)
at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:595)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:581)
Caused by: com.google.cloud.spanner.SpannerException: UNAVAILABLE: com.google.api.gax.rpc.UnavailableException: io.grpc.StatusRuntimeException: UNAVAILABLE: Credentials failed to obtain metadata
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:291)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:297)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:61)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:181)
at com.google.cloud.spanner.spi.v1.GapicSpannerRpc.get(GapicSpannerRpc.java:1883)
at com.google.cloud.spanner.spi.v1.GapicSpannerRpc.lambda$listDatabases$8(GapicSpannerRpc.java:1168)
at com.google.api.gax.retrying.DirectRetryingExecutor.submit(DirectRetryingExecutor.java:103)
at com.google.cloud.RetryHelper.run(RetryHelper.java:76)
at com.google.cloud.RetryHelper.runWithRetries(RetryHelper.java:50)
at com.google.cloud.spanner.spi.v1.GapicSpannerRpc.runWithRetryOnAdministrativeRequestsExceeded(GapicSpannerRpc.java:688)
at com.google.cloud.spanner.spi.v1.GapicSpannerRpc.listDatabases(GapicSpannerRpc.java:1167)
at com.google.cloud.spanner.DatabaseAdminClientImpl$5.getNextPage(DatabaseAdminClientImpl.java:467)
... 27 more
Caused by: java.util.concurrent.ExecutionException: com.google.api.gax.rpc.UnavailableException: io.grpc.StatusRuntimeException: UNAVAILABLE: Credentials failed to obtain metadata
at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:588)
at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:567)
at com.google.cloud.spanner.spi.v1.GapicSpannerRpc.get(GapicSpannerRpc.java:1877)
... 34 more
Caused by: com.google.api.gax.rpc.UnavailableException: io.grpc.StatusRuntimeException: UNAVAILABLE: Credentials failed to obtain metadata
at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:112)
at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:41)
at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:86)
at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:66)
at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97)
at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:67)
at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1132)
at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31)
at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270)
at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038)
at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808)
at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:574)
at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:544)
at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39)
at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23)
at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40)
at com.google.api.gax.grpc.ChannelPool$ReleasingClientCall$1.onClose(ChannelPool.java:541)
at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39)
at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23)
at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40)
at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39)
at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23)
at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40)
at com.google.cloud.spanner.spi.v1.SpannerErrorInterceptor$1$1.onClose(SpannerErrorInterceptor.java:100)
at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:576)
at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70)
at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:757)
at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:736)
at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Caused by: io.grpc.StatusRuntimeException: UNAVAILABLE: Credentials failed to obtain metadata
at io.grpc.Status.asRuntimeException(Status.java:539)
... 25 more
Caused by: com.google.auth.oauth2.GoogleAuthException: Error getting access token for service account: 400 Bad Request
POST https://oauth2.googleapis.com/token
{"error":"invalid_grant","error_description":"Invalid JWT Signature."}, iss: java-client-testing@span-cloud-testing.iam.gserviceaccount.com
at com.google.auth.oauth2.GoogleAuthException.createWithTokenEndpointResponseException(GoogleAuthException.java:131)
at com.google.auth.oauth2.ServiceAccountCredentials.refreshAccessToken(ServiceAccountCredentials.java:541)
at com.google.auth.oauth2.OAuth2Credentials$1.call(OAuth2Credentials.java:269)
at com.google.auth.oauth2.OAuth2Credentials$1.call(OAuth2Credentials.java:266)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at com.google.auth.oauth2.OAuth2Credentials$RefreshTask.run(OAuth2Credentials.java:633)
... 3 more
Caused by: com.google.api.client.http.HttpResponseException: 400 Bad Request
POST https://oauth2.googleapis.com/token
{"error":"invalid_grant","error_description":"Invalid JWT Signature."}
at com.google.api.client.http.HttpResponseException$Builder.build(HttpResponseException.java:293)
at com.google.api.client.http.HttpRequest.execute(HttpRequest.java:1118)
at com.google.auth.oauth2.ServiceAccountCredentials.refreshAccessToken(ServiceAccountCredentials.java:538)
... 7 more
</pre></details>
|
1.0
|
The build failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 76fc16a7651120ac019ad13ca2c2e4bccd2bd7a1
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/246ce13a-77fb-4817-9759-2338c5b0690b), [Sponge](http://sponge2/246ce13a-77fb-4817-9759-2338c5b0690b)
status: failed
<details><summary>Test output</summary><br><pre>com.google.cloud.spanner.SpannerException: UNAVAILABLE: Failed to list the databases of projects/span-cloud-testing/instances/spanner-testing-east1 with pageToken <null>: UNAVAILABLE: com.google.api.gax.rpc.UnavailableException: io.grpc.StatusRuntimeException: UNAVAILABLE: Credentials failed to obtain metadata
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:291)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:297)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:61)
at com.google.cloud.spanner.DatabaseAdminClientImpl$5.getNextPage(DatabaseAdminClientImpl.java:469)
at com.google.cloud.spanner.SpannerImpl$PageFetcher.getNextPage(SpannerImpl.java:293)
at com.google.cloud.spanner.DatabaseAdminClientImpl.listDatabases(DatabaseAdminClientImpl.java:488)
at com.google.cloud.spanner.IntegrationTestEnv.cleanUpOldDatabases(IntegrationTestEnv.java:185)
at com.google.cloud.spanner.IntegrationTestEnv.before(IntegrationTestEnv.java:114)
at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:50)
at org.junit.rules.RunRules.evaluate(RunRules.java:20)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.junit.runners.Suite.runChild(Suite.java:128)
at org.junit.runners.Suite.runChild(Suite.java:27)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:119)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:87)
at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75)
at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:158)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:456)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:169)
at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:595)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:581)
Caused by: com.google.cloud.spanner.SpannerException: UNAVAILABLE: com.google.api.gax.rpc.UnavailableException: io.grpc.StatusRuntimeException: UNAVAILABLE: Credentials failed to obtain metadata
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:291)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:297)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:61)
at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:181)
at com.google.cloud.spanner.spi.v1.GapicSpannerRpc.get(GapicSpannerRpc.java:1883)
at com.google.cloud.spanner.spi.v1.GapicSpannerRpc.lambda$listDatabases$8(GapicSpannerRpc.java:1168)
at com.google.api.gax.retrying.DirectRetryingExecutor.submit(DirectRetryingExecutor.java:103)
at com.google.cloud.RetryHelper.run(RetryHelper.java:76)
at com.google.cloud.RetryHelper.runWithRetries(RetryHelper.java:50)
at com.google.cloud.spanner.spi.v1.GapicSpannerRpc.runWithRetryOnAdministrativeRequestsExceeded(GapicSpannerRpc.java:688)
at com.google.cloud.spanner.spi.v1.GapicSpannerRpc.listDatabases(GapicSpannerRpc.java:1167)
at com.google.cloud.spanner.DatabaseAdminClientImpl$5.getNextPage(DatabaseAdminClientImpl.java:467)
... 27 more
Caused by: java.util.concurrent.ExecutionException: com.google.api.gax.rpc.UnavailableException: io.grpc.StatusRuntimeException: UNAVAILABLE: Credentials failed to obtain metadata
at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:588)
at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:567)
at com.google.cloud.spanner.spi.v1.GapicSpannerRpc.get(GapicSpannerRpc.java:1877)
... 34 more
Caused by: com.google.api.gax.rpc.UnavailableException: io.grpc.StatusRuntimeException: UNAVAILABLE: Credentials failed to obtain metadata
at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:112)
at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:41)
at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:86)
at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:66)
at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97)
at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:67)
at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1132)
at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31)
at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270)
at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038)
at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808)
at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:574)
at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:544)
at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39)
at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23)
at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40)
at com.google.api.gax.grpc.ChannelPool$ReleasingClientCall$1.onClose(ChannelPool.java:541)
at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39)
at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23)
at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40)
at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39)
at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23)
at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40)
at com.google.cloud.spanner.spi.v1.SpannerErrorInterceptor$1$1.onClose(SpannerErrorInterceptor.java:100)
at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:576)
at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70)
at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:757)
at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:736)
at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Caused by: io.grpc.StatusRuntimeException: UNAVAILABLE: Credentials failed to obtain metadata
at io.grpc.Status.asRuntimeException(Status.java:539)
... 25 more
Caused by: com.google.auth.oauth2.GoogleAuthException: Error getting access token for service account: 400 Bad Request
POST https://oauth2.googleapis.com/token
{"error":"invalid_grant","error_description":"Invalid JWT Signature."}, iss: java-client-testing@span-cloud-testing.iam.gserviceaccount.com
at com.google.auth.oauth2.GoogleAuthException.createWithTokenEndpointResponseException(GoogleAuthException.java:131)
at com.google.auth.oauth2.ServiceAccountCredentials.refreshAccessToken(ServiceAccountCredentials.java:541)
at com.google.auth.oauth2.OAuth2Credentials$1.call(OAuth2Credentials.java:269)
at com.google.auth.oauth2.OAuth2Credentials$1.call(OAuth2Credentials.java:266)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at com.google.auth.oauth2.OAuth2Credentials$RefreshTask.run(OAuth2Credentials.java:633)
... 3 more
Caused by: com.google.api.client.http.HttpResponseException: 400 Bad Request
POST https://oauth2.googleapis.com/token
{"error":"invalid_grant","error_description":"Invalid JWT Signature."}
at com.google.api.client.http.HttpResponseException$Builder.build(HttpResponseException.java:293)
at com.google.api.client.http.HttpRequest.execute(HttpRequest.java:1118)
at com.google.auth.oauth2.ServiceAccountCredentials.refreshAccessToken(ServiceAccountCredentials.java:538)
... 7 more
</pre></details>
|
non_infrastructure
|
the build failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output com google cloud spanner spannerexception unavailable failed to list the databases of projects span cloud testing instances spanner testing with pagetoken unavailable com google api gax rpc unavailableexception io grpc statusruntimeexception unavailable credentials failed to obtain metadata at com google cloud spanner spannerexceptionfactory newspannerexceptionpreformatted spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexceptionpreformatted spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner databaseadminclientimpl getnextpage databaseadminclientimpl java at com google cloud spanner spannerimpl pagefetcher getnextpage spannerimpl java at com google cloud spanner databaseadminclientimpl listdatabases databaseadminclientimpl java at com google cloud spanner integrationtestenv cleanupolddatabases integrationtestenv java at com google cloud spanner integrationtestenv before integrationtestenv java at org junit rules externalresource evaluate externalresource java at org junit rules runrules evaluate runrules java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runners suite runchild suite java at org junit runners suite runchild suite java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore junitcore run junitcore java at org apache maven surefire junitcore junitcorewrapper createrequestandrun junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper executelazy junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcoreprovider invoke junitcoreprovider java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java caused by com google cloud spanner spannerexception unavailable com google api gax rpc unavailableexception io grpc statusruntimeexception unavailable credentials failed to obtain metadata at com google cloud spanner spannerexceptionfactory newspannerexceptionpreformatted spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexceptionpreformatted spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner spi gapicspannerrpc get gapicspannerrpc java at com google cloud spanner spi gapicspannerrpc lambda listdatabases gapicspannerrpc java at com google api gax retrying directretryingexecutor submit directretryingexecutor java at com google cloud retryhelper run retryhelper java at com google cloud retryhelper runwithretries retryhelper java at com google cloud spanner spi gapicspannerrpc runwithretryonadministrativerequestsexceeded gapicspannerrpc java at com google cloud spanner spi gapicspannerrpc listdatabases gapicspannerrpc java at com google cloud spanner databaseadminclientimpl getnextpage databaseadminclientimpl java more caused by java util concurrent executionexception com google api gax rpc unavailableexception io grpc statusruntimeexception unavailable credentials failed to obtain metadata at com google common util concurrent abstractfuture getdonevalue abstractfuture java at com google common util concurrent abstractfuture get abstractfuture java at com google cloud spanner spi gapicspannerrpc get gapicspannerrpc java more caused by com google api gax rpc unavailableexception io grpc statusruntimeexception unavailable credentials failed to obtain metadata at com google api gax rpc apiexceptionfactory createexception apiexceptionfactory java at com google api gax rpc apiexceptionfactory createexception apiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcexceptioncallable exceptiontransformingfuture onfailure grpcexceptioncallable java at com google api core apifutures onfailure apifutures java at com google common util concurrent futures callbacklistener run futures java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture complete abstractfuture java at com google common util concurrent abstractfuture setexception abstractfuture java at io grpc stub clientcalls grpcfuture setexception clientcalls java at io grpc stub clientcalls unarystreamtofuture onclose clientcalls java at io grpc partialforwardingclientcalllistener onclose partialforwardingclientcalllistener java at io grpc forwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc forwardingclientcalllistener simpleforwardingclientcalllistener onclose forwardingclientcalllistener java at com google api gax grpc channelpool releasingclientcall onclose channelpool java at io grpc partialforwardingclientcalllistener onclose partialforwardingclientcalllistener java at io grpc forwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc forwardingclientcalllistener simpleforwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc partialforwardingclientcalllistener onclose partialforwardingclientcalllistener java at io grpc forwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc forwardingclientcalllistener simpleforwardingclientcalllistener onclose forwardingclientcalllistener java at com google cloud spanner spi spannererrorinterceptor onclose spannererrorinterceptor java at io grpc internal clientcallimpl closeobserver clientcallimpl java at io grpc internal clientcallimpl access clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runinternal clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runincontext clientcallimpl java at io grpc internal contextrunnable run contextrunnable java at io grpc internal serializingexecutor run serializingexecutor java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask access scheduledthreadpoolexecutor java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by io grpc statusruntimeexception unavailable credentials failed to obtain metadata at io grpc status asruntimeexception status java more caused by com google auth googleauthexception error getting access token for service account bad request post error invalid grant error description invalid jwt signature iss java client testing span cloud testing iam gserviceaccount com at com google auth googleauthexception createwithtokenendpointresponseexception googleauthexception java at com google auth serviceaccountcredentials refreshaccesstoken serviceaccountcredentials java at com google auth call java at com google auth call java at java util concurrent futuretask run futuretask java at com google auth refreshtask run java more caused by com google api client http httpresponseexception bad request post error invalid grant error description invalid jwt signature at com google api client http httpresponseexception builder build httpresponseexception java at com google api client http httprequest execute httprequest java at com google auth serviceaccountcredentials refreshaccesstoken serviceaccountcredentials java more
| 0
|
18,732
| 13,087,481,938
|
IssuesEvent
|
2020-08-02 12:30:57
|
Reccy/MIDIPiano
|
https://api.github.com/repos/Reccy/MIDIPiano
|
closed
|
Optimise script to generate patch version number
|
infrastructure
|
Current patch version number script uses a for-loop, which is a simple naive implementation.
A better way would be to get the list of all patch numbers for a major.minor version, and then sort them to get the largest patch version.
Increment that largest version to get the patch version.
|
1.0
|
Optimise script to generate patch version number - Current patch version number script uses a for-loop, which is a simple naive implementation.
A better way would be to get the list of all patch numbers for a major.minor version, and then sort them to get the largest patch version.
Increment that largest version to get the patch version.
|
infrastructure
|
optimise script to generate patch version number current patch version number script uses a for loop which is a simple naive implementation a better way would be to get the list of all patch numbers for a major minor version and then sort them to get the largest patch version increment that largest version to get the patch version
| 1
|
28,205
| 23,083,552,928
|
IssuesEvent
|
2022-07-26 09:21:54
|
havok25103/character-sheet
|
https://api.github.com/repos/havok25103/character-sheet
|
closed
|
BIG DEVOPS TICKET, BLEGH
|
Infrastructure
|
- [x] Generate local production mysql user and password
- [x] Manage local secret and env variable management
- [x] Bisect MySQL data storage
- [x] Bisect system level backups MySQL
- [x] Bisect system level backups JSON
- [x] Bisect system level blob storage
- [ ] Automated MySQL Backups
- [ ] Automated JSON Backups
- [x] Configure production asset bundling
- [x] Build production asset bundling job
- [x] Check general Performance
- [x] Create production level application admin user
|
1.0
|
BIG DEVOPS TICKET, BLEGH - - [x] Generate local production mysql user and password
- [x] Manage local secret and env variable management
- [x] Bisect MySQL data storage
- [x] Bisect system level backups MySQL
- [x] Bisect system level backups JSON
- [x] Bisect system level blob storage
- [ ] Automated MySQL Backups
- [ ] Automated JSON Backups
- [x] Configure production asset bundling
- [x] Build production asset bundling job
- [x] Check general Performance
- [x] Create production level application admin user
|
infrastructure
|
big devops ticket blegh generate local production mysql user and password manage local secret and env variable management bisect mysql data storage bisect system level backups mysql bisect system level backups json bisect system level blob storage automated mysql backups automated json backups configure production asset bundling build production asset bundling job check general performance create production level application admin user
| 1
|
25,243
| 18,295,024,895
|
IssuesEvent
|
2021-10-05 19:34:32
|
botpress/botpress
|
https://api.github.com/repos/botpress/botpress
|
closed
|
[BUG] Heroku One click deploy timing issue
|
bug infrastructure
|
**Describe the bug**
When you use Heroku one click deploy, the deploy shows it as deployed about a minute before it can be accessed, with no error / waiting message.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to https://github.com/botpress/botpress
2. Click on deploy to Heroku
3. Click Deploy App
4. When the app is deployed click "View" immediately
5. The page will keep loading. About a minute later if you refresh it will show the correct page.
**Expected behavior**
The view button should only appear when the sign-up page is viewable.
|
1.0
|
[BUG] Heroku One click deploy timing issue - **Describe the bug**
When you use Heroku one click deploy, the deploy shows it as deployed about a minute before it can be accessed, with no error / waiting message.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to https://github.com/botpress/botpress
2. Click on deploy to Heroku
3. Click Deploy App
4. When the app is deployed click "View" immediately
5. The page will keep loading. About a minute later if you refresh it will show the correct page.
**Expected behavior**
The view button should only appear when the sign-up page is viewable.
|
infrastructure
|
heroku one click deploy timing issue describe the bug when you use heroku one click deploy the deploy shows it as deployed about a minute before it can be accessed with no error waiting message to reproduce steps to reproduce the behavior go to click on deploy to heroku click deploy app when the app is deployed click view immediately the page will keep loading about a minute later if you refresh it will show the correct page expected behavior the view button should only appear when the sign up page is viewable
| 1
|
18,423
| 12,976,299,181
|
IssuesEvent
|
2020-07-21 18:31:31
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
Microsoft.Bcl.AsyncInterfaces and Microsoft.Bcl.HashCode expose reference assemblies
|
area-Infrastructure-libraries
|
Both of these packages are exposing reference assemblies to desktop projects. This makes them break when used in a project that uses the workflow compiler, as well as others. See https://github.com/dotnet/sdk/issues/1522.
We should service 3.1 to fix this by removing the reference assemblies from the package and look for any other new instances of this that may have been introduced in 3.1.
cc @joperezr
|
1.0
|
Microsoft.Bcl.AsyncInterfaces and Microsoft.Bcl.HashCode expose reference assemblies - Both of these packages are exposing reference assemblies to desktop projects. This makes them break when used in a project that uses the workflow compiler, as well as others. See https://github.com/dotnet/sdk/issues/1522.
We should service 3.1 to fix this by removing the reference assemblies from the package and look for any other new instances of this that may have been introduced in 3.1.
cc @joperezr
|
infrastructure
|
microsoft bcl asyncinterfaces and microsoft bcl hashcode expose reference assemblies both of these packages are exposing reference assemblies to desktop projects this makes them break when used in a project that uses the workflow compiler as well as others see we should service to fix this by removing the reference assemblies from the package and look for any other new instances of this that may have been introduced in cc joperezr
| 1
|
117
| 2,528,176,060
|
IssuesEvent
|
2015-01-22 00:15:01
|
facebook/osquery
|
https://api.github.com/repos/facebook/osquery
|
closed
|
Make "make install" match the file layout of packages
|
deployment infrastructure up for grabs
|
For example, `make install` will install the example config to `/var/osquery/osquery.example.conf`, but the deb will install it to `/usr/share/osquery/osquery.example.conf`.
`make packages` is our source of truth, so let's match that or remove the "install" target.
Deb:
```
root@vagrant-ubuntu-trusty-64:/vagrant# dpkg -c build/linux/osquery-1.3.0-6-g9f937a9.deb
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./var/
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./var/osquery/
-rw-r--r-- 0/0 2713 2015-01-21 23:34 ./var/osquery/osquery.example.conf
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./var/log/
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./var/log/osquery/
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./etc/
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./etc/init.d/
-rw-r--r-- 0/0 2832 2015-01-21 23:34 ./etc/init.d/osqueryd
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./usr/
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./usr/bin/
-rwxr-xr-x 0/0 3987640 2015-01-21 23:34 ./usr/bin/osqueryd
-rwxr-xr-x 0/0 3979448 2015-01-21 23:34 ./usr/bin/osqueryi
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./usr/share/
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./usr/share/doc/
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./usr/share/doc/osquery/
-rw-r--r-- 0/0 144 2015-01-21 23:34 ./usr/share/doc/osquery/changelog.Debian.gz
```
Make:
```
root@vagrant-ubuntu-trusty-64:/vagrant# make install
-- Building for Ubuntu
-- Configuring done
-- Generating done
-- Build files have been written to: /vagrant/build/ubuntu
Install the project...
-- Install configuration: ""
-- Installing: /usr/local/lib/libosquery.a
-- Installing: /usr/local/include
-- Installing: /usr/local/include/osquery
-- Installing: /usr/local/include/osquery/config
-- Installing: /usr/local/include/osquery/config/plugin.h
-- Installing: /usr/local/include/osquery/config.h
-- Installing: /usr/local/include/osquery/core.h
-- Installing: /usr/local/include/osquery/database
-- Installing: /usr/local/include/osquery/database/db_handle.h
-- Installing: /usr/local/include/osquery/database/query.h
-- Installing: /usr/local/include/osquery/database/results.h
-- Installing: /usr/local/include/osquery/database.h
-- Installing: /usr/local/include/osquery/devtools.h
-- Installing: /usr/local/include/osquery/dispatcher.h
-- Installing: /usr/local/include/osquery/events.h
-- Installing: /usr/local/include/osquery/filesystem.h
-- Installing: /usr/local/include/osquery/flags.h
-- Installing: /usr/local/include/osquery/hash.h
-- Installing: /usr/local/include/osquery/logger
-- Installing: /usr/local/include/osquery/logger/plugin.h
-- Installing: /usr/local/include/osquery/logger.h
-- Installing: /usr/local/include/osquery/registry.h
-- Installing: /usr/local/include/osquery/scheduler.h
-- Installing: /usr/local/include/osquery/sql.h
-- Installing: /usr/local/include/osquery/status.h
-- Installing: /usr/local/include/osquery/tables.h
-- Installing: /usr/local/bin/osqueryi
-- Removed runtime path from "/usr/local/bin/osqueryi"
-- Installing: /usr/local/bin/osqueryd
-- Removed runtime path from "/usr/local/bin/osqueryd"
-- Installing: /var/osquery/osquery.example.conf
-- Installing: /etc/init.d/osqueryd
```
|
1.0
|
Make "make install" match the file layout of packages - For example, `make install` will install the example config to `/var/osquery/osquery.example.conf`, but the deb will install it to `/usr/share/osquery/osquery.example.conf`.
`make packages` is our source of truth, so let's match that or remove the "install" target.
Deb:
```
root@vagrant-ubuntu-trusty-64:/vagrant# dpkg -c build/linux/osquery-1.3.0-6-g9f937a9.deb
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./var/
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./var/osquery/
-rw-r--r-- 0/0 2713 2015-01-21 23:34 ./var/osquery/osquery.example.conf
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./var/log/
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./var/log/osquery/
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./etc/
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./etc/init.d/
-rw-r--r-- 0/0 2832 2015-01-21 23:34 ./etc/init.d/osqueryd
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./usr/
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./usr/bin/
-rwxr-xr-x 0/0 3987640 2015-01-21 23:34 ./usr/bin/osqueryd
-rwxr-xr-x 0/0 3979448 2015-01-21 23:34 ./usr/bin/osqueryi
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./usr/share/
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./usr/share/doc/
drwxr-xr-x 0/0 0 2015-01-21 23:34 ./usr/share/doc/osquery/
-rw-r--r-- 0/0 144 2015-01-21 23:34 ./usr/share/doc/osquery/changelog.Debian.gz
```
Make:
```
root@vagrant-ubuntu-trusty-64:/vagrant# make install
-- Building for Ubuntu
-- Configuring done
-- Generating done
-- Build files have been written to: /vagrant/build/ubuntu
Install the project...
-- Install configuration: ""
-- Installing: /usr/local/lib/libosquery.a
-- Installing: /usr/local/include
-- Installing: /usr/local/include/osquery
-- Installing: /usr/local/include/osquery/config
-- Installing: /usr/local/include/osquery/config/plugin.h
-- Installing: /usr/local/include/osquery/config.h
-- Installing: /usr/local/include/osquery/core.h
-- Installing: /usr/local/include/osquery/database
-- Installing: /usr/local/include/osquery/database/db_handle.h
-- Installing: /usr/local/include/osquery/database/query.h
-- Installing: /usr/local/include/osquery/database/results.h
-- Installing: /usr/local/include/osquery/database.h
-- Installing: /usr/local/include/osquery/devtools.h
-- Installing: /usr/local/include/osquery/dispatcher.h
-- Installing: /usr/local/include/osquery/events.h
-- Installing: /usr/local/include/osquery/filesystem.h
-- Installing: /usr/local/include/osquery/flags.h
-- Installing: /usr/local/include/osquery/hash.h
-- Installing: /usr/local/include/osquery/logger
-- Installing: /usr/local/include/osquery/logger/plugin.h
-- Installing: /usr/local/include/osquery/logger.h
-- Installing: /usr/local/include/osquery/registry.h
-- Installing: /usr/local/include/osquery/scheduler.h
-- Installing: /usr/local/include/osquery/sql.h
-- Installing: /usr/local/include/osquery/status.h
-- Installing: /usr/local/include/osquery/tables.h
-- Installing: /usr/local/bin/osqueryi
-- Removed runtime path from "/usr/local/bin/osqueryi"
-- Installing: /usr/local/bin/osqueryd
-- Removed runtime path from "/usr/local/bin/osqueryd"
-- Installing: /var/osquery/osquery.example.conf
-- Installing: /etc/init.d/osqueryd
```
|
infrastructure
|
make make install match the file layout of packages for example make install will install the example config to var osquery osquery example conf but the deb will install it to usr share osquery osquery example conf make packages is our source of truth so let s match that or remove the install target deb root vagrant ubuntu trusty vagrant dpkg c build linux osquery deb drwxr xr x drwxr xr x var drwxr xr x var osquery rw r r var osquery osquery example conf drwxr xr x var log drwxr xr x var log osquery drwxr xr x etc drwxr xr x etc init d rw r r etc init d osqueryd drwxr xr x usr drwxr xr x usr bin rwxr xr x usr bin osqueryd rwxr xr x usr bin osqueryi drwxr xr x usr share drwxr xr x usr share doc drwxr xr x usr share doc osquery rw r r usr share doc osquery changelog debian gz make root vagrant ubuntu trusty vagrant make install building for ubuntu configuring done generating done build files have been written to vagrant build ubuntu install the project install configuration installing usr local lib libosquery a installing usr local include installing usr local include osquery installing usr local include osquery config installing usr local include osquery config plugin h installing usr local include osquery config h installing usr local include osquery core h installing usr local include osquery database installing usr local include osquery database db handle h installing usr local include osquery database query h installing usr local include osquery database results h installing usr local include osquery database h installing usr local include osquery devtools h installing usr local include osquery dispatcher h installing usr local include osquery events h installing usr local include osquery filesystem h installing usr local include osquery flags h installing usr local include osquery hash h installing usr local include osquery logger installing usr local include osquery logger plugin h installing usr local include osquery logger h installing usr local include osquery registry h installing usr local include osquery scheduler h installing usr local include osquery sql h installing usr local include osquery status h installing usr local include osquery tables h installing usr local bin osqueryi removed runtime path from usr local bin osqueryi installing usr local bin osqueryd removed runtime path from usr local bin osqueryd installing var osquery osquery example conf installing etc init d osqueryd
| 1
|
35,292
| 30,913,509,337
|
IssuesEvent
|
2023-08-05 02:06:14
|
SciTools/cartopy
|
https://api.github.com/repos/SciTools/cartopy
|
closed
|
Automate release upload
|
Type: Infrastructure
|
We should make our lives easier and automatically upload the sdist to PyPI when a new release is tagged. Stealing from our MetPy Travis config:
```yml
deploy:
- provider: pypi
user: dopplershift
password:
secure: <ENCRYPTED_PW>
distributions: sdist
upload_docs: no
on:
repo: SciTools/cartopy
python: 3.6
condition: [MAKE SURE WE ONLY DO IT ONCE]
tags: true
```
We might also want to automatically update the docs, but I’m not clear on what the current process for that is.
|
1.0
|
Automate release upload - We should make our lives easier and automatically upload the sdist to PyPI when a new release is tagged. Stealing from our MetPy Travis config:
```yml
deploy:
- provider: pypi
user: dopplershift
password:
secure: <ENCRYPTED_PW>
distributions: sdist
upload_docs: no
on:
repo: SciTools/cartopy
python: 3.6
condition: [MAKE SURE WE ONLY DO IT ONCE]
tags: true
```
We might also want to automatically update the docs, but I’m not clear on what the current process for that is.
|
infrastructure
|
automate release upload we should make our lives easier and automatically upload the sdist to pypi when a new release is tagged stealing from our metpy travis config yml deploy provider pypi user dopplershift password secure distributions sdist upload docs no on repo scitools cartopy python condition tags true we might also want to automatically update the docs but i’m not clear on what the current process for that is
| 1
|
2,920
| 3,964,099,702
|
IssuesEvent
|
2016-05-02 23:01:47
|
catapult-project/catapult
|
https://api.github.com/repos/catapult-project/catapult
|
closed
|
BattOr smoke test in the CQ fails but the "smoke test" CQ step remains yellow
|
Bug Infrastructure P1
|
It looks like even after the BattOr smoke test fails:

the step in the try job continues to be yellow:

Randy, could you look into why this might be happening? [Here's the try job](https://codereview.chromium.org/1932373002/) in question: The good news is that it fails!
|
1.0
|
BattOr smoke test in the CQ fails but the "smoke test" CQ step remains yellow - It looks like even after the BattOr smoke test fails:

the step in the try job continues to be yellow:

Randy, could you look into why this might be happening? [Here's the try job](https://codereview.chromium.org/1932373002/) in question: The good news is that it fails!
|
infrastructure
|
battor smoke test in the cq fails but the smoke test cq step remains yellow it looks like even after the battor smoke test fails the step in the try job continues to be yellow randy could you look into why this might be happening in question the good news is that it fails
| 1
|
31,142
| 25,360,519,611
|
IssuesEvent
|
2022-11-20 20:53:31
|
ceydaduzgec/SWE573
|
https://api.github.com/repos/ceydaduzgec/SWE573
|
closed
|
Create Github actions
|
Type: Bug Fix Component: Infrastructure Type: Research
|
- Research Github actions
- Implement initial setup for Github actions
- Test if it works
- Add it to documentation
|
1.0
|
Create Github actions - - Research Github actions
- Implement initial setup for Github actions
- Test if it works
- Add it to documentation
|
infrastructure
|
create github actions research github actions implement initial setup for github actions test if it works add it to documentation
| 1
|
59,068
| 24,646,260,556
|
IssuesEvent
|
2022-10-17 15:03:24
|
hashicorp/terraform-provider-azurerm
|
https://api.github.com/repos/hashicorp/terraform-provider-azurerm
|
closed
|
Data Source: azurerm_shared_image_version outputs random version ID
|
question service/virtual-machine
|
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Community Note
<!--- Please keep this note for the community --->
* Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform Version
1.0.5
### AzureRM Provider Version
3.26.0
### Affected Resource(s)/Data Source(s)
resource "azurerm_linux_virtual_machine" "linux_vm"
### Terraform Configuration Files
```hcl
variable "shared_image" {
default = {
gallery_name = "XXXXXXXXX"
gallery_resource_group = "XXXXXXXXX"
image_name = "XXXXXXXXX"
"version" = "latest"
}
}
data "azurerm_shared_image_version" "shared_image" {
name = var.shared_image.version
resource_group_name = var.shared_image.gallery_resource_group
gallery_name = var.shared_image.gallery_name
image_name = var.shared_image.image_name
}
resource "azurerm_linux_virtual_machine" "linux_vm" {
...
source_image_id = data.azurerm_shared_image_version.shared_image.id
... (other parameters)
}
```
### Debug Output/Panic Output
```shell
Will obtain the debug info if necessary
```
### Expected Behaviour
ID of the latest version of the specified image should be used for creating the VM.
### Actual Behaviour
ID of the latest version of the specified image should be used, but some random version of the image is being used for creating the vm.
Below is the output of terraform plan
resource "azurerm_windows_virtual_machine" "windows_vm" {
...
source_image_id = "/subscriptions/XXXXXXXX/resourceGroups/XXXXXXXXXX/providers/Microsoft.Compute/galleries/XXXXXXXXXX/images/XXXXXXXXXX/versions/1.0.9"
...
}
### Steps to Reproduce
terraform init
terraform plan
### Important Factoids
_No response_
### References
_No response_
|
1.0
|
Data Source: azurerm_shared_image_version outputs random version ID - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Community Note
<!--- Please keep this note for the community --->
* Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform Version
1.0.5
### AzureRM Provider Version
3.26.0
### Affected Resource(s)/Data Source(s)
resource "azurerm_linux_virtual_machine" "linux_vm"
### Terraform Configuration Files
```hcl
variable "shared_image" {
default = {
gallery_name = "XXXXXXXXX"
gallery_resource_group = "XXXXXXXXX"
image_name = "XXXXXXXXX"
"version" = "latest"
}
}
data "azurerm_shared_image_version" "shared_image" {
name = var.shared_image.version
resource_group_name = var.shared_image.gallery_resource_group
gallery_name = var.shared_image.gallery_name
image_name = var.shared_image.image_name
}
resource "azurerm_linux_virtual_machine" "linux_vm" {
...
source_image_id = data.azurerm_shared_image_version.shared_image.id
... (other parameters)
}
```
### Debug Output/Panic Output
```shell
Will obtain the debug info if necessary
```
### Expected Behaviour
ID of the latest version of the specified image should be used for creating the VM.
### Actual Behaviour
ID of the latest version of the specified image should be used, but some random version of the image is being used for creating the vm.
Below is the output of terraform plan
resource "azurerm_windows_virtual_machine" "windows_vm" {
...
source_image_id = "/subscriptions/XXXXXXXX/resourceGroups/XXXXXXXXXX/providers/Microsoft.Compute/galleries/XXXXXXXXXX/images/XXXXXXXXXX/versions/1.0.9"
...
}
### Steps to Reproduce
terraform init
terraform plan
### Important Factoids
_No response_
### References
_No response_
|
non_infrastructure
|
data source azurerm shared image version outputs random version id is there an existing issue for this i have searched the existing issues community note please vote on this issue by adding a thumbsup to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform version azurerm provider version affected resource s data source s resource azurerm linux virtual machine linux vm terraform configuration files hcl variable shared image default gallery name xxxxxxxxx gallery resource group xxxxxxxxx image name xxxxxxxxx version latest data azurerm shared image version shared image name var shared image version resource group name var shared image gallery resource group gallery name var shared image gallery name image name var shared image image name resource azurerm linux virtual machine linux vm source image id data azurerm shared image version shared image id other parameters debug output panic output shell will obtain the debug info if necessary expected behaviour id of the latest version of the specified image should be used for creating the vm actual behaviour id of the latest version of the specified image should be used but some random version of the image is being used for creating the vm below is the output of terraform plan resource azurerm windows virtual machine windows vm source image id subscriptions xxxxxxxx resourcegroups xxxxxxxxxx providers microsoft compute galleries xxxxxxxxxx images xxxxxxxxxx versions steps to reproduce terraform init terraform plan important factoids no response references no response
| 0
|
14,249
| 10,726,092,689
|
IssuesEvent
|
2019-10-28 08:35:25
|
aarhusstadsarkiv/digital-archive
|
https://api.github.com/repos/aarhusstadsarkiv/digital-archive
|
closed
|
Update pyproject with extra metadata for PyPI
|
infrastructure
|
<!--
Hi! :)
If applicable, please link guides/articles when submitting infrastructure issues.
The markdown syntax for adding links to text is `[text](url)`
-->
Add readme.md, repository/homepage/docs info to `pyproject.toml` - like this:
```toml
readme = "README.md"
homepage = "link"
repository = "link"
documentation = "link"
keywords = ["x", "y", "z"]
```
|
1.0
|
Update pyproject with extra metadata for PyPI - <!--
Hi! :)
If applicable, please link guides/articles when submitting infrastructure issues.
The markdown syntax for adding links to text is `[text](url)`
-->
Add readme.md, repository/homepage/docs info to `pyproject.toml` - like this:
```toml
readme = "README.md"
homepage = "link"
repository = "link"
documentation = "link"
keywords = ["x", "y", "z"]
```
|
infrastructure
|
update pyproject with extra metadata for pypi hi if applicable please link guides articles when submitting infrastructure issues the markdown syntax for adding links to text is url add readme md repository homepage docs info to pyproject toml like this toml readme readme md homepage link repository link documentation link keywords
| 1
|
109,040
| 4,366,769,486
|
IssuesEvent
|
2016-08-03 15:14:15
|
openshift/origin-web-console
|
https://api.github.com/repos/openshift/origin-web-console
|
closed
|
Events sidebar header doesn't wrap well at some widths
|
area/styles kind/bug priority/P3
|
When there are warnings, only the number wraps to the next line.

|
1.0
|
Events sidebar header doesn't wrap well at some widths - When there are warnings, only the number wraps to the next line.

|
non_infrastructure
|
events sidebar header doesn t wrap well at some widths when there are warnings only the number wraps to the next line
| 0
|
124,740
| 16,656,706,359
|
IssuesEvent
|
2021-06-05 17:01:21
|
tarantool/httpng
|
https://api.github.com/repos/tarantool/httpng
|
closed
|
Implement multilisten support
|
8sp backlog enhancement in design teamE
|
Ability to configure:
IPv4/IPv6 addresses to listen;
HTTP/HTTPS;
more than one TCP port;
unix sockets;
|
1.0
|
Implement multilisten support - Ability to configure:
IPv4/IPv6 addresses to listen;
HTTP/HTTPS;
more than one TCP port;
unix sockets;
|
non_infrastructure
|
implement multilisten support ability to configure addresses to listen http https more than one tcp port unix sockets
| 0
|
6,971
| 6,688,985,563
|
IssuesEvent
|
2017-10-08 20:45:25
|
Seaal/Pug
|
https://api.github.com/repos/Seaal/Pug
|
closed
|
Update to .net core 2.0
|
area: infrastructure blocker effort: high feature priority: high
|
Once VS2017 supports .net core 2.0 without a preview version (or when I get back on my desktop and installing a preview is more feasible), the project needs to be updated to use .net core 2.0 and libraries to use .net standard 2.0
|
1.0
|
Update to .net core 2.0 - Once VS2017 supports .net core 2.0 without a preview version (or when I get back on my desktop and installing a preview is more feasible), the project needs to be updated to use .net core 2.0 and libraries to use .net standard 2.0
|
infrastructure
|
update to net core once supports net core without a preview version or when i get back on my desktop and installing a preview is more feasible the project needs to be updated to use net core and libraries to use net standard
| 1
|
9,893
| 8,227,629,661
|
IssuesEvent
|
2018-09-07 00:07:38
|
Dallas-Makerspace/tracker
|
https://api.github.com/repos/Dallas-Makerspace/tracker
|
closed
|
Motherboard offline - no dhcp lease
|
CR/Maintence Committee/Infrastructure Committee/VCC Priority/HIGH System/Community Grid System/Overcloud System/Undercloud
|
## Expected Behavior
Undercloud server motherboard is accessable via all services
## Actual Behavior
Layer 2 passes tests
Layer 3 does not get dhcp lease
## Steps to Reproduce the Problem
1. ping motherboard.dms.local
1. login to motherboard console
1. ping 1.1.1.1
## Specifications
(The version of the project, operating system, hardware etc.)
|
1.0
|
Motherboard offline - no dhcp lease - ## Expected Behavior
Undercloud server motherboard is accessable via all services
## Actual Behavior
Layer 2 passes tests
Layer 3 does not get dhcp lease
## Steps to Reproduce the Problem
1. ping motherboard.dms.local
1. login to motherboard console
1. ping 1.1.1.1
## Specifications
(The version of the project, operating system, hardware etc.)
|
infrastructure
|
motherboard offline no dhcp lease expected behavior undercloud server motherboard is accessable via all services actual behavior layer passes tests layer does not get dhcp lease steps to reproduce the problem ping motherboard dms local login to motherboard console ping specifications the version of the project operating system hardware etc
| 1
|
102,744
| 11,306,530,074
|
IssuesEvent
|
2020-01-18 14:47:24
|
frc5024/InfiniteRecharge
|
https://api.github.com/repos/frc5024/InfiniteRecharge
|
opened
|
Add your technical information
|
documentation feature
|
We want to be like 254: https://media.team254.com/resources/Team_254_Tech_Binder_2019.pdf
When you are finished designing your subsystem, add your documentation to the docs/technicalbinder folder.
|
1.0
|
Add your technical information - We want to be like 254: https://media.team254.com/resources/Team_254_Tech_Binder_2019.pdf
When you are finished designing your subsystem, add your documentation to the docs/technicalbinder folder.
|
non_infrastructure
|
add your technical information we want to be like when you are finished designing your subsystem add your documentation to the docs technicalbinder folder
| 0
|
8,809
| 23,517,470,491
|
IssuesEvent
|
2022-08-18 23:36:38
|
bcgov/foi-flow
|
https://api.github.com/repos/bcgov/foi-flow
|
opened
|
Analysis on Redaction Enhancement Automation Tool/Framework selection
|
architecture Task dev
|
Title of ticket:
#### Description
The purpose of this task is to find a tool or framework to Automate redaction process or in other words to create an Enhancement UX to provide suggestions/indication to IAO during redaction process. This is initial step up before we jump into POC phase. At this moment, RPA tools like RoboCorp(Open Source), UiPath etc. is in consideration. But will expand it to other possibilities based on the analysis.
#### In scope
- Built-in PDF and other Office files reading capability
- Support for programming or customization using Python or any High Level Programming language
- Built in support for Redaction or Highlight
- Cost Analysis or Possible Open Source options with feature support
- Community Support and other Licensing implications
#### Out of scope
Considering the scope of this EPIC or Goal, Infrastructure which we need to run is out of scope, but will consider once a few possibilities has been found.
#### Dependencies
Are there any dependencies?
#### DOD
- [ ] Analyzed RPA tools which can read PDF and other basic document format, which has take script based automate logic
- [ ] Analyzed options for basic REgex expressions on those script that can be applied to Documents for highlighting or indications
- [ ] Has everything to start a POC
|
1.0
|
Analysis on Redaction Enhancement Automation Tool/Framework selection - Title of ticket:
#### Description
The purpose of this task is to find a tool or framework to Automate redaction process or in other words to create an Enhancement UX to provide suggestions/indication to IAO during redaction process. This is initial step up before we jump into POC phase. At this moment, RPA tools like RoboCorp(Open Source), UiPath etc. is in consideration. But will expand it to other possibilities based on the analysis.
#### In scope
- Built-in PDF and other Office files reading capability
- Support for programming or customization using Python or any High Level Programming language
- Built in support for Redaction or Highlight
- Cost Analysis or Possible Open Source options with feature support
- Community Support and other Licensing implications
#### Out of scope
Considering the scope of this EPIC or Goal, Infrastructure which we need to run is out of scope, but will consider once a few possibilities has been found.
#### Dependencies
Are there any dependencies?
#### DOD
- [ ] Analyzed RPA tools which can read PDF and other basic document format, which has take script based automate logic
- [ ] Analyzed options for basic REgex expressions on those script that can be applied to Documents for highlighting or indications
- [ ] Has everything to start a POC
|
non_infrastructure
|
analysis on redaction enhancement automation tool framework selection title of ticket description the purpose of this task is to find a tool or framework to automate redaction process or in other words to create an enhancement ux to provide suggestions indication to iao during redaction process this is initial step up before we jump into poc phase at this moment rpa tools like robocorp open source uipath etc is in consideration but will expand it to other possibilities based on the analysis in scope built in pdf and other office files reading capability support for programming or customization using python or any high level programming language built in support for redaction or highlight cost analysis or possible open source options with feature support community support and other licensing implications out of scope considering the scope of this epic or goal infrastructure which we need to run is out of scope but will consider once a few possibilities has been found dependencies are there any dependencies dod analyzed rpa tools which can read pdf and other basic document format which has take script based automate logic analyzed options for basic regex expressions on those script that can be applied to documents for highlighting or indications has everything to start a poc
| 0
|
9,820
| 3,073,295,035
|
IssuesEvent
|
2015-08-19 21:11:42
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
Three TypeExtensions tests failing on Linux with xUnit 2.1.0
|
2 - In Progress System.Reflection test bug
|
A few of the reflection tests seem to be failing on Linux when run with xUnit 2.1.0. These tests are doing fairly simple stuff, just subscribing some events and then triggering them, and checking the state afterwards.
System.Reflection.Compatibility.UnitTests.EventInfoAddEventHandler.PosTest2
System.Reflection.Compatibility.UnitTests.EventInfoAddEventHandler.PosTest3
System.Reflection.Compatibility.UnitTests.EventInfoAddEventHandler.PosTest1
|
1.0
|
Three TypeExtensions tests failing on Linux with xUnit 2.1.0 - A few of the reflection tests seem to be failing on Linux when run with xUnit 2.1.0. These tests are doing fairly simple stuff, just subscribing some events and then triggering them, and checking the state afterwards.
System.Reflection.Compatibility.UnitTests.EventInfoAddEventHandler.PosTest2
System.Reflection.Compatibility.UnitTests.EventInfoAddEventHandler.PosTest3
System.Reflection.Compatibility.UnitTests.EventInfoAddEventHandler.PosTest1
|
non_infrastructure
|
three typeextensions tests failing on linux with xunit a few of the reflection tests seem to be failing on linux when run with xunit these tests are doing fairly simple stuff just subscribing some events and then triggering them and checking the state afterwards system reflection compatibility unittests eventinfoaddeventhandler system reflection compatibility unittests eventinfoaddeventhandler system reflection compatibility unittests eventinfoaddeventhandler
| 0
|
254,040
| 27,343,192,651
|
IssuesEvent
|
2023-02-27 01:02:10
|
UpendoVentures/ClearCache
|
https://api.github.com/repos/UpendoVentures/ClearCache
|
opened
|
CVE-2022-23535 (High) detected in LiteDB-3.1.0.0.dll
|
security vulnerability
|
## CVE-2022-23535 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>LiteDB-3.1.0.0.dll</b></p></summary>
<p>LiteDB</p>
<p>Library home page: <a href="https://api.nuget.org/packages/litedb.3.1.0.nupkg">https://api.nuget.org/packages/litedb.3.1.0.nupkg</a></p>
<p>Path to vulnerable library: /References/DNN/09.02.01/LiteDB.dll</p>
<p>
Dependency Hierarchy:
- :x: **LiteDB-3.1.0.0.dll** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/UpendoVentures/ClearCache/commit/9b2fdadcb0ce9dacb6e87e0b604cf35d751503be">9b2fdadcb0ce9dacb6e87e0b604cf35d751503be</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
LiteDB is a small, fast and lightweight .NET NoSQL embedded database. Versions prior to 5.0.13 are subject to Deserialization of Untrusted Data. LiteDB uses a special field in JSON documents to cast different types from `BsonDocument` to POCO classes. When instances of an object are not the same of class, `BsonMapper` use a special field `_type` string info with full class name with assembly to be loaded and fit into your model. If your end-user can send to your app a plain JSON string, deserialization can load an unsafe object to fit into your model. This issue is patched in version 5.0.13 with some basic fixes to avoid this, but is not 100% guaranteed when using `Object` type. The next major version will contain an allow-list to select what kind of Assembly can be loaded. Workarounds are detailed in the vendor advisory.
<p>Publish Date: 2023-02-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23535>CVE-2022-23535</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-23535">https://www.cve.org/CVERecord?id=CVE-2022-23535</a></p>
<p>Release Date: 2023-02-24</p>
<p>Fix Resolution: LiteDB - 5.0.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-23535 (High) detected in LiteDB-3.1.0.0.dll - ## CVE-2022-23535 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>LiteDB-3.1.0.0.dll</b></p></summary>
<p>LiteDB</p>
<p>Library home page: <a href="https://api.nuget.org/packages/litedb.3.1.0.nupkg">https://api.nuget.org/packages/litedb.3.1.0.nupkg</a></p>
<p>Path to vulnerable library: /References/DNN/09.02.01/LiteDB.dll</p>
<p>
Dependency Hierarchy:
- :x: **LiteDB-3.1.0.0.dll** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/UpendoVentures/ClearCache/commit/9b2fdadcb0ce9dacb6e87e0b604cf35d751503be">9b2fdadcb0ce9dacb6e87e0b604cf35d751503be</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
LiteDB is a small, fast and lightweight .NET NoSQL embedded database. Versions prior to 5.0.13 are subject to Deserialization of Untrusted Data. LiteDB uses a special field in JSON documents to cast different types from `BsonDocument` to POCO classes. When instances of an object are not the same of class, `BsonMapper` use a special field `_type` string info with full class name with assembly to be loaded and fit into your model. If your end-user can send to your app a plain JSON string, deserialization can load an unsafe object to fit into your model. This issue is patched in version 5.0.13 with some basic fixes to avoid this, but is not 100% guaranteed when using `Object` type. The next major version will contain an allow-list to select what kind of Assembly can be loaded. Workarounds are detailed in the vendor advisory.
<p>Publish Date: 2023-02-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23535>CVE-2022-23535</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-23535">https://www.cve.org/CVERecord?id=CVE-2022-23535</a></p>
<p>Release Date: 2023-02-24</p>
<p>Fix Resolution: LiteDB - 5.0.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_infrastructure
|
cve high detected in litedb dll cve high severity vulnerability vulnerable library litedb dll litedb library home page a href path to vulnerable library references dnn litedb dll dependency hierarchy x litedb dll vulnerable library found in head commit a href found in base branch master vulnerability details litedb is a small fast and lightweight net nosql embedded database versions prior to are subject to deserialization of untrusted data litedb uses a special field in json documents to cast different types from bsondocument to poco classes when instances of an object are not the same of class bsonmapper use a special field type string info with full class name with assembly to be loaded and fit into your model if your end user can send to your app a plain json string deserialization can load an unsafe object to fit into your model this issue is patched in version with some basic fixes to avoid this but is not guaranteed when using object type the next major version will contain an allow list to select what kind of assembly can be loaded workarounds are detailed in the vendor advisory publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution litedb step up your open source security game with mend
| 0
|
117,730
| 9,957,806,933
|
IssuesEvent
|
2019-07-05 18:19:17
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
HTTP2 code coverage: add tests for methods other than GET and POST
|
area-System.Net.Http.SocketsHttpHandler test enhancement
|
See code here, line 978: file://clrmain/public/writable/users/geoffrek/http2codecoverage/System.Net.Http_Http2Connection.htm#file0_line958
We probably have some HTTP/1.1 tests we can convert for this.
|
1.0
|
HTTP2 code coverage: add tests for methods other than GET and POST - See code here, line 978: file://clrmain/public/writable/users/geoffrek/http2codecoverage/System.Net.Http_Http2Connection.htm#file0_line958
We probably have some HTTP/1.1 tests we can convert for this.
|
non_infrastructure
|
code coverage add tests for methods other than get and post see code here line file clrmain public writable users geoffrek system net http htm we probably have some http tests we can convert for this
| 0
|
22,912
| 11,802,630,207
|
IssuesEvent
|
2020-03-18 21:59:35
|
Azure/azure-cli
|
https://api.github.com/repos/Azure/azure-cli
|
closed
|
Unclear of what "server" and "name" should be when using azure sql db import
|
SQL Service Attention
|
When using `azure sql db import` to create a new db from a bacpac file from storage, I get the "Can not perform requested operation on nested resource." as it would require the to-be database to already be present. "Can not perform requested operation on nested resource. Parent resource '<server>/<db-name>' not found."
When trying to omit it, it also fails. When running with --debug I can see I get a 400 on /Microsoft.Sql/servers/testenv-westeurope/databases/principal-portal-db/extensions/import which makes sense as this database doesn't exist yet. But how should I then be able to create it? Create first and then import?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: c1bdd674-b454-e863-c943-9c979b466a92
* Version Independent ID: b15341f9-3f1b-00e0-a1c2-970f25a6ab03
* Content: [az sql db](https://docs.microsoft.com/en-us/cli/azure/sql/db?view=azure-cli-latest#az-sql-db-import)
* Content Source: [src/azure-cli/azure/cli/command_modules/sql/_help.py](https://github.com/Azure/azure-cli/blob/dev/src/azure-cli/azure/cli/command_modules/sql/_help.py)
* Service: **sql-database**
* GitHub Login: @rloutlaw
* Microsoft Alias: **routlaw**
|
1.0
|
Unclear of what "server" and "name" should be when using azure sql db import - When using `azure sql db import` to create a new db from a bacpac file from storage, I get the "Can not perform requested operation on nested resource." as it would require the to-be database to already be present. "Can not perform requested operation on nested resource. Parent resource '<server>/<db-name>' not found."
When trying to omit it, it also fails. When running with --debug I can see I get a 400 on /Microsoft.Sql/servers/testenv-westeurope/databases/principal-portal-db/extensions/import which makes sense as this database doesn't exist yet. But how should I then be able to create it? Create first and then import?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: c1bdd674-b454-e863-c943-9c979b466a92
* Version Independent ID: b15341f9-3f1b-00e0-a1c2-970f25a6ab03
* Content: [az sql db](https://docs.microsoft.com/en-us/cli/azure/sql/db?view=azure-cli-latest#az-sql-db-import)
* Content Source: [src/azure-cli/azure/cli/command_modules/sql/_help.py](https://github.com/Azure/azure-cli/blob/dev/src/azure-cli/azure/cli/command_modules/sql/_help.py)
* Service: **sql-database**
* GitHub Login: @rloutlaw
* Microsoft Alias: **routlaw**
|
non_infrastructure
|
unclear of what server and name should be when using azure sql db import when using azure sql db import to create a new db from a bacpac file from storage i get the can not perform requested operation on nested resource as it would require the to be database to already be present can not perform requested operation on nested resource parent resource lt server gt lt db name gt not found when trying to omit it it also fails when running with debug i can see i get a on microsoft sql servers testenv westeurope databases principal portal db extensions import which makes sense as this database doesn t exist yet but how should i then be able to create it create first and then import document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service sql database github login rloutlaw microsoft alias routlaw
| 0
|
20,641
| 14,099,297,425
|
IssuesEvent
|
2020-11-06 01:01:37
|
noahtalerman/test-issues-kolide
|
https://api.github.com/repos/noahtalerman/test-issues-kolide
|
opened
|
[CLOSED] Turn down logging in tests except after failure
|
Component: Development Infrastructure
|
<a href="https://github.com/zwass"><img src="https://avatars0.githubusercontent.com/u/575602?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [zwass](https://github.com/zwass)**
_Thursday Aug 04, 2016 at 23:00 GMT_
_Originally opened as https://github.com/kolide/fleet/pull/41_
----
When tests succeed, there is now no logging of DB queries or HTTP
requests. If a test fails, the logs will be output.
----
_**[zwass](https://github.com/zwass)** included the following code: https://github.com/kolide/fleet/pull/41/commits_
|
1.0
|
[CLOSED] Turn down logging in tests except after failure - <a href="https://github.com/zwass"><img src="https://avatars0.githubusercontent.com/u/575602?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [zwass](https://github.com/zwass)**
_Thursday Aug 04, 2016 at 23:00 GMT_
_Originally opened as https://github.com/kolide/fleet/pull/41_
----
When tests succeed, there is now no logging of DB queries or HTTP
requests. If a test fails, the logs will be output.
----
_**[zwass](https://github.com/zwass)** included the following code: https://github.com/kolide/fleet/pull/41/commits_
|
infrastructure
|
turn down logging in tests except after failure issue by thursday aug at gmt originally opened as when tests succeed there is now no logging of db queries or http requests if a test fails the logs will be output included the following code
| 1
|
437,515
| 30,601,990,427
|
IssuesEvent
|
2023-07-22 14:00:44
|
disorderedmaterials/dissolve
|
https://api.github.com/repos/disorderedmaterials/dissolve
|
closed
|
Silica example missing setting of density
|
Scope: Documentation
|
**Describe the problem**
The bulk silica example does not tell users to set the density any more, so lots of unhappy simulations result.
|
1.0
|
Silica example missing setting of density - **Describe the problem**
The bulk silica example does not tell users to set the density any more, so lots of unhappy simulations result.
|
non_infrastructure
|
silica example missing setting of density describe the problem the bulk silica example does not tell users to set the density any more so lots of unhappy simulations result
| 0
|
9,205
| 7,868,155,663
|
IssuesEvent
|
2018-06-23 17:51:01
|
kryptokay/kwallet
|
https://api.github.com/repos/kryptokay/kwallet
|
opened
|
Deploy kwallet to production! #17
|
Infrastructure & DevOps
|
Historically speaking, all the feedback we've received came from our internal / beta customers.
They're quite happy with the current state of the art, so now it's time to make `kwallet` publicly available!
That would imply we'll have a public URL access, so everybody around the world could check it.
|
1.0
|
Deploy kwallet to production! #17 - Historically speaking, all the feedback we've received came from our internal / beta customers.
They're quite happy with the current state of the art, so now it's time to make `kwallet` publicly available!
That would imply we'll have a public URL access, so everybody around the world could check it.
|
infrastructure
|
deploy kwallet to production historically speaking all the feedback we ve received came from our internal beta customers they re quite happy with the current state of the art so now it s time to make kwallet publicly available that would imply we ll have a public url access so everybody around the world could check it
| 1
|
17,011
| 12,192,838,898
|
IssuesEvent
|
2020-04-29 13:33:45
|
google-org/new-project-ny
|
https://api.github.com/repos/google-org/new-project-ny
|
opened
|
Merge .gitignore and .dockerignore files into top level file
|
infrastructure
|
Issue first raised in https://github.com/google-org/new-project-ny/pull/19
We have a lot of .gitignore and .dockerignore files that make sense to put at the top level.
|
1.0
|
Merge .gitignore and .dockerignore files into top level file - Issue first raised in https://github.com/google-org/new-project-ny/pull/19
We have a lot of .gitignore and .dockerignore files that make sense to put at the top level.
|
infrastructure
|
merge gitignore and dockerignore files into top level file issue first raised in we have a lot of gitignore and dockerignore files that make sense to put at the top level
| 1
|
462,435
| 13,247,307,080
|
IssuesEvent
|
2020-08-19 17:02:10
|
neuropoly/spinalcordtoolbox
|
https://api.github.com/repos/neuropoly/spinalcordtoolbox
|
closed
|
GLIBCXX_3.4.15 not found on CentOS 6
|
installation priority:HIGH
|
After installing SCT (5f09905515c78aefbaed7781e2d0664427e9cf79) on a centos6 station, I ran into the following problem while running `sct_testing`:
~~~
Checking sct_fmri_moco..............................[FAIL]
====================================================================================================
sct_fmri_moco -i fmri/fmri_r.nii.gz -g 5 -x nn -r 0
====================================================================================================
--
Spinal Cord Toolbox (git-master-5f09905515c78aefbaed7781e2d0664427e9cf79)
Merge within groups: 0%| | 0/6 [00:00<?, ?iter/s]
Merge within groups: 17%|###6 | 1/6 [00:01<00:08, 1.63s/iter]
Merge within groups: 33%|#######3 | 2/6 [00:03<00:06, 1.60s/iter]
Merge within groups: 50%|########### | 3/6 [00:04<00:04, 1.55s/iter]
Merge within groups: 67%|##############6 | 4/6 [00:06<00:03, 1.50s/iter]
Merge within groups: 83%|##################3 | 5/6 [00:07<00:01, 1.48s/iter]
Merge within groups: 100%|######################| 6/6 [00:08<00:00, 1.50s/iter]
Merge within groups: 100%|######################| 6/6 [00:08<00:00, 1.50s/iter]
Z=0/0: 0%| | 0/6 [00:00<?, ?iter/s]
Z=0/0: 0%| | 0/6 [00:00<?, ?iter/s]
Input parameters:
input file ............fmri/fmri_r.nii.gz
Create temporary folder (/tmp/sct-20200214135325.614152-fmri_moco-0129le9n)...
Copying input data to tmp folder and convert to nii...
sct_convert -i /home/jcohen-adad/sct_testing_data/fmri/fmri_r.nii.gz -o /tmp/sct-20200214135325.614152-fmri_moco-0129le9n/fmri.nii
Get dimensions of data...
34 x 34 x 6 x 30
Data orientation: RPI
Treated as axial
Split along T dimension...
Merging volumes...
-------------------------------------------------------------------------------
Estimating motion...
-------------------------------------------------------------------------------
Input parameters:
Input file ............fmri_averaged_groups.nii
Reference file ........fmri_mean_0.nii.gz
Polynomial degree .....2
Smoothing kernel ......0
Gradient step .........1
Metric ................MeanSquares
Sampling ..............0.2
Todo ..................estimate_and_apply
Mask .................
Output mat folder .....mat_groups
Data dimensions:
34 x 34 x 6 x 6
Copy file_target to a temporary file...
sct_convert -i fmri_mean_0.nii.gz -o target.nii.gz
Register. Loop across Z (note: there is only one Z if orientation is axial
Traceback (most recent call last):
File "/home/jcohen-adad/sct/scripts/sct_fmri_moco.py", line 453, in <module>
main()
File "/home/jcohen-adad/sct/scripts/sct_fmri_moco.py", line 213, in main
fmri_moco(param)
File "/home/jcohen-adad/sct/scripts/sct_fmri_moco.py", line 350, in fmri_moco
file_mat = moco.moco(param_moco)
File "/home/jcohen-adad/sct/scripts/msct_moco.py", line 152, in moco
file_data_splitZ_splitT_moco[it], im_mask=input_mask)
File "/home/jcohen-adad/sct/scripts/msct_moco.py", line 276, in register
status, output = sct.run(cmd, verbose=0, **kw)
File "/home/jcohen-adad/sct/scripts/sct_utils.py", line 311, in run
raise RunError(output)
sct_utils.RunError: /home/jcohen-adad/sct/bin/isct_antsSliceRegularizedRegistration: /usr/lib64/libstdc++.so.6: version `GLIBCXX_3.4.15' not found (required by /home/jcohen-adad/sct/bin/isct_antsSliceRegularizedRegistration)
~~~
This issue might be related to https://github.com/neuropoly/spinalcordtoolbox/issues/711, https://github.com/neuropoly/spinalcordtoolbox/issues/158, https://github.com/neuropoly/spinalcordtoolbox/issues/1704
|
1.0
|
GLIBCXX_3.4.15 not found on CentOS 6 - After installing SCT (5f09905515c78aefbaed7781e2d0664427e9cf79) on a centos6 station, I ran into the following problem while running `sct_testing`:
~~~
Checking sct_fmri_moco..............................[FAIL]
====================================================================================================
sct_fmri_moco -i fmri/fmri_r.nii.gz -g 5 -x nn -r 0
====================================================================================================
--
Spinal Cord Toolbox (git-master-5f09905515c78aefbaed7781e2d0664427e9cf79)
Merge within groups: 0%| | 0/6 [00:00<?, ?iter/s]
Merge within groups: 17%|###6 | 1/6 [00:01<00:08, 1.63s/iter]
Merge within groups: 33%|#######3 | 2/6 [00:03<00:06, 1.60s/iter]
Merge within groups: 50%|########### | 3/6 [00:04<00:04, 1.55s/iter]
Merge within groups: 67%|##############6 | 4/6 [00:06<00:03, 1.50s/iter]
Merge within groups: 83%|##################3 | 5/6 [00:07<00:01, 1.48s/iter]
Merge within groups: 100%|######################| 6/6 [00:08<00:00, 1.50s/iter]
Merge within groups: 100%|######################| 6/6 [00:08<00:00, 1.50s/iter]
Z=0/0: 0%| | 0/6 [00:00<?, ?iter/s]
Z=0/0: 0%| | 0/6 [00:00<?, ?iter/s]
Input parameters:
input file ............fmri/fmri_r.nii.gz
Create temporary folder (/tmp/sct-20200214135325.614152-fmri_moco-0129le9n)...
Copying input data to tmp folder and convert to nii...
sct_convert -i /home/jcohen-adad/sct_testing_data/fmri/fmri_r.nii.gz -o /tmp/sct-20200214135325.614152-fmri_moco-0129le9n/fmri.nii
Get dimensions of data...
34 x 34 x 6 x 30
Data orientation: RPI
Treated as axial
Split along T dimension...
Merging volumes...
-------------------------------------------------------------------------------
Estimating motion...
-------------------------------------------------------------------------------
Input parameters:
Input file ............fmri_averaged_groups.nii
Reference file ........fmri_mean_0.nii.gz
Polynomial degree .....2
Smoothing kernel ......0
Gradient step .........1
Metric ................MeanSquares
Sampling ..............0.2
Todo ..................estimate_and_apply
Mask .................
Output mat folder .....mat_groups
Data dimensions:
34 x 34 x 6 x 6
Copy file_target to a temporary file...
sct_convert -i fmri_mean_0.nii.gz -o target.nii.gz
Register. Loop across Z (note: there is only one Z if orientation is axial
Traceback (most recent call last):
File "/home/jcohen-adad/sct/scripts/sct_fmri_moco.py", line 453, in <module>
main()
File "/home/jcohen-adad/sct/scripts/sct_fmri_moco.py", line 213, in main
fmri_moco(param)
File "/home/jcohen-adad/sct/scripts/sct_fmri_moco.py", line 350, in fmri_moco
file_mat = moco.moco(param_moco)
File "/home/jcohen-adad/sct/scripts/msct_moco.py", line 152, in moco
file_data_splitZ_splitT_moco[it], im_mask=input_mask)
File "/home/jcohen-adad/sct/scripts/msct_moco.py", line 276, in register
status, output = sct.run(cmd, verbose=0, **kw)
File "/home/jcohen-adad/sct/scripts/sct_utils.py", line 311, in run
raise RunError(output)
sct_utils.RunError: /home/jcohen-adad/sct/bin/isct_antsSliceRegularizedRegistration: /usr/lib64/libstdc++.so.6: version `GLIBCXX_3.4.15' not found (required by /home/jcohen-adad/sct/bin/isct_antsSliceRegularizedRegistration)
~~~
This issue might be related to https://github.com/neuropoly/spinalcordtoolbox/issues/711, https://github.com/neuropoly/spinalcordtoolbox/issues/158, https://github.com/neuropoly/spinalcordtoolbox/issues/1704
|
non_infrastructure
|
glibcxx not found on centos after installing sct on a station i ran into the following problem while running sct testing checking sct fmri moco sct fmri moco i fmri fmri r nii gz g x nn r spinal cord toolbox git master merge within groups merge within groups merge within groups merge within groups merge within groups merge within groups merge within groups merge within groups z z input parameters input file fmri fmri r nii gz create temporary folder tmp sct fmri moco copying input data to tmp folder and convert to nii sct convert i home jcohen adad sct testing data fmri fmri r nii gz o tmp sct fmri moco fmri nii get dimensions of data x x x data orientation rpi treated as axial split along t dimension merging volumes estimating motion input parameters input file fmri averaged groups nii reference file fmri mean nii gz polynomial degree smoothing kernel gradient step metric meansquares sampling todo estimate and apply mask output mat folder mat groups data dimensions x x x copy file target to a temporary file sct convert i fmri mean nii gz o target nii gz register loop across z note there is only one z if orientation is axial traceback most recent call last file home jcohen adad sct scripts sct fmri moco py line in main file home jcohen adad sct scripts sct fmri moco py line in main fmri moco param file home jcohen adad sct scripts sct fmri moco py line in fmri moco file mat moco moco param moco file home jcohen adad sct scripts msct moco py line in moco file data splitz splitt moco im mask input mask file home jcohen adad sct scripts msct moco py line in register status output sct run cmd verbose kw file home jcohen adad sct scripts sct utils py line in run raise runerror output sct utils runerror home jcohen adad sct bin isct antssliceregularizedregistration usr libstdc so version glibcxx not found required by home jcohen adad sct bin isct antssliceregularizedregistration this issue might be related to
| 0
|
29,782
| 24,270,355,328
|
IssuesEvent
|
2022-09-28 09:46:21
|
PedestrianDynamics/pedestrian-trajectory-analyzer
|
https://api.github.com/repos/PedestrianDynamics/pedestrian-trajectory-analyzer
|
closed
|
Automatically upload all changes in main to test PyPi
|
tooling infrastructure
|
As a user, I want to have access to new features, bug fixes before a release is made. Hence, it is desirable to upload the latest version to https://test.pypi.org or similar.
**Current issue:**
PyPi doesn't allow to overwrite a file with the same version. Thus, to upload a new version the version number needs to be increased everytime (preferably automatically in the CI). Suggestion add a "build" to the version, e.g., `1.0.0-10` for each push to the main branch.
|
1.0
|
Automatically upload all changes in main to test PyPi - As a user, I want to have access to new features, bug fixes before a release is made. Hence, it is desirable to upload the latest version to https://test.pypi.org or similar.
**Current issue:**
PyPi doesn't allow to overwrite a file with the same version. Thus, to upload a new version the version number needs to be increased everytime (preferably automatically in the CI). Suggestion add a "build" to the version, e.g., `1.0.0-10` for each push to the main branch.
|
infrastructure
|
automatically upload all changes in main to test pypi as a user i want to have access to new features bug fixes before a release is made hence it is desirable to upload the latest version to or similar current issue pypi doesn t allow to overwrite a file with the same version thus to upload a new version the version number needs to be increased everytime preferably automatically in the ci suggestion add a build to the version e g for each push to the main branch
| 1
|
203,569
| 15,885,054,557
|
IssuesEvent
|
2021-04-09 19:54:37
|
ConnorAndersonLarson/PlanetParty
|
https://api.github.com/repos/ConnorAndersonLarson/PlanetParty
|
opened
|
User Story Template
|
documentation
|
PUT THIS AS TITLE: As a user, I can <INSERT ACTION HERE>
Detailed User Story:
As a(n) [user type]
In order to [extract business value]
When I [take some action]
(and [take some other action])
Then I [observe an outcome]
(and I [observe another outcome])
Further Information:
<What steps need to be taken to accomplish this task>
<When the task is executed, what happens? What does the user see?>
<How will a developer know when the user story has been successfully accomplished?>
|
1.0
|
User Story Template - PUT THIS AS TITLE: As a user, I can <INSERT ACTION HERE>
Detailed User Story:
As a(n) [user type]
In order to [extract business value]
When I [take some action]
(and [take some other action])
Then I [observe an outcome]
(and I [observe another outcome])
Further Information:
<What steps need to be taken to accomplish this task>
<When the task is executed, what happens? What does the user see?>
<How will a developer know when the user story has been successfully accomplished?>
|
non_infrastructure
|
user story template put this as title as a user i can detailed user story as a n in order to when i and then i and i further information
| 0
|
19,862
| 6,779,650,742
|
IssuesEvent
|
2017-10-29 02:28:20
|
apache/bookkeeper
|
https://api.github.com/repos/apache/bookkeeper
|
closed
|
Enable checkstyle in a few packages
|
area/build release/4.6.0 type/task
|
This is part of #230
- feature
- processor
- shims
- stats
- streaming
- versioning
- zookeeper
|
1.0
|
Enable checkstyle in a few packages - This is part of #230
- feature
- processor
- shims
- stats
- streaming
- versioning
- zookeeper
|
non_infrastructure
|
enable checkstyle in a few packages this is part of feature processor shims stats streaming versioning zookeeper
| 0
|
5,768
| 5,946,631,397
|
IssuesEvent
|
2017-05-26 04:41:10
|
oppia/oppia
|
https://api.github.com/repos/oppia/oppia
|
closed
|
Audit and fix accessibility and low-contrast issues (for, e.g., users of screen readers).
|
loc: frontend owner: @anthkris TODO: other type: infrastructure
|
WAVE is a web accessibility tool that can be used to check how accessible a website is for screen reader users: http://wave.webaim.org/ We can use this to test Oppia pages for accessibility, and make the appropriate fixes (e.g. adding ARIA tags). The WAVE site can also be used to check for contrast issues.
|
1.0
|
Audit and fix accessibility and low-contrast issues (for, e.g., users of screen readers). - WAVE is a web accessibility tool that can be used to check how accessible a website is for screen reader users: http://wave.webaim.org/ We can use this to test Oppia pages for accessibility, and make the appropriate fixes (e.g. adding ARIA tags). The WAVE site can also be used to check for contrast issues.
|
infrastructure
|
audit and fix accessibility and low contrast issues for e g users of screen readers wave is a web accessibility tool that can be used to check how accessible a website is for screen reader users we can use this to test oppia pages for accessibility and make the appropriate fixes e g adding aria tags the wave site can also be used to check for contrast issues
| 1
|
132,094
| 10,730,608,070
|
IssuesEvent
|
2019-10-28 17:45:17
|
xspec/xspec
|
https://api.github.com/repos/xspec/xspec
|
opened
|
ci: make test\ci\install-deps.cmd work with 7zip
|
test
|
It would be nice to make `test\ci\install-deps.cmd` work with 7zip.
https://github.com/xspec/xspec/issues/665#issuecomment-546504468
> I managed to install the dependencies correctly by using a modified version of test\ci\install-deps.cmd , as I have an older Windows version, see https://github.com/heidivanparys/xspec/commit/ddb721ef908aac8c57d4619a57aecda92d7dde68
|
1.0
|
ci: make test\ci\install-deps.cmd work with 7zip - It would be nice to make `test\ci\install-deps.cmd` work with 7zip.
https://github.com/xspec/xspec/issues/665#issuecomment-546504468
> I managed to install the dependencies correctly by using a modified version of test\ci\install-deps.cmd , as I have an older Windows version, see https://github.com/heidivanparys/xspec/commit/ddb721ef908aac8c57d4619a57aecda92d7dde68
|
non_infrastructure
|
ci make test ci install deps cmd work with it would be nice to make test ci install deps cmd work with i managed to install the dependencies correctly by using a modified version of test ci install deps cmd as i have an older windows version see
| 0
|
50,606
| 26,701,464,601
|
IssuesEvent
|
2023-01-27 14:43:20
|
IMA-WorldHealth/bhima
|
https://api.github.com/repos/IMA-WorldHealth/bhima
|
opened
|
Stock Movements Registry Performance Issues
|
performance Stock
|
The stock movements registry is noticeably slower than other registries (patient, cash payment, etc). This is likely due to the complicated queries on the back-end to retrieve data. It may be important that we work on refactoring these queries going forwards to ensure that users who need to look up stock movements can do so quickly and efficiently.
|
True
|
Stock Movements Registry Performance Issues - The stock movements registry is noticeably slower than other registries (patient, cash payment, etc). This is likely due to the complicated queries on the back-end to retrieve data. It may be important that we work on refactoring these queries going forwards to ensure that users who need to look up stock movements can do so quickly and efficiently.
|
non_infrastructure
|
stock movements registry performance issues the stock movements registry is noticeably slower than other registries patient cash payment etc this is likely due to the complicated queries on the back end to retrieve data it may be important that we work on refactoring these queries going forwards to ensure that users who need to look up stock movements can do so quickly and efficiently
| 0
|
531,601
| 15,500,902,066
|
IssuesEvent
|
2021-03-11 09:53:14
|
gruntwork-io/terragrunt
|
https://api.github.com/repos/gruntwork-io/terragrunt
|
closed
|
AWS CLIv2 and AWS SSO auth fails
|
enhancement help wanted high-priority
|
We are using new aws cli v2 SSO auth feature. This works with aws cli, but terragrunt fails to get credentials.
```bash
$ aws --version
aws-cli/2.0.6 Python/3.8.2 Darwin/18.7.0 botocore/2.0.0dev10
```
`.aws/config`
```ini
[profile 123456789012-AWSAdministratorAccess]
sso_start_url = https://d-123456789.awsapps.com/start
sso_region = eu-west-1
sso_account_id = 123456789012
sso_role_name = AWSAdministratorAccess
region = eu-west-1
output = json
```
Check aws cli works
```bash
$ export AWS_PROFILE=123456789012-AWSAdministratorAccess
$ aws sso login
Attempting to automatically open the SSO authorization page in your default browser.
If the browser does not open or you wish to use a different device to authorize this request, open the following URL:
https://device.sso.eu-west-1.amazonaws.com/
Then enter the code:
REDACTED
Successully logged into Start URL: https://d-123456789.awsapps.com/start
$ aws s3 ls
2020-04-08 08:09:44 tf-states.lab.REDACTED
$ aws sts get-caller-identity
{
"UserId": "REDACTED:redacted@redacted",
"Account": "123456789012",
"Arn": "arn:aws:sts::123456789012:assumed-role/AWSReservedSSO_AWSAdministratorAccess_abcdef12345/redacted@redacted"
}
$ terragrunt plan --terragrunt-source ~/REDACTED/git/tf-modules//aws/bootstrap
[terragrunt] 2020/04/09 10:38:44 Terragrunt Version: v0.23.6
[terragrunt] 2020/04/09 10:38:44 Reading Terragrunt config file at /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap/terragrunt.hcl
[terragrunt] 2020/04/09 10:38:44 Did not find any locals block: skipping evaluation.
[terragrunt] 2020/04/09 10:38:44 Running command: /Users/REDACTED/git/cloud-config/terraform/../scripts/aws/get_admin_roles.sh
AWSReservedSSO_AWSAdministratorAccess_abcdef12345[terragrunt] 2020/04/09 10:38:46 run_cmd output: [AWSReservedSSO_AWSAdministratorAccess_abcdef12345]
[terragrunt] 2020/04/09 10:38:46 Found locals block: evaluating the expressions.
[terragrunt] 2020/04/09 10:38:46 Evaluated 5 locals (remaining 8): cloud, environment, default_empty_yaml, api_domain, aws_vars
[terragrunt] 2020/04/09 10:38:46 Evaluated 3 locals (remaining 5): terraform_module_local_path, module_settings_path, aws_root_domain
[terragrunt] 2020/04/09 10:38:46 Evaluated 2 locals (remaining 3): terraform_module_path, terraform_module_version
[terragrunt] 2020/04/09 10:38:46 Evaluated 3 locals (remaining 0): terraform_module_name, default_tags, terraform_source_default
[terragrunt] [/Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap] 2020/04/09 10:38:51 Running command: terraform --version
[terragrunt] 2020/04/09 10:38:51 Downloading Terraform configurations from file:///Users/REDACTED/git/tf-modules into /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap/.terragrunt-cache/bL_WKXycpKiko0WERZJHEmdsjyA/fnop9qC_WS7CKETpjnoJj_MM_ro
[terragrunt] 2020/04/09 10:38:51 Copying files from /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap into /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap/.terragrunt-cache/bL_WKXycpKiko0WERZJHEmdsjyA/fnop9qC_WS7CKETpjnoJj_MM_ro/aws/bootstrap
[terragrunt] 2020/04/09 10:38:51 Setting working directory to /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap/.terragrunt-cache/bL_WKXycpKiko0WERZJHEmdsjyA/fnop9qC_WS7CKETpjnoJj_MM_ro/aws/bootstrap
[terragrunt] 2020/04/09 10:38:51 The file path /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap/.terragrunt-cache/bL_WKXycpKiko0WERZJHEmdsjyA/fnop9qC_WS7CKETpjnoJj_MM_ro/aws/bootstrap/provider.tf already exists and if_exists for code generation set to "overwrite". Regenerating file.
[terragrunt] 2020/04/09 10:38:51 Generated file /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap/.terragrunt-cache/bL_WKXycpKiko0WERZJHEmdsjyA/fnop9qC_WS7CKETpjnoJj_MM_ro/aws/bootstrap/provider.tf.
[terragrunt] 2020/04/09 10:38:51 The file path /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap/.terragrunt-cache/bL_WKXycpKiko0WERZJHEmdsjyA/fnop9qC_WS7CKETpjnoJj_MM_ro/aws/bootstrap/backend.tf already exists, but was a previously generated file by terragrunt. Since if_exists for code generation is set to "overwrite_terragrunt", regenerating file.
[terragrunt] 2020/04/09 10:38:51 Generated file /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap/.terragrunt-cache/bL_WKXycpKiko0WERZJHEmdsjyA/fnop9qC_WS7CKETpjnoJj_MM_ro/aws/bootstrap/backend.tf.
[terragrunt] [/Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap] 2020/04/09 10:38:51 Initializing remote state for the s3 backend
[terragrunt] 2020/04/09 10:38:51 Error finding AWS credentials (did you set the AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY environment variables?): NoCredentialProviders: no valid providers in chain. Deprecated.
For verbose messaging see aws.Config.CredentialsChainVerboseErrors
[terragrunt] 2020/04/09 10:38:51 Unable to determine underlying exit code, so Terragrunt will exit with error code 1
```
|
1.0
|
AWS CLIv2 and AWS SSO auth fails - We are using new aws cli v2 SSO auth feature. This works with aws cli, but terragrunt fails to get credentials.
```bash
$ aws --version
aws-cli/2.0.6 Python/3.8.2 Darwin/18.7.0 botocore/2.0.0dev10
```
`.aws/config`
```ini
[profile 123456789012-AWSAdministratorAccess]
sso_start_url = https://d-123456789.awsapps.com/start
sso_region = eu-west-1
sso_account_id = 123456789012
sso_role_name = AWSAdministratorAccess
region = eu-west-1
output = json
```
Check aws cli works
```bash
$ export AWS_PROFILE=123456789012-AWSAdministratorAccess
$ aws sso login
Attempting to automatically open the SSO authorization page in your default browser.
If the browser does not open or you wish to use a different device to authorize this request, open the following URL:
https://device.sso.eu-west-1.amazonaws.com/
Then enter the code:
REDACTED
Successully logged into Start URL: https://d-123456789.awsapps.com/start
$ aws s3 ls
2020-04-08 08:09:44 tf-states.lab.REDACTED
$ aws sts get-caller-identity
{
"UserId": "REDACTED:redacted@redacted",
"Account": "123456789012",
"Arn": "arn:aws:sts::123456789012:assumed-role/AWSReservedSSO_AWSAdministratorAccess_abcdef12345/redacted@redacted"
}
$ terragrunt plan --terragrunt-source ~/REDACTED/git/tf-modules//aws/bootstrap
[terragrunt] 2020/04/09 10:38:44 Terragrunt Version: v0.23.6
[terragrunt] 2020/04/09 10:38:44 Reading Terragrunt config file at /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap/terragrunt.hcl
[terragrunt] 2020/04/09 10:38:44 Did not find any locals block: skipping evaluation.
[terragrunt] 2020/04/09 10:38:44 Running command: /Users/REDACTED/git/cloud-config/terraform/../scripts/aws/get_admin_roles.sh
AWSReservedSSO_AWSAdministratorAccess_abcdef12345[terragrunt] 2020/04/09 10:38:46 run_cmd output: [AWSReservedSSO_AWSAdministratorAccess_abcdef12345]
[terragrunt] 2020/04/09 10:38:46 Found locals block: evaluating the expressions.
[terragrunt] 2020/04/09 10:38:46 Evaluated 5 locals (remaining 8): cloud, environment, default_empty_yaml, api_domain, aws_vars
[terragrunt] 2020/04/09 10:38:46 Evaluated 3 locals (remaining 5): terraform_module_local_path, module_settings_path, aws_root_domain
[terragrunt] 2020/04/09 10:38:46 Evaluated 2 locals (remaining 3): terraform_module_path, terraform_module_version
[terragrunt] 2020/04/09 10:38:46 Evaluated 3 locals (remaining 0): terraform_module_name, default_tags, terraform_source_default
[terragrunt] [/Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap] 2020/04/09 10:38:51 Running command: terraform --version
[terragrunt] 2020/04/09 10:38:51 Downloading Terraform configurations from file:///Users/REDACTED/git/tf-modules into /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap/.terragrunt-cache/bL_WKXycpKiko0WERZJHEmdsjyA/fnop9qC_WS7CKETpjnoJj_MM_ro
[terragrunt] 2020/04/09 10:38:51 Copying files from /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap into /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap/.terragrunt-cache/bL_WKXycpKiko0WERZJHEmdsjyA/fnop9qC_WS7CKETpjnoJj_MM_ro/aws/bootstrap
[terragrunt] 2020/04/09 10:38:51 Setting working directory to /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap/.terragrunt-cache/bL_WKXycpKiko0WERZJHEmdsjyA/fnop9qC_WS7CKETpjnoJj_MM_ro/aws/bootstrap
[terragrunt] 2020/04/09 10:38:51 The file path /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap/.terragrunt-cache/bL_WKXycpKiko0WERZJHEmdsjyA/fnop9qC_WS7CKETpjnoJj_MM_ro/aws/bootstrap/provider.tf already exists and if_exists for code generation set to "overwrite". Regenerating file.
[terragrunt] 2020/04/09 10:38:51 Generated file /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap/.terragrunt-cache/bL_WKXycpKiko0WERZJHEmdsjyA/fnop9qC_WS7CKETpjnoJj_MM_ro/aws/bootstrap/provider.tf.
[terragrunt] 2020/04/09 10:38:51 The file path /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap/.terragrunt-cache/bL_WKXycpKiko0WERZJHEmdsjyA/fnop9qC_WS7CKETpjnoJj_MM_ro/aws/bootstrap/backend.tf already exists, but was a previously generated file by terragrunt. Since if_exists for code generation is set to "overwrite_terragrunt", regenerating file.
[terragrunt] 2020/04/09 10:38:51 Generated file /Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap/.terragrunt-cache/bL_WKXycpKiko0WERZJHEmdsjyA/fnop9qC_WS7CKETpjnoJj_MM_ro/aws/bootstrap/backend.tf.
[terragrunt] [/Users/REDACTED/git/cloud-config/terraform/lab/aws/bootstrap] 2020/04/09 10:38:51 Initializing remote state for the s3 backend
[terragrunt] 2020/04/09 10:38:51 Error finding AWS credentials (did you set the AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY environment variables?): NoCredentialProviders: no valid providers in chain. Deprecated.
For verbose messaging see aws.Config.CredentialsChainVerboseErrors
[terragrunt] 2020/04/09 10:38:51 Unable to determine underlying exit code, so Terragrunt will exit with error code 1
```
|
non_infrastructure
|
aws and aws sso auth fails we are using new aws cli sso auth feature this works with aws cli but terragrunt fails to get credentials bash aws version aws cli python darwin botocore aws config ini sso start url sso region eu west sso account id sso role name awsadministratoraccess region eu west output json check aws cli works bash export aws profile awsadministratoraccess aws sso login attempting to automatically open the sso authorization page in your default browser if the browser does not open or you wish to use a different device to authorize this request open the following url then enter the code redacted successully logged into start url aws ls tf states lab redacted aws sts get caller identity userid redacted redacted redacted account arn arn aws sts assumed role awsreservedsso awsadministratoraccess redacted redacted terragrunt plan terragrunt source redacted git tf modules aws bootstrap terragrunt version reading terragrunt config file at users redacted git cloud config terraform lab aws bootstrap terragrunt hcl did not find any locals block skipping evaluation running command users redacted git cloud config terraform scripts aws get admin roles sh awsreservedsso awsadministratoraccess run cmd output found locals block evaluating the expressions evaluated locals remaining cloud environment default empty yaml api domain aws vars evaluated locals remaining terraform module local path module settings path aws root domain evaluated locals remaining terraform module path terraform module version evaluated locals remaining terraform module name default tags terraform source default running command terraform version downloading terraform configurations from file users redacted git tf modules into users redacted git cloud config terraform lab aws bootstrap terragrunt cache bl mm ro copying files from users redacted git cloud config terraform lab aws bootstrap into users redacted git cloud config terraform lab aws bootstrap terragrunt cache bl mm ro aws bootstrap setting working directory to users redacted git cloud config terraform lab aws bootstrap terragrunt cache bl mm ro aws bootstrap the file path users redacted git cloud config terraform lab aws bootstrap terragrunt cache bl mm ro aws bootstrap provider tf already exists and if exists for code generation set to overwrite regenerating file generated file users redacted git cloud config terraform lab aws bootstrap terragrunt cache bl mm ro aws bootstrap provider tf the file path users redacted git cloud config terraform lab aws bootstrap terragrunt cache bl mm ro aws bootstrap backend tf already exists but was a previously generated file by terragrunt since if exists for code generation is set to overwrite terragrunt regenerating file generated file users redacted git cloud config terraform lab aws bootstrap terragrunt cache bl mm ro aws bootstrap backend tf initializing remote state for the backend error finding aws credentials did you set the aws access key id and aws secret access key environment variables nocredentialproviders no valid providers in chain deprecated for verbose messaging see aws config credentialschainverboseerrors unable to determine underlying exit code so terragrunt will exit with error code
| 0
|
693,359
| 23,773,370,408
|
IssuesEvent
|
2022-09-01 18:24:54
|
LDSSA/portal
|
https://api.github.com/repos/LDSSA/portal
|
closed
|
upgrade python version for setup instructions and SLU 00
|
priority:high admissions
|
1. figure out the most appropriate python version going forward
1. should we test the new python version for
1. test and change the setup instructions for that python version
|
1.0
|
upgrade python version for setup instructions and SLU 00 - 1. figure out the most appropriate python version going forward
1. should we test the new python version for
1. test and change the setup instructions for that python version
|
non_infrastructure
|
upgrade python version for setup instructions and slu figure out the most appropriate python version going forward should we test the new python version for test and change the setup instructions for that python version
| 0
|
3,980
| 4,746,054,794
|
IssuesEvent
|
2016-10-21 09:40:31
|
camptocamp/ngeo
|
https://api.github.com/repos/camptocamp/ngeo
|
opened
|
Removes permalink required dependencies
|
Infrastructure Ready
|
On this file:
https://github.com/camptocamp/ngeo/blob/master/contribs/gmf/src/services/permalink.js
We should:
* Replace nearly all the ingection by the ingector.
* Get the "elements" from the ingector.
* Don't crash if an element isn't present.
|
1.0
|
Removes permalink required dependencies - On this file:
https://github.com/camptocamp/ngeo/blob/master/contribs/gmf/src/services/permalink.js
We should:
* Replace nearly all the ingection by the ingector.
* Get the "elements" from the ingector.
* Don't crash if an element isn't present.
|
infrastructure
|
removes permalink required dependencies on this file we should replace nearly all the ingection by the ingector get the elements from the ingector don t crash if an element isn t present
| 1
|
35,251
| 30,865,637,810
|
IssuesEvent
|
2023-08-03 07:51:16
|
input-output-hk/mithril
|
https://api.github.com/repos/input-output-hk/mithril
|
closed
|
Ingest `mainnet` metrics/logs in Grafana
|
D-medium task dev :muscle: mainnet-infrastructure :rocket: monitoring :eyes:
|
## Issue
In order to have better monitoring and logging, we want to ingest metrics and logs for `release-mainnet`.
## To do
- [x] Add Grafana metrics ingestion with prometheus configuration in infra
- [x] Add Grafana logs ingestion with promtail configuration in infra
- [x] Update CI/CD workflows
## Parent issue
#1091
|
1.0
|
Ingest `mainnet` metrics/logs in Grafana - ## Issue
In order to have better monitoring and logging, we want to ingest metrics and logs for `release-mainnet`.
## To do
- [x] Add Grafana metrics ingestion with prometheus configuration in infra
- [x] Add Grafana logs ingestion with promtail configuration in infra
- [x] Update CI/CD workflows
## Parent issue
#1091
|
infrastructure
|
ingest mainnet metrics logs in grafana issue in order to have better monitoring and logging we want to ingest metrics and logs for release mainnet to do add grafana metrics ingestion with prometheus configuration in infra add grafana logs ingestion with promtail configuration in infra update ci cd workflows parent issue
| 1
|
14,361
| 3,392,362,795
|
IssuesEvent
|
2015-11-30 19:17:11
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
Test failure in CI build 9685
|
test-failure
|
The following test appears to have failed:
[#9685](https://circleci.com/gh/cockroachdb/cockroach/9685):
```
--- PASS: TestStorePoolGetStoreList (0.01s)
=== RUN TestStorePoolGetStoreDetails
--- PASS: TestStorePoolGetStoreDetails (0.01s)
=== RUN TestStorePoolFindDeadReplicas
W1130 19:16:15.683799 972 storage/store_pool.go:58 store 1 on node 1 is now considered offline
--- FAIL: TestStorePoolFindDeadReplicas (0.01s)
store_pool_test.go:357: expected no dead replicas initially, found 1 ([{1 1 1}])
=== RUN TestStoreInitAndBootstrap
I1130 19:16:15.687243 972 multiraft/multiraft.go:557 node 1 starting
I1130 19:16:15.687730 972 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestStoreInitAndBootstrap (0.01s)
=== RUN TestBootstrapOfNonEmptyStore
I1130 19:16:15.697278 972 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestBootstrapOfNonEmptyStore (0.00s)
=== RUN TestStoreAddRemoveRanges
I1130 19:16:15.705894 972 multiraft/multiraft.go:557 node 1 starting
--
I1130 19:16:44.697772 972 storage/replica.go:1369 gossiping system config from store 1, range 2
I1130 19:16:44.711028 972 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestStoreSplitReadRace (0.05s)
=== RUN Example_rebalancing
--- PASS: Example_rebalancing (0.48s)
FAIL
FAIL github.com/cockroachdb/cockroach/storage 31.859s
=== RUN TestBatchBasics
I1130 19:16:13.452760 977 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestBatchBasics (0.01s)
=== RUN TestBatchGet
I1130 19:16:13.458178 977 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestBatchGet (0.01s)
=== RUN TestBatchMerge
I1130 19:16:13.465316 977 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestBatchMerge (0.00s)
=== RUN TestBatchProto
```
Please assign, take a look and update the issue accordingly.
|
1.0
|
Test failure in CI build 9685 - The following test appears to have failed:
[#9685](https://circleci.com/gh/cockroachdb/cockroach/9685):
```
--- PASS: TestStorePoolGetStoreList (0.01s)
=== RUN TestStorePoolGetStoreDetails
--- PASS: TestStorePoolGetStoreDetails (0.01s)
=== RUN TestStorePoolFindDeadReplicas
W1130 19:16:15.683799 972 storage/store_pool.go:58 store 1 on node 1 is now considered offline
--- FAIL: TestStorePoolFindDeadReplicas (0.01s)
store_pool_test.go:357: expected no dead replicas initially, found 1 ([{1 1 1}])
=== RUN TestStoreInitAndBootstrap
I1130 19:16:15.687243 972 multiraft/multiraft.go:557 node 1 starting
I1130 19:16:15.687730 972 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestStoreInitAndBootstrap (0.01s)
=== RUN TestBootstrapOfNonEmptyStore
I1130 19:16:15.697278 972 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestBootstrapOfNonEmptyStore (0.00s)
=== RUN TestStoreAddRemoveRanges
I1130 19:16:15.705894 972 multiraft/multiraft.go:557 node 1 starting
--
I1130 19:16:44.697772 972 storage/replica.go:1369 gossiping system config from store 1, range 2
I1130 19:16:44.711028 972 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestStoreSplitReadRace (0.05s)
=== RUN Example_rebalancing
--- PASS: Example_rebalancing (0.48s)
FAIL
FAIL github.com/cockroachdb/cockroach/storage 31.859s
=== RUN TestBatchBasics
I1130 19:16:13.452760 977 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestBatchBasics (0.01s)
=== RUN TestBatchGet
I1130 19:16:13.458178 977 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestBatchGet (0.01s)
=== RUN TestBatchMerge
I1130 19:16:13.465316 977 storage/engine/rocksdb.go:132 closing in-memory rocksdb instance
--- PASS: TestBatchMerge (0.00s)
=== RUN TestBatchProto
```
Please assign, take a look and update the issue accordingly.
|
non_infrastructure
|
test failure in ci build the following test appears to have failed pass teststorepoolgetstorelist run teststorepoolgetstoredetails pass teststorepoolgetstoredetails run teststorepoolfinddeadreplicas storage store pool go store on node is now considered offline fail teststorepoolfinddeadreplicas store pool test go expected no dead replicas initially found run teststoreinitandbootstrap multiraft multiraft go node starting storage engine rocksdb go closing in memory rocksdb instance pass teststoreinitandbootstrap run testbootstrapofnonemptystore storage engine rocksdb go closing in memory rocksdb instance pass testbootstrapofnonemptystore run teststoreaddremoveranges multiraft multiraft go node starting storage replica go gossiping system config from store range storage engine rocksdb go closing in memory rocksdb instance pass teststoresplitreadrace run example rebalancing pass example rebalancing fail fail github com cockroachdb cockroach storage run testbatchbasics storage engine rocksdb go closing in memory rocksdb instance pass testbatchbasics run testbatchget storage engine rocksdb go closing in memory rocksdb instance pass testbatchget run testbatchmerge storage engine rocksdb go closing in memory rocksdb instance pass testbatchmerge run testbatchproto please assign take a look and update the issue accordingly
| 0
|
33,131
| 15,797,448,718
|
IssuesEvent
|
2021-04-02 16:44:42
|
Azure/azure-sdk-for-python
|
https://api.github.com/repos/Azure/azure-sdk-for-python
|
closed
|
Document perf protocol
|
Client Performance
|
As a service team or a SDK writer, how do I write a test that gets pick-up as "perf" by CI/live and ADO globally.
Blobs:
https://github.com/Azure/azure-sdk-for-python/pull/15833
ServiceBus:
https://github.com/Azure/azure-sdk-for-python/pull/16066
Either in OneNote of the Language Teams channel and/or https://github.com/Azure/azure-sdk-for-python/tree/master/doc/dev
|
True
|
Document perf protocol - As a service team or a SDK writer, how do I write a test that gets pick-up as "perf" by CI/live and ADO globally.
Blobs:
https://github.com/Azure/azure-sdk-for-python/pull/15833
ServiceBus:
https://github.com/Azure/azure-sdk-for-python/pull/16066
Either in OneNote of the Language Teams channel and/or https://github.com/Azure/azure-sdk-for-python/tree/master/doc/dev
|
non_infrastructure
|
document perf protocol as a service team or a sdk writer how do i write a test that gets pick up as perf by ci live and ado globally blobs servicebus either in onenote of the language teams channel and or
| 0
|
136,827
| 11,088,973,728
|
IssuesEvent
|
2019-12-14 15:12:52
|
microsoft/azure-pipelines-tasks
|
https://api.github.com/repos/microsoft/azure-pipelines-tasks
|
closed
|
VsTestv2 does not fail if it cannot find tests
|
Area: Test enhancement
|
VsTestv2 emits only one warning when no test assemblies were found:
```
##[warning]No results found to publish.
```
This leads to a green build/releease and therefore it is not easy to spot that tests are not being run any more by mistake (e.g. due to a rename of assemblies where the build/release definition was not updated accordingly).
The expectation of the task should be that there are test assemblies with matching tests to test for by default and otherwise fail.
N.B.: I am using VSTS, not TFS
|
1.0
|
VsTestv2 does not fail if it cannot find tests - VsTestv2 emits only one warning when no test assemblies were found:
```
##[warning]No results found to publish.
```
This leads to a green build/releease and therefore it is not easy to spot that tests are not being run any more by mistake (e.g. due to a rename of assemblies where the build/release definition was not updated accordingly).
The expectation of the task should be that there are test assemblies with matching tests to test for by default and otherwise fail.
N.B.: I am using VSTS, not TFS
|
non_infrastructure
|
does not fail if it cannot find tests emits only one warning when no test assemblies were found no results found to publish this leads to a green build releease and therefore it is not easy to spot that tests are not being run any more by mistake e g due to a rename of assemblies where the build release definition was not updated accordingly the expectation of the task should be that there are test assemblies with matching tests to test for by default and otherwise fail n b i am using vsts not tfs
| 0
|
7,876
| 7,120,817,789
|
IssuesEvent
|
2018-01-19 03:33:32
|
pytest-dev/pytest
|
https://api.github.com/repos/pytest-dev/pytest
|
closed
|
get an own ci state for linting so linting doesnt shadow into normal tests
|
status: needs information type: infrastructure
|
linting issues give us false negatives for the normal ci tests
lets have a own github status for linting issues
|
1.0
|
get an own ci state for linting so linting doesnt shadow into normal tests - linting issues give us false negatives for the normal ci tests
lets have a own github status for linting issues
|
infrastructure
|
get an own ci state for linting so linting doesnt shadow into normal tests linting issues give us false negatives for the normal ci tests lets have a own github status for linting issues
| 1
|
16,852
| 12,152,124,519
|
IssuesEvent
|
2020-04-24 21:27:44
|
BCDevOps/developer-experience
|
https://api.github.com/repos/BCDevOps/developer-experience
|
opened
|
Sysdig dashboard issues
|
Infrastructure action-required
|
https://trello.com/c/YTtgpHQX/121-sysdig-dashboard-issues
Ticket logged about dashboard timeout, metrics flapping, and PVC count dashbaord
Sysdig: 8051 - Metrics flapping or unavailable
|
1.0
|
Sysdig dashboard issues - https://trello.com/c/YTtgpHQX/121-sysdig-dashboard-issues
Ticket logged about dashboard timeout, metrics flapping, and PVC count dashbaord
Sysdig: 8051 - Metrics flapping or unavailable
|
infrastructure
|
sysdig dashboard issues ticket logged about dashboard timeout metrics flapping and pvc count dashbaord sysdig metrics flapping or unavailable
| 1
|
146,838
| 23,130,265,208
|
IssuesEvent
|
2022-07-28 09:43:58
|
HYF-Class19/agile-devlopment-group-3
|
https://api.github.com/repos/HYF-Class19/agile-devlopment-group-3
|
closed
|
draw the site design
|
planning Design
|
- [x] draw wireframe using https://excalidraw.com/
- [x] edit design.md
- [x] add source link
- [x] add png and SVG screenshots
|
1.0
|
draw the site design - - [x] draw wireframe using https://excalidraw.com/
- [x] edit design.md
- [x] add source link
- [x] add png and SVG screenshots
|
non_infrastructure
|
draw the site design draw wireframe using edit design md add source link add png and svg screenshots
| 0
|
761,257
| 26,673,695,324
|
IssuesEvent
|
2023-01-26 12:36:20
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
mgmt: mcumgr: bluetooth transport: Inability to use refactored transport as a library in some circumstances
|
bug priority: low area: Bluetooth area: mcumgr
|
**Describe the bug**
The old MCUmgr code did not automatically register anything for Bluetooth, it was expected that user applications would call the functions themselves. With the recent refactor, this has changed and the bluetooth transport can now automatically register itself - this works fine in the samples but causes issues when it is used as a library, if an application also tries to enable bluetooth, one of these registrations is going to fail. This can be worked around by disabling MCUmgr from automatically registering itself at bootup, and the user code needing to call the setup function manually.
It would be a good improvement if this registration code could be setup automatically when the bluetooth stack has been enabled, the other libraries could similarly have services added, none of which would need to enable the bluetooth stack themselves.
**Expected behavior**
Bluetooth and sample applications to work with automatic registration
**Impact**
Showstopper
|
1.0
|
mgmt: mcumgr: bluetooth transport: Inability to use refactored transport as a library in some circumstances - **Describe the bug**
The old MCUmgr code did not automatically register anything for Bluetooth, it was expected that user applications would call the functions themselves. With the recent refactor, this has changed and the bluetooth transport can now automatically register itself - this works fine in the samples but causes issues when it is used as a library, if an application also tries to enable bluetooth, one of these registrations is going to fail. This can be worked around by disabling MCUmgr from automatically registering itself at bootup, and the user code needing to call the setup function manually.
It would be a good improvement if this registration code could be setup automatically when the bluetooth stack has been enabled, the other libraries could similarly have services added, none of which would need to enable the bluetooth stack themselves.
**Expected behavior**
Bluetooth and sample applications to work with automatic registration
**Impact**
Showstopper
|
non_infrastructure
|
mgmt mcumgr bluetooth transport inability to use refactored transport as a library in some circumstances describe the bug the old mcumgr code did not automatically register anything for bluetooth it was expected that user applications would call the functions themselves with the recent refactor this has changed and the bluetooth transport can now automatically register itself this works fine in the samples but causes issues when it is used as a library if an application also tries to enable bluetooth one of these registrations is going to fail this can be worked around by disabling mcumgr from automatically registering itself at bootup and the user code needing to call the setup function manually it would be a good improvement if this registration code could be setup automatically when the bluetooth stack has been enabled the other libraries could similarly have services added none of which would need to enable the bluetooth stack themselves expected behavior bluetooth and sample applications to work with automatic registration impact showstopper
| 0
|
96,220
| 27,782,704,973
|
IssuesEvent
|
2023-03-16 22:45:12
|
webstudio-is/webstudio-builder
|
https://api.github.com/repos/webstudio-is/webstudio-builder
|
opened
|
Add Panel Design Update
|
type:enhancement area:builder prio:1
|
Redesign the Add Panel according to the design spec in the [Webstudio Design Docs Figma
](https://www.figma.com/file/xCBegXEWxROLqA1Y31z2Xo/%F0%9F%93%96-Webstudio-Design-Docs?node-id=3217%3A84631&t=xiLMFeu2JynC19VU-1)
Current Add Panel:

Redesigned Add Panel:

|
1.0
|
Add Panel Design Update - Redesign the Add Panel according to the design spec in the [Webstudio Design Docs Figma
](https://www.figma.com/file/xCBegXEWxROLqA1Y31z2Xo/%F0%9F%93%96-Webstudio-Design-Docs?node-id=3217%3A84631&t=xiLMFeu2JynC19VU-1)
Current Add Panel:

Redesigned Add Panel:

|
non_infrastructure
|
add panel design update redesign the add panel according to the design spec in the webstudio design docs figma current add panel redesigned add panel
| 0
|
4,036
| 4,779,409,153
|
IssuesEvent
|
2016-10-27 22:23:21
|
OpenSCAP/scap-security-guide
|
https://api.github.com/repos/OpenSCAP/scap-security-guide
|
closed
|
scap-security-guide-nist-testsuite Jenkins CI test job fails starting from 2016-07-11
|
BLOCKER bugfix Infrastructure
|
See https://jenkins.open-scap.org/job/scap-security-guide-nist-testsuite/183/
The issue isn't in CI job itself, since there are two errors reported with the testsuite:
https://jenkins.open-scap.org/job/scap-security-guide-nist-testsuite/183/consoleFull
(both for RHEL-6 and RHEL-7 content)
I will inspect the failure. Stay tuned.
|
1.0
|
scap-security-guide-nist-testsuite Jenkins CI test job fails starting from 2016-07-11 - See https://jenkins.open-scap.org/job/scap-security-guide-nist-testsuite/183/
The issue isn't in CI job itself, since there are two errors reported with the testsuite:
https://jenkins.open-scap.org/job/scap-security-guide-nist-testsuite/183/consoleFull
(both for RHEL-6 and RHEL-7 content)
I will inspect the failure. Stay tuned.
|
infrastructure
|
scap security guide nist testsuite jenkins ci test job fails starting from see the issue isn t in ci job itself since there are two errors reported with the testsuite both for rhel and rhel content i will inspect the failure stay tuned
| 1
|
28,609
| 23,388,150,269
|
IssuesEvent
|
2022-08-11 15:20:30
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
closed
|
Offboarding Metrics Automation: Update Offboarding Template to Include Data Formatting
|
infrastructure platform-security
|
## Description
The Platform Security team is attempting to automate the collection of metrics from offboarding requests. Currently the Offboarding issue template does not contain any standardized way to format data fields. To assist in scripting the following changes need to be made:
* Only one individual per ticket.
* Date formatting MM/DD/YYY per: https://design.va.gov/content-style-guide/dates-and-numbers#dates-and-years
* Multiple emails listed in order of relevance to VA business, VA, Company, Personal, etc....
* Github username formatted with @ symbol.
* Additional access field formatted with access titles only. Additional notes may be added on second line after the list.
## Acceptance Criteria
- [x] Pull request containing changes merged.
|
1.0
|
Offboarding Metrics Automation: Update Offboarding Template to Include Data Formatting - ## Description
The Platform Security team is attempting to automate the collection of metrics from offboarding requests. Currently the Offboarding issue template does not contain any standardized way to format data fields. To assist in scripting the following changes need to be made:
* Only one individual per ticket.
* Date formatting MM/DD/YYY per: https://design.va.gov/content-style-guide/dates-and-numbers#dates-and-years
* Multiple emails listed in order of relevance to VA business, VA, Company, Personal, etc....
* Github username formatted with @ symbol.
* Additional access field formatted with access titles only. Additional notes may be added on second line after the list.
## Acceptance Criteria
- [x] Pull request containing changes merged.
|
infrastructure
|
offboarding metrics automation update offboarding template to include data formatting description the platform security team is attempting to automate the collection of metrics from offboarding requests currently the offboarding issue template does not contain any standardized way to format data fields to assist in scripting the following changes need to be made only one individual per ticket date formatting mm dd yyy per multiple emails listed in order of relevance to va business va company personal etc github username formatted with symbol additional access field formatted with access titles only additional notes may be added on second line after the list acceptance criteria pull request containing changes merged
| 1
|
16,222
| 11,871,961,733
|
IssuesEvent
|
2020-03-26 15:08:04
|
grpc/grpc.io
|
https://api.github.com/repos/grpc/grpc.io
|
closed
|
[infra] Support more than just the "bug" issue template
|
infrastructure
|
Not all issues are bugs, so we should support more than the "bug" template, if only an "other" template.
|
1.0
|
[infra] Support more than just the "bug" issue template - Not all issues are bugs, so we should support more than the "bug" template, if only an "other" template.
|
infrastructure
|
support more than just the bug issue template not all issues are bugs so we should support more than the bug template if only an other template
| 1
|
233,057
| 17,841,263,373
|
IssuesEvent
|
2021-09-03 10:23:06
|
keanuwilliams/inspirational-homepage
|
https://api.github.com/repos/keanuwilliams/inspirational-homepage
|
closed
|
Move History into it's own page
|
documentation
|
- Update History in README for v1.1 updates
- Add a page (on the website or as a github pages) to display all the history of Inspirational Homepage
|
1.0
|
Move History into it's own page - - Update History in README for v1.1 updates
- Add a page (on the website or as a github pages) to display all the history of Inspirational Homepage
|
non_infrastructure
|
move history into it s own page update history in readme for updates add a page on the website or as a github pages to display all the history of inspirational homepage
| 0
|
421,682
| 28,352,905,846
|
IssuesEvent
|
2023-04-12 04:37:10
|
open-telemetry/opentelemetry-php
|
https://api.github.com/repos/open-telemetry/opentelemetry-php
|
closed
|
[meta] Documentation for OpenTelemetry PHP
|
documentation examples pinned
|
Meta issue for documentations regarding OpenTelemetry PHP.
- [ ] Add READMEs to individual packages
- [x] https://github.com/open-telemetry/opentelemetry-php/issues/770
- [x] https://github.com/open-telemetry/opentelemetry-php/issues/772
- [x] https://github.com/open-telemetry/opentelemetry-php/issues/771
- [x] https://github.com/open-telemetry/opentelemetry-php/issues/773
- [x] https://github.com/open-telemetry/opentelemetry-php/issues/774
|
1.0
|
[meta] Documentation for OpenTelemetry PHP - Meta issue for documentations regarding OpenTelemetry PHP.
- [ ] Add READMEs to individual packages
- [x] https://github.com/open-telemetry/opentelemetry-php/issues/770
- [x] https://github.com/open-telemetry/opentelemetry-php/issues/772
- [x] https://github.com/open-telemetry/opentelemetry-php/issues/771
- [x] https://github.com/open-telemetry/opentelemetry-php/issues/773
- [x] https://github.com/open-telemetry/opentelemetry-php/issues/774
|
non_infrastructure
|
documentation for opentelemetry php meta issue for documentations regarding opentelemetry php add readmes to individual packages
| 0
|
13,327
| 15,788,296,398
|
IssuesEvent
|
2021-04-01 20:34:06
|
apache/trafficcontrol
|
https://api.github.com/repos/apache/trafficcontrol
|
opened
|
`release.pl` exits success when it fails to update the VERSION file
|
bug process
|
## I'm submitting a ...
- bug report
## Traffic Control components affected ...
None
## Current behavior:
If the `release.pl` script fails to edit the VERSION file, it will plow ahead anyway. Or, presumably, if any step fails it will continue to do things, possibly incorrectly.
## Expected behavior:
If part of the process of creating a release (or candidate) fails, the script should exit immediately, printing what went wrong (ideally to stderr) and use a non-zero exit code.
## Minimal reproduction of the problem with instructions:
I think if you try to create an existing release it won't have any negative side effects. So checkout e.g. the RELEASE-5.0.0-RC1 tag and do `./misc/release.pl --gpg-key={{some key}} --release-no=RELEASE-5.0.0-RC1 cut`.
## Anything else:
You can see it happening in the script's output:
```
Failed to run:git commit -m 'RELEASE: Syncing VERSION file' VERSION
Updating 'VERSION' file
Everything up-to-date
Signing new tag based upon your gpg key
```
|
1.0
|
`release.pl` exits success when it fails to update the VERSION file - ## I'm submitting a ...
- bug report
## Traffic Control components affected ...
None
## Current behavior:
If the `release.pl` script fails to edit the VERSION file, it will plow ahead anyway. Or, presumably, if any step fails it will continue to do things, possibly incorrectly.
## Expected behavior:
If part of the process of creating a release (or candidate) fails, the script should exit immediately, printing what went wrong (ideally to stderr) and use a non-zero exit code.
## Minimal reproduction of the problem with instructions:
I think if you try to create an existing release it won't have any negative side effects. So checkout e.g. the RELEASE-5.0.0-RC1 tag and do `./misc/release.pl --gpg-key={{some key}} --release-no=RELEASE-5.0.0-RC1 cut`.
## Anything else:
You can see it happening in the script's output:
```
Failed to run:git commit -m 'RELEASE: Syncing VERSION file' VERSION
Updating 'VERSION' file
Everything up-to-date
Signing new tag based upon your gpg key
```
|
non_infrastructure
|
release pl exits success when it fails to update the version file i m submitting a bug report traffic control components affected none current behavior if the release pl script fails to edit the version file it will plow ahead anyway or presumably if any step fails it will continue to do things possibly incorrectly expected behavior if part of the process of creating a release or candidate fails the script should exit immediately printing what went wrong ideally to stderr and use a non zero exit code minimal reproduction of the problem with instructions i think if you try to create an existing release it won t have any negative side effects so checkout e g the release tag and do misc release pl gpg key some key release no release cut anything else you can see it happening in the script s output failed to run git commit m release syncing version file version updating version file everything up to date signing new tag based upon your gpg key
| 0
|
28,423
| 23,244,492,117
|
IssuesEvent
|
2022-08-03 18:42:10
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
opened
|
[Application Hosting and Deployment] Validate Helm template MVP (by deploying a service with it)
|
operations devops needs-grooming infrastructure eks
|
## Description
The Platform Infrastructure team has been working to develop an MVP of a standardized template that can be used for deploying applications to EKS using ArgoCD. The template should be validated by using it to deploy a service.
## Background/context
- See parent epic
## Technical notes
_Notes around work that is happening, if applicable (optional, please delete if unused)_
---
## Tasks
- [ ] Decide on a service to deploy
- [ ] Deploy the service using the standardized Helm template
## Acceptance Criteria
- [ ] _What are the specific things that must be true in order for this to be accepted?_
---
### Reminders
- [ ] Please attach your team label and any other appropriate label(s) (operations, devops, and needs-grooming will automatically be applied as part of the template)
- [ ] Please connect to an epic (this will typically be done by the Platform Operations PM or TL)
|
1.0
|
[Application Hosting and Deployment] Validate Helm template MVP (by deploying a service with it) - ## Description
The Platform Infrastructure team has been working to develop an MVP of a standardized template that can be used for deploying applications to EKS using ArgoCD. The template should be validated by using it to deploy a service.
## Background/context
- See parent epic
## Technical notes
_Notes around work that is happening, if applicable (optional, please delete if unused)_
---
## Tasks
- [ ] Decide on a service to deploy
- [ ] Deploy the service using the standardized Helm template
## Acceptance Criteria
- [ ] _What are the specific things that must be true in order for this to be accepted?_
---
### Reminders
- [ ] Please attach your team label and any other appropriate label(s) (operations, devops, and needs-grooming will automatically be applied as part of the template)
- [ ] Please connect to an epic (this will typically be done by the Platform Operations PM or TL)
|
infrastructure
|
validate helm template mvp by deploying a service with it description the platform infrastructure team has been working to develop an mvp of a standardized template that can be used for deploying applications to eks using argocd the template should be validated by using it to deploy a service background context see parent epic technical notes notes around work that is happening if applicable optional please delete if unused tasks decide on a service to deploy deploy the service using the standardized helm template acceptance criteria what are the specific things that must be true in order for this to be accepted reminders please attach your team label and any other appropriate label s operations devops and needs grooming will automatically be applied as part of the template please connect to an epic this will typically be done by the platform operations pm or tl
| 1
|
22,300
| 15,090,810,664
|
IssuesEvent
|
2021-02-06 12:48:33
|
celeritas-project/celeritas
|
https://api.github.com/repos/celeritas-project/celeritas
|
closed
|
Add CUDA CI
|
infrastructure
|
- [x] Create a docker image with CUDA and celeritas spack toolchain
- [x] Create Jenkins file fr launching
- [x] Integrate into https://cloud.cees.ornl.gov/jenkins-ci/
|
1.0
|
Add CUDA CI - - [x] Create a docker image with CUDA and celeritas spack toolchain
- [x] Create Jenkins file fr launching
- [x] Integrate into https://cloud.cees.ornl.gov/jenkins-ci/
|
infrastructure
|
add cuda ci create a docker image with cuda and celeritas spack toolchain create jenkins file fr launching integrate into
| 1
|
8,415
| 7,397,718,339
|
IssuesEvent
|
2018-03-19 01:04:48
|
APSIMInitiative/ApsimX
|
https://api.github.com/repos/APSIMInitiative/ApsimX
|
closed
|
The number of graphs keeping increasing under the linear interpolation function
|
bug interface/infrastructure
|
The number of graphs keeping increasing under the linear interpolation function. Not sure how it happens, presumably increase when running.

|
1.0
|
The number of graphs keeping increasing under the linear interpolation function - The number of graphs keeping increasing under the linear interpolation function. Not sure how it happens, presumably increase when running.

|
infrastructure
|
the number of graphs keeping increasing under the linear interpolation function the number of graphs keeping increasing under the linear interpolation function not sure how it happens presumably increase when running png
| 1
|
13,588
| 10,338,630,403
|
IssuesEvent
|
2019-09-03 17:21:36
|
desktop/desktop
|
https://api.github.com/repos/desktop/desktop
|
closed
|
The autosetuprebase and conflictStyle settings in ~/.gitconfig can cause tests to fail
|
bug infrastructure priority-3
|
I was having trouble getting the test suite to pass on the main branch. This was because I had these lines in my ~/.gitconfig.
```
# If these lines exists a dozen tests will fail!
[branch]
autosetuprebase = always
```
It looks like it is problematic to rely on the user's `.gitconfig` file. A possible solution would be to unset the `HOME` env var in the test scripts, but I'm not sure how well that would work on windows.
-----
## Proposed Solution
@kuychaco @probablycorey @billygriffin and @outofambit discussed this further and agreed upon the following plan for now:
* `conflictStyle = diff3` will be resolved once #8049 is finished (and its currently in progress), so we'll let that go for now
* `autosetuprebase = always` is a value we should set to `false` in the setup of tests that will fail without it. (see #6972 for an example of this approach)
there's still the open question of using a standard config (or a series of them) for our tests, but i think we're punting on that for now!
|
1.0
|
The autosetuprebase and conflictStyle settings in ~/.gitconfig can cause tests to fail - I was having trouble getting the test suite to pass on the main branch. This was because I had these lines in my ~/.gitconfig.
```
# If these lines exists a dozen tests will fail!
[branch]
autosetuprebase = always
```
It looks like it is problematic to rely on the user's `.gitconfig` file. A possible solution would be to unset the `HOME` env var in the test scripts, but I'm not sure how well that would work on windows.
-----
## Proposed Solution
@kuychaco @probablycorey @billygriffin and @outofambit discussed this further and agreed upon the following plan for now:
* `conflictStyle = diff3` will be resolved once #8049 is finished (and its currently in progress), so we'll let that go for now
* `autosetuprebase = always` is a value we should set to `false` in the setup of tests that will fail without it. (see #6972 for an example of this approach)
there's still the open question of using a standard config (or a series of them) for our tests, but i think we're punting on that for now!
|
infrastructure
|
the autosetuprebase and conflictstyle settings in gitconfig can cause tests to fail i was having trouble getting the test suite to pass on the main branch this was because i had these lines in my gitconfig if these lines exists a dozen tests will fail autosetuprebase always it looks like it is problematic to rely on the user s gitconfig file a possible solution would be to unset the home env var in the test scripts but i m not sure how well that would work on windows proposed solution kuychaco probablycorey billygriffin and outofambit discussed this further and agreed upon the following plan for now conflictstyle will be resolved once is finished and its currently in progress so we ll let that go for now autosetuprebase always is a value we should set to false in the setup of tests that will fail without it see for an example of this approach there s still the open question of using a standard config or a series of them for our tests but i think we re punting on that for now
| 1
|
22,530
| 4,817,205,110
|
IssuesEvent
|
2016-11-04 12:56:59
|
zeromq/fszmq
|
https://api.github.com/repos/zeromq/fszmq
|
opened
|
Further Documentation
|
Documentation
|
Based on some community feedback, there are some "gaps" in the documentation which we'd like to address. This is a (potentially) open-ended issue to track such things. So far, we have identified a need for:
- [ ] using the new Socket Options API
- [ ] blocking send/recv (with TimeoutException)
- [ ] handling EINTR in non-blocking code
- [ ] clean application shutdown
|
1.0
|
Further Documentation - Based on some community feedback, there are some "gaps" in the documentation which we'd like to address. This is a (potentially) open-ended issue to track such things. So far, we have identified a need for:
- [ ] using the new Socket Options API
- [ ] blocking send/recv (with TimeoutException)
- [ ] handling EINTR in non-blocking code
- [ ] clean application shutdown
|
non_infrastructure
|
further documentation based on some community feedback there are some gaps in the documentation which we d like to address this is a potentially open ended issue to track such things so far we have identified a need for using the new socket options api blocking send recv with timeoutexception handling eintr in non blocking code clean application shutdown
| 0
|
16,613
| 12,066,279,006
|
IssuesEvent
|
2020-04-16 11:28:22
|
carlomigueldy/dynuxtify
|
https://api.github.com/repos/carlomigueldy/dynuxtify
|
opened
|
add Composition API in code base
|
feature infrastructure
|
so my components will be highly reusable
also have methods that can be used globally
|
1.0
|
add Composition API in code base - so my components will be highly reusable
also have methods that can be used globally
|
infrastructure
|
add composition api in code base so my components will be highly reusable also have methods that can be used globally
| 1
|
14,996
| 8,728,714,184
|
IssuesEvent
|
2018-12-10 18:09:20
|
phetsims/energy-skate-park
|
https://api.github.com/repos/phetsims/energy-skate-park
|
closed
|
Performance degradation from bar graph changes
|
type:performance
|
From #39 and #25, I have noticed that changes to the bar graph have caused the sim to slow down quite a bit on my Windows 10 Laptop in Chrome.
If I remove this line from BarGraphForegroundNode, the sim runs much faster again
```js
bar.children = [ solidBar, imageNode ];
```
|
True
|
Performance degradation from bar graph changes - From #39 and #25, I have noticed that changes to the bar graph have caused the sim to slow down quite a bit on my Windows 10 Laptop in Chrome.
If I remove this line from BarGraphForegroundNode, the sim runs much faster again
```js
bar.children = [ solidBar, imageNode ];
```
|
non_infrastructure
|
performance degradation from bar graph changes from and i have noticed that changes to the bar graph have caused the sim to slow down quite a bit on my windows laptop in chrome if i remove this line from bargraphforegroundnode the sim runs much faster again js bar children
| 0
|
24,179
| 16,989,396,130
|
IssuesEvent
|
2021-06-30 18:19:45
|
dotnet/dotnet-docker
|
https://api.github.com/repos/dotnet/dotnet-docker
|
closed
|
Update unit test for .NET Monitor filter environment variables
|
area-infrastructure
|
The .NET Monitor Dockerfile now sets two environment variables:
* DotnetMonitor_DefaultProcess__Filters__0__Key
* DotnetMonitor_DefaultProcess__Filters__0__Value
A unit test should be updated to ensure that these new environment variables are set within the built image:
https://github.com/dotnet/dotnet-docker/blob/7247fca2a7366eb883d898461aa8c743440eb64f/tests/Microsoft.DotNet.Docker.Tests/MonitorImageTests.cs#L85-L100
|
1.0
|
Update unit test for .NET Monitor filter environment variables - The .NET Monitor Dockerfile now sets two environment variables:
* DotnetMonitor_DefaultProcess__Filters__0__Key
* DotnetMonitor_DefaultProcess__Filters__0__Value
A unit test should be updated to ensure that these new environment variables are set within the built image:
https://github.com/dotnet/dotnet-docker/blob/7247fca2a7366eb883d898461aa8c743440eb64f/tests/Microsoft.DotNet.Docker.Tests/MonitorImageTests.cs#L85-L100
|
infrastructure
|
update unit test for net monitor filter environment variables the net monitor dockerfile now sets two environment variables dotnetmonitor defaultprocess filters key dotnetmonitor defaultprocess filters value a unit test should be updated to ensure that these new environment variables are set within the built image
| 1
|
18,579
| 13,055,925,774
|
IssuesEvent
|
2020-07-30 03:08:23
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
[I3_TESTDATA] Needs an overhaul (Trac #1353)
|
Incomplete Migration Migrated from Trac cleanup infrastructure
|
Migrated from https://code.icecube.wisc.edu/ticket/1353
```json
{
"status": "closed",
"changetime": "2019-07-01T13:30:20",
"description": "There's a lot of old cruft in there we should be able to get rid of.\n\nWe need to think of a good organizational scheme too, like per-project and common locations for files.\n\nAlso, *all* of this should be documented.\n\nFinally, some sort of versioning (it can be really simple) should be implemented, so we can clean periodically while ensuring old releases don't break. \n\nA good way to start is to maybe make a new location and just move the good stuff (I doubt there's much), while cleaning/fixing/adapting/setting-on-fire old, crufty scripts/tests/examples that depend on old, crufty test data.\n\nSeems like a good winter-break project.",
"reporter": "olivas",
"cc": "nega",
"resolution": "fixed",
"_ts": "1561987820131392",
"component": "infrastructure",
"summary": "[I3_TESTDATA] Needs an overhaul",
"priority": "critical",
"keywords": "",
"time": "2015-09-17T15:24:49",
"milestone": "Autumnal Equinox 2019",
"owner": "olivas",
"type": "cleanup"
}
```
|
1.0
|
[I3_TESTDATA] Needs an overhaul (Trac #1353) - Migrated from https://code.icecube.wisc.edu/ticket/1353
```json
{
"status": "closed",
"changetime": "2019-07-01T13:30:20",
"description": "There's a lot of old cruft in there we should be able to get rid of.\n\nWe need to think of a good organizational scheme too, like per-project and common locations for files.\n\nAlso, *all* of this should be documented.\n\nFinally, some sort of versioning (it can be really simple) should be implemented, so we can clean periodically while ensuring old releases don't break. \n\nA good way to start is to maybe make a new location and just move the good stuff (I doubt there's much), while cleaning/fixing/adapting/setting-on-fire old, crufty scripts/tests/examples that depend on old, crufty test data.\n\nSeems like a good winter-break project.",
"reporter": "olivas",
"cc": "nega",
"resolution": "fixed",
"_ts": "1561987820131392",
"component": "infrastructure",
"summary": "[I3_TESTDATA] Needs an overhaul",
"priority": "critical",
"keywords": "",
"time": "2015-09-17T15:24:49",
"milestone": "Autumnal Equinox 2019",
"owner": "olivas",
"type": "cleanup"
}
```
|
infrastructure
|
needs an overhaul trac migrated from json status closed changetime description there s a lot of old cruft in there we should be able to get rid of n nwe need to think of a good organizational scheme too like per project and common locations for files n nalso all of this should be documented n nfinally some sort of versioning it can be really simple should be implemented so we can clean periodically while ensuring old releases don t break n na good way to start is to maybe make a new location and just move the good stuff i doubt there s much while cleaning fixing adapting setting on fire old crufty scripts tests examples that depend on old crufty test data n nseems like a good winter break project reporter olivas cc nega resolution fixed ts component infrastructure summary needs an overhaul priority critical keywords time milestone autumnal equinox owner olivas type cleanup
| 1
|
13,513
| 10,309,873,358
|
IssuesEvent
|
2019-08-29 14:07:51
|
HumanCellAtlas/secondary-analysis
|
https://api.github.com/repos/HumanCellAtlas/secondary-analysis
|
closed
|
Update DCP integration test for updating metadata
|
infrastructure
|
With the "no-op" changes deployed to integration, update the DCP integration test for updating metadata to check that Falcon aborted the update workflow.
┆Issue is synchronized with this [Jira Story](https://broadinstitute.atlassian.net/browse/GH-484)
|
1.0
|
Update DCP integration test for updating metadata - With the "no-op" changes deployed to integration, update the DCP integration test for updating metadata to check that Falcon aborted the update workflow.
┆Issue is synchronized with this [Jira Story](https://broadinstitute.atlassian.net/browse/GH-484)
|
infrastructure
|
update dcp integration test for updating metadata with the no op changes deployed to integration update the dcp integration test for updating metadata to check that falcon aborted the update workflow ┆issue is synchronized with this
| 1
|
170,011
| 13,170,035,101
|
IssuesEvent
|
2020-08-11 14:35:56
|
NationalSecurityAgency/skills-service
|
https://api.github.com/repos/NationalSecurityAgency/skills-service
|
closed
|
Refreshing '/metrics' page causes 404
|
bug test
|
Refreshing '/metrics' page causes 404
- navigate to ``/metrics``:
- F5
|
1.0
|
Refreshing '/metrics' page causes 404 - Refreshing '/metrics' page causes 404
- navigate to ``/metrics``:
- F5
|
non_infrastructure
|
refreshing metrics page causes refreshing metrics page causes navigate to metrics
| 0
|
22,268
| 15,081,213,756
|
IssuesEvent
|
2021-02-05 12:52:45
|
stylelint/stylelint
|
https://api.github.com/repos/stylelint/stylelint
|
opened
|
Fix CodeQL warnings
|
help wanted type: infrastructure
|
The CI is failing on two CodeQL warnings:
```
Check failure on line 541 in lib/rules/indentation/index.js
Inefficient regular expression
This part of the regular expression may cause exponential backtracking on strings starting with '!\n' and containing many repetitions of '\n'.
```
```
Check failure on line 3808 in lib/utils/parseCalcExpression/parser.js
Inefficient regular expression
This part of the regular expression may cause exponential backtracking on strings starting with 'a(' and containing many repetitions of '''.
```
We'll need to fix these. The later is in generated code, so we may need to expedite https://github.com/stylelint/stylelint/issues/4731#issuecomment-623440469 and remove the `function-calc-no-invalid` rule.
|
1.0
|
Fix CodeQL warnings - The CI is failing on two CodeQL warnings:
```
Check failure on line 541 in lib/rules/indentation/index.js
Inefficient regular expression
This part of the regular expression may cause exponential backtracking on strings starting with '!\n' and containing many repetitions of '\n'.
```
```
Check failure on line 3808 in lib/utils/parseCalcExpression/parser.js
Inefficient regular expression
This part of the regular expression may cause exponential backtracking on strings starting with 'a(' and containing many repetitions of '''.
```
We'll need to fix these. The later is in generated code, so we may need to expedite https://github.com/stylelint/stylelint/issues/4731#issuecomment-623440469 and remove the `function-calc-no-invalid` rule.
|
infrastructure
|
fix codeql warnings the ci is failing on two codeql warnings check failure on line in lib rules indentation index js inefficient regular expression this part of the regular expression may cause exponential backtracking on strings starting with n and containing many repetitions of n check failure on line in lib utils parsecalcexpression parser js inefficient regular expression this part of the regular expression may cause exponential backtracking on strings starting with a and containing many repetitions of we ll need to fix these the later is in generated code so we may need to expedite and remove the function calc no invalid rule
| 1
|
166,579
| 12,962,793,643
|
IssuesEvent
|
2020-07-20 17:43:27
|
rancher/rancher
|
https://api.github.com/repos/rancher/rancher
|
closed
|
KDM should be updated for CIS 1.5 security scan to be the default ones run on a cluster
|
[zube]: To Test area/scan-tool kind/bug-qa
|
**What kind of request is this (question/bug/enhancement/feature request):** bug
**Steps to reproduce (least amount of steps as possible):**
- Currently the default scan run on a cluster on 2.4-head or master-head is CIS 1.4
- KDM has to be updated for CIS 1.5 security scan to be the default ones run on a cluster
**Other details that may be helpful:**
UI issue - https://github.com/rancher/rancher/issues/27374
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): 2.4-head and master-head latest
- Installation option (single install/HA): single
**Cluster information**
- Cluster type (Hosted/Infrastructure Provider/Custom/Imported): rke do/custom
- Kubernetes version (use `kubectl version`):
```
1.15 and above
```
|
1.0
|
KDM should be updated for CIS 1.5 security scan to be the default ones run on a cluster - **What kind of request is this (question/bug/enhancement/feature request):** bug
**Steps to reproduce (least amount of steps as possible):**
- Currently the default scan run on a cluster on 2.4-head or master-head is CIS 1.4
- KDM has to be updated for CIS 1.5 security scan to be the default ones run on a cluster
**Other details that may be helpful:**
UI issue - https://github.com/rancher/rancher/issues/27374
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): 2.4-head and master-head latest
- Installation option (single install/HA): single
**Cluster information**
- Cluster type (Hosted/Infrastructure Provider/Custom/Imported): rke do/custom
- Kubernetes version (use `kubectl version`):
```
1.15 and above
```
|
non_infrastructure
|
kdm should be updated for cis security scan to be the default ones run on a cluster what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible currently the default scan run on a cluster on head or master head is cis kdm has to be updated for cis security scan to be the default ones run on a cluster other details that may be helpful ui issue environment information rancher version rancher rancher rancher server image tag or shown bottom left in the ui head and master head latest installation option single install ha single cluster information cluster type hosted infrastructure provider custom imported rke do custom kubernetes version use kubectl version and above
| 0
|
22,753
| 2,650,822,850
|
IssuesEvent
|
2015-03-16 05:15:31
|
cs2103jan2015-t09-3j/main
|
https://api.github.com/repos/cs2103jan2015-t09-3j/main
|
closed
|
Implement "ADD Floating Task" --- Storage --- WriteToFile(str:task)
|
priority.high status.ongoing type.task
|
This method writes the task to the text file and returns the result(boolean).
|
1.0
|
Implement "ADD Floating Task" --- Storage --- WriteToFile(str:task) - This method writes the task to the text file and returns the result(boolean).
|
non_infrastructure
|
implement add floating task storage writetofile str task this method writes the task to the text file and returns the result boolean
| 0
|
18,365
| 12,957,425,140
|
IssuesEvent
|
2020-07-20 09:42:54
|
JMRI/JMRI
|
https://api.github.com/repos/JMRI/JMRI
|
closed
|
Strange error on AppVeyor
|
dev/test/publish infrastructure
|
Recently, the error message:
```
[junit] *** java.lang.instrument ASSERTION FAILED ***: "!errorOutstanding" with message transform method call failed at JPLISAgent.c line: 844
```
has been occurring on AppVeyor, even on successful builds. Googling suggests this is a StackOverflow error in the JVM, and (were we using gradle) might indicate an instrumentation issue, but beyond that, I haven't a clue.
@pabender do you have any suggestions?
|
1.0
|
Strange error on AppVeyor - Recently, the error message:
```
[junit] *** java.lang.instrument ASSERTION FAILED ***: "!errorOutstanding" with message transform method call failed at JPLISAgent.c line: 844
```
has been occurring on AppVeyor, even on successful builds. Googling suggests this is a StackOverflow error in the JVM, and (were we using gradle) might indicate an instrumentation issue, but beyond that, I haven't a clue.
@pabender do you have any suggestions?
|
infrastructure
|
strange error on appveyor recently the error message java lang instrument assertion failed erroroutstanding with message transform method call failed at jplisagent c line has been occurring on appveyor even on successful builds googling suggests this is a stackoverflow error in the jvm and were we using gradle might indicate an instrumentation issue but beyond that i haven t a clue pabender do you have any suggestions
| 1
|
28,379
| 23,204,511,154
|
IssuesEvent
|
2022-08-02 02:57:45
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
Create a kusto queries to monitor runtime tests related lanes on rolling build
|
area-Infrastructure-mono
|
Refer to WASM one for some idea:
https://dataexplorer.azure.com/dashboards/e5917b3e-7d01-4f78-a605-81edaa44acbe?startDate=7days&endDate=now&Lane=Tracking+PR&testNames=System.Runtime.Tests&prNumber=54053
|
1.0
|
Create a kusto queries to monitor runtime tests related lanes on rolling build - Refer to WASM one for some idea:
https://dataexplorer.azure.com/dashboards/e5917b3e-7d01-4f78-a605-81edaa44acbe?startDate=7days&endDate=now&Lane=Tracking+PR&testNames=System.Runtime.Tests&prNumber=54053
|
infrastructure
|
create a kusto queries to monitor runtime tests related lanes on rolling build refer to wasm one for some idea
| 1
|
125,514
| 10,345,223,522
|
IssuesEvent
|
2019-09-04 13:03:06
|
ballerina-platform/ballerina-lang
|
https://api.github.com/repos/ballerina-platform/ballerina-lang
|
closed
|
Local variables are not suggested as function arguments
|
Area/Tooling BetaTesting Component/LanguageServer Type/Bug
|
**Description:**
My code is as follows.
```ballerina
import ballerina/math;
function equationOfCenter(float solarAnomaly) returns float {
float annomalyInRad = solarAnomaly * ( math:PI / 180 );
float anomalySin = math:sin(annomalyInRad);
float anomaly2Sin = math:sin( 2 * )
return annomalyInRad;
}
```
Local variables such as `annomalyInRad` , `anomalySin` etc are not gives as suggestions as function arguments.
<img width="616" alt="image" src="https://user-images.githubusercontent.com/16300038/63092659-11640f00-bf80-11e9-8644-ebadba469803.png">
```
[Error - 5:12:16 PM] Operation 'text/completion' failed! {uri: /Volumes/work/Ballerina/Code/BetaHackathon/SunProject/src/suntime/utils.bal, line: 14, col:39, error: java.lang.NullPointerException, org.wso2.ballerinalang.compiler.tree.BLangFunction.accept(BLangFunction.java:71)}
java.lang.NullPointerException
at org.ballerinalang.langserver.completions.TreeVisitor.visit(TreeVisitor.java:222)
at org.wso2.ballerinalang.compiler.tree.BLangFunction.accept(BLangFunction.java:71)
at org.ballerinalang.langserver.completions.TreeVisitor.acceptNode(TreeVisitor.java:905)
at org.ballerinalang.langserver.completions.TreeVisitor.visit(TreeVisitor.java:186)
at org.wso2.ballerinalang.compiler.tree.BLangPackage.accept(BLangPackage.java:160)
at org.ballerinalang.langserver.completions.util.CompletionUtil.resolveSymbols(CompletionUtil.java:58)
at org.ballerinalang.langserver.BallerinaTextDocumentService.lambda$completion$0(BallerinaTextDocumentService.java:180)
at java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1590)
at java.util.concurrent.CompletableFuture$AsyncSupply.exec(CompletableFuture.java:1582)
at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289)
at java.util.concurrent.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1056)
at java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1692)
at java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:157)
```
**Steps to reproduce:**
**Affected Versions:**
**OS, DB, other environment details and versions:**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
|
1.0
|
Local variables are not suggested as function arguments - **Description:**
My code is as follows.
```ballerina
import ballerina/math;
function equationOfCenter(float solarAnomaly) returns float {
float annomalyInRad = solarAnomaly * ( math:PI / 180 );
float anomalySin = math:sin(annomalyInRad);
float anomaly2Sin = math:sin( 2 * )
return annomalyInRad;
}
```
Local variables such as `annomalyInRad` , `anomalySin` etc are not gives as suggestions as function arguments.
<img width="616" alt="image" src="https://user-images.githubusercontent.com/16300038/63092659-11640f00-bf80-11e9-8644-ebadba469803.png">
```
[Error - 5:12:16 PM] Operation 'text/completion' failed! {uri: /Volumes/work/Ballerina/Code/BetaHackathon/SunProject/src/suntime/utils.bal, line: 14, col:39, error: java.lang.NullPointerException, org.wso2.ballerinalang.compiler.tree.BLangFunction.accept(BLangFunction.java:71)}
java.lang.NullPointerException
at org.ballerinalang.langserver.completions.TreeVisitor.visit(TreeVisitor.java:222)
at org.wso2.ballerinalang.compiler.tree.BLangFunction.accept(BLangFunction.java:71)
at org.ballerinalang.langserver.completions.TreeVisitor.acceptNode(TreeVisitor.java:905)
at org.ballerinalang.langserver.completions.TreeVisitor.visit(TreeVisitor.java:186)
at org.wso2.ballerinalang.compiler.tree.BLangPackage.accept(BLangPackage.java:160)
at org.ballerinalang.langserver.completions.util.CompletionUtil.resolveSymbols(CompletionUtil.java:58)
at org.ballerinalang.langserver.BallerinaTextDocumentService.lambda$completion$0(BallerinaTextDocumentService.java:180)
at java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1590)
at java.util.concurrent.CompletableFuture$AsyncSupply.exec(CompletableFuture.java:1582)
at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289)
at java.util.concurrent.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1056)
at java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1692)
at java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:157)
```
**Steps to reproduce:**
**Affected Versions:**
**OS, DB, other environment details and versions:**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
|
non_infrastructure
|
local variables are not suggested as function arguments description my code is as follows ballerina import ballerina math function equationofcenter float solaranomaly returns float float annomalyinrad solaranomaly math pi float anomalysin math sin annomalyinrad float math sin return annomalyinrad local variables such as annomalyinrad anomalysin etc are not gives as suggestions as function arguments img width alt image src operation text completion failed uri volumes work ballerina code betahackathon sunproject src suntime utils bal line col error java lang nullpointerexception org ballerinalang compiler tree blangfunction accept blangfunction java java lang nullpointerexception at org ballerinalang langserver completions treevisitor visit treevisitor java at org ballerinalang compiler tree blangfunction accept blangfunction java at org ballerinalang langserver completions treevisitor acceptnode treevisitor java at org ballerinalang langserver completions treevisitor visit treevisitor java at org ballerinalang compiler tree blangpackage accept blangpackage java at org ballerinalang langserver completions util completionutil resolvesymbols completionutil java at org ballerinalang langserver ballerinatextdocumentservice lambda completion ballerinatextdocumentservice java at java util concurrent completablefuture asyncsupply run completablefuture java at java util concurrent completablefuture asyncsupply exec completablefuture java at java util concurrent forkjointask doexec forkjointask java at java util concurrent forkjoinpool workqueue runtask forkjoinpool java at java util concurrent forkjoinpool runworker forkjoinpool java at java util concurrent forkjoinworkerthread run forkjoinworkerthread java steps to reproduce affected versions os db other environment details and versions related issues optional suggested labels optional suggested assignees optional
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.