Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
14,545
| 17,662,240,565
|
IssuesEvent
|
2021-08-21 18:56:34
|
parcel-bundler/parcel
|
https://api.github.com/repos/parcel-bundler/parcel
|
closed
|
SCSS imports containing ~/ fail
|
CSS Preprocessing
|
<!---
Thanks for filing an issue 😄 ! Before you submit, please read the following:
Search open/closed issues before submitting since someone might have asked the same thing before!
-->
# 🐛 bug report
When using ~/ paths in SCSS @ imports it fails with an error message
<!--- Provide a general summary of the issue here -->
## 🎛 Configuration (.babelrc, package.json, cli command)
See code sample gist, `yarn build`
## 🤔 Expected Behavior
SCSS import works as expected
## 😯 Current Behavior
Imports using ~ in SCSS cause the following error
```
yarn run v1.22.11
$ parcel build index.html
🚨 Build failed.
Error: ENOENT: no such file or directory, open '/home/maciej/Projekte/parcel-issue/file%3A/home/maciej/Projekte/parcel-issue/other.scss'
error Command failed with exit code 1.
info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command.
```
## 💁 Possible Solution
<!--- Not obligatory, but suggest a fix/reason for the bug -->
## 🔦 Context
It breaks build & serve
## 💻 Code Sample
https://gist.github.com/mkg20001/32cb89621833c22485290b13f8032313
Minimum reporduction is using ~ in SCSS @ import
## 🌍 Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
| Software | Version(s) |
| ---------------- | ---------- |
| Parcel | 2.0.0-rc.0
| Node | v16.6.2
| npm/Yarn | yarn 1.22.11
| Operating System | NixOS 21.11 (Porcupine)
<!-- Love parcel? Please consider supporting our collective:
👉 https://opencollective.com/parcel/donate -->
|
1.0
|
SCSS imports containing ~/ fail - <!---
Thanks for filing an issue 😄 ! Before you submit, please read the following:
Search open/closed issues before submitting since someone might have asked the same thing before!
-->
# 🐛 bug report
When using ~/ paths in SCSS @ imports it fails with an error message
<!--- Provide a general summary of the issue here -->
## 🎛 Configuration (.babelrc, package.json, cli command)
See code sample gist, `yarn build`
## 🤔 Expected Behavior
SCSS import works as expected
## 😯 Current Behavior
Imports using ~ in SCSS cause the following error
```
yarn run v1.22.11
$ parcel build index.html
🚨 Build failed.
Error: ENOENT: no such file or directory, open '/home/maciej/Projekte/parcel-issue/file%3A/home/maciej/Projekte/parcel-issue/other.scss'
error Command failed with exit code 1.
info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command.
```
## 💁 Possible Solution
<!--- Not obligatory, but suggest a fix/reason for the bug -->
## 🔦 Context
It breaks build & serve
## 💻 Code Sample
https://gist.github.com/mkg20001/32cb89621833c22485290b13f8032313
Minimum reporduction is using ~ in SCSS @ import
## 🌍 Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
| Software | Version(s) |
| ---------------- | ---------- |
| Parcel | 2.0.0-rc.0
| Node | v16.6.2
| npm/Yarn | yarn 1.22.11
| Operating System | NixOS 21.11 (Porcupine)
<!-- Love parcel? Please consider supporting our collective:
👉 https://opencollective.com/parcel/donate -->
|
process
|
scss imports containing fail thanks for filing an issue 😄 before you submit please read the following search open closed issues before submitting since someone might have asked the same thing before 🐛 bug report when using paths in scss imports it fails with an error message 🎛 configuration babelrc package json cli command see code sample gist yarn build 🤔 expected behavior scss import works as expected 😯 current behavior imports using in scss cause the following error yarn run parcel build index html 🚨 build failed error enoent no such file or directory open home maciej projekte parcel issue file home maciej projekte parcel issue other scss error command failed with exit code info visit for documentation about this command 💁 possible solution 🔦 context it breaks build serve 💻 code sample minimum reporduction is using in scss import 🌍 your environment software version s parcel rc node npm yarn yarn operating system nixos porcupine love parcel please consider supporting our collective 👉
| 1
|
38,857
| 10,256,921,612
|
IssuesEvent
|
2019-08-21 18:49:52
|
tensorflow/tfjs
|
https://api.github.com/repos/tensorflow/tfjs
|
closed
|
tfjs-examples: Simple object detection: yarn train --gpu fails
|
type:build/install
|
#### TensorFlow.js version
1.2.2
#### Browser version
Windows Version 10.0.17134 Build 17134
Node v10.15.0
### Problem description
**Install appears to succeed (with warnings)**
```
yarn install v1.17.3
[1/5] Validating package.json...
[2/5] Resolving packages...
[3/5] Fetching packages...
info fsevents@1.2.4: The platform "win32" is incompatible with this module.
info "fsevents@1.2.4" is an optional dependency and failed compatibility check. Excluding it from installation.
[4/5] Linking dependencies...
warning "@tensorflow/tfjs > @tensorflow/tfjs-data@1.2.2" has unmet peer dependency "seedrandom@~2.4.3".
warning "@tensorflow/tfjs > @tensorflow/tfjs-core > rollup-plugin-visualizer@1.1.1" has unmet peer dependency "rollup@>=0.60.0".
[5/5] Building fresh packages...
Done in 139.97s.
```
**Training with the gpu flag fails**
```
$ yarn train --gpu
yarn run v1.17.3
$ node train.js --gpu
Training using GPU.
cpu backend was already registered. Reusing existing backend factory.
Platform node has already been set. Overwriting the platform with [object Object].
node-pre-gyp info This Node instance does not support builds for N-API version 4
node-pre-gyp info This Node instance does not support builds for N-API version 4
(node:14276) UnhandledPromiseRejectionWarning: Error: The specified module could not be found.
\\?\C:\Users\Ian\projects\tfjs\tfjs-examples\simple-object-detection\node_modules\@tensorflow\tfjs-node-gpu\lib\napi-v3\tfjs_binding.node
at Object.Module._extensions..node (internal/modules/cjs/loader.js:718:18)
at Module.load (internal/modules/cjs/loader.js:599:32)
at tryModuleLoad (internal/modules/cjs/loader.js:538:12)
at Function.Module._load (internal/modules/cjs/loader.js:530:3)
at Module.require (internal/modules/cjs/loader.js:637:17)
at require (internal/modules/cjs/helpers.js:22:18)
at Object.<anonymous> (C:\Users\Ian\projects\tfjs\tfjs-examples\simple-object-detection\node_modules\@tensorflow\tfjs-node-gpu\dist\index.js:44:16)
at Module._compile (internal/modules/cjs/loader.js:689:30)
at Object.Module._extensions..js (internal/modules/cjs/loader.js:700:10)
at Module.load (internal/modules/cjs/loader.js:599:32)
(node:14276) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). (rejection id: 1)
(node:14276) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code.
Done in 1.74s.
```
|
1.0
|
tfjs-examples: Simple object detection: yarn train --gpu fails - #### TensorFlow.js version
1.2.2
#### Browser version
Windows Version 10.0.17134 Build 17134
Node v10.15.0
### Problem description
**Install appears to succeed (with warnings)**
```
yarn install v1.17.3
[1/5] Validating package.json...
[2/5] Resolving packages...
[3/5] Fetching packages...
info fsevents@1.2.4: The platform "win32" is incompatible with this module.
info "fsevents@1.2.4" is an optional dependency and failed compatibility check. Excluding it from installation.
[4/5] Linking dependencies...
warning "@tensorflow/tfjs > @tensorflow/tfjs-data@1.2.2" has unmet peer dependency "seedrandom@~2.4.3".
warning "@tensorflow/tfjs > @tensorflow/tfjs-core > rollup-plugin-visualizer@1.1.1" has unmet peer dependency "rollup@>=0.60.0".
[5/5] Building fresh packages...
Done in 139.97s.
```
**Training with the gpu flag fails**
```
$ yarn train --gpu
yarn run v1.17.3
$ node train.js --gpu
Training using GPU.
cpu backend was already registered. Reusing existing backend factory.
Platform node has already been set. Overwriting the platform with [object Object].
node-pre-gyp info This Node instance does not support builds for N-API version 4
node-pre-gyp info This Node instance does not support builds for N-API version 4
(node:14276) UnhandledPromiseRejectionWarning: Error: The specified module could not be found.
\\?\C:\Users\Ian\projects\tfjs\tfjs-examples\simple-object-detection\node_modules\@tensorflow\tfjs-node-gpu\lib\napi-v3\tfjs_binding.node
at Object.Module._extensions..node (internal/modules/cjs/loader.js:718:18)
at Module.load (internal/modules/cjs/loader.js:599:32)
at tryModuleLoad (internal/modules/cjs/loader.js:538:12)
at Function.Module._load (internal/modules/cjs/loader.js:530:3)
at Module.require (internal/modules/cjs/loader.js:637:17)
at require (internal/modules/cjs/helpers.js:22:18)
at Object.<anonymous> (C:\Users\Ian\projects\tfjs\tfjs-examples\simple-object-detection\node_modules\@tensorflow\tfjs-node-gpu\dist\index.js:44:16)
at Module._compile (internal/modules/cjs/loader.js:689:30)
at Object.Module._extensions..js (internal/modules/cjs/loader.js:700:10)
at Module.load (internal/modules/cjs/loader.js:599:32)
(node:14276) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). (rejection id: 1)
(node:14276) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code.
Done in 1.74s.
```
|
non_process
|
tfjs examples simple object detection yarn train gpu fails tensorflow js version browser version windows version build node problem description install appears to succeed with warnings yarn install validating package json resolving packages fetching packages info fsevents the platform is incompatible with this module info fsevents is an optional dependency and failed compatibility check excluding it from installation linking dependencies warning tensorflow tfjs tensorflow tfjs data has unmet peer dependency seedrandom warning tensorflow tfjs tensorflow tfjs core rollup plugin visualizer has unmet peer dependency rollup building fresh packages done in training with the gpu flag fails yarn train gpu yarn run node train js gpu training using gpu cpu backend was already registered reusing existing backend factory platform node has already been set overwriting the platform with node pre gyp info this node instance does not support builds for n api version node pre gyp info this node instance does not support builds for n api version node unhandledpromiserejectionwarning error the specified module could not be found c users ian projects tfjs tfjs examples simple object detection node modules tensorflow tfjs node gpu lib napi tfjs binding node at object module extensions node internal modules cjs loader js at module load internal modules cjs loader js at trymoduleload internal modules cjs loader js at function module load internal modules cjs loader js at module require internal modules cjs loader js at require internal modules cjs helpers js at object c users ian projects tfjs tfjs examples simple object detection node modules tensorflow tfjs node gpu dist index js at module compile internal modules cjs loader js at object module extensions js internal modules cjs loader js at module load internal modules cjs loader js node unhandledpromiserejectionwarning unhandled promise rejection this error originated either by throwing inside of an async function without a catch block or by rejecting a promise which was not handled with catch rejection id node deprecationwarning unhandled promise rejections are deprecated in the future promise rejections that are not handled will terminate the node js process with a non zero exit code done in
| 0
|
18,678
| 24,594,494,624
|
IssuesEvent
|
2022-10-14 07:04:39
|
NationalSecurityAgency/ghidra
|
https://api.github.com/repos/NationalSecurityAgency/ghidra
|
closed
|
6809: JMP extended indirect does not disassemble
|
Feature: Processor/MC6800 Status: Internal
|
The byte pattern "6E 9F 58 3E" should disassemble to "JMP [DAT_583e]" but instead is reported as a bad instruction.
To reproduce, load a binary containing the byte pattern above and attempt to disassemble using the 6809 processor module.
Expected behavior is stated above.
Problem appears to be in 6x09.sinc, line 325.
- OS: Ubuntu 18.04.6 LTS
- Java Version: 11.0.16
- Ghidra Version: 10.1.5
- Ghidra Origin: official GitHub distro
|
1.0
|
6809: JMP extended indirect does not disassemble - The byte pattern "6E 9F 58 3E" should disassemble to "JMP [DAT_583e]" but instead is reported as a bad instruction.
To reproduce, load a binary containing the byte pattern above and attempt to disassemble using the 6809 processor module.
Expected behavior is stated above.
Problem appears to be in 6x09.sinc, line 325.
- OS: Ubuntu 18.04.6 LTS
- Java Version: 11.0.16
- Ghidra Version: 10.1.5
- Ghidra Origin: official GitHub distro
|
process
|
jmp extended indirect does not disassemble the byte pattern should disassemble to jmp but instead is reported as a bad instruction to reproduce load a binary containing the byte pattern above and attempt to disassemble using the processor module expected behavior is stated above problem appears to be in sinc line os ubuntu lts java version ghidra version ghidra origin official github distro
| 1
|
19,207
| 25,339,400,987
|
IssuesEvent
|
2022-11-18 19:57:38
|
solop-develop/frontend-core
|
https://api.github.com/repos/solop-develop/frontend-core
|
closed
|
[Bug Report] Error Campo de Búsqueda de Registro
|
bug (PRC) Processes (RPT) Reports (SB) Smart Browser
|
<!--
Note: In order to better solve your problem, please refer to the template to provide complete information, accurately describe the problem, and the incomplete information issue will be closed.
-->
## Bug report
El campo de busqueda de registro solo debe aparecer en el panel de ventana
#### Steps to reproduce
- Abra un Reporte y observe que aparece el campo de busqueda de registro
- Abra un Proceso y observe que aparece el campo de busqueda de registro
- Abra un Smart Browser y observe que aparece el campo de busqueda de registro
#### Screenshot or Gif(截图或动态图)
https://user-images.githubusercontent.com/45974454/199028426-c67025eb-f14c-4ed4-9f44-5633138772ee.mp4
|
1.0
|
[Bug Report] Error Campo de Búsqueda de Registro - <!--
Note: In order to better solve your problem, please refer to the template to provide complete information, accurately describe the problem, and the incomplete information issue will be closed.
-->
## Bug report
El campo de busqueda de registro solo debe aparecer en el panel de ventana
#### Steps to reproduce
- Abra un Reporte y observe que aparece el campo de busqueda de registro
- Abra un Proceso y observe que aparece el campo de busqueda de registro
- Abra un Smart Browser y observe que aparece el campo de busqueda de registro
#### Screenshot or Gif(截图或动态图)
https://user-images.githubusercontent.com/45974454/199028426-c67025eb-f14c-4ed4-9f44-5633138772ee.mp4
|
process
|
error campo de búsqueda de registro note in order to better solve your problem please refer to the template to provide complete information accurately describe the problem and the incomplete information issue will be closed bug report el campo de busqueda de registro solo debe aparecer en el panel de ventana steps to reproduce abra un reporte y observe que aparece el campo de busqueda de registro abra un proceso y observe que aparece el campo de busqueda de registro abra un smart browser y observe que aparece el campo de busqueda de registro screenshot or gif(截图或动态图)
| 1
|
316,569
| 27,167,183,030
|
IssuesEvent
|
2023-02-17 16:13:01
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
reopened
|
Fix miscellaneous_ops.test_torch_tensordot
|
PyTorch Frontend Sub Task Failing Test
|
| | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4054925606/jobs/6977411757" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4014333560/jobs/6894720191" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4054925606/jobs/6977419010" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4014333560/jobs/6894720191" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>Not found</summary>
Not found
</details>
<details>
<summary>Not found</summary>
Not found
</details>
|
1.0
|
Fix miscellaneous_ops.test_torch_tensordot - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4054925606/jobs/6977411757" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4014333560/jobs/6894720191" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4054925606/jobs/6977419010" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4014333560/jobs/6894720191" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>Not found</summary>
Not found
</details>
<details>
<summary>Not found</summary>
Not found
</details>
|
non_process
|
fix miscellaneous ops test torch tensordot tensorflow img src torch img src numpy img src jax img src not found not found not found not found
| 0
|
1,064
| 3,332,900,279
|
IssuesEvent
|
2015-11-11 22:14:30
|
rancher/rancher
|
https://api.github.com/repos/rancher/rancher
|
closed
|
Main stack screen is missing the Add Service
|
area/service area/ui kind/bug status/resolved status/to-test
|
Version - Master with reskin UI
Steps:
1. Create a stack
Results:
There is no way to Add a Service from the main stack screen.
Expected:
In previous UI there was a Add Service.
|
1.0
|
Main stack screen is missing the Add Service - Version - Master with reskin UI
Steps:
1. Create a stack
Results:
There is no way to Add a Service from the main stack screen.
Expected:
In previous UI there was a Add Service.
|
non_process
|
main stack screen is missing the add service version master with reskin ui steps create a stack results there is no way to add a service from the main stack screen expected in previous ui there was a add service
| 0
|
5,087
| 7,876,070,927
|
IssuesEvent
|
2018-06-25 22:55:25
|
Great-Hill-Corporation/quickBlocks
|
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
|
closed
|
Display strings are very slow
|
libs-utillib status-inprocess type-bug
|
It would be much easier to use a stream with a state machine to parse the string as I go. It might even make sense to create some sort of pre-parsed string in a table, so that multiple displays of the same type of record (which happens all the time) is faster.
|
1.0
|
Display strings are very slow - It would be much easier to use a stream with a state machine to parse the string as I go. It might even make sense to create some sort of pre-parsed string in a table, so that multiple displays of the same type of record (which happens all the time) is faster.
|
process
|
display strings are very slow it would be much easier to use a stream with a state machine to parse the string as i go it might even make sense to create some sort of pre parsed string in a table so that multiple displays of the same type of record which happens all the time is faster
| 1
|
4,119
| 15,530,216,371
|
IssuesEvent
|
2021-03-13 18:13:18
|
100Automations/Website
|
https://api.github.com/repos/100Automations/Website
|
closed
|
Create a dropdown for needs help button on automation cards
|
feature: automations role: front end
|
### Overview
Each Automation card needs the button in upper lefthand corner to dropdown list of all roles that are needed on the project.
### Action Items
- [x] Make button clickable, and list hidden when page loads.
- [x] Add list of 'needs help' roles that toggles when button is clicked.
### Resources/Instructions
Use Figma if needed.
<img width="352" alt="Screen Shot 2021-02-20 at 8 49 13 AM" src="https://user-images.githubusercontent.com/38964454/108602742-858f2800-7358-11eb-8436-0c507ff4db80.png">
|
1.0
|
Create a dropdown for needs help button on automation cards - ### Overview
Each Automation card needs the button in upper lefthand corner to dropdown list of all roles that are needed on the project.
### Action Items
- [x] Make button clickable, and list hidden when page loads.
- [x] Add list of 'needs help' roles that toggles when button is clicked.
### Resources/Instructions
Use Figma if needed.
<img width="352" alt="Screen Shot 2021-02-20 at 8 49 13 AM" src="https://user-images.githubusercontent.com/38964454/108602742-858f2800-7358-11eb-8436-0c507ff4db80.png">
|
non_process
|
create a dropdown for needs help button on automation cards overview each automation card needs the button in upper lefthand corner to dropdown list of all roles that are needed on the project action items make button clickable and list hidden when page loads add list of needs help roles that toggles when button is clicked resources instructions use figma if needed img width alt screen shot at am src
| 0
|
211,070
| 16,167,839,791
|
IssuesEvent
|
2021-05-01 21:31:03
|
NixOS/nixpkgs
|
https://api.github.com/repos/NixOS/nixpkgs
|
closed
|
Package tests
|
0.kind: question 2.status: stale 6.topic: developer experience 6.topic: policy discussion 6.topic: testing 9.needs: community feedback
|
Unless I got lost in the package tree, I don't see any tests other than those in `nixos/tests`, which are effectively for the NixOS modules.
I want to start adding package tests, primarily for all those package I regularly update/maintain.
a) Where should I put those
b) How would I bootstrap them? (I want them to be run as part of the channel updates, not unlike the `nixos/tests` files)
c) Do we have concerns with Hydra build times if/when people were to start adding lots of tests?
/cc @vcunat @domenkozar @edolstra @shlevy @aszlig @fpletz @globin @joachifm @copumpkin @grahamc (just pinging people who usually have good opinions on these things, anybody is welcome to provide feedback)
|
1.0
|
Package tests - Unless I got lost in the package tree, I don't see any tests other than those in `nixos/tests`, which are effectively for the NixOS modules.
I want to start adding package tests, primarily for all those package I regularly update/maintain.
a) Where should I put those
b) How would I bootstrap them? (I want them to be run as part of the channel updates, not unlike the `nixos/tests` files)
c) Do we have concerns with Hydra build times if/when people were to start adding lots of tests?
/cc @vcunat @domenkozar @edolstra @shlevy @aszlig @fpletz @globin @joachifm @copumpkin @grahamc (just pinging people who usually have good opinions on these things, anybody is welcome to provide feedback)
|
non_process
|
package tests unless i got lost in the package tree i don t see any tests other than those in nixos tests which are effectively for the nixos modules i want to start adding package tests primarily for all those package i regularly update maintain a where should i put those b how would i bootstrap them i want them to be run as part of the channel updates not unlike the nixos tests files c do we have concerns with hydra build times if when people were to start adding lots of tests cc vcunat domenkozar edolstra shlevy aszlig fpletz globin joachifm copumpkin grahamc just pinging people who usually have good opinions on these things anybody is welcome to provide feedback
| 0
|
422,439
| 12,278,369,968
|
IssuesEvent
|
2020-05-08 09:49:50
|
ComPWA/tensorwaves
|
https://api.github.com/repos/ComPWA/tensorwaves
|
opened
|
Implement gradient for UnbinnedNLL
|
Priority: Critical Type: Enhancement Type: Question
|
Implement the analytic gradient.
- [ ] Clarify how to use the gradient in the Estimator with Intensity which hides the backend (I think this comes back to the "ComPWA math language"). So currently I think it is necessary to have a graph structure that describes the computation, which can be converted to an actual computation. Same concept that tensorflow follows. This would solve many problems at once.
- [ ] Actually implement the gradient
Some tensorflow specific code might look like this (copied from amplitf):
```python
for i, p in enumerate(float_pars):
p.update(par[i])
with tf.GradientTape() as gradient:
gradient.watch(float_pars)
nll_val = nll(*args)
g = gradient.gradient(
nll_val, float_pars, unconnected_gradients=tf.UnconnectedGradients.ZERO)
g_val = [i.numpy() for i in g]
return g_val
```
|
1.0
|
Implement gradient for UnbinnedNLL - Implement the analytic gradient.
- [ ] Clarify how to use the gradient in the Estimator with Intensity which hides the backend (I think this comes back to the "ComPWA math language"). So currently I think it is necessary to have a graph structure that describes the computation, which can be converted to an actual computation. Same concept that tensorflow follows. This would solve many problems at once.
- [ ] Actually implement the gradient
Some tensorflow specific code might look like this (copied from amplitf):
```python
for i, p in enumerate(float_pars):
p.update(par[i])
with tf.GradientTape() as gradient:
gradient.watch(float_pars)
nll_val = nll(*args)
g = gradient.gradient(
nll_val, float_pars, unconnected_gradients=tf.UnconnectedGradients.ZERO)
g_val = [i.numpy() for i in g]
return g_val
```
|
non_process
|
implement gradient for unbinnednll implement the analytic gradient clarify how to use the gradient in the estimator with intensity which hides the backend i think this comes back to the compwa math language so currently i think it is necessary to have a graph structure that describes the computation which can be converted to an actual computation same concept that tensorflow follows this would solve many problems at once actually implement the gradient some tensorflow specific code might look like this copied from amplitf python for i p in enumerate float pars p update par with tf gradienttape as gradient gradient watch float pars nll val nll args g gradient gradient nll val float pars unconnected gradients tf unconnectedgradients zero g val return g val
| 0
|
95,817
| 8,579,104,266
|
IssuesEvent
|
2018-11-13 08:07:14
|
humera987/FXLabs-Test-Automation
|
https://api.github.com/repos/humera987/FXLabs-Test-Automation
|
closed
|
testing : ApiV1AccountsGetQueryParamPagesizeNegativeNumber
|
testing
|
Project : testing
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=NTU5OGZjNTYtYTVkNC00NDMzLWI2NDMtZGQ3YzE5ODY2MTlj; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Tue, 13 Nov 2018 08:06:02 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/accounts?pageSize=-1
Request :
Response :
{
"timestamp" : "2018-11-13T08:06:03.044+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/accounts"
}
Logs :
Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot ---
|
1.0
|
testing : ApiV1AccountsGetQueryParamPagesizeNegativeNumber - Project : testing
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=NTU5OGZjNTYtYTVkNC00NDMzLWI2NDMtZGQ3YzE5ODY2MTlj; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Tue, 13 Nov 2018 08:06:02 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/accounts?pageSize=-1
Request :
Response :
{
"timestamp" : "2018-11-13T08:06:03.044+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/accounts"
}
Logs :
Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot ---
|
non_process
|
testing project testing job uat env uat region us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options set cookie content type transfer encoding date endpoint request response timestamp status error not found message no message available path api api accounts logs assertion resolved to result assertion resolved to result fx bot
| 0
|
56,892
| 11,686,249,518
|
IssuesEvent
|
2020-03-05 10:32:06
|
joomla/joomla-cms
|
https://api.github.com/repos/joomla/joomla-cms
|
closed
|
[4.0] admin list views - blurred
|
J4 Backend Template No Code Attached Yet
|
When you hover on a row in a list such as the article manager the background changes to a grey colour. Unfortunately the light blue border merges into the grey and produces a "blurred" image

There are a few options to "fix" this
Change the colour to something darker

Increase the border width

|
1.0
|
[4.0] admin list views - blurred - When you hover on a row in a list such as the article manager the background changes to a grey colour. Unfortunately the light blue border merges into the grey and produces a "blurred" image

There are a few options to "fix" this
Change the colour to something darker

Increase the border width

|
non_process
|
admin list views blurred when you hover on a row in a list such as the article manager the background changes to a grey colour unfortunately the light blue border merges into the grey and produces a blurred image there are a few options to fix this change the colour to something darker increase the border width
| 0
|
1,740
| 4,427,712,243
|
IssuesEvent
|
2016-08-16 22:25:23
|
sysown/proxysql
|
https://api.github.com/repos/sysown/proxysql
|
closed
|
Count CPU time while processing query rules
|
ADMIN QUERY PROCESSOR ROUTING
|
In order to understand how time consuming is to process query rules, metrics should be collected
New status variables:
* [x] Backend_query_time_nsec
* [x] Query_Processor_time_nsec
|
1.0
|
Count CPU time while processing query rules - In order to understand how time consuming is to process query rules, metrics should be collected
New status variables:
* [x] Backend_query_time_nsec
* [x] Query_Processor_time_nsec
|
process
|
count cpu time while processing query rules in order to understand how time consuming is to process query rules metrics should be collected new status variables backend query time nsec query processor time nsec
| 1
|
4,789
| 7,674,408,656
|
IssuesEvent
|
2018-05-15 03:47:25
|
neuropoly/spinalcordtoolbox
|
https://api.github.com/repos/neuropoly/spinalcordtoolbox
|
opened
|
sct_process_segmentation -discfile -p label-vert
|
sct_process_segmentation
|
2 detected issues with the function:
`sct_process_segmentation -discfile -p label-vert`
1. Assume images with RPI orientation:
```
cd /Volumes/temp/charley/i_XX
sct_process_segmentation -i t2_sag_cerv_seg_AIL.nii.gz -discfile label_discs_AIL.nii.gz -p label-vert
```
2. Use the convention "disc labelvalue=3 ==> disc C3/C4" instead of "disc labelvalue=3 ==> disc C2/C3"
```
cd /Volumes/temp/charley/i_XX
sct_process_segmentation -i t2_sag_cerv_seg_RPI.nii.gz -discfile label_discs_RPI.nii.gz -p label-vert
```
Or I am missing something? :/
|
1.0
|
sct_process_segmentation -discfile -p label-vert - 2 detected issues with the function:
`sct_process_segmentation -discfile -p label-vert`
1. Assume images with RPI orientation:
```
cd /Volumes/temp/charley/i_XX
sct_process_segmentation -i t2_sag_cerv_seg_AIL.nii.gz -discfile label_discs_AIL.nii.gz -p label-vert
```
2. Use the convention "disc labelvalue=3 ==> disc C3/C4" instead of "disc labelvalue=3 ==> disc C2/C3"
```
cd /Volumes/temp/charley/i_XX
sct_process_segmentation -i t2_sag_cerv_seg_RPI.nii.gz -discfile label_discs_RPI.nii.gz -p label-vert
```
Or I am missing something? :/
|
process
|
sct process segmentation discfile p label vert detected issues with the function sct process segmentation discfile p label vert assume images with rpi orientation cd volumes temp charley i xx sct process segmentation i sag cerv seg ail nii gz discfile label discs ail nii gz p label vert use the convention disc labelvalue disc instead of disc labelvalue disc cd volumes temp charley i xx sct process segmentation i sag cerv seg rpi nii gz discfile label discs rpi nii gz p label vert or i am missing something
| 1
|
259,631
| 22,503,473,375
|
IssuesEvent
|
2022-06-23 13:44:39
|
Joystream/pioneer
|
https://api.github.com/repos/Joystream/pioneer
|
closed
|
BuyMembershipModal
|
qa-task qa-tested-ready-for-prod qa-effort-1.5h release:stabilisation7
|
Refactor the `BuyMembershipModal` as part of the #2802 epic.
## Dev scope
- [x] Refactor fields validation
- [x] Manage the form state with React Hook Form instead of XState
## QA scope
- [ ] The fields are validated correctly
- [ ] The transaction gets sent with the parameter set in the modal
|
1.0
|
BuyMembershipModal - Refactor the `BuyMembershipModal` as part of the #2802 epic.
## Dev scope
- [x] Refactor fields validation
- [x] Manage the form state with React Hook Form instead of XState
## QA scope
- [ ] The fields are validated correctly
- [ ] The transaction gets sent with the parameter set in the modal
|
non_process
|
buymembershipmodal refactor the buymembershipmodal as part of the epic dev scope refactor fields validation manage the form state with react hook form instead of xstate qa scope the fields are validated correctly the transaction gets sent with the parameter set in the modal
| 0
|
5,463
| 8,328,327,962
|
IssuesEvent
|
2018-09-27 00:10:12
|
ArctosDB/new-collections
|
https://api.github.com/repos/ArctosDB/new-collections
|
closed
|
Ohio Wesleyan University - First Contact
|
Application in process Committed to Arctos
|
Post initial contact email or use [initial questionnaire web form](http://arctos.database.museum/new_collection.cfm).
Record Collection Contact Information and initial conversations.
|
1.0
|
Ohio Wesleyan University - First Contact - Post initial contact email or use [initial questionnaire web form](http://arctos.database.museum/new_collection.cfm).
Record Collection Contact Information and initial conversations.
|
process
|
ohio wesleyan university first contact post initial contact email or use record collection contact information and initial conversations
| 1
|
3,499
| 6,557,206,170
|
IssuesEvent
|
2017-09-06 16:33:13
|
nion-software/nionswift
|
https://api.github.com/repos/nion-software/nionswift
|
opened
|
Snapshot should produce equivalent display data
|
f - displays f - processing p3 - medium type - bug w4 - ready
|
For instance, if you have a sequence it should produce the displayed item in the sequence. If you have a display slice, it should produce the data associated with the display slice.
This probably implies that snapshot will always produce a basic data item (1d, 2d, XY, etc.)
|
1.0
|
Snapshot should produce equivalent display data - For instance, if you have a sequence it should produce the displayed item in the sequence. If you have a display slice, it should produce the data associated with the display slice.
This probably implies that snapshot will always produce a basic data item (1d, 2d, XY, etc.)
|
process
|
snapshot should produce equivalent display data for instance if you have a sequence it should produce the displayed item in the sequence if you have a display slice it should produce the data associated with the display slice this probably implies that snapshot will always produce a basic data item xy etc
| 1
|
57,745
| 11,785,315,537
|
IssuesEvent
|
2020-03-17 10:05:00
|
erlang-ls/erlang_ls
|
https://api.github.com/repos/erlang-ls/erlang_ls
|
closed
|
Find references for type
|
code navigation
|
**Is your feature request related to a problem? Please describe.**
It's sometimes useful to find all places where a type is being used.
**Describe the solution you'd like**
Show a list of references of the type when placing the cursor in the usage or definition of the type.
**Describe alternatives you've considered**
N/A.
**Additional context**
N/A.
|
1.0
|
Find references for type - **Is your feature request related to a problem? Please describe.**
It's sometimes useful to find all places where a type is being used.
**Describe the solution you'd like**
Show a list of references of the type when placing the cursor in the usage or definition of the type.
**Describe alternatives you've considered**
N/A.
**Additional context**
N/A.
|
non_process
|
find references for type is your feature request related to a problem please describe it s sometimes useful to find all places where a type is being used describe the solution you d like show a list of references of the type when placing the cursor in the usage or definition of the type describe alternatives you ve considered n a additional context n a
| 0
|
565,260
| 16,758,018,375
|
IssuesEvent
|
2021-06-13 07:39:46
|
aviferdman/Workshop-on-Software-Engineering-Project
|
https://api.github.com/repos/aviferdman/Workshop-on-Software-Engineering-Project
|
closed
|
מימוש דרישה 5 - חסינות מפני נפילת המערכת בתרחישים לא רצויים
|
features priority- mid
|
[] - make sure relevant data isn't lost in case of disconnection from db
[] - make sure money isn't transferred in case connection with transaction system is lost
[] - shopping cart is saved in case of disconnection
[] - store frequently used information in cache
|
1.0
|
מימוש דרישה 5 - חסינות מפני נפילת המערכת בתרחישים לא רצויים - [] - make sure relevant data isn't lost in case of disconnection from db
[] - make sure money isn't transferred in case connection with transaction system is lost
[] - shopping cart is saved in case of disconnection
[] - store frequently used information in cache
|
non_process
|
מימוש דרישה חסינות מפני נפילת המערכת בתרחישים לא רצויים make sure relevant data isn t lost in case of disconnection from db make sure money isn t transferred in case connection with transaction system is lost shopping cart is saved in case of disconnection store frequently used information in cache
| 0
|
12,735
| 15,102,060,616
|
IssuesEvent
|
2021-02-08 08:28:42
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
Fix XSRF for Participant manager
|
Bug P0 Participant manager datastore Process: Tested QA Unknown backend
|
Setting security.xsrf.enabled=false in the Participant manager datastore application.properties file allows the Participant manager to make PUT requests successfully. It will not otherwise succeed. See #2332 for more context.
|
1.0
|
Fix XSRF for Participant manager - Setting security.xsrf.enabled=false in the Participant manager datastore application.properties file allows the Participant manager to make PUT requests successfully. It will not otherwise succeed. See #2332 for more context.
|
process
|
fix xsrf for participant manager setting security xsrf enabled false in the participant manager datastore application properties file allows the participant manager to make put requests successfully it will not otherwise succeed see for more context
| 1
|
22,507
| 31,559,004,810
|
IssuesEvent
|
2023-09-03 02:20:18
|
tdwg/hc
|
https://api.github.com/repos/tdwg/hc
|
opened
|
New Term - verbatimSiteDescriptions
|
Term - add normative Process - under public review Class - Event
|
## New term
* Submitter: Humboldt Extension Task Group
* Efficacy Justification (why is this term necessary?): Part of a package of terms in support of biological inventory data.
* Demand Justification (name at least two organizations that independently need this term): The Humboldt Extension Task Group proposing this term consists of numerous organizations.
* Stability Justification (what concerns are there that this might affect existing implementations?): None
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: None
Proposed attributes of the new term:
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): verbatimSiteDescriptions
* Term label (English, not normative): Verbatim Site Descriptions
* Organized in Class (e.g., Occurrence, Event, Location, Taxon): Event
* Definition of the term (normative): Original textual description of the site(s).
* Usage comments (recommendations regarding content, etc., not normative): Site refers to the location at which observations are made or samples/measurements are taken. The site can be at any level of hierarchy. Recommended best practice is to separate multiple values in a list with space vertical bar space ( | ).
* Examples (not normative): `Wet flatwoods | Wet depression surrounded by mesic longleaf pine flatwoods | Ground cover of thick Andropogon spp.`; `Sporobolus floridanus, Vaccinium spp, Rhynchospora spp.`; `Centella erecta, Panicum rigidulum`
* Refines (identifier of the broader term this term refines; normative): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
|
1.0
|
New Term - verbatimSiteDescriptions - ## New term
* Submitter: Humboldt Extension Task Group
* Efficacy Justification (why is this term necessary?): Part of a package of terms in support of biological inventory data.
* Demand Justification (name at least two organizations that independently need this term): The Humboldt Extension Task Group proposing this term consists of numerous organizations.
* Stability Justification (what concerns are there that this might affect existing implementations?): None
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: None
Proposed attributes of the new term:
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): verbatimSiteDescriptions
* Term label (English, not normative): Verbatim Site Descriptions
* Organized in Class (e.g., Occurrence, Event, Location, Taxon): Event
* Definition of the term (normative): Original textual description of the site(s).
* Usage comments (recommendations regarding content, etc., not normative): Site refers to the location at which observations are made or samples/measurements are taken. The site can be at any level of hierarchy. Recommended best practice is to separate multiple values in a list with space vertical bar space ( | ).
* Examples (not normative): `Wet flatwoods | Wet depression surrounded by mesic longleaf pine flatwoods | Ground cover of thick Andropogon spp.`; `Sporobolus floridanus, Vaccinium spp, Rhynchospora spp.`; `Centella erecta, Panicum rigidulum`
* Refines (identifier of the broader term this term refines; normative): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
|
process
|
new term verbatimsitedescriptions new term submitter humboldt extension task group efficacy justification why is this term necessary part of a package of terms in support of biological inventory data demand justification name at least two organizations that independently need this term the humboldt extension task group proposing this term consists of numerous organizations stability justification what concerns are there that this might affect existing implementations none implications for dwciri namespace does this change affect a dwciri term version none proposed attributes of the new term term name in lowercamelcase for properties uppercamelcase for classes verbatimsitedescriptions term label english not normative verbatim site descriptions organized in class e g occurrence event location taxon event definition of the term normative original textual description of the site s usage comments recommendations regarding content etc not normative site refers to the location at which observations are made or samples measurements are taken the site can be at any level of hierarchy recommended best practice is to separate multiple values in a list with space vertical bar space examples not normative wet flatwoods wet depression surrounded by mesic longleaf pine flatwoods ground cover of thick andropogon spp sporobolus floridanus vaccinium spp rhynchospora spp centella erecta panicum rigidulum refines identifier of the broader term this term refines normative none replaces identifier of the existing term that would be deprecated and replaced by this term normative none abcd xpath of the equivalent term in abcd or efg not normative not in abcd
| 1
|
12,124
| 14,740,796,433
|
IssuesEvent
|
2021-01-07 09:38:35
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Newsletter Annoucment
|
anc-process anp-0.5 ant-support
|
In GitLab by @kdjstudios on Dec 4, 2018, 08:46
Hello Team,
We need to review with Gary setting up a newsletter announcement. This will insure our clients and internal users are made aware of any changes or modifications to SA Billing. The topics and concerns are as follows:
- Post and/or Pre release announcements
- What schedule do they announcements need to follow?
- What items are announced and how much detail we will provide?
- Will this be a feature built into SAB or shall we setup a separate email address specifically for this?
- What is the format we are looking for on these announcements?
- Do we want to make any new update or modification able for the user to turn on or off?
- Any other topics we will address as they are mentioned.
|
1.0
|
Newsletter Annoucment - In GitLab by @kdjstudios on Dec 4, 2018, 08:46
Hello Team,
We need to review with Gary setting up a newsletter announcement. This will insure our clients and internal users are made aware of any changes or modifications to SA Billing. The topics and concerns are as follows:
- Post and/or Pre release announcements
- What schedule do they announcements need to follow?
- What items are announced and how much detail we will provide?
- Will this be a feature built into SAB or shall we setup a separate email address specifically for this?
- What is the format we are looking for on these announcements?
- Do we want to make any new update or modification able for the user to turn on or off?
- Any other topics we will address as they are mentioned.
|
process
|
newsletter annoucment in gitlab by kdjstudios on dec hello team we need to review with gary setting up a newsletter announcement this will insure our clients and internal users are made aware of any changes or modifications to sa billing the topics and concerns are as follows post and or pre release announcements what schedule do they announcements need to follow what items are announced and how much detail we will provide will this be a feature built into sab or shall we setup a separate email address specifically for this what is the format we are looking for on these announcements do we want to make any new update or modification able for the user to turn on or off any other topics we will address as they are mentioned
| 1
|
126,983
| 26,958,453,796
|
IssuesEvent
|
2023-02-08 16:25:23
|
nopSolutions/nopCommerce
|
https://api.github.com/repos/nopSolutions/nopCommerce
|
closed
|
Default svg image is scaled large at many places
|
refactoring / source code
|
nopCommerce version: 4.60.1
Recently view image is scaled when svg default image used
Steps to reproduce the problem:
1 Upload SVG image in /Admin/Setting/Media > Product tab > Default image
2 Dont upload main product image that you viewed recently(Make sure /Admin/Setting/Catalog > Additional section > 'Recently viewed products' enabled is checked )
3 Observe recently view product image in Grid page below Filters
attached sample svg

Expected
Recently view image should Not be scaled when svg default image used

|
1.0
|
Default svg image is scaled large at many places - nopCommerce version: 4.60.1
Recently view image is scaled when svg default image used
Steps to reproduce the problem:
1 Upload SVG image in /Admin/Setting/Media > Product tab > Default image
2 Dont upload main product image that you viewed recently(Make sure /Admin/Setting/Catalog > Additional section > 'Recently viewed products' enabled is checked )
3 Observe recently view product image in Grid page below Filters
attached sample svg

Expected
Recently view image should Not be scaled when svg default image used

|
non_process
|
default svg image is scaled large at many places nopcommerce version recently view image is scaled when svg default image used steps to reproduce the problem upload svg image in admin setting media product tab default image dont upload main product image that you viewed recently make sure admin setting catalog additional section recently viewed products enabled is checked observe recently view product image in grid page below filters attached sample svg expected recently view image should not be scaled when svg default image used
| 0
|
4,467
| 7,332,724,583
|
IssuesEvent
|
2018-03-05 17:06:51
|
cedardevs/psi
|
https://api.github.com/repos/cedardevs/psi
|
closed
|
IsoLite parsing happens in script-wrapper
|
EPIC: PSI DSCOVR datastream psi-processor ready script-wrapper
|
Instead of a topic container IsoLite and a service to transform those messages into the OS ES json payload, we want all of that to happen in one service.
|
1.0
|
IsoLite parsing happens in script-wrapper - Instead of a topic container IsoLite and a service to transform those messages into the OS ES json payload, we want all of that to happen in one service.
|
process
|
isolite parsing happens in script wrapper instead of a topic container isolite and a service to transform those messages into the os es json payload we want all of that to happen in one service
| 1
|
4,671
| 7,511,224,285
|
IssuesEvent
|
2018-04-11 05:28:07
|
AmpersandTarski/Ampersand
|
https://api.github.com/repos/AmpersandTarski/Ampersand
|
closed
|
Installation of the Ampersand Compiler on OSX
|
priority:normal software process
|
I am trying to install an Ampersand compiler on a MacBook running OS X. I have opened a terminal (/bin/bash) and installed stack. I have also cloned the Ampersand git repository in ~/git/Ampersand. This is where Ampersand.cabal is found.

As you can see, `stack install` starts as expected.
After some time, the process comes to a halt because of a configuration error. The system says that the C compiler cannot create executables.

Anyone got an idea what's wrong? Or should we ask @Oblosys ?
|
1.0
|
Installation of the Ampersand Compiler on OSX - I am trying to install an Ampersand compiler on a MacBook running OS X. I have opened a terminal (/bin/bash) and installed stack. I have also cloned the Ampersand git repository in ~/git/Ampersand. This is where Ampersand.cabal is found.

As you can see, `stack install` starts as expected.
After some time, the process comes to a halt because of a configuration error. The system says that the C compiler cannot create executables.

Anyone got an idea what's wrong? Or should we ask @Oblosys ?
|
process
|
installation of the ampersand compiler on osx i am trying to install an ampersand compiler on a macbook running os x i have opened a terminal bin bash and installed stack i have also cloned the ampersand git repository in git ampersand this is where ampersand cabal is found as you can see stack install starts as expected after some time the process comes to a halt because of a configuration error the system says that the c compiler cannot create executables anyone got an idea what s wrong or should we ask oblosys
| 1
|
144,814
| 22,571,185,786
|
IssuesEvent
|
2022-06-28 00:20:58
|
openrewrite/rewrite
|
https://api.github.com/repos/openrewrite/rewrite
|
closed
|
Javadoc printing is not correctly handling an edge case when the `@author` tag is used.
|
bug design-partner-1
|
```
/**
* @author FirstName LastName 42
*
*/
public class A {
void method() {}
}
```
Is being printed as:
```
/**
* @author FirstName LastName 42
**/
public class A {
void method() {}
}"
```
|
1.0
|
Javadoc printing is not correctly handling an edge case when the `@author` tag is used. - ```
/**
* @author FirstName LastName 42
*
*/
public class A {
void method() {}
}
```
Is being printed as:
```
/**
* @author FirstName LastName 42
**/
public class A {
void method() {}
}"
```
|
non_process
|
javadoc printing is not correctly handling an edge case when the author tag is used author firstname lastname public class a void method is being printed as author firstname lastname public class a void method
| 0
|
451,721
| 13,040,551,556
|
IssuesEvent
|
2020-07-28 18:43:01
|
OpenSRP/opensrp-client-chw-anc
|
https://api.github.com/repos/OpenSRP/opensrp-client-chw-anc
|
opened
|
Once registered for ANC, the woman's location and ID change
|
BA-specific High Priority bug
|
### Issue
When a woman is registered for ANC, her profile is updated to the ANC profile. When this occurs, the locations changes to Lavingone #445 and the ID changes to 123456 for ALL profiles.
Profile before ANC Registration

Profile after ANC Registration

## To replicate
1. Add a woman to a family
2. Look at her ID and location in the text under her avatar
2. Register her for ANC
3. Open her profile
|
1.0
|
Once registered for ANC, the woman's location and ID change - ### Issue
When a woman is registered for ANC, her profile is updated to the ANC profile. When this occurs, the locations changes to Lavingone #445 and the ID changes to 123456 for ALL profiles.
Profile before ANC Registration

Profile after ANC Registration

## To replicate
1. Add a woman to a family
2. Look at her ID and location in the text under her avatar
2. Register her for ANC
3. Open her profile
|
non_process
|
once registered for anc the woman s location and id change issue when a woman is registered for anc her profile is updated to the anc profile when this occurs the locations changes to lavingone and the id changes to for all profiles profile before anc registration profile after anc registration to replicate add a woman to a family look at her id and location in the text under her avatar register her for anc open her profile
| 0
|
9,299
| 12,309,843,707
|
IssuesEvent
|
2020-05-12 09:37:00
|
threefoldtech/home
|
https://api.github.com/repos/threefoldtech/home
|
closed
|
GE: Deploying min.io/S3 does not complete
|
process_duplicate type_bug
|
I created a Network, and imported the details into Wireguard, then attempted deployment of an min.io S3 solution. I tried this on both the Free farm and GreenEdge-Vienna2. Both resulted in a the following screen spinning forever...


|
1.0
|
GE: Deploying min.io/S3 does not complete - I created a Network, and imported the details into Wireguard, then attempted deployment of an min.io S3 solution. I tried this on both the Free farm and GreenEdge-Vienna2. Both resulted in a the following screen spinning forever...


|
process
|
ge deploying min io does not complete i created a network and imported the details into wireguard then attempted deployment of an min io solution i tried this on both the free farm and greenedge both resulted in a the following screen spinning forever
| 1
|
10,633
| 13,441,586,869
|
IssuesEvent
|
2020-09-08 04:36:49
|
LLNL/axom
|
https://api.github.com/repos/LLNL/axom
|
opened
|
Add check for clang-format version 10
|
Software process
|
Our clang-format config file, introduced in #291 depends on clang-10.
We should add a check to ensure that the version of clang is at least 10.
(This was suggested by @gzagaris)
|
1.0
|
Add check for clang-format version 10 - Our clang-format config file, introduced in #291 depends on clang-10.
We should add a check to ensure that the version of clang is at least 10.
(This was suggested by @gzagaris)
|
process
|
add check for clang format version our clang format config file introduced in depends on clang we should add a check to ensure that the version of clang is at least this was suggested by gzagaris
| 1
|
705,703
| 24,245,285,761
|
IssuesEvent
|
2022-09-27 10:02:10
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
Bluetooth: controller: llcp: limited nr. of simultaneous connections
|
bug priority: low area: Bluetooth area: Bluetooth Controller area: Bluetooth LLCP
|
**Describe the bug**
the babblesim test tests/bluetooth/bsim_bt/bsim_test_multiple fails if the maximum number of connections is higher than circa 170.
The test should pass with a maximum number of connections set to 250
**To Reproduce**
1. change the prj.conf file, so that CONFIG_BT_MAX_CONN=250
2. run the script 'compile.sh' or alternatively only compile the bsim_test_multiple sample
3. run the script tests/bluetooth/bsim_bt/bsim_test_multiple/test_scripts/multiple.sh
4. there will be error-messages saying that connecting failed after about 170 succesfull connections
Possible cause might be related to the allocated buffers, using an uint8_t instead of uint16_t causing memory corruption
|
1.0
|
Bluetooth: controller: llcp: limited nr. of simultaneous connections - **Describe the bug**
the babblesim test tests/bluetooth/bsim_bt/bsim_test_multiple fails if the maximum number of connections is higher than circa 170.
The test should pass with a maximum number of connections set to 250
**To Reproduce**
1. change the prj.conf file, so that CONFIG_BT_MAX_CONN=250
2. run the script 'compile.sh' or alternatively only compile the bsim_test_multiple sample
3. run the script tests/bluetooth/bsim_bt/bsim_test_multiple/test_scripts/multiple.sh
4. there will be error-messages saying that connecting failed after about 170 succesfull connections
Possible cause might be related to the allocated buffers, using an uint8_t instead of uint16_t causing memory corruption
|
non_process
|
bluetooth controller llcp limited nr of simultaneous connections describe the bug the babblesim test tests bluetooth bsim bt bsim test multiple fails if the maximum number of connections is higher than circa the test should pass with a maximum number of connections set to to reproduce change the prj conf file so that config bt max conn run the script compile sh or alternatively only compile the bsim test multiple sample run the script tests bluetooth bsim bt bsim test multiple test scripts multiple sh there will be error messages saying that connecting failed after about succesfull connections possible cause might be related to the allocated buffers using an t instead of t causing memory corruption
| 0
|
3,168
| 6,224,089,584
|
IssuesEvent
|
2017-07-10 13:33:53
|
dzhw/zofar
|
https://api.github.com/repos/dzhw/zofar
|
opened
|
missing definition
|
category: service.processes prio: 9999 status: discussion type: backlog.item
|
**Participant:**
- not involved
**User:**
- clear definition on missing values
**Service:**
- clear definition on missing values
**Dev:**
- not involved
|
1.0
|
missing definition - **Participant:**
- not involved
**User:**
- clear definition on missing values
**Service:**
- clear definition on missing values
**Dev:**
- not involved
|
process
|
missing definition participant not involved user clear definition on missing values service clear definition on missing values dev not involved
| 1
|
783,388
| 27,528,383,373
|
IssuesEvent
|
2023-03-06 19:56:10
|
googleapis/nodejs-storage
|
https://api.github.com/repos/googleapis/nodejs-storage
|
closed
|
`file.acl.get` malformed request if `entity` is not provided
|
type: docs api: storage priority: p3
|
#### Environment details
- OS: Alpine
- Node.js version: 18.12.0
- npm version: 8.19.2
- `@google-cloud/storage` version: 6.9.0
#### Steps to reproduce
1. Create a file in a bucket.
2. Run
```
let fileAcls = [];
await file.acl.get().then((data) => {fileAcls = data[0]});
console.log(fileAcls);
```
3. Note that, while entity ([a required option!](https://github.com/googleapis/nodejs-storage/blob/9741a7a2acb2d4b75020d79e34a7d661ab866c63/src/acl.ts#L42)) was not provided, we listed all ACLs for the file.
4. Change your bucket to requester pays.
5. Run
```
let fileAcls = [];
await file.acl.get({userProject: "YOUR_PROJECT_HERE"}).then((data) => {fileAcls = data[0]});
```
6. Note malformed request error.
The reason for this is [here](https://github.com/googleapis/nodejs-storage/blob/main/src/acl.ts#L719). If `entity` is not provided but any other option is, then our path is `/undefined`. On the other hand, if no options at all are provided, `entity` effectively defaults to `''`. The issue can straightforwardly be worked around by providing the option `{entity: ''}`.
I propose one of two solutions:
1. If entity is not provided, default `entity` to empty string. This effectively makes `entity` not a required field.
2. Enforce `entity` as a required field. Fail `file.acl.get(...)` if `entity` is not provided. Document that users wishing to retrieve all ACLs should pass `{entity: ''}`.
|
1.0
|
`file.acl.get` malformed request if `entity` is not provided - #### Environment details
- OS: Alpine
- Node.js version: 18.12.0
- npm version: 8.19.2
- `@google-cloud/storage` version: 6.9.0
#### Steps to reproduce
1. Create a file in a bucket.
2. Run
```
let fileAcls = [];
await file.acl.get().then((data) => {fileAcls = data[0]});
console.log(fileAcls);
```
3. Note that, while entity ([a required option!](https://github.com/googleapis/nodejs-storage/blob/9741a7a2acb2d4b75020d79e34a7d661ab866c63/src/acl.ts#L42)) was not provided, we listed all ACLs for the file.
4. Change your bucket to requester pays.
5. Run
```
let fileAcls = [];
await file.acl.get({userProject: "YOUR_PROJECT_HERE"}).then((data) => {fileAcls = data[0]});
```
6. Note malformed request error.
The reason for this is [here](https://github.com/googleapis/nodejs-storage/blob/main/src/acl.ts#L719). If `entity` is not provided but any other option is, then our path is `/undefined`. On the other hand, if no options at all are provided, `entity` effectively defaults to `''`. The issue can straightforwardly be worked around by providing the option `{entity: ''}`.
I propose one of two solutions:
1. If entity is not provided, default `entity` to empty string. This effectively makes `entity` not a required field.
2. Enforce `entity` as a required field. Fail `file.acl.get(...)` if `entity` is not provided. Document that users wishing to retrieve all ACLs should pass `{entity: ''}`.
|
non_process
|
file acl get malformed request if entity is not provided environment details os alpine node js version npm version google cloud storage version steps to reproduce create a file in a bucket run let fileacls await file acl get then data fileacls data console log fileacls note that while entity was not provided we listed all acls for the file change your bucket to requester pays run let fileacls await file acl get userproject your project here then data fileacls data note malformed request error the reason for this is if entity is not provided but any other option is then our path is undefined on the other hand if no options at all are provided entity effectively defaults to the issue can straightforwardly be worked around by providing the option entity i propose one of two solutions if entity is not provided default entity to empty string this effectively makes entity not a required field enforce entity as a required field fail file acl get if entity is not provided document that users wishing to retrieve all acls should pass entity
| 0
|
12,281
| 14,791,354,588
|
IssuesEvent
|
2021-01-12 13:23:28
|
weberlab-hhu/Helixer
|
https://api.github.com/repos/weberlab-hhu/Helixer
|
opened
|
Move overlapping to seperate post-processing script
|
post-processing
|
Should have been done like this from the beginning
|
1.0
|
Move overlapping to seperate post-processing script - Should have been done like this from the beginning
|
process
|
move overlapping to seperate post processing script should have been done like this from the beginning
| 1
|
18,738
| 24,642,700,055
|
IssuesEvent
|
2022-10-17 12:51:51
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
junit_test is incompatible with JDK 18+ - "java.lang.UnsupportedOperationException: The Security Manager is deprecated"
|
type: support / not a bug (process) team-Rules-Java untriaged
|
### Description of the bug:
Running a trivial `java_test` with JDK 18+ yields the following (with the test failing):
```
JUnit4 Test Runner
java.lang.UnsupportedOperationException: The Security Manager is deprecated and will be removed in a future release
at java.base/java.lang.System.setSecurityManager(System.java:418)
at com.google.testing.junit.runner.junit4.JUnit4Runner.installSecurityManager(JUnit4Runner.java:256)
at com.google.testing.junit.runner.junit4.JUnit4Runner.run(JUnit4Runner.java:113)
at com.google.testing.junit.runner.BazelTestRunner.runTestsInSuite(BazelTestRunner.java:148)
at com.google.testing.junit.runner.BazelTestRunner.main(BazelTestRunner.java:75)
BazelTestRunner exiting with a return value of 1
```
Having JDK 17 (or older) on the PATH fixes the issue above.
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
#### Sources
`WORKSPACE.bazel`:
```
(Empty)
```
`java/example/BUILD.bazel`:
```python
java_test(
name = "ExampleTest",
srcs = ["ExampleTest.java"],
)
```
`java/example/ExampleTest.java`:
```java
package example;
import org.junit.Test;
public class ExampleTest {
@Test
public void passes() {
}
}
```
#### Repro command
```bash
$ java --version
openjdk 18.0.1
[...]
$ bazel test ... --test_output=errors
[...]
JUnit4 Test Runner
java.lang.UnsupportedOperationException: The Security Manager is deprecated and will be removed in a future release
at java.base/java.lang.System.setSecurityManager(System.java:418)
at com.google.testing.junit.runner.junit4.JUnit4Runner.installSecurityManager(JUnit4Runner.java:256)
at com.google.testing.junit.runner.junit4.JUnit4Runner.run(JUnit4Runner.java:113)
at com.google.testing.junit.runner.BazelTestRunner.runTestsInSuite(BazelTestRunner.java:148)
at com.google.testing.junit.runner.BazelTestRunner.main(BazelTestRunner.java:75)
BazelTestRunner exiting with a return value of 1
[...]
INFO: Build completed, 1 test FAILED
[...]
```
### Which operating system are you running Bazel on?=
Macos 12.6
### What is the output of `bazel info release`?
release 5.1.1
|
1.0
|
junit_test is incompatible with JDK 18+ - "java.lang.UnsupportedOperationException: The Security Manager is deprecated" - ### Description of the bug:
Running a trivial `java_test` with JDK 18+ yields the following (with the test failing):
```
JUnit4 Test Runner
java.lang.UnsupportedOperationException: The Security Manager is deprecated and will be removed in a future release
at java.base/java.lang.System.setSecurityManager(System.java:418)
at com.google.testing.junit.runner.junit4.JUnit4Runner.installSecurityManager(JUnit4Runner.java:256)
at com.google.testing.junit.runner.junit4.JUnit4Runner.run(JUnit4Runner.java:113)
at com.google.testing.junit.runner.BazelTestRunner.runTestsInSuite(BazelTestRunner.java:148)
at com.google.testing.junit.runner.BazelTestRunner.main(BazelTestRunner.java:75)
BazelTestRunner exiting with a return value of 1
```
Having JDK 17 (or older) on the PATH fixes the issue above.
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
#### Sources
`WORKSPACE.bazel`:
```
(Empty)
```
`java/example/BUILD.bazel`:
```python
java_test(
name = "ExampleTest",
srcs = ["ExampleTest.java"],
)
```
`java/example/ExampleTest.java`:
```java
package example;
import org.junit.Test;
public class ExampleTest {
@Test
public void passes() {
}
}
```
#### Repro command
```bash
$ java --version
openjdk 18.0.1
[...]
$ bazel test ... --test_output=errors
[...]
JUnit4 Test Runner
java.lang.UnsupportedOperationException: The Security Manager is deprecated and will be removed in a future release
at java.base/java.lang.System.setSecurityManager(System.java:418)
at com.google.testing.junit.runner.junit4.JUnit4Runner.installSecurityManager(JUnit4Runner.java:256)
at com.google.testing.junit.runner.junit4.JUnit4Runner.run(JUnit4Runner.java:113)
at com.google.testing.junit.runner.BazelTestRunner.runTestsInSuite(BazelTestRunner.java:148)
at com.google.testing.junit.runner.BazelTestRunner.main(BazelTestRunner.java:75)
BazelTestRunner exiting with a return value of 1
[...]
INFO: Build completed, 1 test FAILED
[...]
```
### Which operating system are you running Bazel on?=
Macos 12.6
### What is the output of `bazel info release`?
release 5.1.1
|
process
|
junit test is incompatible with jdk java lang unsupportedoperationexception the security manager is deprecated description of the bug running a trivial java test with jdk yields the following with the test failing test runner java lang unsupportedoperationexception the security manager is deprecated and will be removed in a future release at java base java lang system setsecuritymanager system java at com google testing junit runner installsecuritymanager java at com google testing junit runner run java at com google testing junit runner bazeltestrunner runtestsinsuite bazeltestrunner java at com google testing junit runner bazeltestrunner main bazeltestrunner java bazeltestrunner exiting with a return value of having jdk or older on the path fixes the issue above what s the simplest easiest way to reproduce this bug please provide a minimal example if possible sources workspace bazel empty java example build bazel python java test name exampletest srcs java example exampletest java java package example import org junit test public class exampletest test public void passes repro command bash java version openjdk bazel test test output errors test runner java lang unsupportedoperationexception the security manager is deprecated and will be removed in a future release at java base java lang system setsecuritymanager system java at com google testing junit runner installsecuritymanager java at com google testing junit runner run java at com google testing junit runner bazeltestrunner runtestsinsuite bazeltestrunner java at com google testing junit runner bazeltestrunner main bazeltestrunner java bazeltestrunner exiting with a return value of info build completed test failed which operating system are you running bazel on macos what is the output of bazel info release release
| 1
|
16,815
| 22,060,919,840
|
IssuesEvent
|
2022-05-30 17:41:58
|
bitPogo/kmock
|
https://api.github.com/repos/bitPogo/kmock
|
closed
|
Don't expose `kspy` needlessly
|
enhancement kmock-processor
|
## Description
<!--- Provide a detailed introduction to the issue itself, and why you consider it to be a bug -->
Currently `kspy` is generated always. Even if it is just a minor issue, it would be nice if kspy is not generated at all when not in use.
|
1.0
|
Don't expose `kspy` needlessly - ## Description
<!--- Provide a detailed introduction to the issue itself, and why you consider it to be a bug -->
Currently `kspy` is generated always. Even if it is just a minor issue, it would be nice if kspy is not generated at all when not in use.
|
process
|
don t expose kspy needlessly description currently kspy is generated always even if it is just a minor issue it would be nice if kspy is not generated at all when not in use
| 1
|
77,983
| 10,028,710,543
|
IssuesEvent
|
2019-07-17 12:22:53
|
Onkarn92/facts
|
https://api.github.com/repos/Onkarn92/facts
|
opened
|
Create user-interface for fact-list item.
|
documentation enhancement
|
1. Use recycler-view and card-view.
2. Item layout should be constructed using constraint-layout.
|
1.0
|
Create user-interface for fact-list item. - 1. Use recycler-view and card-view.
2. Item layout should be constructed using constraint-layout.
|
non_process
|
create user interface for fact list item use recycler view and card view item layout should be constructed using constraint layout
| 0
|
152,072
| 5,832,438,188
|
IssuesEvent
|
2017-05-08 21:49:30
|
duckduckgo/zeroclickinfo-spice
|
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
|
closed
|
npm: Improve detail view UX
|
Category: Highest Impact Tasks Improvement Priority: High Status: Work In Progress Suggestion Topic: Other
|
## Problem
<!-- Describe the bug or suggestion in detail -->
In #3257 we added the single detail view back to improve engagement. But we need to make some changes to improve the UX.

## Solution
<!-- Describe the steps, or provide a link to an example search -->
1. Add the description from the api call
2. Remove the separator beside the _More at_ link
3. Send a PR
## People to notify
<!-- Please @mention any relevant people/organizations here:-->
@duckduckgo/duckduckhack-contributors
## Get Started
- [x] 1) Claim this issue by commenting below
- [x] 2) Review our [Contributing Guide](https://github.com/duckduckgo/zeroclickinfo-goodies/blob/master/CONTRIBUTING.md)
- [x] 3) [Set up your development environment](https://docs.duckduckhack.com/welcome/setup-dev-environment.html), and fork this repository
- [ ] 4) Create a Pull Request
## Resources
- Join [DuckDuckHack Slack](https://quackslack.herokuapp.com/) to ask questions
- Join the [DuckDuckHack Forum](https://forum.duckduckhack.com/) to discuss project planning and Instant Answer metrics
- Read the [DuckDuckHack Documentation](https://docs.duckduckhack.com/) for technical help
<!-- DO NOT REMOVE -->
---
<!-- The Instant Answer ID can be found by clicking the `?` icon beside the Instant Answer result on DuckDuckGo.com -->
Instant Answer Page: https://duck.co/ia/view/npm
<!-- FILL THIS IN: ^^^^ -->
|
1.0
|
npm: Improve detail view UX - ## Problem
<!-- Describe the bug or suggestion in detail -->
In #3257 we added the single detail view back to improve engagement. But we need to make some changes to improve the UX.

## Solution
<!-- Describe the steps, or provide a link to an example search -->
1. Add the description from the api call
2. Remove the separator beside the _More at_ link
3. Send a PR
## People to notify
<!-- Please @mention any relevant people/organizations here:-->
@duckduckgo/duckduckhack-contributors
## Get Started
- [x] 1) Claim this issue by commenting below
- [x] 2) Review our [Contributing Guide](https://github.com/duckduckgo/zeroclickinfo-goodies/blob/master/CONTRIBUTING.md)
- [x] 3) [Set up your development environment](https://docs.duckduckhack.com/welcome/setup-dev-environment.html), and fork this repository
- [ ] 4) Create a Pull Request
## Resources
- Join [DuckDuckHack Slack](https://quackslack.herokuapp.com/) to ask questions
- Join the [DuckDuckHack Forum](https://forum.duckduckhack.com/) to discuss project planning and Instant Answer metrics
- Read the [DuckDuckHack Documentation](https://docs.duckduckhack.com/) for technical help
<!-- DO NOT REMOVE -->
---
<!-- The Instant Answer ID can be found by clicking the `?` icon beside the Instant Answer result on DuckDuckGo.com -->
Instant Answer Page: https://duck.co/ia/view/npm
<!-- FILL THIS IN: ^^^^ -->
|
non_process
|
npm improve detail view ux problem in we added the single detail view back to improve engagement but we need to make some changes to improve the ux solution add the description from the api call remove the separator beside the more at link send a pr people to notify duckduckgo duckduckhack contributors get started claim this issue by commenting below review our and fork this repository create a pull request resources join to ask questions join the to discuss project planning and instant answer metrics read the for technical help instant answer page
| 0
|
145,340
| 22,666,337,344
|
IssuesEvent
|
2022-07-03 00:20:15
|
aws/aws-cdk
|
https://api.github.com/repos/aws/aws-cdk
|
closed
|
Staged S3 assets might be maliciously replaced before being used
|
p1 feature-request @aws-cdk/aws-s3 effort/medium needs-design closed-for-staleness
|
CDK uses a shared "staging bucket" that is created for an account as part of `cdk bootstrap`. I believe the intention is to be able to use this bucket for all staged assets across all stacks in the same account/region. However, unless all these stacks share the same security level, this creates a vulnerability window where a malicious actor might replace the contents of an asset after it has been uploaded but before it is used by CloudFormation. This means that through the assets the malicious user can get code running with the credentials of the original user.
### Reproduction Steps
- User starts `cdk deploy`
- CDK uploads assets to staging bucket
- Malicious user notices uploaded asset and replaces the contents with their own
- CDK creates CloudFormation change set
- CloudFormation reads the modified contents of the assets instead of the original
### Environment
- **CLI Version :** 1.12.0
- **Framework Version:** 1.12.0
- **OS :** Linux
- **Language :** TypeScript
### Other
The use of a shared staging bucket should not cause any security problems for the user. The way to achieve this is to ensure the staging bucket has versioning enabled, and meticulously use the "version id" obtained during asset upload in every usage.
- CloudFormation supports adding `?versionId=foobar` to the template URL, at stack create, update and nested stacks.
- Lambda supports specifying code and layers with `S3Bucket`, `S3Key` and `S3ObjectVersion`
After this, the worst a malicious user with access to the bucket can do is DeleteObjectVersion on the object before it is used, which is a denial of service problem only, and limiting access to that operation is simple, as it is not needed in normal operation unlike PutObject.
I believe Docker assets automatically use the sha256 hash from the image being pushed, so they do not share the same problem.
---
This is :bug: Bug Report
|
1.0
|
Staged S3 assets might be maliciously replaced before being used - CDK uses a shared "staging bucket" that is created for an account as part of `cdk bootstrap`. I believe the intention is to be able to use this bucket for all staged assets across all stacks in the same account/region. However, unless all these stacks share the same security level, this creates a vulnerability window where a malicious actor might replace the contents of an asset after it has been uploaded but before it is used by CloudFormation. This means that through the assets the malicious user can get code running with the credentials of the original user.
### Reproduction Steps
- User starts `cdk deploy`
- CDK uploads assets to staging bucket
- Malicious user notices uploaded asset and replaces the contents with their own
- CDK creates CloudFormation change set
- CloudFormation reads the modified contents of the assets instead of the original
### Environment
- **CLI Version :** 1.12.0
- **Framework Version:** 1.12.0
- **OS :** Linux
- **Language :** TypeScript
### Other
The use of a shared staging bucket should not cause any security problems for the user. The way to achieve this is to ensure the staging bucket has versioning enabled, and meticulously use the "version id" obtained during asset upload in every usage.
- CloudFormation supports adding `?versionId=foobar` to the template URL, at stack create, update and nested stacks.
- Lambda supports specifying code and layers with `S3Bucket`, `S3Key` and `S3ObjectVersion`
After this, the worst a malicious user with access to the bucket can do is DeleteObjectVersion on the object before it is used, which is a denial of service problem only, and limiting access to that operation is simple, as it is not needed in normal operation unlike PutObject.
I believe Docker assets automatically use the sha256 hash from the image being pushed, so they do not share the same problem.
---
This is :bug: Bug Report
|
non_process
|
staged assets might be maliciously replaced before being used cdk uses a shared staging bucket that is created for an account as part of cdk bootstrap i believe the intention is to be able to use this bucket for all staged assets across all stacks in the same account region however unless all these stacks share the same security level this creates a vulnerability window where a malicious actor might replace the contents of an asset after it has been uploaded but before it is used by cloudformation this means that through the assets the malicious user can get code running with the credentials of the original user reproduction steps user starts cdk deploy cdk uploads assets to staging bucket malicious user notices uploaded asset and replaces the contents with their own cdk creates cloudformation change set cloudformation reads the modified contents of the assets instead of the original environment cli version framework version os linux language typescript other the use of a shared staging bucket should not cause any security problems for the user the way to achieve this is to ensure the staging bucket has versioning enabled and meticulously use the version id obtained during asset upload in every usage cloudformation supports adding versionid foobar to the template url at stack create update and nested stacks lambda supports specifying code and layers with and after this the worst a malicious user with access to the bucket can do is deleteobjectversion on the object before it is used which is a denial of service problem only and limiting access to that operation is simple as it is not needed in normal operation unlike putobject i believe docker assets automatically use the hash from the image being pushed so they do not share the same problem this is bug bug report
| 0
|
10,171
| 13,044,162,744
|
IssuesEvent
|
2020-07-29 03:47:35
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `GetVar` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `GetVar` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `GetVar` from TiDB -
## Description
Port the scalar function `GetVar` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function getvar from tidb description port the scalar function getvar from tidb to coprocessor score mentor s breeswish recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
12,162
| 4,382,637,212
|
IssuesEvent
|
2016-08-07 01:18:36
|
langbakk/cntrl
|
https://api.github.com/repos/langbakk/cntrl
|
closed
|
CODEREVIEW - shiftlistreport needs a bit of a shiftnames etc. upgrade
|
codereview designchange
|
Currently, shiftnames aren't shown, hence there is a bit of ambiguity going on with shifts.
|
1.0
|
CODEREVIEW - shiftlistreport needs a bit of a shiftnames etc. upgrade - Currently, shiftnames aren't shown, hence there is a bit of ambiguity going on with shifts.
|
non_process
|
codereview shiftlistreport needs a bit of a shiftnames etc upgrade currently shiftnames aren t shown hence there is a bit of ambiguity going on with shifts
| 0
|
16,680
| 21,784,052,506
|
IssuesEvent
|
2022-05-13 23:11:52
|
MicrosoftDocs/windows-uwp
|
https://api.github.com/repos/MicrosoftDocs/windows-uwp
|
closed
|
Not applicable for VS 2022
|
uwp/prod processes-and-threading/tech Pri2
|
There is no template for a console app for C++/WinRT (UWP) in VS 2022.
[Enter feedback here]
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: e739c976-323f-bdf7-14a4-9b2d4872c62c
* Version Independent ID: c580c6b6-dc0d-296e-ffbd-0d866b34779f
* Content: [Create a Universal Windows Platform console app - UWP applications](https://docs.microsoft.com/en-us/windows/uwp/launch-resume/console-uwp)
* Content Source: [windows-apps-src/launch-resume/console-uwp.md](https://github.com/MicrosoftDocs/windows-uwp/blob/docs/windows-apps-src/launch-resume/console-uwp.md)
* Product: **uwp**
* Technology: **processes-and-threading**
* GitHub Login: @alvinashcraft
* Microsoft Alias: **aashcraft**
|
1.0
|
Not applicable for VS 2022 - There is no template for a console app for C++/WinRT (UWP) in VS 2022.
[Enter feedback here]
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: e739c976-323f-bdf7-14a4-9b2d4872c62c
* Version Independent ID: c580c6b6-dc0d-296e-ffbd-0d866b34779f
* Content: [Create a Universal Windows Platform console app - UWP applications](https://docs.microsoft.com/en-us/windows/uwp/launch-resume/console-uwp)
* Content Source: [windows-apps-src/launch-resume/console-uwp.md](https://github.com/MicrosoftDocs/windows-uwp/blob/docs/windows-apps-src/launch-resume/console-uwp.md)
* Product: **uwp**
* Technology: **processes-and-threading**
* GitHub Login: @alvinashcraft
* Microsoft Alias: **aashcraft**
|
process
|
not applicable for vs there is no template for a console app for c winrt uwp in vs document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id ffbd content content source product uwp technology processes and threading github login alvinashcraft microsoft alias aashcraft
| 1
|
136,909
| 20,017,503,078
|
IssuesEvent
|
2022-02-01 13:33:04
|
Elttob/Fusion
|
https://api.github.com/repos/Elttob/Fusion
|
opened
|
Hydrate by name / Unresolved child hydration
|
kind: enhancement status: needs design
|
Following on from #34 - the original issue where `Hydrate` was introduced. Alongside it, a separate construct was also proposed (tentatively called `WithChild`, open to bikeshedding later) which would allow for hydrating an child by name:
```Lua
local message = Value("Hello, world")
local ui = Hydrate (ReplicatedStorage.GameGui:Clone()) {
Parent = Players.LocalPlayer.PlayerGui,
[Children] = {
-- Here, we assume there exists some Message instance as a child of GameGui.
-- This will be located by Hydrate and have the property table applied.
WithChild "Message" {
Text = message
}
}
}
```
The original motive for implementing such a feature is to reduce the burden of working with deeply nested Hydrate trees:
```Lua
-- when dealing with New, the syntax is clean and your code has strong visual hierachy
local gui = New "Thing" {
[Children] = New "Thing" {
[Children] = New "Thing" {...}
}
}
-- dealing with Hydrate is much more verbose and lacks the same visual structure
local gui = PlayerGui.MyGui
Hydrate(gui) {
Value = ...
}
Hydrate(gui.Thing) {
Value = ...
}
Hydrate(gui.Thing.OtherThing) {
Value = ...
}
-- you could nest them but this doesn't solve the problem with indexing getting long
Hydrate(gui) {
[Children] = Hydrate(gui.Thing) {
[Children] = Hydrate(gui.Thing.OtherThing) {
Value = ...
}
}
}
-- this furthermore becomes quickly horrendous with FFC
Hydrate(gui) {
[Children] = Hydrate(gui:FindFirstChild("Thing")) {
[Children] = Hydrate(gui.:FindFirstChild("Thing"):FindFirstChild("OtherThing")) {
Value = ...
}
}
}
-- compared to unresolved children, which are much more in line with New in terms of complexity and structure
Hydrate(gui) {
[Children] = Hydrate "Thing" {
[Children] = Hydrate "OtherThing" {
Value = ...
}
}
}
```
There are some questions about whether this is a good idea to implement, or whether it's the correct problem to solve:
> I don't think you should make assumptions about how the user wants to traverse the tree (`FindFirstChild` vs `WaitForChild` for example). It also goes against the "everything is just an instance" model.
> When we say something like `WithChild "Thing"`, what do we mean? Is it the "Thing" at the current time, or whatever "Thing" will be at any point in time? The latter is more complicated, but also more Fusiony. Now it's no longer about FindFirstChild, but incorporating the names of children into the dependency graph. After that, we're suddenly thinking about incorporating any arbitrary criteria. Sounds a lot like CSS.
There were also some alternate propositions:
> How this might work:
>
> * A means to build a "selector". Might constructed out of symbols somehow, or could be parsed from a string (I hear you already have something like this).
> * Symbols that evaluate a selector, then apply properties to anything that matches. Has two flavors:
>
> * A "query" symbol that evaluates a selector once. Less expensive. Mainly used to target specific objects. This implements what we've been considering so far.
> * A "rule" symbol that creates a living selector. More expensive. Targets anything matching the selector at any time. Implements what I described above.
> * Existing symbols like Children could be used to limit the scope.
>
> The problem of getting an existing instance to bind to could be reinterpreted as a selector where the operand is an actual instance that must be matched. Or a selector that _is_ an instance just selects that instance.
Either way, while the idea seems to be palatable to Fusion users, the exact design is highly up for debate. For this reason, it was decided to ship `Hydrate` separately and leave this out until we could decide on a reasonable design direction.
Feel free to share your thoughts on this below - all thoughts are welcome, since this is still a large unresolved question :)
|
1.0
|
Hydrate by name / Unresolved child hydration - Following on from #34 - the original issue where `Hydrate` was introduced. Alongside it, a separate construct was also proposed (tentatively called `WithChild`, open to bikeshedding later) which would allow for hydrating an child by name:
```Lua
local message = Value("Hello, world")
local ui = Hydrate (ReplicatedStorage.GameGui:Clone()) {
Parent = Players.LocalPlayer.PlayerGui,
[Children] = {
-- Here, we assume there exists some Message instance as a child of GameGui.
-- This will be located by Hydrate and have the property table applied.
WithChild "Message" {
Text = message
}
}
}
```
The original motive for implementing such a feature is to reduce the burden of working with deeply nested Hydrate trees:
```Lua
-- when dealing with New, the syntax is clean and your code has strong visual hierachy
local gui = New "Thing" {
[Children] = New "Thing" {
[Children] = New "Thing" {...}
}
}
-- dealing with Hydrate is much more verbose and lacks the same visual structure
local gui = PlayerGui.MyGui
Hydrate(gui) {
Value = ...
}
Hydrate(gui.Thing) {
Value = ...
}
Hydrate(gui.Thing.OtherThing) {
Value = ...
}
-- you could nest them but this doesn't solve the problem with indexing getting long
Hydrate(gui) {
[Children] = Hydrate(gui.Thing) {
[Children] = Hydrate(gui.Thing.OtherThing) {
Value = ...
}
}
}
-- this furthermore becomes quickly horrendous with FFC
Hydrate(gui) {
[Children] = Hydrate(gui:FindFirstChild("Thing")) {
[Children] = Hydrate(gui.:FindFirstChild("Thing"):FindFirstChild("OtherThing")) {
Value = ...
}
}
}
-- compared to unresolved children, which are much more in line with New in terms of complexity and structure
Hydrate(gui) {
[Children] = Hydrate "Thing" {
[Children] = Hydrate "OtherThing" {
Value = ...
}
}
}
```
There are some questions about whether this is a good idea to implement, or whether it's the correct problem to solve:
> I don't think you should make assumptions about how the user wants to traverse the tree (`FindFirstChild` vs `WaitForChild` for example). It also goes against the "everything is just an instance" model.
> When we say something like `WithChild "Thing"`, what do we mean? Is it the "Thing" at the current time, or whatever "Thing" will be at any point in time? The latter is more complicated, but also more Fusiony. Now it's no longer about FindFirstChild, but incorporating the names of children into the dependency graph. After that, we're suddenly thinking about incorporating any arbitrary criteria. Sounds a lot like CSS.
There were also some alternate propositions:
> How this might work:
>
> * A means to build a "selector". Might constructed out of symbols somehow, or could be parsed from a string (I hear you already have something like this).
> * Symbols that evaluate a selector, then apply properties to anything that matches. Has two flavors:
>
> * A "query" symbol that evaluates a selector once. Less expensive. Mainly used to target specific objects. This implements what we've been considering so far.
> * A "rule" symbol that creates a living selector. More expensive. Targets anything matching the selector at any time. Implements what I described above.
> * Existing symbols like Children could be used to limit the scope.
>
> The problem of getting an existing instance to bind to could be reinterpreted as a selector where the operand is an actual instance that must be matched. Or a selector that _is_ an instance just selects that instance.
Either way, while the idea seems to be palatable to Fusion users, the exact design is highly up for debate. For this reason, it was decided to ship `Hydrate` separately and leave this out until we could decide on a reasonable design direction.
Feel free to share your thoughts on this below - all thoughts are welcome, since this is still a large unresolved question :)
|
non_process
|
hydrate by name unresolved child hydration following on from the original issue where hydrate was introduced alongside it a separate construct was also proposed tentatively called withchild open to bikeshedding later which would allow for hydrating an child by name lua local message value hello world local ui hydrate replicatedstorage gamegui clone parent players localplayer playergui here we assume there exists some message instance as a child of gamegui this will be located by hydrate and have the property table applied withchild message text message the original motive for implementing such a feature is to reduce the burden of working with deeply nested hydrate trees lua when dealing with new the syntax is clean and your code has strong visual hierachy local gui new thing new thing new thing dealing with hydrate is much more verbose and lacks the same visual structure local gui playergui mygui hydrate gui value hydrate gui thing value hydrate gui thing otherthing value you could nest them but this doesn t solve the problem with indexing getting long hydrate gui hydrate gui thing hydrate gui thing otherthing value this furthermore becomes quickly horrendous with ffc hydrate gui hydrate gui findfirstchild thing hydrate gui findfirstchild thing findfirstchild otherthing value compared to unresolved children which are much more in line with new in terms of complexity and structure hydrate gui hydrate thing hydrate otherthing value there are some questions about whether this is a good idea to implement or whether it s the correct problem to solve i don t think you should make assumptions about how the user wants to traverse the tree findfirstchild vs waitforchild for example it also goes against the everything is just an instance model when we say something like withchild thing what do we mean is it the thing at the current time or whatever thing will be at any point in time the latter is more complicated but also more fusiony now it s no longer about findfirstchild but incorporating the names of children into the dependency graph after that we re suddenly thinking about incorporating any arbitrary criteria sounds a lot like css there were also some alternate propositions how this might work a means to build a selector might constructed out of symbols somehow or could be parsed from a string i hear you already have something like this symbols that evaluate a selector then apply properties to anything that matches has two flavors a query symbol that evaluates a selector once less expensive mainly used to target specific objects this implements what we ve been considering so far a rule symbol that creates a living selector more expensive targets anything matching the selector at any time implements what i described above existing symbols like children could be used to limit the scope the problem of getting an existing instance to bind to could be reinterpreted as a selector where the operand is an actual instance that must be matched or a selector that is an instance just selects that instance either way while the idea seems to be palatable to fusion users the exact design is highly up for debate for this reason it was decided to ship hydrate separately and leave this out until we could decide on a reasonable design direction feel free to share your thoughts on this below all thoughts are welcome since this is still a large unresolved question
| 0
|
65,480
| 19,537,110,028
|
IssuesEvent
|
2021-12-31 09:59:10
|
dkfans/keeperfx
|
https://api.github.com/repos/dkfans/keeperfx
|
opened
|
Chickens escape from hatchery and wander around the map
|
Type-Defect
|
In recent versions, chickens can simply walk out of the hatchery.
To reproduce, make a hatchery next to a bridge and wait:

This saved game from r2557 does shows the issue:
[fx1g0004.zip](https://github.com/dkfans/keeperfx/files/7795858/fx1g0004.zip)
|
1.0
|
Chickens escape from hatchery and wander around the map - In recent versions, chickens can simply walk out of the hatchery.
To reproduce, make a hatchery next to a bridge and wait:

This saved game from r2557 does shows the issue:
[fx1g0004.zip](https://github.com/dkfans/keeperfx/files/7795858/fx1g0004.zip)
|
non_process
|
chickens escape from hatchery and wander around the map in recent versions chickens can simply walk out of the hatchery to reproduce make a hatchery next to a bridge and wait this saved game from does shows the issue
| 0
|
2,989
| 5,967,294,419
|
IssuesEvent
|
2017-05-30 15:40:28
|
openvstorage/framework
|
https://api.github.com/repos/openvstorage/framework
|
reopened
|
ovs-workers log show invalid literal for int() with base 10 during urlparse of memcache servers
|
process_cantreproduce
|
Environment: OVH
Only occurs on multiple host envs
```
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: [2016-12-20 09:14:39,454: ERROR/MainProcess] pidbox command error: ValueError("invalid literal for int() with base 10: '11211;172.20.20.101'",)
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: Traceback (most recent call last):
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/kombu/pidbox.py", line 105, in dispatch
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: reply = handle(method, kwdict(arguments))
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/kombu/pidbox.py", line 123, in handle_call
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return self.handle(method, arguments)
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/kombu/pidbox.py", line 120, in handle
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return self.handlers[method](self.state, **arguments)
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/celery/worker/control.py", line 377, in dump_conf
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return jsonify(state.app.conf.table(with_defaults=with_defaults),
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/celery/app/utils.py", line 130, in table
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: self if with_defaults else self.without_defaults())
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/celery/app/utils.py", line 197, in filter_hidden_settings
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return dict((k, maybe_censor(k, v)) for k, v in items(conf))
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/celery/app/utils.py", line 197, in <genexpr>
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return dict((k, maybe_censor(k, v)) for k, v in items(conf))
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/celery/app/utils.py", line 193, in maybe_censor
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return maybe_sanitize_url(value, mask=mask)
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/kombu/utils/url.py", line 63, in maybe_sanitize_url
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return sanitize_url(url, mask)
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/kombu/utils/url.py", line 58, in sanitize_url
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return as_url(*_parse_url(url), sanitize=True, mask=mask)
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/kombu/utils/url.py", line 24, in _parse_url
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return (scheme, unquote(parts.hostname or '') or None, parts.port,
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/urlparse.py", line 113, in port
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: port = int(port, 10)
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: ValueError: invalid literal for int() with base 10: '11211;172.20.20.101'
```
|
1.0
|
ovs-workers log show invalid literal for int() with base 10 during urlparse of memcache servers - Environment: OVH
Only occurs on multiple host envs
```
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: [2016-12-20 09:14:39,454: ERROR/MainProcess] pidbox command error: ValueError("invalid literal for int() with base 10: '11211;172.20.20.101'",)
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: Traceback (most recent call last):
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/kombu/pidbox.py", line 105, in dispatch
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: reply = handle(method, kwdict(arguments))
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/kombu/pidbox.py", line 123, in handle_call
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return self.handle(method, arguments)
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/kombu/pidbox.py", line 120, in handle
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return self.handlers[method](self.state, **arguments)
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/celery/worker/control.py", line 377, in dump_conf
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return jsonify(state.app.conf.table(with_defaults=with_defaults),
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/celery/app/utils.py", line 130, in table
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: self if with_defaults else self.without_defaults())
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/celery/app/utils.py", line 197, in filter_hidden_settings
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return dict((k, maybe_censor(k, v)) for k, v in items(conf))
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/celery/app/utils.py", line 197, in <genexpr>
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return dict((k, maybe_censor(k, v)) for k, v in items(conf))
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/celery/app/utils.py", line 193, in maybe_censor
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return maybe_sanitize_url(value, mask=mask)
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/kombu/utils/url.py", line 63, in maybe_sanitize_url
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return sanitize_url(url, mask)
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/kombu/utils/url.py", line 58, in sanitize_url
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return as_url(*_parse_url(url), sanitize=True, mask=mask)
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/dist-packages/kombu/utils/url.py", line 24, in _parse_url
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: return (scheme, unquote(parts.hostname or '') or None, parts.port,
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: File "/usr/lib/python2.7/urlparse.py", line 113, in port
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: port = int(port, 10)
Dec 20 09:14:39 ctrl-grav-04 celery[5694]: ValueError: invalid literal for int() with base 10: '11211;172.20.20.101'
```
|
process
|
ovs workers log show invalid literal for int with base during urlparse of memcache servers environment ovh only occurs on multiple host envs dec ctrl grav celery pidbox command error valueerror invalid literal for int with base dec ctrl grav celery traceback most recent call last dec ctrl grav celery file usr lib dist packages kombu pidbox py line in dispatch dec ctrl grav celery reply handle method kwdict arguments dec ctrl grav celery file usr lib dist packages kombu pidbox py line in handle call dec ctrl grav celery return self handle method arguments dec ctrl grav celery file usr lib dist packages kombu pidbox py line in handle dec ctrl grav celery return self handlers self state arguments dec ctrl grav celery file usr lib dist packages celery worker control py line in dump conf dec ctrl grav celery return jsonify state app conf table with defaults with defaults dec ctrl grav celery file usr lib dist packages celery app utils py line in table dec ctrl grav celery self if with defaults else self without defaults dec ctrl grav celery file usr lib dist packages celery app utils py line in filter hidden settings dec ctrl grav celery return dict k maybe censor k v for k v in items conf dec ctrl grav celery file usr lib dist packages celery app utils py line in dec ctrl grav celery return dict k maybe censor k v for k v in items conf dec ctrl grav celery file usr lib dist packages celery app utils py line in maybe censor dec ctrl grav celery return maybe sanitize url value mask mask dec ctrl grav celery file usr lib dist packages kombu utils url py line in maybe sanitize url dec ctrl grav celery return sanitize url url mask dec ctrl grav celery file usr lib dist packages kombu utils url py line in sanitize url dec ctrl grav celery return as url parse url url sanitize true mask mask dec ctrl grav celery file usr lib dist packages kombu utils url py line in parse url dec ctrl grav celery return scheme unquote parts hostname or or none parts port dec ctrl grav celery file usr lib urlparse py line in port dec ctrl grav celery port int port dec ctrl grav celery valueerror invalid literal for int with base
| 1
|
14,423
| 17,475,233,412
|
IssuesEvent
|
2021-08-08 01:46:11
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Incorrect CRS conversion
|
Feedback stale Processing Bug
|
When trying to use the processing algorithm "Clip Raster by Mask Layer" with a mask layer CRS being of type WGS84 pseudo-Mercator (EPSG:3857) and a raster CRS being of type World_Mollweide - Projected (ESRI:54009) then there seems to be a mismatch with the borders at the moment of projection. The WGS84 pseudo-Mercator (EPSG:3857) is used in order to easily define the sunbelt region between 40°N and 40°S.
I don't know if I am doing something completely wrong and hope this really is a bug report and not a bad user misunderstanding (otherwise I will put this question somewhere else). But my expectation is that the borders of the two layers with even different CRSes do coarsely end up at the same geological locations. Yet some huge areas are left out (green areas within the blue circles).

Steps to reproduce:
1. Load the vector layer [sunbelt_region.zip](https://github.com/qgis/QGIS/files/6693229/sunbelt_region.zip) and the raster layer [GHS_POP_E2015_GLOBE_R2019A_54009_1K.zip](https://cidportal.jrc.ec.europa.eu/ftp/jrc-opendata/GHSL/GHS_POP_MT_GLOBE_R2019A/GHS_POP_E2015_GLOBE_R2019A_54009_1K/V1-0/GHS_POP_E2015_GLOBE_R2019A_54009_1K_V1_0.zip) into the Layers list. For better visualisation you may want to make the vector layer half transparent and add XYZ map tiles.

2. Call the processing algorithm "Clip Raster by Mask Layer" and select the two layers with default options

3. To get the same visualisation select "Display no data as" and select a color

Even if I change the mask layer to CRS World_Mollweide in order to have mask layer and raster layer in accordance with each other nothing changes

It becomes clearer why if I **change the Project CRS to World_Mollweide** and see that the mask layer _sunbelt_region_ and _sunbelt_region_mollweide_ don't seem to be projected (correctly?) onto the project or destination CRS:

I don't know, maybe the projection/conversion/transformation I am expecting/missing here would be a feature or am I totally on the wrong path?
Regarding the QGIS version I am using:
QGIS version | 3.18.3-Zürich | QGIS code revision | 735cc85be9
-- | -- | -- | --
Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2
Compiled against GDAL/OGR | 3.1.4 | Running against GDAL/OGR | 3.1.4
Compiled against GEOS | 3.8.1-CAPI-1.13.3 | Running against GEOS | 3.8.1-CAPI-1.13.3
Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0
PostgreSQL Client Version | 11.5 | SpatiaLite Version | 4.3.0
QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8
Compiled against PROJ | 6.3.2 | Running against PROJ | Rel. 6.3.2, May 1st, 2020
OS Version | Windows 10 (10.0)
Active python plugins | GroupStats; latlontools; nominatim; QuickOSM; db_manager; processing
|
1.0
|
Incorrect CRS conversion - When trying to use the processing algorithm "Clip Raster by Mask Layer" with a mask layer CRS being of type WGS84 pseudo-Mercator (EPSG:3857) and a raster CRS being of type World_Mollweide - Projected (ESRI:54009) then there seems to be a mismatch with the borders at the moment of projection. The WGS84 pseudo-Mercator (EPSG:3857) is used in order to easily define the sunbelt region between 40°N and 40°S.
I don't know if I am doing something completely wrong and hope this really is a bug report and not a bad user misunderstanding (otherwise I will put this question somewhere else). But my expectation is that the borders of the two layers with even different CRSes do coarsely end up at the same geological locations. Yet some huge areas are left out (green areas within the blue circles).

Steps to reproduce:
1. Load the vector layer [sunbelt_region.zip](https://github.com/qgis/QGIS/files/6693229/sunbelt_region.zip) and the raster layer [GHS_POP_E2015_GLOBE_R2019A_54009_1K.zip](https://cidportal.jrc.ec.europa.eu/ftp/jrc-opendata/GHSL/GHS_POP_MT_GLOBE_R2019A/GHS_POP_E2015_GLOBE_R2019A_54009_1K/V1-0/GHS_POP_E2015_GLOBE_R2019A_54009_1K_V1_0.zip) into the Layers list. For better visualisation you may want to make the vector layer half transparent and add XYZ map tiles.

2. Call the processing algorithm "Clip Raster by Mask Layer" and select the two layers with default options

3. To get the same visualisation select "Display no data as" and select a color

Even if I change the mask layer to CRS World_Mollweide in order to have mask layer and raster layer in accordance with each other nothing changes

It becomes clearer why if I **change the Project CRS to World_Mollweide** and see that the mask layer _sunbelt_region_ and _sunbelt_region_mollweide_ don't seem to be projected (correctly?) onto the project or destination CRS:

I don't know, maybe the projection/conversion/transformation I am expecting/missing here would be a feature or am I totally on the wrong path?
Regarding the QGIS version I am using:
QGIS version | 3.18.3-Zürich | QGIS code revision | 735cc85be9
-- | -- | -- | --
Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2
Compiled against GDAL/OGR | 3.1.4 | Running against GDAL/OGR | 3.1.4
Compiled against GEOS | 3.8.1-CAPI-1.13.3 | Running against GEOS | 3.8.1-CAPI-1.13.3
Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0
PostgreSQL Client Version | 11.5 | SpatiaLite Version | 4.3.0
QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8
Compiled against PROJ | 6.3.2 | Running against PROJ | Rel. 6.3.2, May 1st, 2020
OS Version | Windows 10 (10.0)
Active python plugins | GroupStats; latlontools; nominatim; QuickOSM; db_manager; processing
|
process
|
incorrect crs conversion when trying to use the processing algorithm clip raster by mask layer with a mask layer crs being of type pseudo mercator epsg and a raster crs being of type world mollweide projected esri then there seems to be a mismatch with the borders at the moment of projection the pseudo mercator epsg is used in order to easily define the sunbelt region between °n and °s i don t know if i am doing something completely wrong and hope this really is a bug report and not a bad user misunderstanding otherwise i will put this question somewhere else but my expectation is that the borders of the two layers with even different crses do coarsely end up at the same geological locations yet some huge areas are left out green areas within the blue circles steps to reproduce load the vector layer and the raster layer into the layers list for better visualisation you may want to make the vector layer half transparent and add xyz map tiles call the processing algorithm clip raster by mask layer and select the two layers with default options to get the same visualisation select display no data as and select a color even if i change the mask layer to crs world mollweide in order to have mask layer and raster layer in accordance with each other nothing changes it becomes clearer why if i change the project crs to world mollweide and see that the mask layer sunbelt region and sunbelt region mollweide don t seem to be projected correctly onto the project or destination crs i don t know maybe the projection conversion transformation i am expecting missing here would be a feature or am i totally on the wrong path regarding the qgis version i am using qgis version zürich qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel may os version windows active python plugins groupstats latlontools nominatim quickosm db manager processing
| 1
|
12,851
| 15,238,459,944
|
IssuesEvent
|
2021-02-19 01:58:59
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Add possibility to choose between seperate line and polylines in QGIS "Split lines"
|
Feature Request Feedback Processing stale
|
Author Name: **Lene Fischer** (@LeneFischer)
Original Redmine Issue: [21534](https://issues.qgis.org/issues/21534)
Redmine category:processing/qgis
---
I have used the two functions which initially looks the same. I miss the possibility to choose between seperate line and polylines
|
1.0
|
Add possibility to choose between seperate line and polylines in QGIS "Split lines" - Author Name: **Lene Fischer** (@LeneFischer)
Original Redmine Issue: [21534](https://issues.qgis.org/issues/21534)
Redmine category:processing/qgis
---
I have used the two functions which initially looks the same. I miss the possibility to choose between seperate line and polylines
|
process
|
add possibility to choose between seperate line and polylines in qgis split lines author name lene fischer lenefischer original redmine issue redmine category processing qgis i have used the two functions which initially looks the same i miss the possibility to choose between seperate line and polylines
| 1
|
16,957
| 22,318,724,258
|
IssuesEvent
|
2022-06-14 02:44:15
|
openslide/openslide
|
https://api.github.com/repos/openslide/openslide
|
opened
|
Require contributors to assent to the Developer Certificate of Origin
|
enhancement development-process
|
As part of the ongoing effort to modernize OpenSlide project infrastructure and make it easier to contribute, we should formalize our protections against contributions that may be legally encumbered, including contributions based on file format documentation obtained under an NDA or a restrictive license.
Some projects accomplish this via a Contributor License Agreement, but that approach is controversial. CLAs typically require contributors to either assign their copyright to the project, or at least grant the project an unlimited right to relicense the code, allowing it to be used in proprietary products. I'm among the people who believe that open-source projects are healthier when their copyrights are held by a variety of people, ensuring that relicensing the project is difficult or impossible.
Instead, we can use the [Developer Certificate of Origin](https://developercertificate.org/), originated by the Linux kernel and also used by other projects. It would protect the project by requiring contributors to certify that they have the right to submit their contribution, but it does not grant the project any additional rights, nor require the contributor to give up any. Making this certification would be a lightweight process: contributors would just add `Signed-off-by: Developer Name <developer.email@example.com>` to their commit messages.
DCO paragraph (a) prohibits contributions that the contributor does not have the right to submit, which would provide the desired protection against code derived from encumbered documentation. Since this point is crucial for OpenSlide, I think the contribution guide and CI notifications should explicitly call it out to avoid any confusion.
We can implement the new requirement by:
- [ ] Adding a contribution guide with the full text of the DCO and instructions for signing off on commits.
- [ ] Adding CI checks that fail if commits are not properly signed off.
- [ ] [Requiring signoff on web-based commits](https://github.blog/changelog/2022-06-08-admins-can-require-sign-off-on-web-based-commits/).
- [ ] Strengthening GitHub branch protection so that everyone, including repository maintainers, must submit a PR that passes CI before landing code. (In practice, this would affect mostly me.)
To be clear, this change isn't the result of any particular incident. Potentially-encumbered code has been a concern since the beginning of the project, and it's been clear for a while that we should formalize our existing processes for keeping such material out of OpenSlide.
I'll open a PR soon in the [.github metadata repo](https://github.com/openslide/.github) with the exact CONTRIBUTING text and CI checks I'm proposing, and will link the PR here for visibility.
|
1.0
|
Require contributors to assent to the Developer Certificate of Origin - As part of the ongoing effort to modernize OpenSlide project infrastructure and make it easier to contribute, we should formalize our protections against contributions that may be legally encumbered, including contributions based on file format documentation obtained under an NDA or a restrictive license.
Some projects accomplish this via a Contributor License Agreement, but that approach is controversial. CLAs typically require contributors to either assign their copyright to the project, or at least grant the project an unlimited right to relicense the code, allowing it to be used in proprietary products. I'm among the people who believe that open-source projects are healthier when their copyrights are held by a variety of people, ensuring that relicensing the project is difficult or impossible.
Instead, we can use the [Developer Certificate of Origin](https://developercertificate.org/), originated by the Linux kernel and also used by other projects. It would protect the project by requiring contributors to certify that they have the right to submit their contribution, but it does not grant the project any additional rights, nor require the contributor to give up any. Making this certification would be a lightweight process: contributors would just add `Signed-off-by: Developer Name <developer.email@example.com>` to their commit messages.
DCO paragraph (a) prohibits contributions that the contributor does not have the right to submit, which would provide the desired protection against code derived from encumbered documentation. Since this point is crucial for OpenSlide, I think the contribution guide and CI notifications should explicitly call it out to avoid any confusion.
We can implement the new requirement by:
- [ ] Adding a contribution guide with the full text of the DCO and instructions for signing off on commits.
- [ ] Adding CI checks that fail if commits are not properly signed off.
- [ ] [Requiring signoff on web-based commits](https://github.blog/changelog/2022-06-08-admins-can-require-sign-off-on-web-based-commits/).
- [ ] Strengthening GitHub branch protection so that everyone, including repository maintainers, must submit a PR that passes CI before landing code. (In practice, this would affect mostly me.)
To be clear, this change isn't the result of any particular incident. Potentially-encumbered code has been a concern since the beginning of the project, and it's been clear for a while that we should formalize our existing processes for keeping such material out of OpenSlide.
I'll open a PR soon in the [.github metadata repo](https://github.com/openslide/.github) with the exact CONTRIBUTING text and CI checks I'm proposing, and will link the PR here for visibility.
|
process
|
require contributors to assent to the developer certificate of origin as part of the ongoing effort to modernize openslide project infrastructure and make it easier to contribute we should formalize our protections against contributions that may be legally encumbered including contributions based on file format documentation obtained under an nda or a restrictive license some projects accomplish this via a contributor license agreement but that approach is controversial clas typically require contributors to either assign their copyright to the project or at least grant the project an unlimited right to relicense the code allowing it to be used in proprietary products i m among the people who believe that open source projects are healthier when their copyrights are held by a variety of people ensuring that relicensing the project is difficult or impossible instead we can use the originated by the linux kernel and also used by other projects it would protect the project by requiring contributors to certify that they have the right to submit their contribution but it does not grant the project any additional rights nor require the contributor to give up any making this certification would be a lightweight process contributors would just add signed off by developer name to their commit messages dco paragraph a prohibits contributions that the contributor does not have the right to submit which would provide the desired protection against code derived from encumbered documentation since this point is crucial for openslide i think the contribution guide and ci notifications should explicitly call it out to avoid any confusion we can implement the new requirement by adding a contribution guide with the full text of the dco and instructions for signing off on commits adding ci checks that fail if commits are not properly signed off strengthening github branch protection so that everyone including repository maintainers must submit a pr that passes ci before landing code in practice this would affect mostly me to be clear this change isn t the result of any particular incident potentially encumbered code has been a concern since the beginning of the project and it s been clear for a while that we should formalize our existing processes for keeping such material out of openslide i ll open a pr soon in the with the exact contributing text and ci checks i m proposing and will link the pr here for visibility
| 1
|
671,532
| 22,765,052,132
|
IssuesEvent
|
2022-07-08 02:51:48
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
opened
|
unexpected hehaviour for async GRPC in platform k8s+aarch64
|
kind/bug lang/c++ priority/P2 untriaged
|
<!--
PLEASE DO NOT POST A QUESTION HERE.
This form is for bug reports and feature requests ONLY!
For general questions and troubleshooting, please ask/look for answers at StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc
For questions that specifically need to be answered by gRPC team members, please ask/look for answers at grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io
Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new)
-->
### What version of gRPC and what language are you using?
v1.26.0,v1.48.0
### What operating system (Linux, Windows,...) and version?
Linux version 4.14.0-115.el7a.0.1.aarch64
### What runtime / compiler are you using (e.g. python version or version of gcc)
clang-13
### What did you do?
1. The streaming grpc api define like this:
`rpc EstablishMPPConnection(mpp.EstablishMPPConnectionRequest) returns (stream mpp.MPPDataPacket) {}`
2. The server send some mpp.MPPDataPackets and then invoke `ServerAsyncWriter::Finish ` to tell client to finish the read.
3. The corresponding CompletionQueue `cq.Next(&tag, &ok)` got `ok == true` for that event.
4. I confirmed that the `ServerAsyncWriter::Finish` is called and its `ok` is returned and its value is `true`
### What did you expect to see?
the client receive the `finish` signal like `reader->Read` return `false`;
### What did you see instead?
the client just block until the connection timeout or reset by peer. It seems the client didn't receive the `finish` signal.
Make sure you include information that can help us debug (full error message, exception listing, stack trace, logs).
See [TROUBLESHOOTING.md](https://github.com/grpc/grpc/blob/master/TROUBLESHOOTING.md) for how to diagnose problems better.
### Anything else we should know about your project / environment?
the grpc source code says:
```
/// Client-side Write, Client-side WritesDone, Server-side Write,
/// Server-side Finish, Server-side SendInitialMetadata (which is
/// typically included in Write or Finish when not done explicitly):
/// \a ok means that the data/metadata/status/etc is going to go to the
/// wire. If it is false, it not going to the wire because the call
/// is already dead (i.e., canceled, deadline expired, other side
/// dropped the channel, etc).
```
Since it says "for Server-side Finish, ok means that the data/metadata/status/etc is going to go to the wire. ", so I think it's an unpected hehaviour for k8s+aarch64 environment.
|
1.0
|
unexpected hehaviour for async GRPC in platform k8s+aarch64 - <!--
PLEASE DO NOT POST A QUESTION HERE.
This form is for bug reports and feature requests ONLY!
For general questions and troubleshooting, please ask/look for answers at StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc
For questions that specifically need to be answered by gRPC team members, please ask/look for answers at grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io
Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new)
-->
### What version of gRPC and what language are you using?
v1.26.0,v1.48.0
### What operating system (Linux, Windows,...) and version?
Linux version 4.14.0-115.el7a.0.1.aarch64
### What runtime / compiler are you using (e.g. python version or version of gcc)
clang-13
### What did you do?
1. The streaming grpc api define like this:
`rpc EstablishMPPConnection(mpp.EstablishMPPConnectionRequest) returns (stream mpp.MPPDataPacket) {}`
2. The server send some mpp.MPPDataPackets and then invoke `ServerAsyncWriter::Finish ` to tell client to finish the read.
3. The corresponding CompletionQueue `cq.Next(&tag, &ok)` got `ok == true` for that event.
4. I confirmed that the `ServerAsyncWriter::Finish` is called and its `ok` is returned and its value is `true`
### What did you expect to see?
the client receive the `finish` signal like `reader->Read` return `false`;
### What did you see instead?
the client just block until the connection timeout or reset by peer. It seems the client didn't receive the `finish` signal.
Make sure you include information that can help us debug (full error message, exception listing, stack trace, logs).
See [TROUBLESHOOTING.md](https://github.com/grpc/grpc/blob/master/TROUBLESHOOTING.md) for how to diagnose problems better.
### Anything else we should know about your project / environment?
the grpc source code says:
```
/// Client-side Write, Client-side WritesDone, Server-side Write,
/// Server-side Finish, Server-side SendInitialMetadata (which is
/// typically included in Write or Finish when not done explicitly):
/// \a ok means that the data/metadata/status/etc is going to go to the
/// wire. If it is false, it not going to the wire because the call
/// is already dead (i.e., canceled, deadline expired, other side
/// dropped the channel, etc).
```
Since it says "for Server-side Finish, ok means that the data/metadata/status/etc is going to go to the wire. ", so I think it's an unpected hehaviour for k8s+aarch64 environment.
|
non_process
|
unexpected hehaviour for async grpc in platform please do not post a question here this form is for bug reports and feature requests only for general questions and troubleshooting please ask look for answers at stackoverflow with grpc tag for questions that specifically need to be answered by grpc team members please ask look for answers at grpc io mailing list issues specific to grpc java grpc go grpc node grpc dart grpc web should be created in the repository they belong to e g what version of grpc and what language are you using , what operating system linux windows and version linux version what runtime compiler are you using e g python version or version of gcc clang what did you do the streaming grpc api define like this rpc establishmppconnection mpp establishmppconnectionrequest returns stream mpp mppdatapacket the server send some mpp mppdatapackets and then invoke serverasyncwriter finish to tell client to finish the read the corresponding completionqueue cq next tag ok got ok true for that event i confirmed that the serverasyncwriter finish is called and its ok is returned and its value is true what did you expect to see the client receive the finish signal like reader read return false what did you see instead the client just block until the connection timeout or reset by peer it seems the client didn t receive the finish signal make sure you include information that can help us debug full error message exception listing stack trace logs see for how to diagnose problems better anything else we should know about your project environment the grpc source code says client side write client side writesdone server side write server side finish server side sendinitialmetadata which is typically included in write or finish when not done explicitly a ok means that the data metadata status etc is going to go to the wire if it is false it not going to the wire because the call is already dead i e canceled deadline expired other side dropped the channel etc since it says for server side finish ok means that the data metadata status etc is going to go to the wire so i think it s an unpected hehaviour for environment
| 0
|
3,191
| 6,259,656,546
|
IssuesEvent
|
2017-07-14 18:35:24
|
PeaceGeeksSociety/salesforce
|
https://api.github.com/repos/PeaceGeeksSociety/salesforce
|
opened
|
Create project in SF
|
Community Processes Salesforce Fields
|
We would like to create new projects within SF so that we can collect time against project activities and collect contacts.
Eg. Services Advisor Pathways --> collects time against project activities
Google Impact Challenge (create campaign) --> collects contacts, contact campaign relationship?
Done when: have created new projects or campaigns in SF that allow us to track time and contacts against them.
|
1.0
|
Create project in SF - We would like to create new projects within SF so that we can collect time against project activities and collect contacts.
Eg. Services Advisor Pathways --> collects time against project activities
Google Impact Challenge (create campaign) --> collects contacts, contact campaign relationship?
Done when: have created new projects or campaigns in SF that allow us to track time and contacts against them.
|
process
|
create project in sf we would like to create new projects within sf so that we can collect time against project activities and collect contacts eg services advisor pathways collects time against project activities google impact challenge create campaign collects contacts contact campaign relationship done when have created new projects or campaigns in sf that allow us to track time and contacts against them
| 1
|
8,945
| 12,058,425,531
|
IssuesEvent
|
2020-04-15 17:26:18
|
googleapis/nodejs-bigtable
|
https://api.github.com/repos/googleapis/nodejs-bigtable
|
opened
|
Refactor streaming of table rows and remove through2 dependency
|
priority: p2 type: process
|
The code for table.ts is a bit complicated and can likely be streamlined. In order to retry, we nest two streams.
It is likely worth removing the two instances of through and subclassing stream. This should help to shutdown on termination and not attempt writes past that point.
This also will remove dependencies on `through2` and `readable-stream`
|
1.0
|
Refactor streaming of table rows and remove through2 dependency - The code for table.ts is a bit complicated and can likely be streamlined. In order to retry, we nest two streams.
It is likely worth removing the two instances of through and subclassing stream. This should help to shutdown on termination and not attempt writes past that point.
This also will remove dependencies on `through2` and `readable-stream`
|
process
|
refactor streaming of table rows and remove dependency the code for table ts is a bit complicated and can likely be streamlined in order to retry we nest two streams it is likely worth removing the two instances of through and subclassing stream this should help to shutdown on termination and not attempt writes past that point this also will remove dependencies on and readable stream
| 1
|
22,689
| 31,992,523,832
|
IssuesEvent
|
2023-09-21 07:04:59
|
X-Sharp/XSharpPublic
|
https://api.github.com/repos/X-Sharp/XSharpPublic
|
closed
|
Preprocessor can't handle UDCs in code spanning in multiple lines
|
bug Preprocessor
|
Following code leads to syntax errors, for the UDCs in the code that span in multiple lines:
```
FUNCTION Start() AS VOID
LOCAL n := 5,m := 3 AS INT
LOCAL i AS INT
#xtranslate TestUDC <expr1> <expr2> => <expr1>-<expr2>
// OK:
i := TestUDC n m
? i
i := ;
TestUDC n m
// error XS9002: Parser: unexpected input 'n'
i := TestUDC ;
n m
i := TestUDC n ;
m
// original sample:
#xtranslate WITH <expr> {<v1>[, <vN>]} => {<expr>:<v1> [, <expr>:<vN>]}
// OK:
LOCAL arr := WITH Example() {A, B, C}
arr := WITH Example() {A, B, C}
// error XS9002: Parser: unexpected input 'WITH'
arr := WITH Example() {;
A, B, C;
}
arr := WITH Example() ;
{A, B, C}
AEval(arr, {|c| QOut(c) })
FUNCTION Example() AS TestClass
RETURN TestClass{}
CLASS TestClass
ACCESS A; RETURN "A"
ACCESS B; RETURN "B"
ACCESS C; RETURN "C"
END CLASS
```
|
1.0
|
Preprocessor can't handle UDCs in code spanning in multiple lines - Following code leads to syntax errors, for the UDCs in the code that span in multiple lines:
```
FUNCTION Start() AS VOID
LOCAL n := 5,m := 3 AS INT
LOCAL i AS INT
#xtranslate TestUDC <expr1> <expr2> => <expr1>-<expr2>
// OK:
i := TestUDC n m
? i
i := ;
TestUDC n m
// error XS9002: Parser: unexpected input 'n'
i := TestUDC ;
n m
i := TestUDC n ;
m
// original sample:
#xtranslate WITH <expr> {<v1>[, <vN>]} => {<expr>:<v1> [, <expr>:<vN>]}
// OK:
LOCAL arr := WITH Example() {A, B, C}
arr := WITH Example() {A, B, C}
// error XS9002: Parser: unexpected input 'WITH'
arr := WITH Example() {;
A, B, C;
}
arr := WITH Example() ;
{A, B, C}
AEval(arr, {|c| QOut(c) })
FUNCTION Example() AS TestClass
RETURN TestClass{}
CLASS TestClass
ACCESS A; RETURN "A"
ACCESS B; RETURN "B"
ACCESS C; RETURN "C"
END CLASS
```
|
process
|
preprocessor can t handle udcs in code spanning in multiple lines following code leads to syntax errors for the udcs in the code that span in multiple lines function start as void local n m as int local i as int xtranslate testudc ok i testudc n m i i testudc n m error parser unexpected input n i testudc n m i testudc n m original sample xtranslate with ok local arr with example a b c arr with example a b c error parser unexpected input with arr with example a b c arr with example a b c aeval arr c qout c function example as testclass return testclass class testclass access a return a access b return b access c return c end class
| 1
|
15,640
| 19,824,628,182
|
IssuesEvent
|
2022-01-20 04:04:18
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
RunBook description invalid
|
automation/svc triaged cxp product-feedback process-automation/subsvc Pri2
|
On [Target the start and stop actions against a subscription and resource group](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management-config#target-the-start-and-stop-actions-against-a-subscription-and-resource-group) step it refers to 'ScheduledStartStop_Parent' runbook , but having run this Resource 3 times the only runbooks created are called: 'AzureAutomationTutorialWithIdentity' and 'AzureAutomationTutorialWithIdentityGraphical'
Running the runbook does not prompt nor give any instructions about setting the "ACTION" or "WHATIF" parameters.
The output of the runbook 'as is' are:
```
Please enable appropriate RBAC permissions to the system identity of this automation account. Otherwise, the runbook may fail...
Logging in to Azure...
Environments
------------
{[AzureChinaCloud, AzureChinaCloud], [AzureCloud, AzureCloud], [AzureGermanCloud, AzureGermanCloud], [AzureUSGovernme...
```
With the error: ```'this.Client.SubscriptionId' cannot be null.```
Can I provide any more information to assist in improving the product?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 9aafa67b-4919-6773-d55c-c30e4e3ba703
* Version Independent ID: 51bec2e2-1984-d605-2c6e-6286e40262b8
* Content: [Configure Azure Automation Start/Stop VMs during off-hours](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management-config)
* Content Source: [articles/automation/automation-solution-vm-management-config.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-solution-vm-management-config.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SGSneha
* Microsoft Alias: **v-ssudhir**
|
1.0
|
RunBook description invalid - On [Target the start and stop actions against a subscription and resource group](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management-config#target-the-start-and-stop-actions-against-a-subscription-and-resource-group) step it refers to 'ScheduledStartStop_Parent' runbook , but having run this Resource 3 times the only runbooks created are called: 'AzureAutomationTutorialWithIdentity' and 'AzureAutomationTutorialWithIdentityGraphical'
Running the runbook does not prompt nor give any instructions about setting the "ACTION" or "WHATIF" parameters.
The output of the runbook 'as is' are:
```
Please enable appropriate RBAC permissions to the system identity of this automation account. Otherwise, the runbook may fail...
Logging in to Azure...
Environments
------------
{[AzureChinaCloud, AzureChinaCloud], [AzureCloud, AzureCloud], [AzureGermanCloud, AzureGermanCloud], [AzureUSGovernme...
```
With the error: ```'this.Client.SubscriptionId' cannot be null.```
Can I provide any more information to assist in improving the product?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 9aafa67b-4919-6773-d55c-c30e4e3ba703
* Version Independent ID: 51bec2e2-1984-d605-2c6e-6286e40262b8
* Content: [Configure Azure Automation Start/Stop VMs during off-hours](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management-config)
* Content Source: [articles/automation/automation-solution-vm-management-config.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-solution-vm-management-config.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SGSneha
* Microsoft Alias: **v-ssudhir**
|
process
|
runbook description invalid on step it refers to scheduledstartstop parent runbook but having run this resource times the only runbooks created are called azureautomationtutorialwithidentity and azureautomationtutorialwithidentitygraphical running the runbook does not prompt nor give any instructions about setting the action or whatif parameters the output of the runbook as is are please enable appropriate rbac permissions to the system identity of this automation account otherwise the runbook may fail logging in to azure environments azureusgovernme with the error this client subscriptionid cannot be null can i provide any more information to assist in improving the product document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login sgsneha microsoft alias v ssudhir
| 1
|
527,767
| 15,352,382,229
|
IssuesEvent
|
2021-03-01 06:55:03
|
magento/magento2
|
https://api.github.com/repos/magento/magento2
|
closed
|
Updating Customer through rest/all/V1/customers/:id resets group_id if group_id not passed in payload
|
Event: dmcdindia1 Fixed in 2.1.x Issue: Confirmed Priority: P4 Progress: ready for dev Reproduced on 2.1.x Reproduced on 2.2.x Reproduced on 2.3.x Severity: S4 Triage: Dev.Experience bug report stale issue
|
<!--- Provide a general summary of the issue in the Title above -->
<!--- Before adding new issues, please, check this article https://github.com/magento/magento2/wiki/Issue-reporting-guidelines-->
### Preconditions
<!--- Provide a more detailed information of environment you use -->
<!--- Magento version, tag, HEAD, etc., PHP & MySQL version, etc.. -->
1. Magento Version 2.2.3 using rest API
2. Passing a JSON payload with minimum required data (email, first/last name, website_id, id)
### Steps to reproduce
<!--- Provide a set of unambiguous steps to reproduce this bug include code, if relevant -->
1. Update any customer who's group_id isn't the first group_id in the source list
2. Pass the customer payload with email, firstname, lastname, website_id, id
### Expected result
<!--- Tell us what should happen -->
1. I expect that only the data passed in the payload would be modified
2. Example: {"customer":{"email":"test@gmail.com","firstname":"magento","lastname":"user","website_id":1,"id":77486}} should leave the customer's full data structure unchanged.
### Actual result
<!--- Tell us what happens instead -->
1. If the above customer had a group_id of 5, and our full group_id consisted of [1,2,3,4,5], the customer's group_id after passing the above payload would be set to 1:
Before:

PUT:

Results:

<!--- (This may be platform independent comment) -->
|
1.0
|
Updating Customer through rest/all/V1/customers/:id resets group_id if group_id not passed in payload - <!--- Provide a general summary of the issue in the Title above -->
<!--- Before adding new issues, please, check this article https://github.com/magento/magento2/wiki/Issue-reporting-guidelines-->
### Preconditions
<!--- Provide a more detailed information of environment you use -->
<!--- Magento version, tag, HEAD, etc., PHP & MySQL version, etc.. -->
1. Magento Version 2.2.3 using rest API
2. Passing a JSON payload with minimum required data (email, first/last name, website_id, id)
### Steps to reproduce
<!--- Provide a set of unambiguous steps to reproduce this bug include code, if relevant -->
1. Update any customer who's group_id isn't the first group_id in the source list
2. Pass the customer payload with email, firstname, lastname, website_id, id
### Expected result
<!--- Tell us what should happen -->
1. I expect that only the data passed in the payload would be modified
2. Example: {"customer":{"email":"test@gmail.com","firstname":"magento","lastname":"user","website_id":1,"id":77486}} should leave the customer's full data structure unchanged.
### Actual result
<!--- Tell us what happens instead -->
1. If the above customer had a group_id of 5, and our full group_id consisted of [1,2,3,4,5], the customer's group_id after passing the above payload would be set to 1:
Before:

PUT:

Results:

<!--- (This may be platform independent comment) -->
|
non_process
|
updating customer through rest all customers id resets group id if group id not passed in payload before adding new issues please check this article preconditions magento version using rest api passing a json payload with minimum required data email first last name website id id steps to reproduce update any customer who s group id isn t the first group id in the source list pass the customer payload with email firstname lastname website id id expected result i expect that only the data passed in the payload would be modified example customer email test gmail com firstname magento lastname user website id id should leave the customer s full data structure unchanged actual result if the above customer had a group id of and our full group id consisted of the customer s group id after passing the above payload would be set to before put results
| 0
|
8,086
| 11,257,665,374
|
IssuesEvent
|
2020-01-13 00:20:15
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Azure Automation Start/Stop VM
|
Pri3 automation/svc cxp process-automation/subsvc product-question triaged
|
After i enabled the Solution i see a list of Schedules and Wanted to Know more about this Schedules.
I have a Start /Stop VM Schedule every Weekday M-F 5:00 AM PT to 8:00 PM PT setup. Please provide details as i cannot understand the below 3 schedules and why are they setup.
Schedule_AutoStop_CreateAlert_Parent - Status Off
Sequenced-StartVM -Status Off
Sequenced StopVM - Status Off
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 225c9d05-83dd-b006-0025-3753f5ab25bf
* Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096
* Content: [Start/Stop VMs during off-hours solution](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management)
* Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
1.0
|
Azure Automation Start/Stop VM - After i enabled the Solution i see a list of Schedules and Wanted to Know more about this Schedules.
I have a Start /Stop VM Schedule every Weekday M-F 5:00 AM PT to 8:00 PM PT setup. Please provide details as i cannot understand the below 3 schedules and why are they setup.
Schedule_AutoStop_CreateAlert_Parent - Status Off
Sequenced-StartVM -Status Off
Sequenced StopVM - Status Off
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 225c9d05-83dd-b006-0025-3753f5ab25bf
* Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096
* Content: [Start/Stop VMs during off-hours solution](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management)
* Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
process
|
azure automation start stop vm after i enabled the solution i see a list of schedules and wanted to know more about this schedules i have a start stop vm schedule every weekday m f am pt to pm pt setup please provide details as i cannot understand the below schedules and why are they setup schedule autostop createalert parent status off sequenced startvm status off sequenced stopvm status off document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
| 1
|
33,625
| 7,738,510,469
|
IssuesEvent
|
2018-05-28 12:20:54
|
Yoast/yoast-components
|
https://api.github.com/repos/Yoast/yoast-components
|
closed
|
Snippet editor: fields shouldn't look active when not focused
|
a11y code-review ★ snippet preview
|
In the new React snippet editor:
- click on a field
- then click outside of the field or on any other focusable element that is not a snippet editor field
- the field still looks "focused" while actually it isn't:
<img width="711" alt="screen shot 2018-05-07 at 10 47 49" src="https://user-images.githubusercontent.com/1682452/39693434-4a12151e-51e4-11e8-9529-aa54fea44f31.png">
|
1.0
|
Snippet editor: fields shouldn't look active when not focused - In the new React snippet editor:
- click on a field
- then click outside of the field or on any other focusable element that is not a snippet editor field
- the field still looks "focused" while actually it isn't:
<img width="711" alt="screen shot 2018-05-07 at 10 47 49" src="https://user-images.githubusercontent.com/1682452/39693434-4a12151e-51e4-11e8-9529-aa54fea44f31.png">
|
non_process
|
snippet editor fields shouldn t look active when not focused in the new react snippet editor click on a field then click outside of the field or on any other focusable element that is not a snippet editor field the field still looks focused while actually it isn t img width alt screen shot at src
| 0
|
90,032
| 10,651,881,713
|
IssuesEvent
|
2019-10-17 11:23:59
|
OpenEnergyPlatform/open-MaStR
|
https://api.github.com/repos/OpenEnergyPlatform/open-MaStR
|
opened
|
Function download_parallel_power_unit() has no proper documentation (docstring)
|
documentation
|
docstring is missing
|
1.0
|
Function download_parallel_power_unit() has no proper documentation (docstring) - docstring is missing
|
non_process
|
function download parallel power unit has no proper documentation docstring docstring is missing
| 0
|
8,513
| 11,695,621,065
|
IssuesEvent
|
2020-03-06 08:00:28
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `CharLength` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `CharLength` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `CharLength` from TiDB -
## Description
Port the scalar function `CharLength` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function charlength from tidb description port the scalar function charlength from tidb to coprocessor score mentor s breeswish recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
60,206
| 17,023,368,973
|
IssuesEvent
|
2021-07-03 01:39:55
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Merkaartor.pro broken in revision 13799
|
Component: merkaartor Priority: trivial Resolution: duplicate Type: defect
|
**[Submitted to the original trac issue database at 10.17pm, Thursday, 19th February 2009]**
Index: Merkaartor.pro
===================================================================
--- Merkaartor.pro (revision 13799)
+++ Merkaartor.pro (working copy)
@@ -70,7 +70,7 @@
translations/merkaartor_ar.ts \
translations/merkaartor_cs.ts \
translations/merkaartor_de.ts \
- translations/merkaartor_es.ts
+ translations/merkaartor_es.ts \
translations/merkaartor_fr.ts \
translations/merkaartor_it.ts \
translations/merkaartor_pl.ts \
@@ -80,7 +80,7 @@
translations/merkaartor_ar.qm \
translations/merkaartor_cs.qm \
translations/merkaartor_de.qm \
- translations/merkaartor_es.ts
+ translations/merkaartor_es.ts \
translations/merkaartor_fr.qm \
translations/merkaartor_it.qm \
translations/merkaartor_pl.qm \
|
1.0
|
Merkaartor.pro broken in revision 13799 - **[Submitted to the original trac issue database at 10.17pm, Thursday, 19th February 2009]**
Index: Merkaartor.pro
===================================================================
--- Merkaartor.pro (revision 13799)
+++ Merkaartor.pro (working copy)
@@ -70,7 +70,7 @@
translations/merkaartor_ar.ts \
translations/merkaartor_cs.ts \
translations/merkaartor_de.ts \
- translations/merkaartor_es.ts
+ translations/merkaartor_es.ts \
translations/merkaartor_fr.ts \
translations/merkaartor_it.ts \
translations/merkaartor_pl.ts \
@@ -80,7 +80,7 @@
translations/merkaartor_ar.qm \
translations/merkaartor_cs.qm \
translations/merkaartor_de.qm \
- translations/merkaartor_es.ts
+ translations/merkaartor_es.ts \
translations/merkaartor_fr.qm \
translations/merkaartor_it.qm \
translations/merkaartor_pl.qm \
|
non_process
|
merkaartor pro broken in revision index merkaartor pro merkaartor pro revision merkaartor pro working copy translations merkaartor ar ts translations merkaartor cs ts translations merkaartor de ts translations merkaartor es ts translations merkaartor es ts translations merkaartor fr ts translations merkaartor it ts translations merkaartor pl ts translations merkaartor ar qm translations merkaartor cs qm translations merkaartor de qm translations merkaartor es ts translations merkaartor es ts translations merkaartor fr qm translations merkaartor it qm translations merkaartor pl qm
| 0
|
16,839
| 22,088,713,722
|
IssuesEvent
|
2022-06-01 02:54:49
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
"Address already in use" from DataLoader on different process with num_workers > 1
|
module: multiprocessing module: dataloader triaged
|
### 🐛 Describe the bug
On a system with multiple GPUs, several independent PyTorch jobs can be run. If a first job has a DataLoader with num_workers > 1, and a second job is launched with num_workers > 1, the latter will throw the following exception:
```
Traceback (most recent call last):
File "/opt/python/lib/python3.9/multiprocessing/queues.py", line 245, in _feed
obj = _ForkingPickler.dumps(obj)
File "/opt/python/lib/python3.9/multiprocessing/reduction.py", line 51, in dumps
cls(buf, protocol).dump(obj)
File "/opt/python/lib/python3.9/site-packages/torch/multiprocessing/reductions.py", line 348, in reduce_storage
df = multiprocessing.reduction.DupFd(fd)
File "/opt/python/lib/python3.9/multiprocessing/reduction.py", line 198, in DupFd
return resource_sharer.DupFd(fd)
File "/opt/python/lib/python3.9/multiprocessing/resource_sharer.py", line 53, in __init__
self._id = _resource_sharer.register(send, close)
File "/opt/python/lib/python3.9/multiprocessing/resource_sharer.py", line 76, in register
self._start()
File "/opt/python/lib/python3.9/multiprocessing/resource_sharer.py", line 126, in _start
self._listener = Listener(authkey=process.current_process().authkey)
File "/opt/python/lib/python3.9/multiprocessing/connection.py", line 453, in __init__
self._listener = SocketListener(address, family, backlog)
File "/opt/python/lib/python3.9/multiprocessing/connection.py", line 596, in __init__
self._socket.bind(address)
OSError: [Errno 98] Address already in use
```
The process will become a zombie process that does not die and does not do anything.
### Versions
1.11.0
cc @SsnL @VitalyFedyunin @ejguan @NivekT
|
1.0
|
"Address already in use" from DataLoader on different process with num_workers > 1 - ### 🐛 Describe the bug
On a system with multiple GPUs, several independent PyTorch jobs can be run. If a first job has a DataLoader with num_workers > 1, and a second job is launched with num_workers > 1, the latter will throw the following exception:
```
Traceback (most recent call last):
File "/opt/python/lib/python3.9/multiprocessing/queues.py", line 245, in _feed
obj = _ForkingPickler.dumps(obj)
File "/opt/python/lib/python3.9/multiprocessing/reduction.py", line 51, in dumps
cls(buf, protocol).dump(obj)
File "/opt/python/lib/python3.9/site-packages/torch/multiprocessing/reductions.py", line 348, in reduce_storage
df = multiprocessing.reduction.DupFd(fd)
File "/opt/python/lib/python3.9/multiprocessing/reduction.py", line 198, in DupFd
return resource_sharer.DupFd(fd)
File "/opt/python/lib/python3.9/multiprocessing/resource_sharer.py", line 53, in __init__
self._id = _resource_sharer.register(send, close)
File "/opt/python/lib/python3.9/multiprocessing/resource_sharer.py", line 76, in register
self._start()
File "/opt/python/lib/python3.9/multiprocessing/resource_sharer.py", line 126, in _start
self._listener = Listener(authkey=process.current_process().authkey)
File "/opt/python/lib/python3.9/multiprocessing/connection.py", line 453, in __init__
self._listener = SocketListener(address, family, backlog)
File "/opt/python/lib/python3.9/multiprocessing/connection.py", line 596, in __init__
self._socket.bind(address)
OSError: [Errno 98] Address already in use
```
The process will become a zombie process that does not die and does not do anything.
### Versions
1.11.0
cc @SsnL @VitalyFedyunin @ejguan @NivekT
|
process
|
address already in use from dataloader on different process with num workers 🐛 describe the bug on a system with multiple gpus several independent pytorch jobs can be run if a first job has a dataloader with num workers and a second job is launched with num workers the latter will throw the following exception traceback most recent call last file opt python lib multiprocessing queues py line in feed obj forkingpickler dumps obj file opt python lib multiprocessing reduction py line in dumps cls buf protocol dump obj file opt python lib site packages torch multiprocessing reductions py line in reduce storage df multiprocessing reduction dupfd fd file opt python lib multiprocessing reduction py line in dupfd return resource sharer dupfd fd file opt python lib multiprocessing resource sharer py line in init self id resource sharer register send close file opt python lib multiprocessing resource sharer py line in register self start file opt python lib multiprocessing resource sharer py line in start self listener listener authkey process current process authkey file opt python lib multiprocessing connection py line in init self listener socketlistener address family backlog file opt python lib multiprocessing connection py line in init self socket bind address oserror address already in use the process will become a zombie process that does not die and does not do anything versions cc ssnl vitalyfedyunin ejguan nivekt
| 1
|
10,365
| 13,186,106,047
|
IssuesEvent
|
2020-08-12 23:06:31
|
MatthewVandyke/CPW213-eCommerceSite
|
https://api.github.com/repos/MatthewVandyke/CPW213-eCommerceSite
|
closed
|
Add CI Pipeline
|
developer process
|
Add a continuous integration pipeline that will check to make sure code in a pull request compiles successfully.
|
1.0
|
Add CI Pipeline - Add a continuous integration pipeline that will check to make sure code in a pull request compiles successfully.
|
process
|
add ci pipeline add a continuous integration pipeline that will check to make sure code in a pull request compiles successfully
| 1
|
78,891
| 22,496,235,458
|
IssuesEvent
|
2022-06-23 07:50:32
|
OpenModelica/OpenModelica
|
https://api.github.com/repos/OpenModelica/OpenModelica
|
closed
|
Windows installers fail SmartScreen checks
|
enhancement COMP/Build System
|
When installing OMC on Windows, the SmartScreen filter identifies the OMC installer as suspicious software from unidentified authors, and requires to give explicit consent to perform a potentially dangerous installation.
This may be ok for hardened hackers that know about the OSMC, but it's not projecting an image of quality and dependability on the sofware, particularly for industrial and corporate use. Looking like potential malware is not a very good marketing strategy :)
I would recommend that from 1.13.0 we start signing the installer with a certificate, so that we avoid this kind of problems. More information on how to do this is found [here](https://blogs.msdn.microsoft.com/ie/2011/03/22/smartscreen-application-reputation-building-reputation/).
----------
From https://trac.openmodelica.org/OpenModelica/ticket/4829
|
1.0
|
Windows installers fail SmartScreen checks - When installing OMC on Windows, the SmartScreen filter identifies the OMC installer as suspicious software from unidentified authors, and requires to give explicit consent to perform a potentially dangerous installation.
This may be ok for hardened hackers that know about the OSMC, but it's not projecting an image of quality and dependability on the sofware, particularly for industrial and corporate use. Looking like potential malware is not a very good marketing strategy :)
I would recommend that from 1.13.0 we start signing the installer with a certificate, so that we avoid this kind of problems. More information on how to do this is found [here](https://blogs.msdn.microsoft.com/ie/2011/03/22/smartscreen-application-reputation-building-reputation/).
----------
From https://trac.openmodelica.org/OpenModelica/ticket/4829
|
non_process
|
windows installers fail smartscreen checks when installing omc on windows the smartscreen filter identifies the omc installer as suspicious software from unidentified authors and requires to give explicit consent to perform a potentially dangerous installation this may be ok for hardened hackers that know about the osmc but it s not projecting an image of quality and dependability on the sofware particularly for industrial and corporate use looking like potential malware is not a very good marketing strategy i would recommend that from we start signing the installer with a certificate so that we avoid this kind of problems more information on how to do this is found from
| 0
|
8,900
| 11,993,516,465
|
IssuesEvent
|
2020-04-08 12:10:00
|
errbotio/errbot
|
https://api.github.com/repos/errbotio/errbot
|
closed
|
6.1.2 not released
|
#release-process type: support/question
|
### I am...
* [x] Reporting a bug
* [ ] Suggesting a new feature
* [ ] Requesting help with running my bot
* [ ] Requesting help writing plugins
* [ ] Here about something else
### I am running...
* Errbot version:
* OS version:
* Python version:
* Using a virtual environment: yes/no
### Issue description
It seems errbot 6.1.2 isn't released as a tag or release on the github page, even though there's release notes on the website.
|
1.0
|
6.1.2 not released -
### I am...
* [x] Reporting a bug
* [ ] Suggesting a new feature
* [ ] Requesting help with running my bot
* [ ] Requesting help writing plugins
* [ ] Here about something else
### I am running...
* Errbot version:
* OS version:
* Python version:
* Using a virtual environment: yes/no
### Issue description
It seems errbot 6.1.2 isn't released as a tag or release on the github page, even though there's release notes on the website.
|
process
|
not released i am reporting a bug suggesting a new feature requesting help with running my bot requesting help writing plugins here about something else i am running errbot version os version python version using a virtual environment yes no issue description it seems errbot isn t released as a tag or release on the github page even though there s release notes on the website
| 1
|
47,520
| 25,040,079,053
|
IssuesEvent
|
2022-11-04 19:45:09
|
rancher/dashboard
|
https://api.github.com/repos/rancher/dashboard
|
closed
|
Project & Cluster Members: Move loading indicators to tables
|
[zube]: QA Working kind/enhancement area/performance
|
These 2 places don't use table loading indicators, so it looks like the pages are slow to load.
|
True
|
Project & Cluster Members: Move loading indicators to tables - These 2 places don't use table loading indicators, so it looks like the pages are slow to load.
|
non_process
|
project cluster members move loading indicators to tables these places don t use table loading indicators so it looks like the pages are slow to load
| 0
|
307,809
| 9,422,505,067
|
IssuesEvent
|
2019-04-11 09:31:08
|
OpenNebula/one
|
https://api.github.com/repos/OpenNebula/one
|
closed
|
Creating persistent image via sunstone (web)
|
Category: Sunstone Community Priority: Normal Status: Pending Type: Bug
|
**Description**
Creating new persistent image from sunstone does not work.
**To Reproduce**
Just create a new empty datablock and set persistent to YES.
After creating check out details about the image (Info tab, Persistent - No).
**Expected behavior**
The new image should be persistent.
**Details**
- Affected Component: Sunstone
- Hypervisor: KVM
- Version: 5.6 to 5.8
**Additional context**
Creating new persistent image via cli works fine.
It seems like it is because of misspelling in html template for select element:
sunstone/public/app/tabs/images-tab/form-panels/create/wizard.hbs
Section:
```
<div class="large-4 columns">
<select id="img_persistent">
<option val=""></option>
<option val="YES"> {{tr "YES"}}</option>
<option val="NO"> {{tr "NO"}}</option>
</select>
</div>
```
We changed all "val=" to "value=" and rebuild sunstone - this fixed the problem.
The issue started when dropdown list appeared in code instead of checkbox.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [x] Branch created
- [x] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
|
1.0
|
Creating persistent image via sunstone (web) - **Description**
Creating new persistent image from sunstone does not work.
**To Reproduce**
Just create a new empty datablock and set persistent to YES.
After creating check out details about the image (Info tab, Persistent - No).
**Expected behavior**
The new image should be persistent.
**Details**
- Affected Component: Sunstone
- Hypervisor: KVM
- Version: 5.6 to 5.8
**Additional context**
Creating new persistent image via cli works fine.
It seems like it is because of misspelling in html template for select element:
sunstone/public/app/tabs/images-tab/form-panels/create/wizard.hbs
Section:
```
<div class="large-4 columns">
<select id="img_persistent">
<option val=""></option>
<option val="YES"> {{tr "YES"}}</option>
<option val="NO"> {{tr "NO"}}</option>
</select>
</div>
```
We changed all "val=" to "value=" and rebuild sunstone - this fixed the problem.
The issue started when dropdown list appeared in code instead of checkbox.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [x] Branch created
- [x] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
|
non_process
|
creating persistent image via sunstone web description creating new persistent image from sunstone does not work to reproduce just create a new empty datablock and set persistent to yes after creating check out details about the image info tab persistent no expected behavior the new image should be persistent details affected component sunstone hypervisor kvm version to additional context creating new persistent image via cli works fine it seems like it is because of misspelling in html template for select element sunstone public app tabs images tab form panels create wizard hbs section tr yes tr no we changed all val to value and rebuild sunstone this fixed the problem the issue started when dropdown list appeared in code instead of checkbox progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches
| 0
|
98,029
| 16,343,716,228
|
IssuesEvent
|
2021-05-13 03:29:11
|
samq-wsdemo/easybuggy
|
https://api.github.com/repos/samq-wsdemo/easybuggy
|
opened
|
CVE-2019-2692 (Medium) detected in mysql-connector-java-5.1.25.jar
|
security vulnerability
|
## CVE-2019-2692 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-connector-java-5.1.25.jar</b></p></summary>
<p>MySQL JDBC Type 4 driver</p>
<p>Library home page: <a href="http://dev.mysql.com/doc/connector-j/en/">http://dev.mysql.com/doc/connector-j/en/</a></p>
<p>Path to dependency file: easybuggy/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/mysql/mysql-connector-java/5.1.25/mysql-connector-java-5.1.25.jar,easybuggy/target/easybuggy-1-SNAPSHOT/WEB-INF/lib/mysql-connector-java-5.1.25.jar,easybuggy/.extract/webapps/ROOT/WEB-INF/lib/mysql-connector-java-5.1.25.jar</p>
<p>
Dependency Hierarchy:
- :x: **mysql-connector-java-5.1.25.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-wsdemo/easybuggy/commit/e5c5214a65668c2d872761f1094636c04a2c100d">e5c5214a65668c2d872761f1094636c04a2c100d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Vulnerability in the MySQL Connectors component of Oracle MySQL (subcomponent: Connector/J). Supported versions that are affected are 8.0.15 and prior. Difficult to exploit vulnerability allows high privileged attacker with logon to the infrastructure where MySQL Connectors executes to compromise MySQL Connectors. Successful attacks require human interaction from a person other than the attacker. Successful attacks of this vulnerability can result in takeover of MySQL Connectors. CVSS 3.0 Base Score 6.3 (Confidentiality, Integrity and Availability impacts). CVSS Vector: (CVSS:3.0/AV:L/AC:H/PR:H/UI:R/S:U/C:H/I:H/A:H).
<p>Publish Date: 2019-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-2692>CVE-2019-2692</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: High
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-jcq3-cprp-m333">https://github.com/advisories/GHSA-jcq3-cprp-m333</a></p>
<p>Release Date: 2019-04-23</p>
<p>Fix Resolution: mysql:mysql-connector-java:8.0.16</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"mysql","packageName":"mysql-connector-java","packageVersion":"5.1.25","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"mysql:mysql-connector-java:5.1.25","isMinimumFixVersionAvailable":true,"minimumFixVersion":"mysql:mysql-connector-java:8.0.16"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-2692","vulnerabilityDetails":"Vulnerability in the MySQL Connectors component of Oracle MySQL (subcomponent: Connector/J). Supported versions that are affected are 8.0.15 and prior. Difficult to exploit vulnerability allows high privileged attacker with logon to the infrastructure where MySQL Connectors executes to compromise MySQL Connectors. Successful attacks require human interaction from a person other than the attacker. Successful attacks of this vulnerability can result in takeover of MySQL Connectors. CVSS 3.0 Base Score 6.3 (Confidentiality, Integrity and Availability impacts). CVSS Vector: (CVSS:3.0/AV:L/AC:H/PR:H/UI:R/S:U/C:H/I:H/A:H).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-2692","cvss3Severity":"medium","cvss3Score":"6.3","cvss3Metrics":{"A":"High","AC":"High","PR":"High","S":"Unchanged","C":"High","UI":"Required","AV":"Local","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2019-2692 (Medium) detected in mysql-connector-java-5.1.25.jar - ## CVE-2019-2692 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-connector-java-5.1.25.jar</b></p></summary>
<p>MySQL JDBC Type 4 driver</p>
<p>Library home page: <a href="http://dev.mysql.com/doc/connector-j/en/">http://dev.mysql.com/doc/connector-j/en/</a></p>
<p>Path to dependency file: easybuggy/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/mysql/mysql-connector-java/5.1.25/mysql-connector-java-5.1.25.jar,easybuggy/target/easybuggy-1-SNAPSHOT/WEB-INF/lib/mysql-connector-java-5.1.25.jar,easybuggy/.extract/webapps/ROOT/WEB-INF/lib/mysql-connector-java-5.1.25.jar</p>
<p>
Dependency Hierarchy:
- :x: **mysql-connector-java-5.1.25.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-wsdemo/easybuggy/commit/e5c5214a65668c2d872761f1094636c04a2c100d">e5c5214a65668c2d872761f1094636c04a2c100d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Vulnerability in the MySQL Connectors component of Oracle MySQL (subcomponent: Connector/J). Supported versions that are affected are 8.0.15 and prior. Difficult to exploit vulnerability allows high privileged attacker with logon to the infrastructure where MySQL Connectors executes to compromise MySQL Connectors. Successful attacks require human interaction from a person other than the attacker. Successful attacks of this vulnerability can result in takeover of MySQL Connectors. CVSS 3.0 Base Score 6.3 (Confidentiality, Integrity and Availability impacts). CVSS Vector: (CVSS:3.0/AV:L/AC:H/PR:H/UI:R/S:U/C:H/I:H/A:H).
<p>Publish Date: 2019-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-2692>CVE-2019-2692</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: High
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-jcq3-cprp-m333">https://github.com/advisories/GHSA-jcq3-cprp-m333</a></p>
<p>Release Date: 2019-04-23</p>
<p>Fix Resolution: mysql:mysql-connector-java:8.0.16</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"mysql","packageName":"mysql-connector-java","packageVersion":"5.1.25","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"mysql:mysql-connector-java:5.1.25","isMinimumFixVersionAvailable":true,"minimumFixVersion":"mysql:mysql-connector-java:8.0.16"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-2692","vulnerabilityDetails":"Vulnerability in the MySQL Connectors component of Oracle MySQL (subcomponent: Connector/J). Supported versions that are affected are 8.0.15 and prior. Difficult to exploit vulnerability allows high privileged attacker with logon to the infrastructure where MySQL Connectors executes to compromise MySQL Connectors. Successful attacks require human interaction from a person other than the attacker. Successful attacks of this vulnerability can result in takeover of MySQL Connectors. CVSS 3.0 Base Score 6.3 (Confidentiality, Integrity and Availability impacts). CVSS Vector: (CVSS:3.0/AV:L/AC:H/PR:H/UI:R/S:U/C:H/I:H/A:H).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-2692","cvss3Severity":"medium","cvss3Score":"6.3","cvss3Metrics":{"A":"High","AC":"High","PR":"High","S":"Unchanged","C":"High","UI":"Required","AV":"Local","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve medium detected in mysql connector java jar cve medium severity vulnerability vulnerable library mysql connector java jar mysql jdbc type driver library home page a href path to dependency file easybuggy pom xml path to vulnerable library canner repository mysql mysql connector java mysql connector java jar easybuggy target easybuggy snapshot web inf lib mysql connector java jar easybuggy extract webapps root web inf lib mysql connector java jar dependency hierarchy x mysql connector java jar vulnerable library found in head commit a href found in base branch master vulnerability details vulnerability in the mysql connectors component of oracle mysql subcomponent connector j supported versions that are affected are and prior difficult to exploit vulnerability allows high privileged attacker with logon to the infrastructure where mysql connectors executes to compromise mysql connectors successful attacks require human interaction from a person other than the attacker successful attacks of this vulnerability can result in takeover of mysql connectors cvss base score confidentiality integrity and availability impacts cvss vector cvss av l ac h pr h ui r s u c h i h a h publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required high user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution mysql mysql connector java rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree mysql mysql connector java isminimumfixversionavailable true minimumfixversion mysql mysql connector java basebranches vulnerabilityidentifier cve vulnerabilitydetails vulnerability in the mysql connectors component of oracle mysql subcomponent connector j supported versions that are affected are and prior difficult to exploit vulnerability allows high privileged attacker with logon to the infrastructure where mysql connectors executes to compromise mysql connectors successful attacks require human interaction from a person other than the attacker successful attacks of this vulnerability can result in takeover of mysql connectors cvss base score confidentiality integrity and availability impacts cvss vector cvss av l ac h pr h ui r s u c h i h a h vulnerabilityurl
| 0
|
20,490
| 13,940,824,060
|
IssuesEvent
|
2020-10-22 18:28:26
|
joshsoftware/code-curiosity
|
https://api.github.com/repos/joshsoftware/code-curiosity
|
closed
|
Remove hard-coded devise secret key base
|
infrastructure
|
`config/initializers/devise.rb` specifies `config.secret_key` as a hardcoded hex string. We shouldn't be doing this and should move that to an environment variable at the very least.
|
1.0
|
Remove hard-coded devise secret key base - `config/initializers/devise.rb` specifies `config.secret_key` as a hardcoded hex string. We shouldn't be doing this and should move that to an environment variable at the very least.
|
non_process
|
remove hard coded devise secret key base config initializers devise rb specifies config secret key as a hardcoded hex string we shouldn t be doing this and should move that to an environment variable at the very least
| 0
|
761
| 3,244,917,024
|
IssuesEvent
|
2015-10-16 07:16:36
|
superroma/testcafe-hammerhead
|
https://api.github.com/repos/superroma/testcafe-hammerhead
|
closed
|
Implement API to add a ShadowUI suffix to css rules for the injectable stylesheets
|
!IMPORTANT! AREA: server SYSTEM: resource processing SYSTEM: sandbox TYPE: enhancement
|
Hammehread should have API to process css files (add a suffix to css rules).
It's necessary to fix the https://github.com/superroma/testcafe-phoenix/issues/54 issue
|
1.0
|
Implement API to add a ShadowUI suffix to css rules for the injectable stylesheets - Hammehread should have API to process css files (add a suffix to css rules).
It's necessary to fix the https://github.com/superroma/testcafe-phoenix/issues/54 issue
|
process
|
implement api to add a shadowui suffix to css rules for the injectable stylesheets hammehread should have api to process css files add a suffix to css rules it s necessary to fix the issue
| 1
|
182,532
| 21,672,914,386
|
IssuesEvent
|
2022-05-08 08:40:18
|
turkdevops/core
|
https://api.github.com/repos/turkdevops/core
|
reopened
|
CVE-2020-7720 (High) detected in node-forge-0.7.6.tgz
|
security vulnerability
|
## CVE-2020-7720 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.7.6.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.7.6.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.7.6.tgz</a></p>
<p>Path to dependency file: core/hyperdeploy/package.json</p>
<p>Path to vulnerable library: core/hyperdeploy/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- firebase-tools-6.0.0.tgz (Root Library)
- google-auto-auth-0.7.2.tgz
- google-auth-library-0.10.0.tgz
- gtoken-1.2.3.tgz
- google-p12-pem-0.1.2.tgz
- :x: **node-forge-0.7.6.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/core/commit/cbe729212a3605e9a1f31ba8a602d4b9fd758403">cbe729212a3605e9a1f31ba8a602d4b9fd758403</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package node-forge before 0.10.0 is vulnerable to Prototype Pollution via the util.setPath function. Note: Version 0.10.0 is a breaking change removing the vulnerable functions.
<p>Publish Date: 2020-09-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7720>CVE-2020-7720</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/digitalbazaar/forge/blob/master/CHANGELOG.md">https://github.com/digitalbazaar/forge/blob/master/CHANGELOG.md</a></p>
<p>Release Date: 2020-09-13</p>
<p>Fix Resolution: node-forge - 0.10.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-7720 (High) detected in node-forge-0.7.6.tgz - ## CVE-2020-7720 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.7.6.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.7.6.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.7.6.tgz</a></p>
<p>Path to dependency file: core/hyperdeploy/package.json</p>
<p>Path to vulnerable library: core/hyperdeploy/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- firebase-tools-6.0.0.tgz (Root Library)
- google-auto-auth-0.7.2.tgz
- google-auth-library-0.10.0.tgz
- gtoken-1.2.3.tgz
- google-p12-pem-0.1.2.tgz
- :x: **node-forge-0.7.6.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/core/commit/cbe729212a3605e9a1f31ba8a602d4b9fd758403">cbe729212a3605e9a1f31ba8a602d4b9fd758403</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package node-forge before 0.10.0 is vulnerable to Prototype Pollution via the util.setPath function. Note: Version 0.10.0 is a breaking change removing the vulnerable functions.
<p>Publish Date: 2020-09-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7720>CVE-2020-7720</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/digitalbazaar/forge/blob/master/CHANGELOG.md">https://github.com/digitalbazaar/forge/blob/master/CHANGELOG.md</a></p>
<p>Release Date: 2020-09-13</p>
<p>Fix Resolution: node-forge - 0.10.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in node forge tgz cve high severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file core hyperdeploy package json path to vulnerable library core hyperdeploy node modules node forge package json dependency hierarchy firebase tools tgz root library google auto auth tgz google auth library tgz gtoken tgz google pem tgz x node forge tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package node forge before is vulnerable to prototype pollution via the util setpath function note version is a breaking change removing the vulnerable functions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge step up your open source security game with whitesource
| 0
|
22,754
| 32,075,611,656
|
IssuesEvent
|
2023-09-25 10:47:34
|
h4sh5/npm-auto-scanner
|
https://api.github.com/repos/h4sh5/npm-auto-scanner
|
opened
|
@azure-tools/cadl-ranch 0.8.2 has 2 guarddog issues
|
npm-silent-process-execution
|
```{"npm-silent-process-execution":[{"code":" const cp = spawn(nodeExe, [\n entrypoint,\n \"serve\",\n config.scenariosPath,\n \"--port\",\n config.port.toString(),\n \"--coverageFile\",\n config.coverageFile,\n ... });","location":"package/dist/actions/serve.js:20","message":"This package is silently executing another executable"},{"code":" const cp = spawn(\n nodeExe,\n [\n entrypoint,\n \"serve\",\n config.scenariosPath,\n \"--port\",\n config.port.toString(),\n \"--coverageFile\",\n config.coverageFile,\n ],\n {\n ... },\n );","location":"package/src/actions/serve.ts:33","message":"This package is silently executing another executable"}]}```
|
1.0
|
@azure-tools/cadl-ranch 0.8.2 has 2 guarddog issues - ```{"npm-silent-process-execution":[{"code":" const cp = spawn(nodeExe, [\n entrypoint,\n \"serve\",\n config.scenariosPath,\n \"--port\",\n config.port.toString(),\n \"--coverageFile\",\n config.coverageFile,\n ... });","location":"package/dist/actions/serve.js:20","message":"This package is silently executing another executable"},{"code":" const cp = spawn(\n nodeExe,\n [\n entrypoint,\n \"serve\",\n config.scenariosPath,\n \"--port\",\n config.port.toString(),\n \"--coverageFile\",\n config.coverageFile,\n ],\n {\n ... },\n );","location":"package/src/actions/serve.ts:33","message":"This package is silently executing another executable"}]}```
|
process
|
azure tools cadl ranch has guarddog issues npm silent process execution n n n location package src actions serve ts message this package is silently executing another executable
| 1
|
706,267
| 24,263,562,400
|
IssuesEvent
|
2022-09-28 02:44:00
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
closed
|
Unusually high CPU consumption
|
kind/bug lang/Python priority/P2 disposition/requires reporter action
|
<!--
PLEASE DO NOT POST A QUESTION HERE.
This form is for bug reports and feature requests ONLY!
For general questions and troubleshooting, please ask/look for answers at StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc
For questions that specifically need to be answered by gRPC team members, please ask/look for answers at grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io
Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new)
-->
### What version of gRPC and what language are you using?
grpcio 1.42.0 py37hce63b2e_0 defaults
grpcio-tools 1.16.1 py37h2531618_0 defaults
using python
### What operating system (Linux, Windows,...) and version?
Linux version 3.10.0-693.el7.x86_64 (builder@kbuilder.dev.centos.org) (gcc version 4.8.5 20150623 (Red Hat 4.8.5-16) (GCC) ) #1 SMP Tue Aug 22 21:09:27 UTC 2017
And my machine gets 80 cores of CPU
### What runtime / compiler are you using (e.g. python version or version of gcc)
python 3.7.13
### What did you do?
I tried to use grpc framework to create a server. When ran the script that covers grpc, I found the related process always using approximately all cpus (like 7666 %CPU from command [top], given by 80 cores) and affect others badly.
### What I expect to know?
I am wondering whether there is a controlled parameter to limit the use of cpu, like only using certain 40% cpu?
|
1.0
|
Unusually high CPU consumption - <!--
PLEASE DO NOT POST A QUESTION HERE.
This form is for bug reports and feature requests ONLY!
For general questions and troubleshooting, please ask/look for answers at StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc
For questions that specifically need to be answered by gRPC team members, please ask/look for answers at grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io
Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new)
-->
### What version of gRPC and what language are you using?
grpcio 1.42.0 py37hce63b2e_0 defaults
grpcio-tools 1.16.1 py37h2531618_0 defaults
using python
### What operating system (Linux, Windows,...) and version?
Linux version 3.10.0-693.el7.x86_64 (builder@kbuilder.dev.centos.org) (gcc version 4.8.5 20150623 (Red Hat 4.8.5-16) (GCC) ) #1 SMP Tue Aug 22 21:09:27 UTC 2017
And my machine gets 80 cores of CPU
### What runtime / compiler are you using (e.g. python version or version of gcc)
python 3.7.13
### What did you do?
I tried to use grpc framework to create a server. When ran the script that covers grpc, I found the related process always using approximately all cpus (like 7666 %CPU from command [top], given by 80 cores) and affect others badly.
### What I expect to know?
I am wondering whether there is a controlled parameter to limit the use of cpu, like only using certain 40% cpu?
|
non_process
|
unusually high cpu consumption please do not post a question here this form is for bug reports and feature requests only for general questions and troubleshooting please ask look for answers at stackoverflow with grpc tag for questions that specifically need to be answered by grpc team members please ask look for answers at grpc io mailing list issues specific to grpc java grpc go grpc node grpc dart grpc web should be created in the repository they belong to e g what version of grpc and what language are you using grpcio defaults grpcio tools defaults using python what operating system linux windows and version linux version builder kbuilder dev centos org gcc version red hat gcc smp tue aug utc and my machine gets cores of cpu what runtime compiler are you using e g python version or version of gcc python what did you do i tried to use grpc framework to create a server when ran the script that covers grpc i found the related process always using approximately all cpus like cpu from command given by cores and affect others badly what i expect to know i am wondering whether there is a controlled parameter to limit the use of cpu like only using certain cpu
| 0
|
1,334
| 3,886,420,088
|
IssuesEvent
|
2016-04-14 00:50:44
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
Process start doesn't resolve environment variables in the executable path.
|
System.Diagnostics.Process X-Plat
|
If you use Process.Start(ProcessStartInfo) method on OSX (or any Unix OS) and a path to executable contains environment variables in it (this environment variables are passed to ProcessStartInfo.Environment). This variables will not be resolved and "System.ComponentModel.Win32Exception: No such file or directory" exception will be thrown.
This behavior is inconsistent with the Windows version.
Note: path looks like "%EnvVar%\test.exe
|
1.0
|
Process start doesn't resolve environment variables in the executable path. - If you use Process.Start(ProcessStartInfo) method on OSX (or any Unix OS) and a path to executable contains environment variables in it (this environment variables are passed to ProcessStartInfo.Environment). This variables will not be resolved and "System.ComponentModel.Win32Exception: No such file or directory" exception will be thrown.
This behavior is inconsistent with the Windows version.
Note: path looks like "%EnvVar%\test.exe
|
process
|
process start doesn t resolve environment variables in the executable path if you use process start processstartinfo method on osx or any unix os and a path to executable contains environment variables in it this environment variables are passed to processstartinfo environment this variables will not be resolved and system componentmodel no such file or directory exception will be thrown this behavior is inconsistent with the windows version note path looks like envvar test exe
| 1
|
229,964
| 7,602,556,880
|
IssuesEvent
|
2018-04-29 02:49:16
|
RobotLocomotion/drake
|
https://api.github.com/repos/RobotLocomotion/drake
|
closed
|
DRAKE_SPDLOG_TRACE macros lack do..while guards
|
priority: medium team: software core type: bug
|
See https://github.com/RobotLocomotion/drake/pull/8087#issuecomment-385210772. Our macros have a bare `if`, which is a no-no in macro implementation.
|
1.0
|
DRAKE_SPDLOG_TRACE macros lack do..while guards - See https://github.com/RobotLocomotion/drake/pull/8087#issuecomment-385210772. Our macros have a bare `if`, which is a no-no in macro implementation.
|
non_process
|
drake spdlog trace macros lack do while guards see our macros have a bare if which is a no no in macro implementation
| 0
|
15,559
| 10,135,625,317
|
IssuesEvent
|
2019-08-02 10:39:04
|
hyperledger/burrow
|
https://api.github.com/repos/hyperledger/burrow
|
closed
|
[Deploy] mempool-signing in "burrow deploy --mempool-signing" needs a better name
|
area/usability
|
Someone new to blockchain is not going to know what mempool-signing is, and mempool is a very overloaded term.
Please make suggestions for a better name.
burrow deploy --server-signing
burrow deploy --burrow-signing
|
True
|
[Deploy] mempool-signing in "burrow deploy --mempool-signing" needs a better name - Someone new to blockchain is not going to know what mempool-signing is, and mempool is a very overloaded term.
Please make suggestions for a better name.
burrow deploy --server-signing
burrow deploy --burrow-signing
|
non_process
|
mempool signing in burrow deploy mempool signing needs a better name someone new to blockchain is not going to know what mempool signing is and mempool is a very overloaded term please make suggestions for a better name burrow deploy server signing burrow deploy burrow signing
| 0
|
31,226
| 6,470,056,708
|
IssuesEvent
|
2017-08-17 08:02:57
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
closed
|
[lock] lease lock not released under cluster shutdown test, 3.8-SANPSHOT
|
Team: Core Type: Defect
|
nightly builds
https://hazelcast-l337.ci.cloudbees.com/view/shutdown/job/shutdown-lease-lock/
https://hazelcast-l337.ci.cloudbees.com/view/rollup/job/rollup-lease-lock/
some times fail, and the lease locks are not released.
members are being shutdown and restarted.
https://hazelcast-l337.ci.cloudbees.com/view/rollup/job/rollup-lease-lock/8/console
http://54.87.52.100/~jenkins/workspace/rollup-lease-lock/3.7.3/2016_11_15-03_08_36
```
HzMember1HZ validate hz.lock.validate.LeaseLockAssert threadId=0 global.AssertionException: lease-lock292.isLocked()==true
```
however this fail is rare,
|
1.0
|
[lock] lease lock not released under cluster shutdown test, 3.8-SANPSHOT - nightly builds
https://hazelcast-l337.ci.cloudbees.com/view/shutdown/job/shutdown-lease-lock/
https://hazelcast-l337.ci.cloudbees.com/view/rollup/job/rollup-lease-lock/
some times fail, and the lease locks are not released.
members are being shutdown and restarted.
https://hazelcast-l337.ci.cloudbees.com/view/rollup/job/rollup-lease-lock/8/console
http://54.87.52.100/~jenkins/workspace/rollup-lease-lock/3.7.3/2016_11_15-03_08_36
```
HzMember1HZ validate hz.lock.validate.LeaseLockAssert threadId=0 global.AssertionException: lease-lock292.isLocked()==true
```
however this fail is rare,
|
non_process
|
lease lock not released under cluster shutdown test sanpshot nightly builds some times fail and the lease locks are not released members are being shutdown and restarted validate hz lock validate leaselockassert threadid global assertionexception lease islocked true however this fail is rare
| 0
|
8,774
| 11,893,463,430
|
IssuesEvent
|
2020-03-29 11:42:22
|
gojek/feast
|
https://api.github.com/repos/gojek/feast
|
closed
|
Exclude documentation PRs from end-to-end tests
|
area/infra kind/process priority/p3
|
## Expected Behavior
Documentation PRs should not trigger end-to-end tests (or require them to be merged in)
## Possible Solution
Exclude the following from the prow configuration:
- .github
- docs
- README.md
- OWNERS
- CHANGELOG.md
- LICENSE
|
1.0
|
Exclude documentation PRs from end-to-end tests - ## Expected Behavior
Documentation PRs should not trigger end-to-end tests (or require them to be merged in)
## Possible Solution
Exclude the following from the prow configuration:
- .github
- docs
- README.md
- OWNERS
- CHANGELOG.md
- LICENSE
|
process
|
exclude documentation prs from end to end tests expected behavior documentation prs should not trigger end to end tests or require them to be merged in possible solution exclude the following from the prow configuration github docs readme md owners changelog md license
| 1
|
3,947
| 2,543,704,557
|
IssuesEvent
|
2015-01-29 01:01:10
|
GoogleCloudPlatform/kubernetes
|
https://api.github.com/repos/GoogleCloudPlatform/kubernetes
|
closed
|
kubelet fails to start with the --api_servers flag
|
kind/support priority/P0
|
Kubernetes Version: 0.8.2
Kubelet config:
```
/opt/bin/kubelet --address=0.0.0.0 --port=10250
--hostname_override=192.168.12.100 \
--api_servers=http://192.168.12.20:8080 \
--logtostderr=true
```
Kubelet Logs:
```
Jan 28 22:25:56 node0 systemd[1]: Started Kubernetes Kubelet.
Jan 28 22:25:56 node0 kubelet[1134]: I0128 22:25:56.068549 1134 node.go:52] Connecting to docker on unix:///var/run/docker.sock
Jan 28 22:25:56 node0 kubelet[1134]: E0128 22:25:56.068810 1134 util.go:131] Unable to make apiserver client: stat : no such file or directory
Jan 28 22:25:56 node0 kubelet[1134]: panic: runtime error: invalid memory address or nil pointer dereference
Jan 28 22:25:56 node0 kubelet[1134]: [signal 0xb code=0x1 addr=0x50 pc=0x5943e5]
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 16 [running]:
Jan 28 22:25:56 node0 kubelet[1134]: runtime.panic(0x9e2220, 0xf47893)
Jan 28 22:25:56 node0 kubelet[1134]: /usr/src/go/src/pkg/runtime/panic.c:279 +0xf5
Jan 28 22:25:56 node0 kubelet[1134]: github.com/coreos/go-etcd/etcd.(*Client).GetCluster(0x0, 0xc2080c6c30, 0x10, 0x4961fb)
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/Godeps/_workspace/src/github.com/coreos/go-etcd/etcd/client.go:284 +0x5
Jan 28 22:25:56 node0 kubelet[1134]: github.com/GoogleCloudPlatform/kubernetes/pkg/standalone.makePodSourceConfig(0xc20805a2c0, 0x10)
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/pkg/standalone/standalone.go:205 +0x2a2
Jan 28 22:25:56 node0 kubelet[1134]: github.com/GoogleCloudPlatform/kubernetes/pkg/standalone.RunKubelet(0xc20805a2c0)
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/pkg/standalone/standalone.go:164 +0xf3
Jan 28 22:25:56 node0 kubelet[1134]: main.main()
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/cmd/kubelet/kubelet.go:131 +0x530
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 17 [runnable]:
Jan 28 22:25:56 node0 kubelet[1134]: runtime.MHeap_Scavenger()
Jan 28 22:25:56 node0 kubelet[1134]: /usr/src/go/src/pkg/runtime/mheap.c:507
Jan 28 22:25:56 node0 kubelet[1134]: runtime.goexit()
Jan 28 22:25:56 node0 kubelet[1134]: /usr/src/go/src/pkg/runtime/proc.c:1445
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 18 [runnable]:
Jan 28 22:25:56 node0 kubelet[1134]: bgsweep()
Jan 28 22:25:56 node0 kubelet[1134]: /usr/src/go/src/pkg/runtime/mgc0.c:1976
Jan 28 22:25:56 node0 kubelet[1134]: runtime.goexit()
Jan 28 22:25:56 node0 kubelet[1134]: /usr/src/go/src/pkg/runtime/proc.c:1445
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 19 [runnable]:
Jan 28 22:25:56 node0 kubelet[1134]: runfinq()
Jan 28 22:25:56 node0 kubelet[1134]: /usr/src/go/src/pkg/runtime/mgc0.c:2606
Jan 28 22:25:56 node0 kubelet[1134]: runtime.goexit()
Jan 28 22:25:56 node0 kubelet[1134]: /usr/src/go/src/pkg/runtime/proc.c:1445
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 20 [runnable]:
Jan 28 22:25:56 node0 kubelet[1134]: github.com/golang/glog.(*loggingT).flushDaemon(0x10036c0)
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/Godeps/_workspace/src/github.com/golang/glog/glog.go:878
Jan 28 22:25:56 node0 kubelet[1134]: created by github.com/golang/glog.init<C2><B7>1
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/Godeps/_workspace/src/github.com/golang/glog/glog.go:410 +0x2b2
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 17 [syscall]:
Jan 28 22:25:56 node0 kubelet[1134]: runtime.goexit()
Jan 28 22:25:56 node0 kubelet[1134]: /usr/src/go/src/pkg/runtime/proc.c:1445
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 21 [runnable]:
Jan 28 22:25:56 node0 kubelet[1134]: github.com/GoogleCloudPlatform/kubernetes/pkg/watch.(*Broadcaster).loop(0xc208025020)
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/pkg/watch/mux.go:105
Jan 28 22:25:56 node0 kubelet[1134]: created by github.com/GoogleCloudPlatform/kubernetes/pkg/watch.NewBroadcaster
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/pkg/watch/mux.go:46 +0x9c
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 22 [runnable]:
Jan 28 22:25:56 node0 kubelet[1134]: github.com/GoogleCloudPlatform/kubernetes/pkg/util.Forever(0xd2d7a0, 0x12a05f200)
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/pkg/util/util.go:56
Jan 28 22:25:56 node0 kubelet[1134]: created by github.com/GoogleCloudPlatform/kubernetes/pkg/util.InitLogs
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/pkg/util/logs.go:48 +0x9a
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 23 [runnable]:
Jan 28 22:25:56 node0 kubelet[1134]: github.com/GoogleCloudPlatform/kubernetes/pkg/client/record.func<C2><B7>003()
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/pkg/client/record/event.go:100
Jan 28 22:25:56 node0 kubelet[1134]: created by github.com/GoogleCloudPlatform/kubernetes/pkg/client/record.GetEvents
Jan 28 22:25:56 node0 systemd[1]: kube-kubelet.service: main process exited, code=exited, status=2/INVALIDARGUMENT
Jan 28 22:25:56 node0 systemd[1]: Unit kube-kubelet.service entered failed state.
Jan 28 22:25:56 node0 systemd[1]: kube-kubelet.service failed.
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/pkg/client/record/event.go:115 +0xbc
```
|
1.0
|
kubelet fails to start with the --api_servers flag - Kubernetes Version: 0.8.2
Kubelet config:
```
/opt/bin/kubelet --address=0.0.0.0 --port=10250
--hostname_override=192.168.12.100 \
--api_servers=http://192.168.12.20:8080 \
--logtostderr=true
```
Kubelet Logs:
```
Jan 28 22:25:56 node0 systemd[1]: Started Kubernetes Kubelet.
Jan 28 22:25:56 node0 kubelet[1134]: I0128 22:25:56.068549 1134 node.go:52] Connecting to docker on unix:///var/run/docker.sock
Jan 28 22:25:56 node0 kubelet[1134]: E0128 22:25:56.068810 1134 util.go:131] Unable to make apiserver client: stat : no such file or directory
Jan 28 22:25:56 node0 kubelet[1134]: panic: runtime error: invalid memory address or nil pointer dereference
Jan 28 22:25:56 node0 kubelet[1134]: [signal 0xb code=0x1 addr=0x50 pc=0x5943e5]
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 16 [running]:
Jan 28 22:25:56 node0 kubelet[1134]: runtime.panic(0x9e2220, 0xf47893)
Jan 28 22:25:56 node0 kubelet[1134]: /usr/src/go/src/pkg/runtime/panic.c:279 +0xf5
Jan 28 22:25:56 node0 kubelet[1134]: github.com/coreos/go-etcd/etcd.(*Client).GetCluster(0x0, 0xc2080c6c30, 0x10, 0x4961fb)
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/Godeps/_workspace/src/github.com/coreos/go-etcd/etcd/client.go:284 +0x5
Jan 28 22:25:56 node0 kubelet[1134]: github.com/GoogleCloudPlatform/kubernetes/pkg/standalone.makePodSourceConfig(0xc20805a2c0, 0x10)
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/pkg/standalone/standalone.go:205 +0x2a2
Jan 28 22:25:56 node0 kubelet[1134]: github.com/GoogleCloudPlatform/kubernetes/pkg/standalone.RunKubelet(0xc20805a2c0)
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/pkg/standalone/standalone.go:164 +0xf3
Jan 28 22:25:56 node0 kubelet[1134]: main.main()
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/cmd/kubelet/kubelet.go:131 +0x530
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 17 [runnable]:
Jan 28 22:25:56 node0 kubelet[1134]: runtime.MHeap_Scavenger()
Jan 28 22:25:56 node0 kubelet[1134]: /usr/src/go/src/pkg/runtime/mheap.c:507
Jan 28 22:25:56 node0 kubelet[1134]: runtime.goexit()
Jan 28 22:25:56 node0 kubelet[1134]: /usr/src/go/src/pkg/runtime/proc.c:1445
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 18 [runnable]:
Jan 28 22:25:56 node0 kubelet[1134]: bgsweep()
Jan 28 22:25:56 node0 kubelet[1134]: /usr/src/go/src/pkg/runtime/mgc0.c:1976
Jan 28 22:25:56 node0 kubelet[1134]: runtime.goexit()
Jan 28 22:25:56 node0 kubelet[1134]: /usr/src/go/src/pkg/runtime/proc.c:1445
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 19 [runnable]:
Jan 28 22:25:56 node0 kubelet[1134]: runfinq()
Jan 28 22:25:56 node0 kubelet[1134]: /usr/src/go/src/pkg/runtime/mgc0.c:2606
Jan 28 22:25:56 node0 kubelet[1134]: runtime.goexit()
Jan 28 22:25:56 node0 kubelet[1134]: /usr/src/go/src/pkg/runtime/proc.c:1445
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 20 [runnable]:
Jan 28 22:25:56 node0 kubelet[1134]: github.com/golang/glog.(*loggingT).flushDaemon(0x10036c0)
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/Godeps/_workspace/src/github.com/golang/glog/glog.go:878
Jan 28 22:25:56 node0 kubelet[1134]: created by github.com/golang/glog.init<C2><B7>1
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/Godeps/_workspace/src/github.com/golang/glog/glog.go:410 +0x2b2
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 17 [syscall]:
Jan 28 22:25:56 node0 kubelet[1134]: runtime.goexit()
Jan 28 22:25:56 node0 kubelet[1134]: /usr/src/go/src/pkg/runtime/proc.c:1445
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 21 [runnable]:
Jan 28 22:25:56 node0 kubelet[1134]: github.com/GoogleCloudPlatform/kubernetes/pkg/watch.(*Broadcaster).loop(0xc208025020)
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/pkg/watch/mux.go:105
Jan 28 22:25:56 node0 kubelet[1134]: created by github.com/GoogleCloudPlatform/kubernetes/pkg/watch.NewBroadcaster
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/pkg/watch/mux.go:46 +0x9c
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 22 [runnable]:
Jan 28 22:25:56 node0 kubelet[1134]: github.com/GoogleCloudPlatform/kubernetes/pkg/util.Forever(0xd2d7a0, 0x12a05f200)
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/pkg/util/util.go:56
Jan 28 22:25:56 node0 kubelet[1134]: created by github.com/GoogleCloudPlatform/kubernetes/pkg/util.InitLogs
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/pkg/util/logs.go:48 +0x9a
Jan 28 22:25:56 node0 kubelet[1134]: goroutine 23 [runnable]:
Jan 28 22:25:56 node0 kubelet[1134]: github.com/GoogleCloudPlatform/kubernetes/pkg/client/record.func<C2><B7>003()
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/pkg/client/record/event.go:100
Jan 28 22:25:56 node0 kubelet[1134]: created by github.com/GoogleCloudPlatform/kubernetes/pkg/client/record.GetEvents
Jan 28 22:25:56 node0 systemd[1]: kube-kubelet.service: main process exited, code=exited, status=2/INVALIDARGUMENT
Jan 28 22:25:56 node0 systemd[1]: Unit kube-kubelet.service entered failed state.
Jan 28 22:25:56 node0 systemd[1]: kube-kubelet.service failed.
Jan 28 22:25:56 node0 kubelet[1134]: /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/pkg/client/record/event.go:115 +0xbc
```
|
non_process
|
kubelet fails to start with the api servers flag kubernetes version kubelet config opt bin kubelet address port hostname override api servers logtostderr true kubelet logs jan systemd started kubernetes kubelet jan kubelet node go connecting to docker on unix var run docker sock jan kubelet util go unable to make apiserver client stat no such file or directory jan kubelet panic runtime error invalid memory address or nil pointer dereference jan kubelet jan kubelet goroutine jan kubelet runtime panic jan kubelet usr src go src pkg runtime panic c jan kubelet github com coreos go etcd etcd client getcluster jan kubelet go src github com googlecloudplatform kubernetes godeps workspace src github com coreos go etcd etcd client go jan kubelet github com googlecloudplatform kubernetes pkg standalone makepodsourceconfig jan kubelet go src github com googlecloudplatform kubernetes output dockerized go src github com googlecloudplatform kubernetes pkg standalone standalone go jan kubelet github com googlecloudplatform kubernetes pkg standalone runkubelet jan kubelet go src github com googlecloudplatform kubernetes output dockerized go src github com googlecloudplatform kubernetes pkg standalone standalone go jan kubelet main main jan kubelet go src github com googlecloudplatform kubernetes output dockerized go src github com googlecloudplatform kubernetes cmd kubelet kubelet go jan kubelet goroutine jan kubelet runtime mheap scavenger jan kubelet usr src go src pkg runtime mheap c jan kubelet runtime goexit jan kubelet usr src go src pkg runtime proc c jan kubelet goroutine jan kubelet bgsweep jan kubelet usr src go src pkg runtime c jan kubelet runtime goexit jan kubelet usr src go src pkg runtime proc c jan kubelet goroutine jan kubelet runfinq jan kubelet usr src go src pkg runtime c jan kubelet runtime goexit jan kubelet usr src go src pkg runtime proc c jan kubelet goroutine jan kubelet github com golang glog loggingt flushdaemon jan kubelet go src github com googlecloudplatform kubernetes godeps workspace src github com golang glog glog go jan kubelet created by github com golang glog init jan kubelet go src github com googlecloudplatform kubernetes godeps workspace src github com golang glog glog go jan kubelet goroutine jan kubelet runtime goexit jan kubelet usr src go src pkg runtime proc c jan kubelet goroutine jan kubelet github com googlecloudplatform kubernetes pkg watch broadcaster loop jan kubelet go src github com googlecloudplatform kubernetes output dockerized go src github com googlecloudplatform kubernetes pkg watch mux go jan kubelet created by github com googlecloudplatform kubernetes pkg watch newbroadcaster jan kubelet go src github com googlecloudplatform kubernetes output dockerized go src github com googlecloudplatform kubernetes pkg watch mux go jan kubelet goroutine jan kubelet github com googlecloudplatform kubernetes pkg util forever jan kubelet go src github com googlecloudplatform kubernetes output dockerized go src github com googlecloudplatform kubernetes pkg util util go jan kubelet created by github com googlecloudplatform kubernetes pkg util initlogs jan kubelet go src github com googlecloudplatform kubernetes output dockerized go src github com googlecloudplatform kubernetes pkg util logs go jan kubelet goroutine jan kubelet github com googlecloudplatform kubernetes pkg client record func jan kubelet go src github com googlecloudplatform kubernetes output dockerized go src github com googlecloudplatform kubernetes pkg client record event go jan kubelet created by github com googlecloudplatform kubernetes pkg client record getevents jan systemd kube kubelet service main process exited code exited status invalidargument jan systemd unit kube kubelet service entered failed state jan systemd kube kubelet service failed jan kubelet go src github com googlecloudplatform kubernetes output dockerized go src github com googlecloudplatform kubernetes pkg client record event go
| 0
|
599,176
| 18,266,866,823
|
IssuesEvent
|
2021-10-04 09:27:29
|
Project-Easter/Flutter-UI
|
https://api.github.com/repos/Project-Easter/Flutter-UI
|
closed
|
Delete all the commented files
|
easy high-priority Hacktoberfest good first issue
|
## Description
*Delete* all the files that are *fully* commented. An example has been shown in the screenshot.
## Screenshots

## Additional data
List the names of all the files you deleted and write them in your pull request.
Feel free to ask anything in the comments and do not *delete* any folder before asking here.
|
1.0
|
Delete all the commented files - ## Description
*Delete* all the files that are *fully* commented. An example has been shown in the screenshot.
## Screenshots

## Additional data
List the names of all the files you deleted and write them in your pull request.
Feel free to ask anything in the comments and do not *delete* any folder before asking here.
|
non_process
|
delete all the commented files description delete all the files that are fully commented an example has been shown in the screenshot screenshots additional data list the names of all the files you deleted and write them in your pull request feel free to ask anything in the comments and do not delete any folder before asking here
| 0
|
6,048
| 8,870,817,797
|
IssuesEvent
|
2019-01-11 10:36:40
|
enKryptIO/ethvm
|
https://api.github.com/repos/enKryptIO/ethvm
|
closed
|
Apply appropriate resource constraints to the various docker services to prevent resource starvation/conflicts when running a local mainnet test
|
priority:medium project:processing
|
When running everything locally and syncing against mainnet eventually MongoDB attempts to consume too many resources. We can however constrain the resources allocated to each service with the appropriate docker compose config
|
1.0
|
Apply appropriate resource constraints to the various docker services to prevent resource starvation/conflicts when running a local mainnet test - When running everything locally and syncing against mainnet eventually MongoDB attempts to consume too many resources. We can however constrain the resources allocated to each service with the appropriate docker compose config
|
process
|
apply appropriate resource constraints to the various docker services to prevent resource starvation conflicts when running a local mainnet test when running everything locally and syncing against mainnet eventually mongodb attempts to consume too many resources we can however constrain the resources allocated to each service with the appropriate docker compose config
| 1
|
43,900
| 2,893,939,415
|
IssuesEvent
|
2015-06-15 20:37:05
|
PresConsUIUC/PSAP
|
https://api.github.com/repos/PresConsUIUC/PSAP
|
closed
|
change where link for help in upper right of application links to
|
priority-high
|
@jamieviva @adolski we need to change the landing page for the help button - currently goes to digested advanced help (for lack of a better term). change to HTML landing page for all things help - with intro text providing context for each mode of help - ie. user manual, basic help, advanced help, camtasia videos, and format ID guide. Jamie will draft content and get ready for alex to implement for next milestone.
|
1.0
|
change where link for help in upper right of application links to - @jamieviva @adolski we need to change the landing page for the help button - currently goes to digested advanced help (for lack of a better term). change to HTML landing page for all things help - with intro text providing context for each mode of help - ie. user manual, basic help, advanced help, camtasia videos, and format ID guide. Jamie will draft content and get ready for alex to implement for next milestone.
|
non_process
|
change where link for help in upper right of application links to jamieviva adolski we need to change the landing page for the help button currently goes to digested advanced help for lack of a better term change to html landing page for all things help with intro text providing context for each mode of help ie user manual basic help advanced help camtasia videos and format id guide jamie will draft content and get ready for alex to implement for next milestone
| 0
|
167,778
| 6,346,400,520
|
IssuesEvent
|
2017-07-28 01:59:42
|
openshift/origin
|
https://api.github.com/repos/openshift/origin
|
closed
|
Error message from admission controller contains something that is wrong
|
priority/P2
|
```
Standard error from the command:
Error from server (InternalError): Internal error occurred: [exec operation is not allowed because the pod's security context exceeds your permissions: pods "ruby-sample-build-1-build" is forbidden: unable to validate against any security context constraint: [spec.containers[0].securityContext.privileged: Invalid value: true: Privileged containers are not allowed spec.containers[0].securityContext.volumes[0]: Invalid value: "hostPath": hostPath volumes are not allowed to be used], object does not implement the Object interfaces]
Running test/end-to-end/core.sh:449: executing 'oc logs build/ruby-sample-build-1 --namespace test -f > '/tmp/openshift/test-end-to-end-docker//logs/test-build.log'' expecting success...
```
`object does not implement the Object interfaces` should *not* be in here.
|
1.0
|
Error message from admission controller contains something that is wrong - ```
Standard error from the command:
Error from server (InternalError): Internal error occurred: [exec operation is not allowed because the pod's security context exceeds your permissions: pods "ruby-sample-build-1-build" is forbidden: unable to validate against any security context constraint: [spec.containers[0].securityContext.privileged: Invalid value: true: Privileged containers are not allowed spec.containers[0].securityContext.volumes[0]: Invalid value: "hostPath": hostPath volumes are not allowed to be used], object does not implement the Object interfaces]
Running test/end-to-end/core.sh:449: executing 'oc logs build/ruby-sample-build-1 --namespace test -f > '/tmp/openshift/test-end-to-end-docker//logs/test-build.log'' expecting success...
```
`object does not implement the Object interfaces` should *not* be in here.
|
non_process
|
error message from admission controller contains something that is wrong standard error from the command error from server internalerror internal error occurred securitycontext privileged invalid value true privileged containers are not allowed spec containers securitycontext volumes invalid value hostpath hostpath volumes are not allowed to be used object does not implement the object interfaces running test end to end core sh executing oc logs build ruby sample build namespace test f tmp openshift test end to end docker logs test build log expecting success object does not implement the object interfaces should not be in here
| 0
|
12,841
| 15,223,395,247
|
IssuesEvent
|
2021-02-18 02:34:01
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
MongoDB shouldn't do an initial projection for :default date bucketing
|
Database/Mongo Querying/Processor Type:Performance
|
Given a query like
```clj
(mt/dataset attempted-murders
(mt/mbql-query attempts
{:aggregation [[:count]]
:filter [:time-interval $datetime :last :month]}))
```
Which end up with a native query like this: (including #14835 optimizations):
```clj
[{"$project" {"datetime~~~default"
{:$let {:vars {:column "$datetime"}, :in :$$column}}}}
{"$match"
{"$and"
[{"datetime~~~default" {"$gte" {:___date "2021-01"}}}
{"datetime~~~default" {"$lt" {:___date "2021-02"}}}]}}
{"$group" {"_id" nil, "count" {"$sum" 1}}}
{"$sort" {"_id" 1}}
{"$project" {"_id" false, "count" true}}]
```
That initial projection is going to destroy potential usage of indexes, and we don't need it at all, since `:default` just means we explicitly do not want any extra temporal bucketing.
This is a subset of #4216. I fixed most of #4216 in #11832, but cases such as these are still causing initial projections... by solving this specific case we *may* be able to finally close #4216 entirely.
|
1.0
|
MongoDB shouldn't do an initial projection for :default date bucketing - Given a query like
```clj
(mt/dataset attempted-murders
(mt/mbql-query attempts
{:aggregation [[:count]]
:filter [:time-interval $datetime :last :month]}))
```
Which end up with a native query like this: (including #14835 optimizations):
```clj
[{"$project" {"datetime~~~default"
{:$let {:vars {:column "$datetime"}, :in :$$column}}}}
{"$match"
{"$and"
[{"datetime~~~default" {"$gte" {:___date "2021-01"}}}
{"datetime~~~default" {"$lt" {:___date "2021-02"}}}]}}
{"$group" {"_id" nil, "count" {"$sum" 1}}}
{"$sort" {"_id" 1}}
{"$project" {"_id" false, "count" true}}]
```
That initial projection is going to destroy potential usage of indexes, and we don't need it at all, since `:default` just means we explicitly do not want any extra temporal bucketing.
This is a subset of #4216. I fixed most of #4216 in #11832, but cases such as these are still causing initial projections... by solving this specific case we *may* be able to finally close #4216 entirely.
|
process
|
mongodb shouldn t do an initial projection for default date bucketing given a query like clj mt dataset attempted murders mt mbql query attempts aggregation filter which end up with a native query like this including optimizations clj project datetime default let vars column datetime in column match and datetime default gte date datetime default lt date group id nil count sum sort id project id false count true that initial projection is going to destroy potential usage of indexes and we don t need it at all since default just means we explicitly do not want any extra temporal bucketing this is a subset of i fixed most of in but cases such as these are still causing initial projections by solving this specific case we may be able to finally close entirely
| 1
|
20,345
| 27,002,528,007
|
IssuesEvent
|
2023-02-10 09:04:33
|
tradingstrategy-ai/frontend
|
https://api.github.com/repos/tradingstrategy-ai/frontend
|
closed
|
Add compatibility glue to the /tokens endpoint consumers
|
process priority: P1 size: XS
|
As discussed offline, the current implementation of the `/tokens` API endpoint deviates from its specs. The paginated response object contains a field named "result" containing a list of search results, while the the actual naming should be "results" - both to match the spec and for consistency with paginated responses from other endpoints.
When this temporary compatibility glue is in place, we can change the API implementation, after which the glue can be removed.
|
1.0
|
Add compatibility glue to the /tokens endpoint consumers - As discussed offline, the current implementation of the `/tokens` API endpoint deviates from its specs. The paginated response object contains a field named "result" containing a list of search results, while the the actual naming should be "results" - both to match the spec and for consistency with paginated responses from other endpoints.
When this temporary compatibility glue is in place, we can change the API implementation, after which the glue can be removed.
|
process
|
add compatibility glue to the tokens endpoint consumers as discussed offline the current implementation of the tokens api endpoint deviates from its specs the paginated response object contains a field named result containing a list of search results while the the actual naming should be results both to match the spec and for consistency with paginated responses from other endpoints when this temporary compatibility glue is in place we can change the api implementation after which the glue can be removed
| 1
|
62,195
| 17,023,870,091
|
IssuesEvent
|
2021-07-03 04:16:59
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Incorrect city for postal code 77429
|
Component: nominatim Priority: minor Resolution: wontfix Type: defect
|
**[Submitted to the original trac issue database at 4.33pm, Thursday, 25th July 2013]**
77429 is one of two US zip codes assigned to Cypress, TX. Nominatim is reporting the city as Stillwater Village, which is a small neighborhood within this zip code.
|
1.0
|
Incorrect city for postal code 77429 - **[Submitted to the original trac issue database at 4.33pm, Thursday, 25th July 2013]**
77429 is one of two US zip codes assigned to Cypress, TX. Nominatim is reporting the city as Stillwater Village, which is a small neighborhood within this zip code.
|
non_process
|
incorrect city for postal code is one of two us zip codes assigned to cypress tx nominatim is reporting the city as stillwater village which is a small neighborhood within this zip code
| 0
|
80,637
| 30,454,362,960
|
IssuesEvent
|
2023-07-16 17:48:01
|
vector-im/element-desktop
|
https://api.github.com/repos/vector-im/element-desktop
|
opened
|
Matrix formatting doesn't handle color "gold" across auto/light/dark/black modes
|
T-Defect
|
### Steps to reproduce
Submit formatted room message:
<b><font color=\"gold\"> hello world</font></b>
Client in Auto mode (light): displays
Client in light mode: displays
Client in Black mode: displays
Client in Dark mode: Text shows up as black and is unreadable.
### Outcome
#### What did you expect?
Text to show up in gold even in dark mode given the color
#### What happened instead?
Text showed up as black instead of gold making it unreadable.
### Operating system
Windows 10 x64
### Application version
Element version: 1.11.35 Olm version: 3.2.14
### How did you install the app?
element.io web site as desktop app
### Homeserver
matrix-synapse==1.84.1 matrix-synapse-ldap3==0.1.4
### Will you send logs?
No
|
1.0
|
Matrix formatting doesn't handle color "gold" across auto/light/dark/black modes - ### Steps to reproduce
Submit formatted room message:
<b><font color=\"gold\"> hello world</font></b>
Client in Auto mode (light): displays
Client in light mode: displays
Client in Black mode: displays
Client in Dark mode: Text shows up as black and is unreadable.
### Outcome
#### What did you expect?
Text to show up in gold even in dark mode given the color
#### What happened instead?
Text showed up as black instead of gold making it unreadable.
### Operating system
Windows 10 x64
### Application version
Element version: 1.11.35 Olm version: 3.2.14
### How did you install the app?
element.io web site as desktop app
### Homeserver
matrix-synapse==1.84.1 matrix-synapse-ldap3==0.1.4
### Will you send logs?
No
|
non_process
|
matrix formatting doesn t handle color gold across auto light dark black modes steps to reproduce submit formatted room message hello world client in auto mode light displays client in light mode displays client in black mode displays client in dark mode text shows up as black and is unreadable outcome what did you expect text to show up in gold even in dark mode given the color what happened instead text showed up as black instead of gold making it unreadable operating system windows application version element version olm version how did you install the app element io web site as desktop app homeserver matrix synapse matrix synapse will you send logs no
| 0
|
1,705
| 4,350,053,762
|
IssuesEvent
|
2016-07-31 00:34:37
|
arianht/meantweets
|
https://api.github.com/repos/arianht/meantweets
|
opened
|
Update README to document development information
|
P2 process
|
- How to setup the project?
- How to run locally?
- What is the development process?
|
1.0
|
Update README to document development information - - How to setup the project?
- How to run locally?
- What is the development process?
|
process
|
update readme to document development information how to setup the project how to run locally what is the development process
| 1
|
26,424
| 5,253,276,773
|
IssuesEvent
|
2017-02-02 09:03:16
|
restic/restic
|
https://api.github.com/repos/restic/restic
|
closed
|
Document creating a new REST backend
|
backend:rest documentation
|
As agreed with @zcalusic: Amend the REST protocol description and add an explicit `POST {path}?create=true` to create the repository on the server side.
|
1.0
|
Document creating a new REST backend - As agreed with @zcalusic: Amend the REST protocol description and add an explicit `POST {path}?create=true` to create the repository on the server side.
|
non_process
|
document creating a new rest backend as agreed with zcalusic amend the rest protocol description and add an explicit post path create true to create the repository on the server side
| 0
|
831,283
| 32,044,309,497
|
IssuesEvent
|
2023-09-22 22:50:59
|
rhunk/SnapEnhance
|
https://api.github.com/repos/rhunk/SnapEnhance
|
closed
|
feat: Add marker next to deleted logged messages
|
enhancement low priority accepted
|
### Description
As the title says, adding a marker (e.g. a small yellow circle to the right of the message) to deleted messages would make it more clear as to which messages a friend has deleted.
### Agreement
- [X] There is no issue already suggesting my feature.
|
1.0
|
feat: Add marker next to deleted logged messages - ### Description
As the title says, adding a marker (e.g. a small yellow circle to the right of the message) to deleted messages would make it more clear as to which messages a friend has deleted.
### Agreement
- [X] There is no issue already suggesting my feature.
|
non_process
|
feat add marker next to deleted logged messages description as the title says adding a marker e g a small yellow circle to the right of the message to deleted messages would make it more clear as to which messages a friend has deleted agreement there is no issue already suggesting my feature
| 0
|
16,827
| 22,060,991,260
|
IssuesEvent
|
2022-05-30 17:48:29
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
I don't know how to deploy extensions using Azure PowerShell and Azure CLI.
|
automation/svc triaged assigned-to-author doc-enhancement process-automation/subsvc Pri2
|
Hi team,
If HRW extension is going through a proxy, we will need to configure the proxy settings.
The proxy settings require deployment using PowerShell or Azure CLI.
However, the following document does not tell me how to execute the command.
If we are using PowerShell, we will probably need to run Set-AzVMExtension command.
It would be helpful to have a sample of Set-AzVMExtension.
https://docs.microsoft.com/en-us/azure/automation/extension-based-hybrid-runbook-worker-install?tabs=windows#proxy-server-use

---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 2854aee3-7933-ee94-bb03-3a6a4c540088
* Version Independent ID: c902ddb9-46bf-c53c-46c4-e1c3d2cec7fd
* Content: [Deploy an extension-based Windows or Linux User Hybrid Runbook Worker in Azure Automation (Preview)](https://docs.microsoft.com/en-us/azure/automation/extension-based-hybrid-runbook-worker-install?tabs=windows#add-a-machine-to-a-hybrid-worker-group)
* Content Source: [articles/automation/extension-based-hybrid-runbook-worker-install.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/extension-based-hybrid-runbook-worker-install.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SGSneha
* Microsoft Alias: **v-ssudhir**
|
1.0
|
I don't know how to deploy extensions using Azure PowerShell and Azure CLI. -
Hi team,
If HRW extension is going through a proxy, we will need to configure the proxy settings.
The proxy settings require deployment using PowerShell or Azure CLI.
However, the following document does not tell me how to execute the command.
If we are using PowerShell, we will probably need to run Set-AzVMExtension command.
It would be helpful to have a sample of Set-AzVMExtension.
https://docs.microsoft.com/en-us/azure/automation/extension-based-hybrid-runbook-worker-install?tabs=windows#proxy-server-use

---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 2854aee3-7933-ee94-bb03-3a6a4c540088
* Version Independent ID: c902ddb9-46bf-c53c-46c4-e1c3d2cec7fd
* Content: [Deploy an extension-based Windows or Linux User Hybrid Runbook Worker in Azure Automation (Preview)](https://docs.microsoft.com/en-us/azure/automation/extension-based-hybrid-runbook-worker-install?tabs=windows#add-a-machine-to-a-hybrid-worker-group)
* Content Source: [articles/automation/extension-based-hybrid-runbook-worker-install.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/extension-based-hybrid-runbook-worker-install.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SGSneha
* Microsoft Alias: **v-ssudhir**
|
process
|
i don t know how to deploy extensions using azure powershell and azure cli hi team if hrw extension is going through a proxy we will need to configure the proxy settings the proxy settings require deployment using powershell or azure cli however the following document does not tell me how to execute the command if we are using powershell we will probably need to run set azvmextension command it would be helpful to have a sample of set azvmextension document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login sgsneha microsoft alias v ssudhir
| 1
|
386,381
| 26,680,378,484
|
IssuesEvent
|
2023-01-26 17:11:55
|
trpc/trpc
|
https://api.github.com/repos/trpc/trpc
|
closed
|
bug: codesandbox repro example not working
|
📚 documentation / examples blocked 🐛 bug: unconfirmed
|
### Provide environment information
https://stackblitz.com/github/trpc/examples-next-minimal-starter
Chrome: 108.0.5359.124
Firefox: 107.0.1
### Describe the bug
Currently, the stackblitz project doesn't start due to some swc error
<img width="702" alt="CleanShot 2022-12-20 at 09 55 21@2x" src="https://user-images.githubusercontent.com/51714798/208625352-c316db4f-3776-4654-9de1-47087a8e10f5.png">
### Link to reproduction
https://stackblitz.com/github/trpc/examples-next-minimal-starter
### To reproduce
Go to link, check console
### Additional information
Solutions:
- Move to CodeSandbox? Slow
- Look into error and figuring out what's causing it
### 👨👧👦 Contributing
- [ ] 🙋♂️ Yes, I'd be down to file a PR fixing this bug!
|
1.0
|
bug: codesandbox repro example not working - ### Provide environment information
https://stackblitz.com/github/trpc/examples-next-minimal-starter
Chrome: 108.0.5359.124
Firefox: 107.0.1
### Describe the bug
Currently, the stackblitz project doesn't start due to some swc error
<img width="702" alt="CleanShot 2022-12-20 at 09 55 21@2x" src="https://user-images.githubusercontent.com/51714798/208625352-c316db4f-3776-4654-9de1-47087a8e10f5.png">
### Link to reproduction
https://stackblitz.com/github/trpc/examples-next-minimal-starter
### To reproduce
Go to link, check console
### Additional information
Solutions:
- Move to CodeSandbox? Slow
- Look into error and figuring out what's causing it
### 👨👧👦 Contributing
- [ ] 🙋♂️ Yes, I'd be down to file a PR fixing this bug!
|
non_process
|
bug codesandbox repro example not working provide environment information chrome firefox describe the bug currently the stackblitz project doesn t start due to some swc error img width alt cleanshot at src link to reproduction to reproduce go to link check console additional information solutions move to codesandbox slow look into error and figuring out what s causing it 👨👧👦 contributing 🙋♂️ yes i d be down to file a pr fixing this bug
| 0
|
15,112
| 18,848,250,902
|
IssuesEvent
|
2021-11-11 17:18:39
|
googleapis/python-bigquery-storage
|
https://api.github.com/repos/googleapis/python-bigquery-storage
|
closed
|
tests: test with pyarrow on Python 3.9
|
status: blocked type: process api: bigquerystorage
|
Once pyarrow publishes Python 3.9 wheels to PyPI, we can remove our hacks from synth.py & noxfile.py to make sure we tests those code paths.
|
1.0
|
tests: test with pyarrow on Python 3.9 - Once pyarrow publishes Python 3.9 wheels to PyPI, we can remove our hacks from synth.py & noxfile.py to make sure we tests those code paths.
|
process
|
tests test with pyarrow on python once pyarrow publishes python wheels to pypi we can remove our hacks from synth py noxfile py to make sure we tests those code paths
| 1
|
63,220
| 12,293,213,605
|
IssuesEvent
|
2020-05-10 17:58:43
|
pi-hole/FTL
|
https://api.github.com/repos/pi-hole/FTL
|
closed
|
Unable to clone/DEV via Windows due to repo directory naming using reserved device names
|
Code maintanance Fixed in next release
|
The use of the directory named `aux` in this repo makes it impossible to successfully fully clone and build/develop at all on a Windows environment due to this being one of the several "reserved" names (for device use) implemented back in the mid 1970's.
I was aware of this so was not surprised when I saw the git error upon cloning on WIN regarding the directory named `AUX` causing the catastrophic git clone failure (my hair is really grey) but for most it is obviously probably not known and/or they are not attempting to clone and develop using a Windows environment.
Here is a bit more info from MS stating as much and some background ("file" also obviously goes for "directory" as they are both just file table/journal entries):
https://docs.microsoft.com/en-us/windows/desktop/fileio/naming-a-file
```
Do not use the following reserved names for the name of a file:
CON, PRN, AUX, NUL, COM1, COM2, COM3, COM4, COM5, COM6, COM7, COM8, COM9, LPT1, LPT2, LPT3, LPT4, LPT5, LPT6, LPT7, LPT8, and LPT9. Also avoid these names followed immediately by an extension; for example, NUL.txt is not recommended. For more information, see Namespaces.
```
Since many DEVs, despite having access to Linux-based systems, perform their DEV workflows in a Windows environment it would be great to have this fixed so the repo can exist on a windows volume. TIA!
|
1.0
|
Unable to clone/DEV via Windows due to repo directory naming using reserved device names - The use of the directory named `aux` in this repo makes it impossible to successfully fully clone and build/develop at all on a Windows environment due to this being one of the several "reserved" names (for device use) implemented back in the mid 1970's.
I was aware of this so was not surprised when I saw the git error upon cloning on WIN regarding the directory named `AUX` causing the catastrophic git clone failure (my hair is really grey) but for most it is obviously probably not known and/or they are not attempting to clone and develop using a Windows environment.
Here is a bit more info from MS stating as much and some background ("file" also obviously goes for "directory" as they are both just file table/journal entries):
https://docs.microsoft.com/en-us/windows/desktop/fileio/naming-a-file
```
Do not use the following reserved names for the name of a file:
CON, PRN, AUX, NUL, COM1, COM2, COM3, COM4, COM5, COM6, COM7, COM8, COM9, LPT1, LPT2, LPT3, LPT4, LPT5, LPT6, LPT7, LPT8, and LPT9. Also avoid these names followed immediately by an extension; for example, NUL.txt is not recommended. For more information, see Namespaces.
```
Since many DEVs, despite having access to Linux-based systems, perform their DEV workflows in a Windows environment it would be great to have this fixed so the repo can exist on a windows volume. TIA!
|
non_process
|
unable to clone dev via windows due to repo directory naming using reserved device names the use of the directory named aux in this repo makes it impossible to successfully fully clone and build develop at all on a windows environment due to this being one of the several reserved names for device use implemented back in the mid s i was aware of this so was not surprised when i saw the git error upon cloning on win regarding the directory named aux causing the catastrophic git clone failure my hair is really grey but for most it is obviously probably not known and or they are not attempting to clone and develop using a windows environment here is a bit more info from ms stating as much and some background file also obviously goes for directory as they are both just file table journal entries do not use the following reserved names for the name of a file con prn aux nul and also avoid these names followed immediately by an extension for example nul txt is not recommended for more information see namespaces since many devs despite having access to linux based systems perform their dev workflows in a windows environment it would be great to have this fixed so the repo can exist on a windows volume tia
| 0
|
19,512
| 25,828,110,127
|
IssuesEvent
|
2022-12-12 14:20:20
|
microsoft/vscode
|
https://api.github.com/repos/microsoft/vscode
|
closed
|
Alpine Linux: /bin/ps: unrecognized option: x
|
bug linux workbench-diagnostics terminal-process
|
Alpine Linux' `ps` does not appear to have the `x` option:
```
[IPC Library: Pty Host] DEBUG ChildProcessMonitor: Fetching process tree failed Error: Command failed: /bin/ps -ax -o pid=,ppid=,pcpu=,pmem=,command=
/bin/ps: unrecognized option: x
BusyBox v1.34.1 (2021-11-23 00:57:35 UTC) multi-call binary.
Usage: ps [-o COL1,COL2=HEADER] [-T]
Show list of processes
-o COL1,COL2=HEADER Select columns for display
-T Show threads
at ChildProcess.exithandler (child_process.js:308:12)
at ChildProcess.emit (events.js:315:20)
at maybeClose (internal/child_process.js:1048:16)
at Socket.<anonymous> (internal/child_process.js:439:11)
at Socket.emit (events.js:315:20)
at Pipe.<anonymous> (net.js:673:12)
```
Version: 1.64.0-insider
Commit: 45060913204279d4ecfe0575ca69d68211fb3798
Date: 2022-01-26T05:15:06.786Z
Electron: 13.5.2
Chromium: 91.0.4472.164
Node.js: 14.16.0
V8: 9.1.269.39-electron.0
OS: Darwin x64 20.6.0
|
1.0
|
Alpine Linux: /bin/ps: unrecognized option: x - Alpine Linux' `ps` does not appear to have the `x` option:
```
[IPC Library: Pty Host] DEBUG ChildProcessMonitor: Fetching process tree failed Error: Command failed: /bin/ps -ax -o pid=,ppid=,pcpu=,pmem=,command=
/bin/ps: unrecognized option: x
BusyBox v1.34.1 (2021-11-23 00:57:35 UTC) multi-call binary.
Usage: ps [-o COL1,COL2=HEADER] [-T]
Show list of processes
-o COL1,COL2=HEADER Select columns for display
-T Show threads
at ChildProcess.exithandler (child_process.js:308:12)
at ChildProcess.emit (events.js:315:20)
at maybeClose (internal/child_process.js:1048:16)
at Socket.<anonymous> (internal/child_process.js:439:11)
at Socket.emit (events.js:315:20)
at Pipe.<anonymous> (net.js:673:12)
```
Version: 1.64.0-insider
Commit: 45060913204279d4ecfe0575ca69d68211fb3798
Date: 2022-01-26T05:15:06.786Z
Electron: 13.5.2
Chromium: 91.0.4472.164
Node.js: 14.16.0
V8: 9.1.269.39-electron.0
OS: Darwin x64 20.6.0
|
process
|
alpine linux bin ps unrecognized option x alpine linux ps does not appear to have the x option debug childprocessmonitor fetching process tree failed error command failed bin ps ax o pid ppid pcpu pmem command bin ps unrecognized option x busybox utc multi call binary usage ps show list of processes o header select columns for display t show threads at childprocess exithandler child process js at childprocess emit events js at maybeclose internal child process js at socket internal child process js at socket emit events js at pipe net js version insider commit date electron chromium node js electron os darwin
| 1
|
720
| 3,206,808,212
|
IssuesEvent
|
2015-10-05 06:00:31
|
e-government-ua/i
|
https://api.github.com/repos/e-government-ua/i
|
closed
|
На главном портале реализовать два новых валидатора - по ОКПО и МФО
|
active bug hi priority In process of testing question test
|
11) EDRPOU //код ЄДРПОУ.
Логика: жестко восемь цифр, тип стринг(чтобы можно было ставить default=” ”)
Сообщение: Такий код ЄДРПОУ не існує - (ви не можете вписувати літери)
Поля: edrpou
12) MFO //код банка.
Логика: жестко шесть цифр.тип стринг.(чтобы можно было ставить default=” ”)
Сообщение: Такого коду банку не існує - (ви не можете вписувати літери)
Поля: mfo
|
1.0
|
На главном портале реализовать два новых валидатора - по ОКПО и МФО - 11) EDRPOU //код ЄДРПОУ.
Логика: жестко восемь цифр, тип стринг(чтобы можно было ставить default=” ”)
Сообщение: Такий код ЄДРПОУ не існує - (ви не можете вписувати літери)
Поля: edrpou
12) MFO //код банка.
Логика: жестко шесть цифр.тип стринг.(чтобы можно было ставить default=” ”)
Сообщение: Такого коду банку не існує - (ви не можете вписувати літери)
Поля: mfo
|
process
|
на главном портале реализовать два новых валидатора по окпо и мфо edrpou код єдрпоу логика жестко восемь цифр тип стринг чтобы можно было ставить default ” ” сообщение такий код єдрпоу не існує ви не можете вписувати літери поля edrpou mfo код банка логика жестко шесть цифр тип стринг чтобы можно было ставить default ” ” сообщение такого коду банку не існує ви не можете вписувати літери поля mfo
| 1
|
16,555
| 21,568,886,226
|
IssuesEvent
|
2022-05-02 04:54:53
|
medic/cht-core
|
https://api.github.com/repos/medic/cht-core
|
opened
|
Release 3.15.0-FR-csv-imports
|
Type: Internal process
|
When development is ready to begin on a [Feature Release](https://docs.communityhealthtoolkit.org/core/releases/feature_releases/#release-names), an engineer on the appropriate Care Team or Allies should be nominated as a Release Engineer. They will be responsible for making sure the following tasks are followed, though not necessarily doing the work themselves.
# Planning
- [ ] Create a GH Milestone for the release.
- [ ] Add all the issues to be worked on to the Milestone.
- [X] Have an actual named deployment and specific end user that will be testing this Feature Release. They need to test in production, on the latest version. No speculative Feature Releases.
- [X] Assign an engineer as Release Engineer for this release. - @latin-panda \o/
# Development
- [ ] Create a new release branch in `cht-core` from the most recent release and call it `<major>.<minor>.<patch>-FR-<FEATURE-NAME>`. If latest is `3.15.0` and the feature is to "allow movies to be uploaded", call it `3.15.0-FR-movie-upload`. Done before the release so all PRs can be set to merge to this branch, and not to `master`.
- [ ] Set the version number in `package.json` and `package-lock.json` and submit a PR. The easiest way to do this is to use `npm --no-git-tag-version version <feature-release>`.
- [ ] Ensure QA is briefed and is partnering with the Trio to ensure early and often checks of the feature are on track to be of production quality from the start.
# Releasing
This is an iterative process and it's assumed there will be multiple numbered releases throughout development of the Feature Release.
- [ ] Build a beta named `<major>.<minor>.<patch>-FR-<FEATURE-NAME>-1` by pushing a git tag and when CI completes successfully notify the QA team that it's ready for release testing. If an updated Feature Release is needed, increment the last `1` by calling it `<major>.<minor>.<patch>-FR-<FEATURE-NAME>-2` etc.
# Close-out
- [ ] Validate with the actual end user that this Feature Release delivers a quantifiable improvement. If yes, plan on adding the feature to the next minor release by creating a new ticket to merge the code to `master`. If no, we leave the code dead in this branch, never to be merged to `master`, but still loved all the same.
- [ ] Mark this issue "done" and close the Milestone.
|
1.0
|
Release 3.15.0-FR-csv-imports - When development is ready to begin on a [Feature Release](https://docs.communityhealthtoolkit.org/core/releases/feature_releases/#release-names), an engineer on the appropriate Care Team or Allies should be nominated as a Release Engineer. They will be responsible for making sure the following tasks are followed, though not necessarily doing the work themselves.
# Planning
- [ ] Create a GH Milestone for the release.
- [ ] Add all the issues to be worked on to the Milestone.
- [X] Have an actual named deployment and specific end user that will be testing this Feature Release. They need to test in production, on the latest version. No speculative Feature Releases.
- [X] Assign an engineer as Release Engineer for this release. - @latin-panda \o/
# Development
- [ ] Create a new release branch in `cht-core` from the most recent release and call it `<major>.<minor>.<patch>-FR-<FEATURE-NAME>`. If latest is `3.15.0` and the feature is to "allow movies to be uploaded", call it `3.15.0-FR-movie-upload`. Done before the release so all PRs can be set to merge to this branch, and not to `master`.
- [ ] Set the version number in `package.json` and `package-lock.json` and submit a PR. The easiest way to do this is to use `npm --no-git-tag-version version <feature-release>`.
- [ ] Ensure QA is briefed and is partnering with the Trio to ensure early and often checks of the feature are on track to be of production quality from the start.
# Releasing
This is an iterative process and it's assumed there will be multiple numbered releases throughout development of the Feature Release.
- [ ] Build a beta named `<major>.<minor>.<patch>-FR-<FEATURE-NAME>-1` by pushing a git tag and when CI completes successfully notify the QA team that it's ready for release testing. If an updated Feature Release is needed, increment the last `1` by calling it `<major>.<minor>.<patch>-FR-<FEATURE-NAME>-2` etc.
# Close-out
- [ ] Validate with the actual end user that this Feature Release delivers a quantifiable improvement. If yes, plan on adding the feature to the next minor release by creating a new ticket to merge the code to `master`. If no, we leave the code dead in this branch, never to be merged to `master`, but still loved all the same.
- [ ] Mark this issue "done" and close the Milestone.
|
process
|
release fr csv imports when development is ready to begin on a an engineer on the appropriate care team or allies should be nominated as a release engineer they will be responsible for making sure the following tasks are followed though not necessarily doing the work themselves planning create a gh milestone for the release add all the issues to be worked on to the milestone have an actual named deployment and specific end user that will be testing this feature release they need to test in production on the latest version no speculative feature releases assign an engineer as release engineer for this release latin panda o development create a new release branch in cht core from the most recent release and call it fr if latest is and the feature is to allow movies to be uploaded call it fr movie upload done before the release so all prs can be set to merge to this branch and not to master set the version number in package json and package lock json and submit a pr the easiest way to do this is to use npm no git tag version version ensure qa is briefed and is partnering with the trio to ensure early and often checks of the feature are on track to be of production quality from the start releasing this is an iterative process and it s assumed there will be multiple numbered releases throughout development of the feature release build a beta named fr by pushing a git tag and when ci completes successfully notify the qa team that it s ready for release testing if an updated feature release is needed increment the last by calling it fr etc close out validate with the actual end user that this feature release delivers a quantifiable improvement if yes plan on adding the feature to the next minor release by creating a new ticket to merge the code to master if no we leave the code dead in this branch never to be merged to master but still loved all the same mark this issue done and close the milestone
| 1
|
58,188
| 11,848,519,165
|
IssuesEvent
|
2020-03-24 13:54:46
|
GooglingTheCancerGenome/CNN
|
https://api.github.com/repos/GooglingTheCancerGenome/CNN
|
closed
|
Error message for chr_array.py
|
bug code
|
```
+ python chr_array.py --bam /hpc/cog_bioinf/ridder/users/lsantuari/Git/DeepSV_refactoring/CNN/data/test/chr22.bam --chr chr22 --twobit /hpc/cog_bioinf/ridder/users/lsantuari/Git/DeepSV_refactoring/CNN/data/test/chr22.2bit --map /hpc/cog_bioinf/ridder/users/lsantuari/Git/DeepSV_refactoring/CNN/data/test/chr22.bw --out chr_array.npy --outputpath . --logfile chr_array.log
Error. nthreads cannot be larger than environment variable "NUMEXPR_MAX_THREADS" (64)Traceback (most recent call last):
File "chr_array.py", line 346, in <module>
main()
File "chr_array.py", line 340, in main
cmd_name=cmd_name
File "chr_array.py", line 191, in create_hdf5
chr_array[idx, channel_index] = vals
File "bcolz/carray_ext.pyx", line 2128, in bcolz.carray_ext.carray.__setitem__
ValueError: could not broadcast input array from shape (1179) into shape (46)
```
|
1.0
|
Error message for chr_array.py - ```
+ python chr_array.py --bam /hpc/cog_bioinf/ridder/users/lsantuari/Git/DeepSV_refactoring/CNN/data/test/chr22.bam --chr chr22 --twobit /hpc/cog_bioinf/ridder/users/lsantuari/Git/DeepSV_refactoring/CNN/data/test/chr22.2bit --map /hpc/cog_bioinf/ridder/users/lsantuari/Git/DeepSV_refactoring/CNN/data/test/chr22.bw --out chr_array.npy --outputpath . --logfile chr_array.log
Error. nthreads cannot be larger than environment variable "NUMEXPR_MAX_THREADS" (64)Traceback (most recent call last):
File "chr_array.py", line 346, in <module>
main()
File "chr_array.py", line 340, in main
cmd_name=cmd_name
File "chr_array.py", line 191, in create_hdf5
chr_array[idx, channel_index] = vals
File "bcolz/carray_ext.pyx", line 2128, in bcolz.carray_ext.carray.__setitem__
ValueError: could not broadcast input array from shape (1179) into shape (46)
```
|
non_process
|
error message for chr array py python chr array py bam hpc cog bioinf ridder users lsantuari git deepsv refactoring cnn data test bam chr twobit hpc cog bioinf ridder users lsantuari git deepsv refactoring cnn data test map hpc cog bioinf ridder users lsantuari git deepsv refactoring cnn data test bw out chr array npy outputpath logfile chr array log error nthreads cannot be larger than environment variable numexpr max threads traceback most recent call last file chr array py line in main file chr array py line in main cmd name cmd name file chr array py line in create chr array vals file bcolz carray ext pyx line in bcolz carray ext carray setitem valueerror could not broadcast input array from shape into shape
| 0
|
9,876
| 12,886,367,578
|
IssuesEvent
|
2020-07-13 09:22:28
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Internal: trigger e2e tests for patch branches
|
kind/improvement process/candidate team/typescript topic: internal
|
For npm publishing script.
Right now the e2e tests are not run on patch branches so it checks the latest dev e2e results. Meaning we could have a broken publishing process in future patches.
|
1.0
|
Internal: trigger e2e tests for patch branches - For npm publishing script.
Right now the e2e tests are not run on patch branches so it checks the latest dev e2e results. Meaning we could have a broken publishing process in future patches.
|
process
|
internal trigger tests for patch branches for npm publishing script right now the tests are not run on patch branches so it checks the latest dev results meaning we could have a broken publishing process in future patches
| 1
|
686,235
| 23,483,460,964
|
IssuesEvent
|
2022-08-17 12:40:03
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
closed
|
Grpc.Tools is not passing OutputOptions or AdditionalProtocArguments to protoc.exe
|
kind/bug lang/C# priority/P2
|
### What version of gRPC and what language are you using?
Grpc.Tools 2.36.4
### What operating system (Linux, Windows,...) and version?
Windows 10 + Linux (current, idk)
### What runtime / compiler are you using (e.g. python version or version of gcc)
Whatever Grpc.Tools uses
### What did you do?
I have included 3 proto files in a VS project. Two are compiled and are simple message payloads. The third is linked from here and compiled by another proj file.
``` xml
<ItemGroup>
<Protobuf Include="Protos\example_data_payloads.proto" GrpcServices="None" ProtoCompile="true" OutputOptions="csharp_opt=file_extension=.g.cs"/>
<Protobuf Include="Protos\example_data_schema.proto" GrpcServices="None" ProtoCompile="true" OutputOptions="file_extension=.g.cs"/>
<Protobuf Include="Protos\example_data_stack.proto" GrpcServices="None" ProtoCompile="false" />
</ItemGroup>
```
For the two that are proto compiled in this project, I am trying to fix the file extension to note that the file has been code generated so it can be excluded from test coverage reports. OutputOptions is described [here](https://github.com/grpc/grpc/blob/master/src/csharp/BUILD-INTEGRATION.md) and I've tried a few variations of it but the file is still generated as a .cs file.
### What did you expect to see?
the code file generated with a .g.cs extension
### What did you see instead?
just the file with a .cs extension in the obj/ directory
The proto file for the first item is
``` protobuf
syntax = "proto3";
option csharp_namespace = "Core.Schema";
package example_data_schema;
message ExampleData {
bytes id = 1;
string name = 3;
int64 code = 5;
string description = 7;
}
```
the code is otherwise generated correctly. I would really prefer the exclusion mechanism to be part of the compiled output, like a [GeneratedCode] attribute, since analyzers may be working from the compiled output and not using source analysis. The current conventions used by the proto generation are not any of the common ways to designate code thats been generated.
|
1.0
|
Grpc.Tools is not passing OutputOptions or AdditionalProtocArguments to protoc.exe -
### What version of gRPC and what language are you using?
Grpc.Tools 2.36.4
### What operating system (Linux, Windows,...) and version?
Windows 10 + Linux (current, idk)
### What runtime / compiler are you using (e.g. python version or version of gcc)
Whatever Grpc.Tools uses
### What did you do?
I have included 3 proto files in a VS project. Two are compiled and are simple message payloads. The third is linked from here and compiled by another proj file.
``` xml
<ItemGroup>
<Protobuf Include="Protos\example_data_payloads.proto" GrpcServices="None" ProtoCompile="true" OutputOptions="csharp_opt=file_extension=.g.cs"/>
<Protobuf Include="Protos\example_data_schema.proto" GrpcServices="None" ProtoCompile="true" OutputOptions="file_extension=.g.cs"/>
<Protobuf Include="Protos\example_data_stack.proto" GrpcServices="None" ProtoCompile="false" />
</ItemGroup>
```
For the two that are proto compiled in this project, I am trying to fix the file extension to note that the file has been code generated so it can be excluded from test coverage reports. OutputOptions is described [here](https://github.com/grpc/grpc/blob/master/src/csharp/BUILD-INTEGRATION.md) and I've tried a few variations of it but the file is still generated as a .cs file.
### What did you expect to see?
the code file generated with a .g.cs extension
### What did you see instead?
just the file with a .cs extension in the obj/ directory
The proto file for the first item is
``` protobuf
syntax = "proto3";
option csharp_namespace = "Core.Schema";
package example_data_schema;
message ExampleData {
bytes id = 1;
string name = 3;
int64 code = 5;
string description = 7;
}
```
the code is otherwise generated correctly. I would really prefer the exclusion mechanism to be part of the compiled output, like a [GeneratedCode] attribute, since analyzers may be working from the compiled output and not using source analysis. The current conventions used by the proto generation are not any of the common ways to designate code thats been generated.
|
non_process
|
grpc tools is not passing outputoptions or additionalprotocarguments to protoc exe what version of grpc and what language are you using grpc tools what operating system linux windows and version windows linux current idk what runtime compiler are you using e g python version or version of gcc whatever grpc tools uses what did you do i have included proto files in a vs project two are compiled and are simple message payloads the third is linked from here and compiled by another proj file xml for the two that are proto compiled in this project i am trying to fix the file extension to note that the file has been code generated so it can be excluded from test coverage reports outputoptions is described and i ve tried a few variations of it but the file is still generated as a cs file what did you expect to see the code file generated with a g cs extension what did you see instead just the file with a cs extension in the obj directory the proto file for the first item is protobuf syntax option csharp namespace core schema package example data schema message exampledata bytes id string name code string description the code is otherwise generated correctly i would really prefer the exclusion mechanism to be part of the compiled output like a attribute since analyzers may be working from the compiled output and not using source analysis the current conventions used by the proto generation are not any of the common ways to designate code thats been generated
| 0
|
14,242
| 17,172,627,552
|
IssuesEvent
|
2021-07-15 07:26:22
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
Add option to set the ellipsoid to use for calculations with qgis_process tool (Request in QGIS)
|
3.16 Processing
|
### Request for documentation
From pull request QGIS/qgis#37992
Author: @nyalldawson
QGIS version: 3.16
**Add option to set the ellipsoid to use for calculations with qgis_process tool**
### PR Description:
- Add explicit settings for ellipsoid/distance unit/area unit to QgsProcessingContext, and prefer to use these instead of trying to get the settings from a QgsProject (which isn't available from the standalone tool).
- Expose a new argument to allow users to specify "--ELLIPSOID=xxxx" when running algorithms via the qgis_process tool.
- Flags algorithms which make no sense to run outside of the GUI environment (such as "select by ..." algorithms), and hides/prevents them from running via qgis_process.
- Also warns when running deprecated or algorithms with known issues from qgis_process
- Flags algorithms which require a valid project to execute so that errors are raised when running these from qgis_process, unless the new "--PROJECT_PATH=xxxx" argument is specified and pointing to a valid project path.
Fixes #37988
### Commits tagged with [need-docs] or [FEATURE]
|
1.0
|
Add option to set the ellipsoid to use for calculations with qgis_process tool (Request in QGIS) - ### Request for documentation
From pull request QGIS/qgis#37992
Author: @nyalldawson
QGIS version: 3.16
**Add option to set the ellipsoid to use for calculations with qgis_process tool**
### PR Description:
- Add explicit settings for ellipsoid/distance unit/area unit to QgsProcessingContext, and prefer to use these instead of trying to get the settings from a QgsProject (which isn't available from the standalone tool).
- Expose a new argument to allow users to specify "--ELLIPSOID=xxxx" when running algorithms via the qgis_process tool.
- Flags algorithms which make no sense to run outside of the GUI environment (such as "select by ..." algorithms), and hides/prevents them from running via qgis_process.
- Also warns when running deprecated or algorithms with known issues from qgis_process
- Flags algorithms which require a valid project to execute so that errors are raised when running these from qgis_process, unless the new "--PROJECT_PATH=xxxx" argument is specified and pointing to a valid project path.
Fixes #37988
### Commits tagged with [need-docs] or [FEATURE]
|
process
|
add option to set the ellipsoid to use for calculations with qgis process tool request in qgis request for documentation from pull request qgis qgis author nyalldawson qgis version add option to set the ellipsoid to use for calculations with qgis process tool pr description add explicit settings for ellipsoid distance unit area unit to qgsprocessingcontext and prefer to use these instead of trying to get the settings from a qgsproject which isn t available from the standalone tool expose a new argument to allow users to specify ellipsoid xxxx when running algorithms via the qgis process tool flags algorithms which make no sense to run outside of the gui environment such as select by algorithms and hides prevents them from running via qgis process also warns when running deprecated or algorithms with known issues from qgis process flags algorithms which require a valid project to execute so that errors are raised when running these from qgis process unless the new project path xxxx argument is specified and pointing to a valid project path fixes commits tagged with or
| 1
|
31,013
| 25,252,731,532
|
IssuesEvent
|
2022-11-15 15:46:50
|
CGAL/cgal
|
https://api.github.com/repos/CGAL/cgal
|
opened
|
.github/workflows/build_doc.yml has an error, when the Doxygen build has errors
|
bug Infrastructure
|
## Issue Details
See that run: https://github.com/CGAL/cgal/actions/runs/3469346229/jobs/5796501149
|
1.0
|
.github/workflows/build_doc.yml has an error, when the Doxygen build has errors - ## Issue Details
See that run: https://github.com/CGAL/cgal/actions/runs/3469346229/jobs/5796501149
|
non_process
|
github workflows build doc yml has an error when the doxygen build has errors issue details see that run
| 0
|
17,980
| 23,997,309,603
|
IssuesEvent
|
2022-09-14 08:36:47
|
BirthdayResearch/contented
|
https://api.github.com/repos/BirthdayResearch/contented
|
closed
|
(feat): Markdown table processor support
|
kind/feature needs/triage area/contented-processor
|
<!-- Please only use this template for submitting enhancement/feature requests -->
#### What would you like to be added:
As per title.
#### Why is this needed:
Useful feature to have since some data is better represented in a tabular form. For example, a glossary, schema, etc.
|
1.0
|
(feat): Markdown table processor support - <!-- Please only use this template for submitting enhancement/feature requests -->
#### What would you like to be added:
As per title.
#### Why is this needed:
Useful feature to have since some data is better represented in a tabular form. For example, a glossary, schema, etc.
|
process
|
feat markdown table processor support what would you like to be added as per title why is this needed useful feature to have since some data is better represented in a tabular form for example a glossary schema etc
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.