Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
70,645
| 7,194,973,813
|
IssuesEvent
|
2018-02-04 12:16:24
|
OAButton/discussion
|
https://api.github.com/repos/OAButton/discussion
|
closed
|
test all the stats from tools
|
1. Admin 1. Bookmarklet 1. Chatbot 1. DeliverOA 1. EmbedOA 1. Institutional 1. Plugin 1. Website 2. Enhancement 3. Blocked: Test
|
Make sure that:
- [x] uses from the frontpage show
- [x] /bulk uses show
- [ ] we can see uses from iframe embedoa
- [ ] we can see uses from different installs of /embedoa
- [x] we can see uses from illiad deliveroa
- [x] we can see uses from chatbot
- [x] need to test on new plugins ( bookmarklet / plugin / ff )
|
1.0
|
test all the stats from tools - Make sure that:
- [x] uses from the frontpage show
- [x] /bulk uses show
- [ ] we can see uses from iframe embedoa
- [ ] we can see uses from different installs of /embedoa
- [x] we can see uses from illiad deliveroa
- [x] we can see uses from chatbot
- [x] need to test on new plugins ( bookmarklet / plugin / ff )
|
non_process
|
test all the stats from tools make sure that uses from the frontpage show bulk uses show we can see uses from iframe embedoa we can see uses from different installs of embedoa we can see uses from illiad deliveroa we can see uses from chatbot need to test on new plugins bookmarklet plugin ff
| 0
|
20,048
| 6,808,623,095
|
IssuesEvent
|
2017-11-04 05:44:22
|
Great-Hill-Corporation/quickBlocks
|
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
|
reopened
|
Release names
|
build-general status-inprocess type-info type-project
|
Releases are named after trolley line stops coming out of Philadelphia. Here are the names:
<img width="627" alt="screen shot 2017-08-25 at 11 25 44 pm" src="https://user-images.githubusercontent.com/5417918/29738274-dc7bdb44-89ec-11e7-8285-f55787132309.png">
|
1.0
|
Release names - Releases are named after trolley line stops coming out of Philadelphia. Here are the names:
<img width="627" alt="screen shot 2017-08-25 at 11 25 44 pm" src="https://user-images.githubusercontent.com/5417918/29738274-dc7bdb44-89ec-11e7-8285-f55787132309.png">
|
non_process
|
release names releases are named after trolley line stops coming out of philadelphia here are the names img width alt screen shot at pm src
| 0
|
3,345
| 6,480,466,727
|
IssuesEvent
|
2017-08-18 13:29:57
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
child_process.on("error") does not function as expected in Ubuntu
|
child_process
|
Not working with:
* **Version**: 8.2.1
* **Platform**: Ubuntu (linux) 16.04
Works as expected with:
* **Version**: 10.0 16251
* **Platform**: Windows
The example in the [documentation](https://nodejs.org/api/child_process.html#child_process_child_process_spawn_command_args_options) cites the following workflow:
```[javascript]
const { spawn } = require('child_process');
const subprocess = spawn('bad_command');
subprocess.on('error', (err) => {
console.log('Failed to start subprocess.');
});
```
This however only works in Windows environment. In Ubuntu the .on("exit", function is called regardless if the command is found in the system path and the on("error", function is never called.
|
1.0
|
child_process.on("error") does not function as expected in Ubuntu - Not working with:
* **Version**: 8.2.1
* **Platform**: Ubuntu (linux) 16.04
Works as expected with:
* **Version**: 10.0 16251
* **Platform**: Windows
The example in the [documentation](https://nodejs.org/api/child_process.html#child_process_child_process_spawn_command_args_options) cites the following workflow:
```[javascript]
const { spawn } = require('child_process');
const subprocess = spawn('bad_command');
subprocess.on('error', (err) => {
console.log('Failed to start subprocess.');
});
```
This however only works in Windows environment. In Ubuntu the .on("exit", function is called regardless if the command is found in the system path and the on("error", function is never called.
|
process
|
child process on error does not function as expected in ubuntu not working with version platform ubuntu linux works as expected with version platform windows the example in the cites the following workflow const spawn require child process const subprocess spawn bad command subprocess on error err console log failed to start subprocess this however only works in windows environment in ubuntu the on exit function is called regardless if the command is found in the system path and the on error function is never called
| 1
|
758,452
| 26,555,833,729
|
IssuesEvent
|
2023-01-20 11:56:41
|
saudalnasser/starlux
|
https://api.github.com/repos/saudalnasser/starlux
|
opened
|
feat: hot reload
|
type: feature priority: medium
|
## Problem
there's no way to deploy commands automatically on changes during development
## Solution(s)
a way to look for changes in specific commands and deploy them accordingly to the specified guild
|
1.0
|
feat: hot reload - ## Problem
there's no way to deploy commands automatically on changes during development
## Solution(s)
a way to look for changes in specific commands and deploy them accordingly to the specified guild
|
non_process
|
feat hot reload problem there s no way to deploy commands automatically on changes during development solution s a way to look for changes in specific commands and deploy them accordingly to the specified guild
| 0
|
183,461
| 6,688,901,299
|
IssuesEvent
|
2017-10-08 19:51:02
|
qutebrowser/qutebrowser
|
https://api.github.com/repos/qutebrowser/qutebrowser
|
closed
|
Have a way to reset bindings to their default
|
component: completion component: config priority: 1 - middle
|
After doing `:unbind H`, there's no way to set it to its default without looking it up.
I can see two solutions:
- Something like `:unbind --default H` should exist (`:config-unset` isn't really a good match for bindings).
- The completion should show the default after `:bind H` (similar to what the `:set` completion already does) (@rcorre - we probably need to give completion functions a parsed commandline for this one to work?)
|
1.0
|
Have a way to reset bindings to their default - After doing `:unbind H`, there's no way to set it to its default without looking it up.
I can see two solutions:
- Something like `:unbind --default H` should exist (`:config-unset` isn't really a good match for bindings).
- The completion should show the default after `:bind H` (similar to what the `:set` completion already does) (@rcorre - we probably need to give completion functions a parsed commandline for this one to work?)
|
non_process
|
have a way to reset bindings to their default after doing unbind h there s no way to set it to its default without looking it up i can see two solutions something like unbind default h should exist config unset isn t really a good match for bindings the completion should show the default after bind h similar to what the set completion already does rcorre we probably need to give completion functions a parsed commandline for this one to work
| 0
|
446,431
| 31,476,099,592
|
IssuesEvent
|
2023-08-30 10:48:31
|
onOffice-Web-Org/oo-wp-plugin
|
https://api.github.com/repos/onOffice-Web-Org/oo-wp-plugin
|
closed
|
Sort property lists
|
needs documentation in review feature
|
### Discussed in https://github.com/onOffice-Web-Org/oo-wp-plugin/discussions/507
<div type='discussions-op-text'>
<sup>Originally posted by **fredericalpers** April 25, 2023</sup>
### Current state
At the moment the created property lists can not be sorted. With a large number of lists, it is difficult to find the correct one.
### Desired state
The ones marked in the screenshot should be sortable, like in the WP page overview. Sorting for "Name of list", "Selected filter", "Template" and "List type".

</div>
|
1.0
|
Sort property lists - ### Discussed in https://github.com/onOffice-Web-Org/oo-wp-plugin/discussions/507
<div type='discussions-op-text'>
<sup>Originally posted by **fredericalpers** April 25, 2023</sup>
### Current state
At the moment the created property lists can not be sorted. With a large number of lists, it is difficult to find the correct one.
### Desired state
The ones marked in the screenshot should be sortable, like in the WP page overview. Sorting for "Name of list", "Selected filter", "Template" and "List type".

</div>
|
non_process
|
sort property lists discussed in originally posted by fredericalpers april current state at the moment the created property lists can not be sorted with a large number of lists it is difficult to find the correct one desired state the ones marked in the screenshot should be sortable like in the wp page overview sorting for name of list selected filter template and list type
| 0
|
782,900
| 27,511,003,187
|
IssuesEvent
|
2023-03-06 08:49:05
|
AY2223S2-CS2103T-T17-3/tp
|
https://api.github.com/repos/AY2223S2-CS2103T-T17-3/tp
|
closed
|
Delete tag
|
type.Story priority.High
|
As a user I can delete tag from a person so that I can update the information overtime.
|
1.0
|
Delete tag - As a user I can delete tag from a person so that I can update the information overtime.
|
non_process
|
delete tag as a user i can delete tag from a person so that i can update the information overtime
| 0
|
390,949
| 11,566,119,511
|
IssuesEvent
|
2020-02-20 11:51:35
|
AugurProject/augur
|
https://api.github.com/repos/AugurProject/augur
|
closed
|
Buy participation tokens > My REP balance deducts after about a minute but my participation tokens don't show up here
|
Bug Needed for V2 launch Priority: High
|
**Read to end of ticket**
Here it should show 11

it is updating in the total PT tokens purchased but not showing that I own them

After pushing time forward a week it's now working correctly
|
1.0
|
Buy participation tokens > My REP balance deducts after about a minute but my participation tokens don't show up here - **Read to end of ticket**
Here it should show 11

it is updating in the total PT tokens purchased but not showing that I own them

After pushing time forward a week it's now working correctly
|
non_process
|
buy participation tokens my rep balance deducts after about a minute but my participation tokens don t show up here read to end of ticket here it should show it is updating in the total pt tokens purchased but not showing that i own them after pushing time forward a week it s now working correctly
| 0
|
16,356
| 21,035,673,021
|
IssuesEvent
|
2022-03-31 07:36:06
|
linuxdeepin/developer-center
|
https://api.github.com/repos/linuxdeepin/developer-center
|
closed
|
Transparent background misplaced
|
bug | interface style bug | porting other | delay processing
|
<!--Please keep one issue focus on one bug, If you have multiple issues to report, try open multiple issues to report each of them.-->
## Describe the bug
<!--A clear and concise description of what the bug is.-->
The transparent background in some applications is not in the correct place.
## To Reproduce
<!--We can fix a bug faster if we can reproduce the bug, so provide steps to reproduce the bug is very helpful for us.-->
Steps to reproduce the behavior:
1. Open notification panel in Arch/Manjaro
2. See the result
Or,
1. Open Deepin terminal.
2. Make sure you have enabled "blur background" from the settings.
3. See the result.
## Expected behavior
<!--A clear and concise description of what you expected to happen.-->
The transparent background should be in the correct place.
## Screenshots
<!--If applicable, add screenshots to help explain your problem.-->


## Environments:
### Distro & Version
<!--e.g. Linux Deepin 15.7-->
Manjaro 20.0.2
DDE v20
### Related package version
<!--e.g. dde-file-manager v1.7 (4.5.6.2-2)-->
<!--Please try to avoid using the word "latest" without providing the actual version number.-->
<!--If you are not sure about which package went wrong, you can provide a list of package and it's version which you thinks is probably related with your issue-->
`kwin v5.18.5-1`
`deepin-kwin v5.0.14.1-3`
`deepin-dock v5.1.0.9-1`
`deepin-terminal v5.0.4.1-1`
### Additional context
<!--Add any other context about the problem here. It will be very helpful for us to isolate the problem. We will ask for more details when needed so feel free to remove this section if you really don't know what to provide.-->
This bug occurs only in Manjaro/Arch. It does not occur in Deepin OS.
|
1.0
|
Transparent background misplaced - <!--Please keep one issue focus on one bug, If you have multiple issues to report, try open multiple issues to report each of them.-->
## Describe the bug
<!--A clear and concise description of what the bug is.-->
The transparent background in some applications is not in the correct place.
## To Reproduce
<!--We can fix a bug faster if we can reproduce the bug, so provide steps to reproduce the bug is very helpful for us.-->
Steps to reproduce the behavior:
1. Open notification panel in Arch/Manjaro
2. See the result
Or,
1. Open Deepin terminal.
2. Make sure you have enabled "blur background" from the settings.
3. See the result.
## Expected behavior
<!--A clear and concise description of what you expected to happen.-->
The transparent background should be in the correct place.
## Screenshots
<!--If applicable, add screenshots to help explain your problem.-->


## Environments:
### Distro & Version
<!--e.g. Linux Deepin 15.7-->
Manjaro 20.0.2
DDE v20
### Related package version
<!--e.g. dde-file-manager v1.7 (4.5.6.2-2)-->
<!--Please try to avoid using the word "latest" without providing the actual version number.-->
<!--If you are not sure about which package went wrong, you can provide a list of package and it's version which you thinks is probably related with your issue-->
`kwin v5.18.5-1`
`deepin-kwin v5.0.14.1-3`
`deepin-dock v5.1.0.9-1`
`deepin-terminal v5.0.4.1-1`
### Additional context
<!--Add any other context about the problem here. It will be very helpful for us to isolate the problem. We will ask for more details when needed so feel free to remove this section if you really don't know what to provide.-->
This bug occurs only in Manjaro/Arch. It does not occur in Deepin OS.
|
process
|
transparent background misplaced describe the bug the transparent background in some applications is not in the correct place to reproduce steps to reproduce the behavior open notification panel in arch manjaro see the result or open deepin terminal make sure you have enabled blur background from the settings see the result expected behavior the transparent background should be in the correct place screenshots environments distro version manjaro dde related package version kwin deepin kwin deepin dock deepin terminal additional context this bug occurs only in manjaro arch it does not occur in deepin os
| 1
|
14,000
| 16,772,297,253
|
IssuesEvent
|
2021-06-14 16:08:57
|
deepset-ai/haystack
|
https://api.github.com/repos/deepset-ai/haystack
|
closed
|
Preprocessing of json files
|
topic:preprocessing type:bug
|
**Question**
Why does the Preprocessing of a large json file shows only 20% progress
**Additional context**
I have a set of 20 files to train when I train them individually the Preprocessing step shows 100% but while i train a combine json file it only shows 20% as though only 20% of the dataset has been preprocessed.
Preprocessing Dataset data/merged_file.json: 20%|██ | 4/20 [00:03<00:14, 1.09 Dicts/s]
Code is picked from the tutorial:-
reader = FARMReader(model_name_or_path="distilbert-base-uncased-distilled-squad", use_gpu=True)
train_data = "data"
reader.train(data_dir=train_data, train_filename="merged_file.json", use_gpu=True, n_epochs=1, save_dir="my_model")
I tried running individual json files but i see 100% for every json file.
But when i try to run a merged file only 20% shows up
I have checked the merged json file and it does not have any special characters, the structure is also according the Squad format although the training completes without any issue this 20% bothers me.
Please help advise on what i can check to troubleshoot this further or if there is some misunderstanding in terms of the functionality
|
1.0
|
Preprocessing of json files - **Question**
Why does the Preprocessing of a large json file shows only 20% progress
**Additional context**
I have a set of 20 files to train when I train them individually the Preprocessing step shows 100% but while i train a combine json file it only shows 20% as though only 20% of the dataset has been preprocessed.
Preprocessing Dataset data/merged_file.json: 20%|██ | 4/20 [00:03<00:14, 1.09 Dicts/s]
Code is picked from the tutorial:-
reader = FARMReader(model_name_or_path="distilbert-base-uncased-distilled-squad", use_gpu=True)
train_data = "data"
reader.train(data_dir=train_data, train_filename="merged_file.json", use_gpu=True, n_epochs=1, save_dir="my_model")
I tried running individual json files but i see 100% for every json file.
But when i try to run a merged file only 20% shows up
I have checked the merged json file and it does not have any special characters, the structure is also according the Squad format although the training completes without any issue this 20% bothers me.
Please help advise on what i can check to troubleshoot this further or if there is some misunderstanding in terms of the functionality
|
process
|
preprocessing of json files question why does the preprocessing of a large json file shows only progress additional context i have a set of files to train when i train them individually the preprocessing step shows but while i train a combine json file it only shows as though only of the dataset has been preprocessed preprocessing dataset data merged file json ██ code is picked from the tutorial reader farmreader model name or path distilbert base uncased distilled squad use gpu true train data data reader train data dir train data train filename merged file json use gpu true n epochs save dir my model i tried running individual json files but i see for every json file but when i try to run a merged file only shows up i have checked the merged json file and it does not have any special characters the structure is also according the squad format although the training completes without any issue this bothers me please help advise on what i can check to troubleshoot this further or if there is some misunderstanding in terms of the functionality
| 1
|
22,031
| 30,545,641,576
|
IssuesEvent
|
2023-07-20 03:29:23
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
roblox-pyc 1.18.53 has 2 GuardDog issues
|
guarddog silent-process-execution
|
https://pypi.org/project/roblox-pyc
https://inspector.pypi.io/project/roblox-pyc
```{
"dependency": "roblox-pyc",
"version": "1.18.53",
"result": {
"issues": 2,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "roblox-pyc-1.18.53/src/robloxpy.py:121",
"code": " subprocess.call([\"luarocks\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-1.18.53/src/robloxpy.py:128",
"code": " subprocess.call([\"moonc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpi0a_fqms/roblox-pyc"
}
}```
|
1.0
|
roblox-pyc 1.18.53 has 2 GuardDog issues - https://pypi.org/project/roblox-pyc
https://inspector.pypi.io/project/roblox-pyc
```{
"dependency": "roblox-pyc",
"version": "1.18.53",
"result": {
"issues": 2,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "roblox-pyc-1.18.53/src/robloxpy.py:121",
"code": " subprocess.call([\"luarocks\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-1.18.53/src/robloxpy.py:128",
"code": " subprocess.call([\"moonc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpi0a_fqms/roblox-pyc"
}
}```
|
process
|
roblox pyc has guarddog issues dependency roblox pyc version result issues errors results silent process execution location roblox pyc src robloxpy py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null location roblox pyc src robloxpy py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp fqms roblox pyc
| 1
|
16,689
| 21,791,068,486
|
IssuesEvent
|
2022-05-14 22:53:26
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
Add Tonight with Katherine Newbury from Late Night
|
suggested title in process
|
Title: Tonight with Katherine Newbury
Type (film/tv show): tv show
Film or show in which it appears: Late Night (2019)
Is the parent film/show streaming anywhere? Amazon
About when in the parent film/show does it appear? Throughout
Actual footage of the film/show can be seen (yes/no)? yes
|
1.0
|
Add Tonight with Katherine Newbury from Late Night - Title: Tonight with Katherine Newbury
Type (film/tv show): tv show
Film or show in which it appears: Late Night (2019)
Is the parent film/show streaming anywhere? Amazon
About when in the parent film/show does it appear? Throughout
Actual footage of the film/show can be seen (yes/no)? yes
|
process
|
add tonight with katherine newbury from late night title tonight with katherine newbury type film tv show tv show film or show in which it appears late night is the parent film show streaming anywhere amazon about when in the parent film show does it appear throughout actual footage of the film show can be seen yes no yes
| 1
|
435,365
| 30,496,200,434
|
IssuesEvent
|
2023-07-18 11:00:16
|
CliMA/Oceananigans.jl
|
https://api.github.com/repos/CliMA/Oceananigans.jl
|
opened
|
Docs don't build for tagged releases
|
testing 🧪 documentation 📜
|
See, e.g., recent v0.85.
Possibly something related with buildkite settings. E.g. for v0.84.1 the tag triggered [a buildkite run](https://buildkite.com/clima/oceananigans/builds/12019)
<img width="1215" alt="Screenshot 2023-07-18 at 1 57 24 pm" src="https://github.com/CliMA/Oceananigans.jl/assets/7112768/f0b0f59a-f909-4858-aaf1-3536ed485d91">
<img width="1234" alt="Screenshot 2023-07-18 at 1 58 19 pm" src="https://github.com/CliMA/Oceananigans.jl/assets/7112768/818d2fb1-086f-40ac-a109-52833375aa40">
|
1.0
|
Docs don't build for tagged releases - See, e.g., recent v0.85.
Possibly something related with buildkite settings. E.g. for v0.84.1 the tag triggered [a buildkite run](https://buildkite.com/clima/oceananigans/builds/12019)
<img width="1215" alt="Screenshot 2023-07-18 at 1 57 24 pm" src="https://github.com/CliMA/Oceananigans.jl/assets/7112768/f0b0f59a-f909-4858-aaf1-3536ed485d91">
<img width="1234" alt="Screenshot 2023-07-18 at 1 58 19 pm" src="https://github.com/CliMA/Oceananigans.jl/assets/7112768/818d2fb1-086f-40ac-a109-52833375aa40">
|
non_process
|
docs don t build for tagged releases see e g recent possibly something related with buildkite settings e g for the tag triggered img width alt screenshot at pm src img width alt screenshot at pm src
| 0
|
216,616
| 7,310,173,459
|
IssuesEvent
|
2018-02-28 14:17:14
|
ballerina-platform/ballerina-message-broker
|
https://api.github.com/repos/ballerina-platform/ballerina-message-broker
|
closed
|
JMS local transaction test cases
|
Complexity/Moderate Module/broker-core Priority/High Type/Task
|
**Description:**
Write JMS local transaction test cases for both queue and topic
|
1.0
|
JMS local transaction test cases - **Description:**
Write JMS local transaction test cases for both queue and topic
|
non_process
|
jms local transaction test cases description write jms local transaction test cases for both queue and topic
| 0
|
14,646
| 17,774,387,421
|
IssuesEvent
|
2021-08-30 17:15:11
|
cagov/design-system
|
https://api.github.com/repos/cagov/design-system
|
closed
|
In statewide header official tagline show only "State of California" in mobile
|
Proposal - Review process CA Design System
|
In statewide header make "Official website of the State of California" to show only "State of California" in mobile. Figma Example is below:
<img width="768" alt="Screen Shot 2021-08-19 at 4 29 39 PM" src="https://user-images.githubusercontent.com/31669748/130156877-cd6bf128-8232-4331-8555-2aa75657836b.png">
|
1.0
|
In statewide header official tagline show only "State of California" in mobile - In statewide header make "Official website of the State of California" to show only "State of California" in mobile. Figma Example is below:
<img width="768" alt="Screen Shot 2021-08-19 at 4 29 39 PM" src="https://user-images.githubusercontent.com/31669748/130156877-cd6bf128-8232-4331-8555-2aa75657836b.png">
|
process
|
in statewide header official tagline show only state of california in mobile in statewide header make official website of the state of california to show only state of california in mobile figma example is below img width alt screen shot at pm src
| 1
|
64,164
| 8,714,210,517
|
IssuesEvent
|
2018-12-07 06:53:34
|
PegaSysEng/pantheon
|
https://api.github.com/repos/PegaSysEng/pantheon
|
closed
|
Document debug_metrics
|
doc next release documentation
|
### Requirements
* Add debug_metrics to JSON-RPC API as added by PR #313
* Add conceptual content about why and how to use debug_metrics
* Include specific use-cases or examples
|
1.0
|
Document debug_metrics - ### Requirements
* Add debug_metrics to JSON-RPC API as added by PR #313
* Add conceptual content about why and how to use debug_metrics
* Include specific use-cases or examples
|
non_process
|
document debug metrics requirements add debug metrics to json rpc api as added by pr add conceptual content about why and how to use debug metrics include specific use cases or examples
| 0
|
19,499
| 25,809,137,132
|
IssuesEvent
|
2022-12-11 17:37:19
|
anitsh/til
|
https://api.github.com/repos/anitsh/til
|
opened
|
Theory of Constraints (TOC)
|
principle process
|
A process improvement methodology that emphasizes the importance of identifying the "system constraint" or bottleneck.
By leveraging this constraint, organizations can achieve their financial goals while delivering on-time-in-full (OTIF) to customers, avoiding stock-outs in the supply chain, reducing lead time, etc. Other common benefits of implementing it includes - better control over operations, less inventory, reduced conflicts between team member and drastically reduced firefighting. Often, additional capacity gets exposed without further capital investment or hiring additional workers.
The Core Idea – Every system has a limiting factor or constraint. Focusing improvement efforts to better utilize this constraint is normally the fastest and most effective way to improve profitability.
No chain can ever be stronger than it's weakest link.
Process and Philosophies:
What to change?
What to change to?
How to cause the change?
Four principles of flow which are as follows:
1. Improving flow (or equivalently lead time) is a primary objective of operations.
2. This primary objective should be translated into a practical mechanism that guides the operation when not to produce (prevents overproduction).
3. Local efficiencies must be abolished.
4. A focusing process to balance flow must be in place.
A successful Theory of Constraints implementation will have the following benefits:
- Increased Profit: the primary goal of TOC for most companies
- Fast Improvement: a result of focusing all attention on one critical area – the system constraint
- Improved Capacity: optimizing the constraint enables more product to be manufactured
- Reduced Lead Times: optimizing the constraint results in smoother and faster product flow
- Reduced Inventory: eliminating bottlenecks means there will be less work-in-process
The Theory of Constraints provides a specific methodology for identifying and eliminating constraints, referred to as the Five Focusing Steps. As shown in the following diagram, it is a cyclical process.
.
Exploit | Make quick improvements to the throughput of the constraint using existing resources (i.e., make the most of what you have).
Subordinate | Review all other activities in the process to ensure that they are aligned with and truly support the needs of the constraint.
Elevate | If the constraint still exists (i.e., it has not moved), consider what further actions can be taken to eliminate it from being the constraint. Normally, actions are continued at this step until the constraint has been “broken” (until it has moved somewhere else). In some cases, capital investment may be required.
Repeat | The Five Focusing Steps are a continuous improvement cycle. Therefore, once a constraint is resolved the next constraint should immediately be addressed. This step is a reminder to never become complacent – aggressively improve the current constraint…and then immediately move on to the next constraint.
# Resource
- https://www.tocinstitute.org/theory-of-constraints.html
- https://www.leanproduction.com/theory-of-constraints
- https://toc-flow.com/four-principles-of-flow
- https://cdn.ymaws.com/www.tocico.org/resource/resmgr/webinar_files/tocico_webinar_4.12.19__4.12.pdf
|
1.0
|
Theory of Constraints (TOC) - A process improvement methodology that emphasizes the importance of identifying the "system constraint" or bottleneck.
By leveraging this constraint, organizations can achieve their financial goals while delivering on-time-in-full (OTIF) to customers, avoiding stock-outs in the supply chain, reducing lead time, etc. Other common benefits of implementing it includes - better control over operations, less inventory, reduced conflicts between team member and drastically reduced firefighting. Often, additional capacity gets exposed without further capital investment or hiring additional workers.
The Core Idea – Every system has a limiting factor or constraint. Focusing improvement efforts to better utilize this constraint is normally the fastest and most effective way to improve profitability.
No chain can ever be stronger than it's weakest link.
Process and Philosophies:
What to change?
What to change to?
How to cause the change?
Four principles of flow which are as follows:
1. Improving flow (or equivalently lead time) is a primary objective of operations.
2. This primary objective should be translated into a practical mechanism that guides the operation when not to produce (prevents overproduction).
3. Local efficiencies must be abolished.
4. A focusing process to balance flow must be in place.
A successful Theory of Constraints implementation will have the following benefits:
- Increased Profit: the primary goal of TOC for most companies
- Fast Improvement: a result of focusing all attention on one critical area – the system constraint
- Improved Capacity: optimizing the constraint enables more product to be manufactured
- Reduced Lead Times: optimizing the constraint results in smoother and faster product flow
- Reduced Inventory: eliminating bottlenecks means there will be less work-in-process
The Theory of Constraints provides a specific methodology for identifying and eliminating constraints, referred to as the Five Focusing Steps. As shown in the following diagram, it is a cyclical process.
.
Exploit | Make quick improvements to the throughput of the constraint using existing resources (i.e., make the most of what you have).
Subordinate | Review all other activities in the process to ensure that they are aligned with and truly support the needs of the constraint.
Elevate | If the constraint still exists (i.e., it has not moved), consider what further actions can be taken to eliminate it from being the constraint. Normally, actions are continued at this step until the constraint has been “broken” (until it has moved somewhere else). In some cases, capital investment may be required.
Repeat | The Five Focusing Steps are a continuous improvement cycle. Therefore, once a constraint is resolved the next constraint should immediately be addressed. This step is a reminder to never become complacent – aggressively improve the current constraint…and then immediately move on to the next constraint.
# Resource
- https://www.tocinstitute.org/theory-of-constraints.html
- https://www.leanproduction.com/theory-of-constraints
- https://toc-flow.com/four-principles-of-flow
- https://cdn.ymaws.com/www.tocico.org/resource/resmgr/webinar_files/tocico_webinar_4.12.19__4.12.pdf
|
process
|
theory of constraints toc a process improvement methodology that emphasizes the importance of identifying the system constraint or bottleneck by leveraging this constraint organizations can achieve their financial goals while delivering on time in full otif to customers avoiding stock outs in the supply chain reducing lead time etc other common benefits of implementing it includes better control over operations less inventory reduced conflicts between team member and drastically reduced firefighting often additional capacity gets exposed without further capital investment or hiring additional workers the core idea – every system has a limiting factor or constraint focusing improvement efforts to better utilize this constraint is normally the fastest and most effective way to improve profitability no chain can ever be stronger than it s weakest link process and philosophies what to change what to change to how to cause the change four principles of flow which are as follows improving flow or equivalently lead time is a primary objective of operations this primary objective should be translated into a practical mechanism that guides the operation when not to produce prevents overproduction local efficiencies must be abolished a focusing process to balance flow must be in place a successful theory of constraints implementation will have the following benefits increased profit the primary goal of toc for most companies fast improvement a result of focusing all attention on one critical area – the system constraint improved capacity optimizing the constraint enables more product to be manufactured reduced lead times optimizing the constraint results in smoother and faster product flow reduced inventory eliminating bottlenecks means there will be less work in process the theory of constraints provides a specific methodology for identifying and eliminating constraints referred to as the five focusing steps as shown in the following diagram it is a cyclical process step objective identify identify the current constraint the single part of the process that limits the rate at which the goal is achieved exploit make quick improvements to the throughput of the constraint using existing resources i e make the most of what you have subordinate review all other activities in the process to ensure that they are aligned with and truly support the needs of the constraint elevate if the constraint still exists i e it has not moved consider what further actions can be taken to eliminate it from being the constraint normally actions are continued at this step until the constraint has been “broken” until it has moved somewhere else in some cases capital investment may be required repeat the five focusing steps are a continuous improvement cycle therefore once a constraint is resolved the next constraint should immediately be addressed this step is a reminder to never become complacent – aggressively improve the current constraint…and then immediately move on to the next constraint resource
| 1
|
4,338
| 7,244,583,027
|
IssuesEvent
|
2018-02-14 15:33:10
|
hardvolk/foodie-journal
|
https://api.github.com/repos/hardvolk/foodie-journal
|
closed
|
Trabajar investigando la funcionalidad de la API de Yelp y como integrarla al programa
|
In process
|
Investigación
|
1.0
|
Trabajar investigando la funcionalidad de la API de Yelp y como integrarla al programa - Investigación
|
process
|
trabajar investigando la funcionalidad de la api de yelp y como integrarla al programa investigación
| 1
|
524,389
| 15,213,004,807
|
IssuesEvent
|
2021-02-17 11:10:55
|
HSLdevcom/bultti
|
https://api.github.com/repos/HSLdevcom/bultti
|
closed
|
Ennakkotarkastusraporttien selailu
|
Epic Priority 1
|
Toimija
HSL pääkäyttäjä, HSL muokkaaja ylempi ja alempi, HSL selaaja, Liikennöitsijä muokkaaja, liikennöitsijä se-laaja
Esiehto
Ennakkotarkastus on tehty
Lopputulos
Raporttien selailu / tarvittavien tietojen näkeminen
Käyttötiheys
Joitain kertoja per liikennöitsijä per aikataulukausi
1
Käyttäjälle näytetään kirjautumissivu. Kirjautumisen jälkeen järjestelmä hakee kirjautuneen käyttäjän tie-dot ja siirtää käyttäjän automaattisesti etusivulle.
2
Käyttäjä kirjautuu järjestelmään ja siirtyy etusivulle. Käyttäjä valitsee osioista ennakkotarkastuksen.
3
Järjestelmä siirtää käyttäjän ennakkotarkastukseen, jossa on kaksi eri osiota: tarkastuksen teko ja raport-tien selailu.
4
Käyttäjä siirtyy raportit osioon ja valitsee tarkasteltavan aikataulukauden (alasvetovalikko) sekä liikennöit-sijän (liikennöitsijä voi valita vain itsensä, alasvetovalikko). Käyttäjä selaa haluamiaan raportteja, sorttaa ja / tai vie niitä Exceliin.
5
Järjestelmä hakee raportin tiedot ja näyttää käyttäjän valitseman raportin. Käyttäjä sorttaa raporttia käyt-täjän antamien käskyjen mukaan käyttöliittymässä. Järjestelmä vie raportin Exceliin / lataa tiedoston, jos käyttäjä aktivoi kyseisen painikkeen.
6
Käyttäjä siirtyy takaisin etusivulle.
Variaatiot
V3
Käyttäjä tekee raporttien selailun valinnat uudelleen. Käyttötapaus D alkaa alusta.
V4
Käyttäjä, jolla ei ole käyttöoikeuksia, yrittää päästä ennakkotarkastukseen. Järjestelmä ilmoittaa rajoite-tuista käyttöoikeuksista.
Poikkeukset
P1
Tekninen vika. Järjestelmä ilmoittaa keskeytymisestä.
|
1.0
|
Ennakkotarkastusraporttien selailu - Toimija
HSL pääkäyttäjä, HSL muokkaaja ylempi ja alempi, HSL selaaja, Liikennöitsijä muokkaaja, liikennöitsijä se-laaja
Esiehto
Ennakkotarkastus on tehty
Lopputulos
Raporttien selailu / tarvittavien tietojen näkeminen
Käyttötiheys
Joitain kertoja per liikennöitsijä per aikataulukausi
1
Käyttäjälle näytetään kirjautumissivu. Kirjautumisen jälkeen järjestelmä hakee kirjautuneen käyttäjän tie-dot ja siirtää käyttäjän automaattisesti etusivulle.
2
Käyttäjä kirjautuu järjestelmään ja siirtyy etusivulle. Käyttäjä valitsee osioista ennakkotarkastuksen.
3
Järjestelmä siirtää käyttäjän ennakkotarkastukseen, jossa on kaksi eri osiota: tarkastuksen teko ja raport-tien selailu.
4
Käyttäjä siirtyy raportit osioon ja valitsee tarkasteltavan aikataulukauden (alasvetovalikko) sekä liikennöit-sijän (liikennöitsijä voi valita vain itsensä, alasvetovalikko). Käyttäjä selaa haluamiaan raportteja, sorttaa ja / tai vie niitä Exceliin.
5
Järjestelmä hakee raportin tiedot ja näyttää käyttäjän valitseman raportin. Käyttäjä sorttaa raporttia käyt-täjän antamien käskyjen mukaan käyttöliittymässä. Järjestelmä vie raportin Exceliin / lataa tiedoston, jos käyttäjä aktivoi kyseisen painikkeen.
6
Käyttäjä siirtyy takaisin etusivulle.
Variaatiot
V3
Käyttäjä tekee raporttien selailun valinnat uudelleen. Käyttötapaus D alkaa alusta.
V4
Käyttäjä, jolla ei ole käyttöoikeuksia, yrittää päästä ennakkotarkastukseen. Järjestelmä ilmoittaa rajoite-tuista käyttöoikeuksista.
Poikkeukset
P1
Tekninen vika. Järjestelmä ilmoittaa keskeytymisestä.
|
non_process
|
ennakkotarkastusraporttien selailu toimija hsl pääkäyttäjä hsl muokkaaja ylempi ja alempi hsl selaaja liikennöitsijä muokkaaja liikennöitsijä se laaja esiehto ennakkotarkastus on tehty lopputulos raporttien selailu tarvittavien tietojen näkeminen käyttötiheys joitain kertoja per liikennöitsijä per aikataulukausi käyttäjälle näytetään kirjautumissivu kirjautumisen jälkeen järjestelmä hakee kirjautuneen käyttäjän tie dot ja siirtää käyttäjän automaattisesti etusivulle käyttäjä kirjautuu järjestelmään ja siirtyy etusivulle käyttäjä valitsee osioista ennakkotarkastuksen järjestelmä siirtää käyttäjän ennakkotarkastukseen jossa on kaksi eri osiota tarkastuksen teko ja raport tien selailu käyttäjä siirtyy raportit osioon ja valitsee tarkasteltavan aikataulukauden alasvetovalikko sekä liikennöit sijän liikennöitsijä voi valita vain itsensä alasvetovalikko käyttäjä selaa haluamiaan raportteja sorttaa ja tai vie niitä exceliin järjestelmä hakee raportin tiedot ja näyttää käyttäjän valitseman raportin käyttäjä sorttaa raporttia käyt täjän antamien käskyjen mukaan käyttöliittymässä järjestelmä vie raportin exceliin lataa tiedoston jos käyttäjä aktivoi kyseisen painikkeen käyttäjä siirtyy takaisin etusivulle variaatiot käyttäjä tekee raporttien selailun valinnat uudelleen käyttötapaus d alkaa alusta käyttäjä jolla ei ole käyttöoikeuksia yrittää päästä ennakkotarkastukseen järjestelmä ilmoittaa rajoite tuista käyttöoikeuksista poikkeukset tekninen vika järjestelmä ilmoittaa keskeytymisestä
| 0
|
21,311
| 28,505,193,651
|
IssuesEvent
|
2023-04-18 20:44:37
|
TUM-Dev/NavigaTUM
|
https://api.github.com/repos/TUM-Dev/NavigaTUM
|
closed
|
[Entry] [0502.EG.221]:
|
entry webform delete-after-processing
|
There's a WC hidden inside this Treppenhaus. I couldn't find the WC number, but the one thing I know from my observation is there's no gender markers for this. LG, DiversiTUM
|
1.0
|
[Entry] [0502.EG.221]: - There's a WC hidden inside this Treppenhaus. I couldn't find the WC number, but the one thing I know from my observation is there's no gender markers for this. LG, DiversiTUM
|
process
|
there s a wc hidden inside this treppenhaus i couldn t find the wc number but the one thing i know from my observation is there s no gender markers for this lg diversitum
| 1
|
49,173
| 6,015,995,909
|
IssuesEvent
|
2017-06-07 04:59:36
|
Microsoft/vsts-tasks
|
https://api.github.com/repos/Microsoft/vsts-tasks
|
closed
|
[DeployVisualStudioTestAgent] Problem with username having more than 8 characters
|
Area: Test
|
Hello,
I've tried to deploy the test agent with Machine's admin username: trinh.pham and got the error as below
```
2017-05-22T05:17:40.6870427Z RemoteDeployerSource Verbose: 1221675 : [RemoteDeployer][22:May:17:12:17:38:8037; 6772, 5](LGVN14143-W10:5985-9005d6f4-9e7f-441d-81be-38f36e1922be)Finished retrying 1 out of 3 times for exception: System.Management.Automation.CmdletProviderInvocationException: An object at the specified path C:\Users\TRINH~1.PHA does not exist. ---> System.Management.Automation.PSArgumentException: An object at the specified path C:\Users\TRINH~1.PHA does not exist.
2017-05-22T05:17:40.6870427Z at Microsoft.PowerShell.Commands.FileSystemProvider.NormalizeThePath(String basepath, Stack`1 tokenizedPathStack)
2017-05-22T05:17:40.6870427Z at Microsoft.PowerShell.Commands.FileSystemProvider.NormalizeRelativePathHelper(String path, String basePath)
2017-05-22T05:17:40.6870427Z --- End of inner exception stack trace ---
2017-05-22T05:17:40.6880438Z at System.Management.Automation.Runspaces.PipelineBase.Invoke(IEnumerable input)
2017-05-22T05:17:40.6880438Z at System.Management.Automation.PowerShell.Worker.ConstructPipelineAndDoWork(Runspace rs, Boolean performSyncInvoke)
2017-05-22T05:17:40.6880438Z at System.Management.Automation.PowerShell.Worker.CreateRunspaceIfNeededAndDoWork(Runspace rsToUse, Boolean isSync)
2017-05-22T05:17:40.6890444Z at System.Management.Automation.PowerShell.CoreInvokeHelper[TInput,TOutput](PSDataCollection`1 input, PSDataCollection`1 output, PSInvocationSettings settings)
2017-05-22T05:17:40.6890444Z at System.Management.Automation.PowerShell.CoreInvoke[TInput,TOutput](PSDataCollection`1 input, PSDataCollection`1 output, PSInvocationSettings settings)
2017-05-22T05:17:40.6900446Z at System.Management.Automation.PowerShell.Invoke(IEnumerable input, PSInvocationSettings settings)
2017-05-22T05:17:40.6900446Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Deployment.PowershellExecutor.<>c__DisplayClass28_0.<Invoke>b__0()
2017-05-22T05:17:40.6900446Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Utilities.RetryExecutor.<>c__DisplayClass19_0`1.<Execute>b__0()
2017-05-22T05:17:40.6910458Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Utilities.RetryExecutor.Execute(Action action)
2017-05-22T05:17:40.6910458Z RemoteDeployerSource Error: 1221741 : [RemoteDeployer][22:May:17:12:17:38:8057; 6772, 5](LGVN14143-W10:5985-9005d6f4-9e7f-441d-81be-38f36e1922be)System.AggregateException: Failed to execute the powershell script. Consult the logs below for details of the error.
2017-05-22T05:17:40.6910458Z An object at the specified path C:\Users\TRINH~1.PHA does not exist. ---> System.Management.Automation.CmdletProviderInvocationException: An object at the specified path C:\Users\TRINH~1.PHA does not exist. ---> System.Management.Automation.PSArgumentException: An object at the specified path C:\Users\TRINH~1.PHA does not exist.
2017-05-22T05:17:40.6920460Z at Microsoft.PowerShell.Commands.FileSystemProvider.NormalizeThePath(String basepath, Stack`1 tokenizedPathStack)
2017-05-22T05:17:40.6920460Z at Microsoft.PowerShell.Commands.FileSystemProvider.NormalizeRelativePathHelper(String path, String basePath)
2017-05-22T05:17:40.6920460Z --- End of inner exception stack trace ---
2017-05-22T05:17:40.6920460Z at System.Management.Automation.Runspaces.PipelineBase.Invoke(IEnumerable input)
2017-05-22T05:17:40.6920460Z at System.Management.Automation.PowerShell.Worker.ConstructPipelineAndDoWork(Runspace rs, Boolean performSyncInvoke)
2017-05-22T05:17:40.6930474Z at System.Management.Automation.PowerShell.Worker.CreateRunspaceIfNeededAndDoWork(Runspace rsToUse, Boolean isSync)
2017-05-22T05:17:40.6930474Z at System.Management.Automation.PowerShell.CoreInvokeHelper[TInput,TOutput](PSDataCollection`1 input, PSDataCollection`1 output, PSInvocationSettings settings)
2017-05-22T05:17:40.6930474Z at System.Management.Automation.PowerShell.CoreInvoke[TInput,TOutput](PSDataCollection`1 input, PSDataCollection`1 output, PSInvocationSettings settings)
2017-05-22T05:17:40.6940483Z at System.Management.Automation.PowerShell.Invoke(IEnumerable input, PSInvocationSettings settings)
2017-05-22T05:17:40.6940483Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Deployment.PowershellExecutor.<>c__DisplayClass28_0.<Invoke>b__0()
2017-05-22T05:17:40.6940483Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Utilities.RetryExecutor.<>c__DisplayClass19_0`1.<Execute>b__0()
2017-05-22T05:17:40.6940483Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Utilities.RetryExecutor.Execute(Action action)
2017-05-22T05:17:40.6940483Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Utilities.RetryExecutor.Execute[T](Func`1 action)
2017-05-22T05:17:40.6950488Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Deployment.PowershellExecutor.Invoke(String errorContextMessage, Boolean writeResultToLog, Boolean isCancellable, ActionPreference errorActionPreference)
2017-05-22T05:17:40.6950488Z --- End of inner exception stack trace ---
2017-05-22T05:17:40.6950488Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Deployment.PowershellExecutor.Invoke(String errorContextMessage, Boolean writeResultToLog, Boolean isCancellable, ActionPreference errorActionPreference)
2017-05-22T05:17:40.7090581Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Deployment.ScriptExecutor.ExecuteScript(IPowerShell powerShell, String script, String arguments, Boolean isScriptDotSourced, Boolean useHttp)
2017-05-22T05:17:40.7100584Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Deployment.DeploymentService.RunPowerShellScripts(DeploymentMachineSpecification deploymentMachine, ScriptSpecification scriptSpecification, ScriptSpecification initializationScriptSpecification, String applicationPath, IPowerShell powerShellSession)
2017-05-22T05:17:40.7100584Z ---> (Inner Exception #0) System.Management.Automation.CmdletProviderInvocationException: An object at the specified path C:\Users\TRINH~1.PHA does not exist. ---> System.Management.Automation.PSArgumentException: An object at the specified path C:\Users\TRINH~1.PHA does not exist.
2017-05-22T05:17:40.7110589Z at Microsoft.PowerShell.Commands.FileSystemProvider.NormalizeThePath(String basepath, Stack`1 tokenizedPathStack)
2017-05-22T05:17:40.7110589Z at Microsoft.PowerShell.Commands.FileSystemProvider.NormalizeRelativePathHelper(String path, String basePath)
2017-05-22T05:17:40.7110589Z --- End of inner exception stack trace ---
2017-05-22T05:17:40.7120607Z at System.Management.Automation.Runspaces.PipelineBase.Invoke(IEnumerable input)
2017-05-22T05:17:40.7120607Z at System.Management.Automation.PowerShell.Worker.ConstructPipelineAndDoWork(Runspace rs, Boolean performSyncInvoke)
2017-05-22T05:17:40.7130606Z at System.Management.Automation.PowerShell.Worker.CreateRunspaceIfNeededAndDoWork(Runspace rsToUse, Boolean isSync)
2017-05-22T05:17:40.7130606Z at System.Management.Automation.PowerShell.CoreInvokeHelper[TInput,TOutput](PSDataCollection`1 input, PSDataCollection`1 output, PSInvocationSettings settings)
2017-05-22T05:17:40.7130606Z at System.Management.Automation.PowerShell.CoreInvoke[TInput,TOutput](PSDataCollection`1 input, PSDataCollection`1 output, PSInvocationSettings settings)
2017-05-22T05:17:40.7140611Z at System.Management.Automation.PowerShell.Invoke(IEnumerable input, PSInvocationSettings settings)
2017-05-22T05:17:40.7140611Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Deployment.PowershellExecutor.<>c__DisplayClass28_0.<Invoke>b__0()
2017-05-22T05:17:40.7140611Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Utilities.RetryExecutor.<>c__DisplayClass19_0`1.<Execute>b__0()
2017-05-22T05:17:40.7150617Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Utilities.RetryExecutor.Execute(Action action)
2017-05-22T05:17:40.7150617Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Utilities.RetryExecutor.Execute[T](Func`1 action)
2017-05-22T05:17:40.7150617Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Deployment.PowershellExecutor.Invoke(String errorContextMessage, Boolean writeResultToLog, Boolean isCancellable, ActionPreference errorActionPreference)<---
```
Trying with a shorter username (less than 8 characters) worked as expected.
Please help review and fix it ASAP.
**Environments**
- Windows 10 x64
- TFS 2017 on-premises, update 1
- VSTS Build Agent: vsts-agent-win7-x64-2.116.0
- VSTS Task: \DeployVisualStudioTestAgent\1.0.42
- My username [trinh.pham] is administrator and is configured for running as build agent service and test agent
|
1.0
|
[DeployVisualStudioTestAgent] Problem with username having more than 8 characters - Hello,
I've tried to deploy the test agent with Machine's admin username: trinh.pham and got the error as below
```
2017-05-22T05:17:40.6870427Z RemoteDeployerSource Verbose: 1221675 : [RemoteDeployer][22:May:17:12:17:38:8037; 6772, 5](LGVN14143-W10:5985-9005d6f4-9e7f-441d-81be-38f36e1922be)Finished retrying 1 out of 3 times for exception: System.Management.Automation.CmdletProviderInvocationException: An object at the specified path C:\Users\TRINH~1.PHA does not exist. ---> System.Management.Automation.PSArgumentException: An object at the specified path C:\Users\TRINH~1.PHA does not exist.
2017-05-22T05:17:40.6870427Z at Microsoft.PowerShell.Commands.FileSystemProvider.NormalizeThePath(String basepath, Stack`1 tokenizedPathStack)
2017-05-22T05:17:40.6870427Z at Microsoft.PowerShell.Commands.FileSystemProvider.NormalizeRelativePathHelper(String path, String basePath)
2017-05-22T05:17:40.6870427Z --- End of inner exception stack trace ---
2017-05-22T05:17:40.6880438Z at System.Management.Automation.Runspaces.PipelineBase.Invoke(IEnumerable input)
2017-05-22T05:17:40.6880438Z at System.Management.Automation.PowerShell.Worker.ConstructPipelineAndDoWork(Runspace rs, Boolean performSyncInvoke)
2017-05-22T05:17:40.6880438Z at System.Management.Automation.PowerShell.Worker.CreateRunspaceIfNeededAndDoWork(Runspace rsToUse, Boolean isSync)
2017-05-22T05:17:40.6890444Z at System.Management.Automation.PowerShell.CoreInvokeHelper[TInput,TOutput](PSDataCollection`1 input, PSDataCollection`1 output, PSInvocationSettings settings)
2017-05-22T05:17:40.6890444Z at System.Management.Automation.PowerShell.CoreInvoke[TInput,TOutput](PSDataCollection`1 input, PSDataCollection`1 output, PSInvocationSettings settings)
2017-05-22T05:17:40.6900446Z at System.Management.Automation.PowerShell.Invoke(IEnumerable input, PSInvocationSettings settings)
2017-05-22T05:17:40.6900446Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Deployment.PowershellExecutor.<>c__DisplayClass28_0.<Invoke>b__0()
2017-05-22T05:17:40.6900446Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Utilities.RetryExecutor.<>c__DisplayClass19_0`1.<Execute>b__0()
2017-05-22T05:17:40.6910458Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Utilities.RetryExecutor.Execute(Action action)
2017-05-22T05:17:40.6910458Z RemoteDeployerSource Error: 1221741 : [RemoteDeployer][22:May:17:12:17:38:8057; 6772, 5](LGVN14143-W10:5985-9005d6f4-9e7f-441d-81be-38f36e1922be)System.AggregateException: Failed to execute the powershell script. Consult the logs below for details of the error.
2017-05-22T05:17:40.6910458Z An object at the specified path C:\Users\TRINH~1.PHA does not exist. ---> System.Management.Automation.CmdletProviderInvocationException: An object at the specified path C:\Users\TRINH~1.PHA does not exist. ---> System.Management.Automation.PSArgumentException: An object at the specified path C:\Users\TRINH~1.PHA does not exist.
2017-05-22T05:17:40.6920460Z at Microsoft.PowerShell.Commands.FileSystemProvider.NormalizeThePath(String basepath, Stack`1 tokenizedPathStack)
2017-05-22T05:17:40.6920460Z at Microsoft.PowerShell.Commands.FileSystemProvider.NormalizeRelativePathHelper(String path, String basePath)
2017-05-22T05:17:40.6920460Z --- End of inner exception stack trace ---
2017-05-22T05:17:40.6920460Z at System.Management.Automation.Runspaces.PipelineBase.Invoke(IEnumerable input)
2017-05-22T05:17:40.6920460Z at System.Management.Automation.PowerShell.Worker.ConstructPipelineAndDoWork(Runspace rs, Boolean performSyncInvoke)
2017-05-22T05:17:40.6930474Z at System.Management.Automation.PowerShell.Worker.CreateRunspaceIfNeededAndDoWork(Runspace rsToUse, Boolean isSync)
2017-05-22T05:17:40.6930474Z at System.Management.Automation.PowerShell.CoreInvokeHelper[TInput,TOutput](PSDataCollection`1 input, PSDataCollection`1 output, PSInvocationSettings settings)
2017-05-22T05:17:40.6930474Z at System.Management.Automation.PowerShell.CoreInvoke[TInput,TOutput](PSDataCollection`1 input, PSDataCollection`1 output, PSInvocationSettings settings)
2017-05-22T05:17:40.6940483Z at System.Management.Automation.PowerShell.Invoke(IEnumerable input, PSInvocationSettings settings)
2017-05-22T05:17:40.6940483Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Deployment.PowershellExecutor.<>c__DisplayClass28_0.<Invoke>b__0()
2017-05-22T05:17:40.6940483Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Utilities.RetryExecutor.<>c__DisplayClass19_0`1.<Execute>b__0()
2017-05-22T05:17:40.6940483Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Utilities.RetryExecutor.Execute(Action action)
2017-05-22T05:17:40.6940483Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Utilities.RetryExecutor.Execute[T](Func`1 action)
2017-05-22T05:17:40.6950488Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Deployment.PowershellExecutor.Invoke(String errorContextMessage, Boolean writeResultToLog, Boolean isCancellable, ActionPreference errorActionPreference)
2017-05-22T05:17:40.6950488Z --- End of inner exception stack trace ---
2017-05-22T05:17:40.6950488Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Deployment.PowershellExecutor.Invoke(String errorContextMessage, Boolean writeResultToLog, Boolean isCancellable, ActionPreference errorActionPreference)
2017-05-22T05:17:40.7090581Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Deployment.ScriptExecutor.ExecuteScript(IPowerShell powerShell, String script, String arguments, Boolean isScriptDotSourced, Boolean useHttp)
2017-05-22T05:17:40.7100584Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Deployment.DeploymentService.RunPowerShellScripts(DeploymentMachineSpecification deploymentMachine, ScriptSpecification scriptSpecification, ScriptSpecification initializationScriptSpecification, String applicationPath, IPowerShell powerShellSession)
2017-05-22T05:17:40.7100584Z ---> (Inner Exception #0) System.Management.Automation.CmdletProviderInvocationException: An object at the specified path C:\Users\TRINH~1.PHA does not exist. ---> System.Management.Automation.PSArgumentException: An object at the specified path C:\Users\TRINH~1.PHA does not exist.
2017-05-22T05:17:40.7110589Z at Microsoft.PowerShell.Commands.FileSystemProvider.NormalizeThePath(String basepath, Stack`1 tokenizedPathStack)
2017-05-22T05:17:40.7110589Z at Microsoft.PowerShell.Commands.FileSystemProvider.NormalizeRelativePathHelper(String path, String basePath)
2017-05-22T05:17:40.7110589Z --- End of inner exception stack trace ---
2017-05-22T05:17:40.7120607Z at System.Management.Automation.Runspaces.PipelineBase.Invoke(IEnumerable input)
2017-05-22T05:17:40.7120607Z at System.Management.Automation.PowerShell.Worker.ConstructPipelineAndDoWork(Runspace rs, Boolean performSyncInvoke)
2017-05-22T05:17:40.7130606Z at System.Management.Automation.PowerShell.Worker.CreateRunspaceIfNeededAndDoWork(Runspace rsToUse, Boolean isSync)
2017-05-22T05:17:40.7130606Z at System.Management.Automation.PowerShell.CoreInvokeHelper[TInput,TOutput](PSDataCollection`1 input, PSDataCollection`1 output, PSInvocationSettings settings)
2017-05-22T05:17:40.7130606Z at System.Management.Automation.PowerShell.CoreInvoke[TInput,TOutput](PSDataCollection`1 input, PSDataCollection`1 output, PSInvocationSettings settings)
2017-05-22T05:17:40.7140611Z at System.Management.Automation.PowerShell.Invoke(IEnumerable input, PSInvocationSettings settings)
2017-05-22T05:17:40.7140611Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Deployment.PowershellExecutor.<>c__DisplayClass28_0.<Invoke>b__0()
2017-05-22T05:17:40.7140611Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Utilities.RetryExecutor.<>c__DisplayClass19_0`1.<Execute>b__0()
2017-05-22T05:17:40.7150617Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Utilities.RetryExecutor.Execute(Action action)
2017-05-22T05:17:40.7150617Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Utilities.RetryExecutor.Execute[T](Func`1 action)
2017-05-22T05:17:40.7150617Z at Microsoft.VisualStudio.Services.DevTestLabs.RemoteDeployer.Deployment.PowershellExecutor.Invoke(String errorContextMessage, Boolean writeResultToLog, Boolean isCancellable, ActionPreference errorActionPreference)<---
```
Trying with a shorter username (less than 8 characters) worked as expected.
Please help review and fix it ASAP.
**Environments**
- Windows 10 x64
- TFS 2017 on-premises, update 1
- VSTS Build Agent: vsts-agent-win7-x64-2.116.0
- VSTS Task: \DeployVisualStudioTestAgent\1.0.42
- My username [trinh.pham] is administrator and is configured for running as build agent service and test agent
|
non_process
|
problem with username having more than characters hello i ve tried to deploy the test agent with machine s admin username trinh pham and got the error as below remotedeployersource verbose finished retrying out of times for exception system management automation cmdletproviderinvocationexception an object at the specified path c users trinh pha does not exist system management automation psargumentexception an object at the specified path c users trinh pha does not exist at microsoft powershell commands filesystemprovider normalizethepath string basepath stack tokenizedpathstack at microsoft powershell commands filesystemprovider normalizerelativepathhelper string path string basepath end of inner exception stack trace at system management automation runspaces pipelinebase invoke ienumerable input at system management automation powershell worker constructpipelineanddowork runspace rs boolean performsyncinvoke at system management automation powershell worker createrunspaceifneededanddowork runspace rstouse boolean issync at system management automation powershell coreinvokehelper psdatacollection input psdatacollection output psinvocationsettings settings at system management automation powershell coreinvoke psdatacollection input psdatacollection output psinvocationsettings settings at system management automation powershell invoke ienumerable input psinvocationsettings settings at microsoft visualstudio services devtestlabs remotedeployer deployment powershellexecutor c b at microsoft visualstudio services devtestlabs remotedeployer utilities retryexecutor c b at microsoft visualstudio services devtestlabs remotedeployer utilities retryexecutor execute action action remotedeployersource error system aggregateexception failed to execute the powershell script consult the logs below for details of the error an object at the specified path c users trinh pha does not exist system management automation cmdletproviderinvocationexception an object at the specified path c users trinh pha does not exist system management automation psargumentexception an object at the specified path c users trinh pha does not exist at microsoft powershell commands filesystemprovider normalizethepath string basepath stack tokenizedpathstack at microsoft powershell commands filesystemprovider normalizerelativepathhelper string path string basepath end of inner exception stack trace at system management automation runspaces pipelinebase invoke ienumerable input at system management automation powershell worker constructpipelineanddowork runspace rs boolean performsyncinvoke at system management automation powershell worker createrunspaceifneededanddowork runspace rstouse boolean issync at system management automation powershell coreinvokehelper psdatacollection input psdatacollection output psinvocationsettings settings at system management automation powershell coreinvoke psdatacollection input psdatacollection output psinvocationsettings settings at system management automation powershell invoke ienumerable input psinvocationsettings settings at microsoft visualstudio services devtestlabs remotedeployer deployment powershellexecutor c b at microsoft visualstudio services devtestlabs remotedeployer utilities retryexecutor c b at microsoft visualstudio services devtestlabs remotedeployer utilities retryexecutor execute action action at microsoft visualstudio services devtestlabs remotedeployer utilities retryexecutor execute func action at microsoft visualstudio services devtestlabs remotedeployer deployment powershellexecutor invoke string errorcontextmessage boolean writeresulttolog boolean iscancellable actionpreference erroractionpreference end of inner exception stack trace at microsoft visualstudio services devtestlabs remotedeployer deployment powershellexecutor invoke string errorcontextmessage boolean writeresulttolog boolean iscancellable actionpreference erroractionpreference at microsoft visualstudio services devtestlabs remotedeployer deployment scriptexecutor executescript ipowershell powershell string script string arguments boolean isscriptdotsourced boolean usehttp at microsoft visualstudio services devtestlabs remotedeployer deployment deploymentservice runpowershellscripts deploymentmachinespecification deploymentmachine scriptspecification scriptspecification scriptspecification initializationscriptspecification string applicationpath ipowershell powershellsession inner exception system management automation cmdletproviderinvocationexception an object at the specified path c users trinh pha does not exist system management automation psargumentexception an object at the specified path c users trinh pha does not exist at microsoft powershell commands filesystemprovider normalizethepath string basepath stack tokenizedpathstack at microsoft powershell commands filesystemprovider normalizerelativepathhelper string path string basepath end of inner exception stack trace at system management automation runspaces pipelinebase invoke ienumerable input at system management automation powershell worker constructpipelineanddowork runspace rs boolean performsyncinvoke at system management automation powershell worker createrunspaceifneededanddowork runspace rstouse boolean issync at system management automation powershell coreinvokehelper psdatacollection input psdatacollection output psinvocationsettings settings at system management automation powershell coreinvoke psdatacollection input psdatacollection output psinvocationsettings settings at system management automation powershell invoke ienumerable input psinvocationsettings settings at microsoft visualstudio services devtestlabs remotedeployer deployment powershellexecutor c b at microsoft visualstudio services devtestlabs remotedeployer utilities retryexecutor c b at microsoft visualstudio services devtestlabs remotedeployer utilities retryexecutor execute action action at microsoft visualstudio services devtestlabs remotedeployer utilities retryexecutor execute func action at microsoft visualstudio services devtestlabs remotedeployer deployment powershellexecutor invoke string errorcontextmessage boolean writeresulttolog boolean iscancellable actionpreference erroractionpreference trying with a shorter username less than characters worked as expected please help review and fix it asap environments windows tfs on premises update vsts build agent vsts agent vsts task deployvisualstudiotestagent my username is administrator and is configured for running as build agent service and test agent
| 0
|
8,397
| 11,567,076,146
|
IssuesEvent
|
2020-02-20 13:42:30
|
GetTerminus/terminus-ui
|
https://api.github.com/repos/GetTerminus/terminus-ui
|
closed
|
Demo: Add new abbreviation pipe to pipes demo page
|
Focus: consumer Goal: Process Improvement Type: chore
|
- [ ] Add simple example to existing pipes demos page.
|
1.0
|
Demo: Add new abbreviation pipe to pipes demo page - - [ ] Add simple example to existing pipes demos page.
|
process
|
demo add new abbreviation pipe to pipes demo page add simple example to existing pipes demos page
| 1
|
9,823
| 12,827,602,697
|
IssuesEvent
|
2020-07-06 18:49:20
|
googleapis/code-suggester
|
https://api.github.com/repos/googleapis/code-suggester
|
opened
|
Create a README for users
|
type: process
|
Create README.md so users know what commands are available and how to install
- [ ] describe the CLI API
- [ ] describe the framework-core library
- [ ] describe how to install the package
- [ ] link to any external docs
|
1.0
|
Create a README for users - Create README.md so users know what commands are available and how to install
- [ ] describe the CLI API
- [ ] describe the framework-core library
- [ ] describe how to install the package
- [ ] link to any external docs
|
process
|
create a readme for users create readme md so users know what commands are available and how to install describe the cli api describe the framework core library describe how to install the package link to any external docs
| 1
|
483,431
| 13,924,694,991
|
IssuesEvent
|
2020-10-21 15:52:21
|
GoogleCloudPlatform/python-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
|
closed
|
IAP special path should be query parameter
|
api: iap priority: p2 samples type: feature request
|
## In which file did you encounter the issue?
https://github.com/GoogleCloudPlatform/python-docs-samples/blob/master/appengine/standard/iap/js/poll.js#L49
### Did you change the file? If so, how?
No
## Describe the issue
Current sample code uses special url path to initiate session refresh:
```js
iapSessionRefreshWindow = window.open("/_gcp_iap/do_session_refresh");
```
IAP interactions should use the preferred method of query parameters moving forward:
```js
iapSessionRefreshWindow = window.open("?gcp-iap-mode=DO_SESSION_REFRESH");
```
|
1.0
|
IAP special path should be query parameter - ## In which file did you encounter the issue?
https://github.com/GoogleCloudPlatform/python-docs-samples/blob/master/appengine/standard/iap/js/poll.js#L49
### Did you change the file? If so, how?
No
## Describe the issue
Current sample code uses special url path to initiate session refresh:
```js
iapSessionRefreshWindow = window.open("/_gcp_iap/do_session_refresh");
```
IAP interactions should use the preferred method of query parameters moving forward:
```js
iapSessionRefreshWindow = window.open("?gcp-iap-mode=DO_SESSION_REFRESH");
```
|
non_process
|
iap special path should be query parameter in which file did you encounter the issue did you change the file if so how no describe the issue current sample code uses special url path to initiate session refresh js iapsessionrefreshwindow window open gcp iap do session refresh iap interactions should use the preferred method of query parameters moving forward js iapsessionrefreshwindow window open gcp iap mode do session refresh
| 0
|
15,549
| 19,703,502,451
|
IssuesEvent
|
2022-01-12 19:07:56
|
googleapis/java-network-security
|
https://api.github.com/repos/googleapis/java-network-security
|
opened
|
Your .repo-metadata.json file has a problem 🤒
|
type: process repo-metadata: lint
|
You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* release_level must be equal to one of the allowed values in .repo-metadata.json
* api_shortname 'network-security' invalid in .repo-metadata.json
☝️ Once you correct these problems, you can close this issue.
Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* release_level must be equal to one of the allowed values in .repo-metadata.json
* api_shortname 'network-security' invalid in .repo-metadata.json
☝️ Once you correct these problems, you can close this issue.
Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 release level must be equal to one of the allowed values in repo metadata json api shortname network security invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
| 1
|
282,459
| 8,706,680,100
|
IssuesEvent
|
2018-12-06 04:03:29
|
magda-io/magda
|
https://api.github.com/repos/magda-io/magda
|
opened
|
Search results help button text is not accessible (critical)
|
priority: high
|
### Problem description
Intopia issue 24 - At the end of the Open Data Quality section, after the star rating, each search results listing provides a help button, a question mark in a circle. This is announced as “graphic help link”. Hovering over the button reveals a tooltip with important information for the user.
### Problem reproduction steps
Make the tooltip accessible to screen reader users or provide an alternative way for them to access the information in the tooltip.
For a detailed explanation of the accessibility issues with tooltips and how to address, see Heydon Pickering’s article Tooltips & Toggletips.
For an example of markup, styling and script for an accessible tooltip, see Heydon’s Codepen.
### Screenshot / Design / File reference
[24] Search results help button text is not accessible
Type: WCAG 2.1 failure
Fails 2.1.1 Keyboard - Level A
Location: Help button, Search page, Dataset page
Platform: All
Severity: Critical
What’s the issue?
At the end of the Open Data Quality section, after the star rating, each search results listing provides a help button, a question mark in a circle. This is announced as “graphic help link”. Hovering over the button reveals a tooltip with important information for the user.
Why it matters The help icon is not keyboard accessible and requires a mouse hover to reveal a tooltip with information and a link related to how the star rating is calculated. The information in the tooltip, which provides a link that in turn explains how the star gradings are calculated, is not accessible in any other way.
Our recommendation
Make the tooltip accessible to screen reader users or provide an alternative way for them to access the information in the tooltip.
For a detailed explanation of the accessibility issues with tooltips and how to address, see Heydon Pickering’s article Tooltips & Toggletips.
For an example of markup, styling and script for an accessible tooltip, see Heydon’s Codepen.
|
1.0
|
Search results help button text is not accessible (critical) - ### Problem description
Intopia issue 24 - At the end of the Open Data Quality section, after the star rating, each search results listing provides a help button, a question mark in a circle. This is announced as “graphic help link”. Hovering over the button reveals a tooltip with important information for the user.
### Problem reproduction steps
Make the tooltip accessible to screen reader users or provide an alternative way for them to access the information in the tooltip.
For a detailed explanation of the accessibility issues with tooltips and how to address, see Heydon Pickering’s article Tooltips & Toggletips.
For an example of markup, styling and script for an accessible tooltip, see Heydon’s Codepen.
### Screenshot / Design / File reference
[24] Search results help button text is not accessible
Type: WCAG 2.1 failure
Fails 2.1.1 Keyboard - Level A
Location: Help button, Search page, Dataset page
Platform: All
Severity: Critical
What’s the issue?
At the end of the Open Data Quality section, after the star rating, each search results listing provides a help button, a question mark in a circle. This is announced as “graphic help link”. Hovering over the button reveals a tooltip with important information for the user.
Why it matters The help icon is not keyboard accessible and requires a mouse hover to reveal a tooltip with information and a link related to how the star rating is calculated. The information in the tooltip, which provides a link that in turn explains how the star gradings are calculated, is not accessible in any other way.
Our recommendation
Make the tooltip accessible to screen reader users or provide an alternative way for them to access the information in the tooltip.
For a detailed explanation of the accessibility issues with tooltips and how to address, see Heydon Pickering’s article Tooltips & Toggletips.
For an example of markup, styling and script for an accessible tooltip, see Heydon’s Codepen.
|
non_process
|
search results help button text is not accessible critical problem description intopia issue at the end of the open data quality section after the star rating each search results listing provides a help button a question mark in a circle this is announced as “graphic help link” hovering over the button reveals a tooltip with important information for the user problem reproduction steps make the tooltip accessible to screen reader users or provide an alternative way for them to access the information in the tooltip for a detailed explanation of the accessibility issues with tooltips and how to address see heydon pickering’s article tooltips toggletips for an example of markup styling and script for an accessible tooltip see heydon’s codepen screenshot design file reference search results help button text is not accessible type wcag failure fails keyboard level a location help button search page dataset page platform all severity critical what’s the issue at the end of the open data quality section after the star rating each search results listing provides a help button a question mark in a circle this is announced as “graphic help link” hovering over the button reveals a tooltip with important information for the user why it matters the help icon is not keyboard accessible and requires a mouse hover to reveal a tooltip with information and a link related to how the star rating is calculated the information in the tooltip which provides a link that in turn explains how the star gradings are calculated is not accessible in any other way our recommendation make the tooltip accessible to screen reader users or provide an alternative way for them to access the information in the tooltip for a detailed explanation of the accessibility issues with tooltips and how to address see heydon pickering’s article tooltips toggletips for an example of markup styling and script for an accessible tooltip see heydon’s codepen
| 0
|
298,439
| 22,499,118,965
|
IssuesEvent
|
2022-06-23 10:09:14
|
MauriceNino/dashdot
|
https://api.github.com/repos/MauriceNino/dashdot
|
opened
|
[Feature] Allow integrations to dash.
|
type: documentation type: feature
|
### Description of the feature
Dash. handles a lot of the overhead when querying system information through docker and displays all that in some charts. Other projects (like https://github.com/ajnart/homarr/) may want to simply include those, instead of going through the hassle of re-implementing everything.
### Additional context
- [ ] Create URL options, that allow to view a single chart with no surroundings
- [ ] Create an API endpoint, that allows to query the config
- [ ] Create a doc, that details on how to include dash. charts in your page
|
1.0
|
[Feature] Allow integrations to dash. - ### Description of the feature
Dash. handles a lot of the overhead when querying system information through docker and displays all that in some charts. Other projects (like https://github.com/ajnart/homarr/) may want to simply include those, instead of going through the hassle of re-implementing everything.
### Additional context
- [ ] Create URL options, that allow to view a single chart with no surroundings
- [ ] Create an API endpoint, that allows to query the config
- [ ] Create a doc, that details on how to include dash. charts in your page
|
non_process
|
allow integrations to dash description of the feature dash handles a lot of the overhead when querying system information through docker and displays all that in some charts other projects like may want to simply include those instead of going through the hassle of re implementing everything additional context create url options that allow to view a single chart with no surroundings create an api endpoint that allows to query the config create a doc that details on how to include dash charts in your page
| 0
|
10,014
| 13,043,884,184
|
IssuesEvent
|
2020-07-29 02:56:51
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `WeekWithMode` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `WeekWithMode` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `WeekWithMode` from TiDB -
## Description
Port the scalar function `WeekWithMode` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function weekwithmode from tidb description port the scalar function weekwithmode from tidb to coprocessor score mentor s breeswish recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
131,941
| 28,057,607,917
|
IssuesEvent
|
2023-03-29 10:24:22
|
TheCoderAdi/GamingBeast
|
https://api.github.com/repos/TheCoderAdi/GamingBeast
|
opened
|
More Games
|
enhancement hackcodex2023
|
It's fun to have more games in a website , free feel to create mini games in javascript.
If you are creating more games don't add in the navigation bar .
You can create a new button in the navigation bar like more games , and add your games.
## Work
- [ ] Add games
|
1.0
|
More Games - It's fun to have more games in a website , free feel to create mini games in javascript.
If you are creating more games don't add in the navigation bar .
You can create a new button in the navigation bar like more games , and add your games.
## Work
- [ ] Add games
|
non_process
|
more games it s fun to have more games in a website free feel to create mini games in javascript if you are creating more games don t add in the navigation bar you can create a new button in the navigation bar like more games and add your games work add games
| 0
|
782,304
| 27,492,779,896
|
IssuesEvent
|
2023-03-04 20:39:37
|
azerothcore/azerothcore-wotlk
|
https://api.github.com/repos/azerothcore/azerothcore-wotlk
|
closed
|
Core/Script: Halls of Reflection
|
Instance - Dungeon - Northrend 80 Priority-Medium
|
<!-- IF YOU DO NOT FILL THIS OUT, WE WILL CLOSE YOUR ISSUE! -->
<!-- IF YOU DO NOT FILL THIS OUT, WE WILL CLOSE YOUR ISSUE! -->
<!-- IF YOU DO NOT FILL THIS OUT, WE WILL CLOSE YOUR ISSUE! -->
<!-- IF YOU DO NOT FILL THIS OUT, WE WILL CLOSE YOUR ISSUE! -->
<!-- IF YOU DO NOT FILL THIS OUT, WE WILL CLOSE YOUR ISSUE! -->
<!-- IF YOU DO NOT FILL THIS OUT, WE WILL CLOSE YOUR ISSUE! -->
<!-- IF YOU DO NOT FILL THIS OUT, WE WILL CLOSE YOUR ISSUE! -->
<!-- IF YOU DO NOT FILL THIS OUT, WE WILL CLOSE YOUR ISSUE! -->
**Description**:
A wrong script in the halls of reflection.
**Current behaviour**: Tell us what happens
When you enter the dungeon, triggered the script and it is not correct!
The sound and text don't match. The text was from the opposite faction.
Example: a Jaina approach to Uther and the conversation (script) triggered an appeal to Silwan.
After the murder of "Marvin" should be assigned the quest:
Alliance: http://wowhead.com/quest=24500 - Auto acceptance
Horde: http://wowhead.com/quest=24802 - Auto acceptance
Missing text and sounds:
- Stupid girl! The one you seek is long dead! Now he is a shadow ... (Something like this) :-)
Простите за мой англиский! :-)
Sorry for my English! :-)
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/54355046-core-script-halls-of-reflection?utm_campaign=plugin&utm_content=tracker%2F40032087&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F40032087&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
1.0
|
Core/Script: Halls of Reflection - <!-- IF YOU DO NOT FILL THIS OUT, WE WILL CLOSE YOUR ISSUE! -->
<!-- IF YOU DO NOT FILL THIS OUT, WE WILL CLOSE YOUR ISSUE! -->
<!-- IF YOU DO NOT FILL THIS OUT, WE WILL CLOSE YOUR ISSUE! -->
<!-- IF YOU DO NOT FILL THIS OUT, WE WILL CLOSE YOUR ISSUE! -->
<!-- IF YOU DO NOT FILL THIS OUT, WE WILL CLOSE YOUR ISSUE! -->
<!-- IF YOU DO NOT FILL THIS OUT, WE WILL CLOSE YOUR ISSUE! -->
<!-- IF YOU DO NOT FILL THIS OUT, WE WILL CLOSE YOUR ISSUE! -->
<!-- IF YOU DO NOT FILL THIS OUT, WE WILL CLOSE YOUR ISSUE! -->
**Description**:
A wrong script in the halls of reflection.
**Current behaviour**: Tell us what happens
When you enter the dungeon, triggered the script and it is not correct!
The sound and text don't match. The text was from the opposite faction.
Example: a Jaina approach to Uther and the conversation (script) triggered an appeal to Silwan.
After the murder of "Marvin" should be assigned the quest:
Alliance: http://wowhead.com/quest=24500 - Auto acceptance
Horde: http://wowhead.com/quest=24802 - Auto acceptance
Missing text and sounds:
- Stupid girl! The one you seek is long dead! Now he is a shadow ... (Something like this) :-)
Простите за мой англиский! :-)
Sorry for my English! :-)
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/54355046-core-script-halls-of-reflection?utm_campaign=plugin&utm_content=tracker%2F40032087&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F40032087&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
non_process
|
core script halls of reflection description a wrong script in the halls of reflection current behaviour tell us what happens when you enter the dungeon triggered the script and it is not correct the sound and text don t match the text was from the opposite faction example a jaina approach to uther and the conversation script triggered an appeal to silwan after the murder of marvin should be assigned the quest alliance auto acceptance horde auto acceptance missing text and sounds stupid girl the one you seek is long dead now he is a shadow something like this простите за мой англиский sorry for my english want to back this issue we accept bounties via
| 0
|
191
| 2,596,535,631
|
IssuesEvent
|
2015-02-20 21:23:43
|
sensiasoft/sensorhub
|
https://api.github.com/repos/sensiasoft/sensorhub
|
opened
|
Janino based processing
|
enhancement processing
|
What about support for processing using simple java algorithms compiled in real-time by Janino?
|
1.0
|
Janino based processing - What about support for processing using simple java algorithms compiled in real-time by Janino?
|
process
|
janino based processing what about support for processing using simple java algorithms compiled in real time by janino
| 1
|
19,432
| 25,600,933,439
|
IssuesEvent
|
2022-12-01 20:07:13
|
dtcenter/MET
|
https://api.github.com/repos/dtcenter/MET
|
opened
|
Enhance the logic for determining quality flags for time summaries of point observations.
|
type: enhancement priority: medium alert: NEED ACCOUNT KEY alert: NEED PROJECT ASSIGNMENT requestor: METplus Team MET: PreProcessing Tools (Point)
|
## Describe the Enhancement ##
While documenting MET's handling of quality flags for dtcenter/MET#2278, @JohnHalleyGotway found that that handling could be improved when computing time summaries of point observations. This step is preformed by calling common library code from the point pre-processing tools, including pb2nc, madis2nc, ascii2nc, and so on. As of MET version 11.0.0, the quality flag of the time summary observations is always stored by `summary_obs.cc` as being NA.
This issue is to enhance the time summary logic to provide more useful quality flag data.
MET supports the computation of several types of time summaries (min, max, range, mean, stdev, median, P##, sum). Recommend the following loigc:
- For min and max: Select the actual matching observation and report its quality flag string.
- For median, and P##: If the percentile corresponds to an actual observation, report its quality flag string. If its the average of 2 values, and those quality strings are numeric, report the maximum of their values.
- For the range: If the quality flags for the min/max are numeric, report the maximum of their values.
- For mean, stdev, and sum: If the quality flags are numeric, report the maximum of all values in the time window.
The question is what to do otherwise? If the quality flags are not numeric, should the output quality flag be "NA" or the most common string?
### Time Estimate ###
*Estimate the amount of work required here.*
*Issues should represent approximately 1 to 3 days of work.*
### Sub-Issues ###
Consider breaking the enhancement down into sub-issues.
- [ ] *Add a checkbox for each sub-issue here.*
### Relevant Deadlines ###
*List relevant project deadlines here or state NONE.*
### Funding Source ###
*Define the source of funding and account keys here or state NONE.*
## Define the Metadata ##
### Assignee ###
- [ ] Select **engineer(s)** or **no engineer** required
- [ ] Select **scientist(s)** or **no scientist** required
### Labels ###
- [x] Select **component(s)**
- [x] Select **priority**
- [x] Select **requestor(s)**
### Projects and Milestone ###
- [x] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED PROJECT ASSIGNMENT** label
- [x] Select **Milestone** as the next official version or **Future Versions**
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [x] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
No impacts.
## Enhancement Checklist ##
See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)** and **Linked issues**
Select: **Repository** level development cycle **Project** for the next official release
Select: **Milestone** as the next official version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
|
1.0
|
Enhance the logic for determining quality flags for time summaries of point observations. - ## Describe the Enhancement ##
While documenting MET's handling of quality flags for dtcenter/MET#2278, @JohnHalleyGotway found that that handling could be improved when computing time summaries of point observations. This step is preformed by calling common library code from the point pre-processing tools, including pb2nc, madis2nc, ascii2nc, and so on. As of MET version 11.0.0, the quality flag of the time summary observations is always stored by `summary_obs.cc` as being NA.
This issue is to enhance the time summary logic to provide more useful quality flag data.
MET supports the computation of several types of time summaries (min, max, range, mean, stdev, median, P##, sum). Recommend the following loigc:
- For min and max: Select the actual matching observation and report its quality flag string.
- For median, and P##: If the percentile corresponds to an actual observation, report its quality flag string. If its the average of 2 values, and those quality strings are numeric, report the maximum of their values.
- For the range: If the quality flags for the min/max are numeric, report the maximum of their values.
- For mean, stdev, and sum: If the quality flags are numeric, report the maximum of all values in the time window.
The question is what to do otherwise? If the quality flags are not numeric, should the output quality flag be "NA" or the most common string?
### Time Estimate ###
*Estimate the amount of work required here.*
*Issues should represent approximately 1 to 3 days of work.*
### Sub-Issues ###
Consider breaking the enhancement down into sub-issues.
- [ ] *Add a checkbox for each sub-issue here.*
### Relevant Deadlines ###
*List relevant project deadlines here or state NONE.*
### Funding Source ###
*Define the source of funding and account keys here or state NONE.*
## Define the Metadata ##
### Assignee ###
- [ ] Select **engineer(s)** or **no engineer** required
- [ ] Select **scientist(s)** or **no scientist** required
### Labels ###
- [x] Select **component(s)**
- [x] Select **priority**
- [x] Select **requestor(s)**
### Projects and Milestone ###
- [x] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED PROJECT ASSIGNMENT** label
- [x] Select **Milestone** as the next official version or **Future Versions**
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [x] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
No impacts.
## Enhancement Checklist ##
See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)** and **Linked issues**
Select: **Repository** level development cycle **Project** for the next official release
Select: **Milestone** as the next official version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
|
process
|
enhance the logic for determining quality flags for time summaries of point observations describe the enhancement while documenting met s handling of quality flags for dtcenter met johnhalleygotway found that that handling could be improved when computing time summaries of point observations this step is preformed by calling common library code from the point pre processing tools including and so on as of met version the quality flag of the time summary observations is always stored by summary obs cc as being na this issue is to enhance the time summary logic to provide more useful quality flag data met supports the computation of several types of time summaries min max range mean stdev median p sum recommend the following loigc for min and max select the actual matching observation and report its quality flag string for median and p if the percentile corresponds to an actual observation report its quality flag string if its the average of values and those quality strings are numeric report the maximum of their values for the range if the quality flags for the min max are numeric report the maximum of their values for mean stdev and sum if the quality flags are numeric report the maximum of all values in the time window the question is what to do otherwise if the quality flags are not numeric should the output quality flag be na or the most common string time estimate estimate the amount of work required here issues should represent approximately to days of work sub issues consider breaking the enhancement down into sub issues add a checkbox for each sub issue here relevant deadlines list relevant project deadlines here or state none funding source define the source of funding and account keys here or state none define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone select repository and or organization level project s or add alert need project assignment label select milestone as the next official version or future versions define related issue s consider the impact to the other metplus components no impacts enhancement checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of develop branch name feature complete the development and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into develop pull request feature define the pull request metadata as permissions allow select reviewer s and linked issues select repository level development cycle project for the next official release select milestone as the next official version iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue
| 1
|
173,792
| 14,438,099,867
|
IssuesEvent
|
2020-12-07 12:32:51
|
pi-top/pi-top-Python-SDK
|
https://api.github.com/repos/pi-top/pi-top-Python-SDK
|
closed
|
Update top image in README
|
documentation
|
Should be a more accurate representation of all of the things that the SDK is useful for.
|
1.0
|
Update top image in README - Should be a more accurate representation of all of the things that the SDK is useful for.
|
non_process
|
update top image in readme should be a more accurate representation of all of the things that the sdk is useful for
| 0
|
8,561
| 11,734,226,532
|
IssuesEvent
|
2020-03-11 08:57:12
|
labuladong/fucking-algorithm
|
https://api.github.com/repos/labuladong/fucking-algorithm
|
closed
|
Any chance of English translation?
|
enhancement processing
|
Hi,
I've few pages using Google Translator, and the translation didn't sound perfect. Still I managed to read this page https://labuladong.gitbook.io/algo/shu-ju-jie-gou-xi-lie/xue-xi-shu-ju-jie-gou-he-suan-fa-de-gao-xiao-fang-fa and you're explanation about the basic DS is very good.
I want to read the whole book, but I wonder if you (or, someone) is planning for English translation? Thanks for the effort.
|
1.0
|
Any chance of English translation? - Hi,
I've few pages using Google Translator, and the translation didn't sound perfect. Still I managed to read this page https://labuladong.gitbook.io/algo/shu-ju-jie-gou-xi-lie/xue-xi-shu-ju-jie-gou-he-suan-fa-de-gao-xiao-fang-fa and you're explanation about the basic DS is very good.
I want to read the whole book, but I wonder if you (or, someone) is planning for English translation? Thanks for the effort.
|
process
|
any chance of english translation hi i ve few pages using google translator and the translation didn t sound perfect still i managed to read this page and you re explanation about the basic ds is very good i want to read the whole book but i wonder if you or someone is planning for english translation thanks for the effort
| 1
|
22,400
| 31,142,289,992
|
IssuesEvent
|
2023-08-16 01:44:34
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
Flaky test: AssertionError: Timed out retrying after 4000ms: expected '<h1.pt-20px.font-medium.text-center.text-32px.text-body-gray-900>' to contain 'Choose a Browser'
|
OS: linux process: flaky test topic: flake ❄️ stage: flake stale
|
### Link to dashboard or CircleCI failure
https://app.circleci.com/pipelines/github/cypress-io/cypress/41465/workflows/5d7f4386-729e-4aba-aeb6-3541a4a9c5b0/jobs/1716910
### Link to failing test in GitHub
https://github.com/cypress-io/cypress/blob/develop/packages/launchpad/cypress/e2e/choose-a-browser.cy.ts#L218
### Analysis
<img width="1147" alt="Screen Shot 2022-08-05 at 7 23 13 PM" src="https://user-images.githubusercontent.com/26726429/183229885-bbe8238a-64ec-44a5-8875-e328c9e0d098.png">
### Cypress Version
10.4.0
### Other
Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
|
1.0
|
Flaky test: AssertionError: Timed out retrying after 4000ms: expected '<h1.pt-20px.font-medium.text-center.text-32px.text-body-gray-900>' to contain 'Choose a Browser' - ### Link to dashboard or CircleCI failure
https://app.circleci.com/pipelines/github/cypress-io/cypress/41465/workflows/5d7f4386-729e-4aba-aeb6-3541a4a9c5b0/jobs/1716910
### Link to failing test in GitHub
https://github.com/cypress-io/cypress/blob/develop/packages/launchpad/cypress/e2e/choose-a-browser.cy.ts#L218
### Analysis
<img width="1147" alt="Screen Shot 2022-08-05 at 7 23 13 PM" src="https://user-images.githubusercontent.com/26726429/183229885-bbe8238a-64ec-44a5-8875-e328c9e0d098.png">
### Cypress Version
10.4.0
### Other
Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
|
process
|
flaky test assertionerror timed out retrying after expected to contain choose a browser link to dashboard or circleci failure link to failing test in github analysis img width alt screen shot at pm src cypress version other search for this issue number in the codebase to find the test s skipped until this issue is fixed
| 1
|
4,404
| 2,852,839,402
|
IssuesEvent
|
2015-06-01 15:32:08
|
scikit-learn/scikit-learn
|
https://api.github.com/repos/scikit-learn/scikit-learn
|
closed
|
Pipeline.named_steps not documented
|
Documentation Easy
|
``Pipeline.named_steps`` is not documented, and there is no example of using either ``named_steps`` or ``steps`` to access any attributes of the estimator.
That question comes up pretty frequently.
``named_steps`` is weird as it is initialized in ``__init__`` and doesn't have a trailing underscore.
Maybe the real fix is #1769, but for the moment we should just document better, I think.
|
1.0
|
Pipeline.named_steps not documented - ``Pipeline.named_steps`` is not documented, and there is no example of using either ``named_steps`` or ``steps`` to access any attributes of the estimator.
That question comes up pretty frequently.
``named_steps`` is weird as it is initialized in ``__init__`` and doesn't have a trailing underscore.
Maybe the real fix is #1769, but for the moment we should just document better, I think.
|
non_process
|
pipeline named steps not documented pipeline named steps is not documented and there is no example of using either named steps or steps to access any attributes of the estimator that question comes up pretty frequently named steps is weird as it is initialized in init and doesn t have a trailing underscore maybe the real fix is but for the moment we should just document better i think
| 0
|
2,667
| 5,446,928,891
|
IssuesEvent
|
2017-03-07 12:04:53
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
child_process exec '<()' unexpected token
|
child_process question
|
<!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Version**: v7.7.1
* **Platform**: Linux ai 4.9.11-1-ARCH x86_64 GNU/Linux
* **Subsystem**: child_process
<!-- Enter your issue details below this comment. -->
core module child_process.exec issue with <() command.
I executed same command in my bash and zsh shell everything works fine. Once I copy same command to code I get this error.
Shell command:
`comm -13 <(sort new.csv | uniq) <(sort old.csv | uniq) > diff.csv`
When executing same shell command it trows me error:
```
/bin/sh: -c: line 0: syntax error near unexpected token `('
/bin/sh: -c: line 0: `comm -13 <(sort new.csv | uniq) <(sort old.csv | uniq) > diff.csv'
```
Code:
```
exec(`comm -13 <(sort new.csv | uniq) <(sort old.csv | uniq) > diff.csv`,
(error, stdout, stderr) => {
if (error) {
return error
}
console.log(`${stdout}`)
console.log(`${stderr}`)
})
```
After troubleshooting I think problem is with `<(` part of the command.
|
1.0
|
child_process exec '<()' unexpected token - <!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Version**: v7.7.1
* **Platform**: Linux ai 4.9.11-1-ARCH x86_64 GNU/Linux
* **Subsystem**: child_process
<!-- Enter your issue details below this comment. -->
core module child_process.exec issue with <() command.
I executed same command in my bash and zsh shell everything works fine. Once I copy same command to code I get this error.
Shell command:
`comm -13 <(sort new.csv | uniq) <(sort old.csv | uniq) > diff.csv`
When executing same shell command it trows me error:
```
/bin/sh: -c: line 0: syntax error near unexpected token `('
/bin/sh: -c: line 0: `comm -13 <(sort new.csv | uniq) <(sort old.csv | uniq) > diff.csv'
```
Code:
```
exec(`comm -13 <(sort new.csv | uniq) <(sort old.csv | uniq) > diff.csv`,
(error, stdout, stderr) => {
if (error) {
return error
}
console.log(`${stdout}`)
console.log(`${stderr}`)
})
```
After troubleshooting I think problem is with `<(` part of the command.
|
process
|
child process exec unexpected token thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify affected core module name if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able version platform linux ai arch gnu linux subsystem child process core module child process exec issue with command i executed same command in my bash and zsh shell everything works fine once i copy same command to code i get this error shell command comm diff csv when executing same shell command it trows me error bin sh c line syntax error near unexpected token bin sh c line comm diff csv code exec comm diff csv error stdout stderr if error return error console log stdout console log stderr after troubleshooting i think problem is with part of the command
| 1
|
1,136
| 3,626,388,581
|
IssuesEvent
|
2016-02-10 00:33:23
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Failed test: test-child-process-fork-regr-gh-2847
|
arm child_process test windows
|
Looks like this test is starting to fail in our windows environment: [fail 1](https://ci.nodejs.org/job/node-test-binary-windows/186/RUN_SUBSET=2,VS_VERSION=vs2013,label=win2008r2/tapTestReport/test.tap-11/), [fail 2](https://ci.nodejs.org/job/node-test-binary-windows/185/RUN_SUBSET=2,VS_VERSION=vs2013,label=win2008r2/tapTestReport/test.tap-11/)
```
not ok 11 test-child-process-fork-regr-gh-2847.js
# events.js:141
# throw er; // Unhandled 'error' event
# ^
#
# Error: read ECONNRESET
# at exports._errnoException (util.js:860:11)
# at TCP.onread (net.js:544:26)
```
|
1.0
|
Failed test: test-child-process-fork-regr-gh-2847 - Looks like this test is starting to fail in our windows environment: [fail 1](https://ci.nodejs.org/job/node-test-binary-windows/186/RUN_SUBSET=2,VS_VERSION=vs2013,label=win2008r2/tapTestReport/test.tap-11/), [fail 2](https://ci.nodejs.org/job/node-test-binary-windows/185/RUN_SUBSET=2,VS_VERSION=vs2013,label=win2008r2/tapTestReport/test.tap-11/)
```
not ok 11 test-child-process-fork-regr-gh-2847.js
# events.js:141
# throw er; // Unhandled 'error' event
# ^
#
# Error: read ECONNRESET
# at exports._errnoException (util.js:860:11)
# at TCP.onread (net.js:544:26)
```
|
process
|
failed test test child process fork regr gh looks like this test is starting to fail in our windows environment not ok test child process fork regr gh js events js throw er unhandled error event error read econnreset at exports errnoexception util js at tcp onread net js
| 1
|
17,304
| 23,122,081,495
|
IssuesEvent
|
2022-07-27 22:58:09
|
CTSRD-CHERI/cheribsd
|
https://api.github.com/repos/CTSRD-CHERI/cheribsd
|
closed
|
vmspaces need identifiers
|
coprocesses
|
We need a way to refer to vmspaces in `ps`, `procstat`, etc output. We might also want to be able to name them (uniquely?).
|
1.0
|
vmspaces need identifiers - We need a way to refer to vmspaces in `ps`, `procstat`, etc output. We might also want to be able to name them (uniquely?).
|
process
|
vmspaces need identifiers we need a way to refer to vmspaces in ps procstat etc output we might also want to be able to name them uniquely
| 1
|
50,972
| 26,863,107,957
|
IssuesEvent
|
2023-02-03 20:22:56
|
bevyengine/bevy
|
https://api.github.com/repos/bevyengine/bevy
|
closed
|
Low level Component Reflection utiliites for Animation
|
C-Enhancement A-ECS C-Performance A-Animation A-Reflection
|
## What problem does this solve or what need does it fill?
For property based animation (i.e. "animate anything"), we need lower level Reflect utilities than `ReflectComponent` for performant writes of the animated properties. For these systems we will likely have on hand a ComponentId, Entity (and thus EntityLocation), and a pre-parsed property path (see #4081).
## What solution would you like?
Add another optional entry to the `ComponentInfo` vtable. Preferably `unsafe fn(Ptr<'a>) -> Option<&'a dyn Reflect>` and `unsafe fn(PtrMut<'a>) -> &'a mut dyn Reflect`. Only the latter is really neededThis can be combined with `Entity{Ref,Mut}::get_by_id`. Realistically these functions are going to just add the vtable pointer metadata to the underlying pointer.
Preferably we should be able to find a way to automatically register these hooks for all types that are registered that implement `Component + Reflect`.
## What alternative(s) have you considered?
Using `ReflectComponent` as is, repeatedly fetching the same `EntityRef` internally and `TypeId ` -> `ComponentId` hashmap lookup for every property on every component on every animated entity.
|
True
|
Low level Component Reflection utiliites for Animation - ## What problem does this solve or what need does it fill?
For property based animation (i.e. "animate anything"), we need lower level Reflect utilities than `ReflectComponent` for performant writes of the animated properties. For these systems we will likely have on hand a ComponentId, Entity (and thus EntityLocation), and a pre-parsed property path (see #4081).
## What solution would you like?
Add another optional entry to the `ComponentInfo` vtable. Preferably `unsafe fn(Ptr<'a>) -> Option<&'a dyn Reflect>` and `unsafe fn(PtrMut<'a>) -> &'a mut dyn Reflect`. Only the latter is really neededThis can be combined with `Entity{Ref,Mut}::get_by_id`. Realistically these functions are going to just add the vtable pointer metadata to the underlying pointer.
Preferably we should be able to find a way to automatically register these hooks for all types that are registered that implement `Component + Reflect`.
## What alternative(s) have you considered?
Using `ReflectComponent` as is, repeatedly fetching the same `EntityRef` internally and `TypeId ` -> `ComponentId` hashmap lookup for every property on every component on every animated entity.
|
non_process
|
low level component reflection utiliites for animation what problem does this solve or what need does it fill for property based animation i e animate anything we need lower level reflect utilities than reflectcomponent for performant writes of the animated properties for these systems we will likely have on hand a componentid entity and thus entitylocation and a pre parsed property path see what solution would you like add another optional entry to the componentinfo vtable preferably unsafe fn ptr option and unsafe fn ptrmut a mut dyn reflect only the latter is really neededthis can be combined with entity ref mut get by id realistically these functions are going to just add the vtable pointer metadata to the underlying pointer preferably we should be able to find a way to automatically register these hooks for all types that are registered that implement component reflect what alternative s have you considered using reflectcomponent as is repeatedly fetching the same entityref internally and typeid componentid hashmap lookup for every property on every component on every animated entity
| 0
|
621,864
| 19,598,132,737
|
IssuesEvent
|
2022-01-05 20:36:43
|
gsbelarus/check-and-cash
|
https://api.github.com/repos/gsbelarus/check-and-cash
|
closed
|
Возврат спецпитания
|
POSitive:Check Priority-Normal Severity - Critical
|
Доделать возврат блюд, отданных в кредит и спецпитанию.
Изменять остатки и правльно подставлять суммы
|
1.0
|
Возврат спецпитания - Доделать возврат блюд, отданных в кредит и спецпитанию.
Изменять остатки и правльно подставлять суммы
|
non_process
|
возврат спецпитания доделать возврат блюд отданных в кредит и спецпитанию изменять остатки и правльно подставлять суммы
| 0
|
19,287
| 25,466,265,480
|
IssuesEvent
|
2022-11-25 04:55:27
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[GCI] [PM] Issue related to error message which is getting displayed for organizational users credentials in the sign in screen
|
Bug P1 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
|
UI of Error message is getting displayed differently for organizational user credentials in the sign in screen
**AR:**

**ER:** UI of Error message should get displayed similar for organizational user credentials in the sign in screen as similar to the non-organizational user

|
3.0
|
[GCI] [PM] Issue related to error message which is getting displayed for organizational users credentials in the sign in screen - UI of Error message is getting displayed differently for organizational user credentials in the sign in screen
**AR:**

**ER:** UI of Error message should get displayed similar for organizational user credentials in the sign in screen as similar to the non-organizational user

|
process
|
issue related to error message which is getting displayed for organizational users credentials in the sign in screen ui of error message is getting displayed differently for organizational user credentials in the sign in screen ar er ui of error message should get displayed similar for organizational user credentials in the sign in screen as similar to the non organizational user
| 1
|
28,219
| 6,969,522,028
|
IssuesEvent
|
2017-12-11 06:02:26
|
brain-labs/brain
|
https://api.github.com/repos/brain-labs/brain
|
opened
|
Rename expression files (and classes)
|
Discussion Enhancement help wanted Priority: Low Related to: Code Refactoring
|
We need to rename all files and classes because they are being described wrongly.
For examply:
`OutputExpr`, should be `OutputStmt`, however, `ArithmeticExpr` should probably be kept the same. We need to discuss it.
Any thoughts @haskellcamargo, @ryukinix, @rafaelcn ?
|
1.0
|
Rename expression files (and classes) - We need to rename all files and classes because they are being described wrongly.
For examply:
`OutputExpr`, should be `OutputStmt`, however, `ArithmeticExpr` should probably be kept the same. We need to discuss it.
Any thoughts @haskellcamargo, @ryukinix, @rafaelcn ?
|
non_process
|
rename expression files and classes we need to rename all files and classes because they are being described wrongly for examply outputexpr should be outputstmt however arithmeticexpr should probably be kept the same we need to discuss it any thoughts haskellcamargo ryukinix rafaelcn
| 0
|
81,945
| 23,622,691,251
|
IssuesEvent
|
2022-08-24 22:31:47
|
expo/expo
|
https://api.github.com/repos/expo/expo
|
closed
|
[expo-dev-menu] is broken for landscape orientation
|
Issue accepted Development Builds
|
### Summary
`expo-dev-menu` is broken for projects that have `"orientation": "landscape"` in `app.json`.
`expo-dev-menu` is hardcoded for `portrait` orientation and is not usable in landscape(button on initial screen is not pressable)
Affects both Expo Go and app with `expo-dev-client`
### Managed or bare workflow? If you have `ios/` or `android/` directories in your project, the answer is bare!
both bare and managed
### What platform(s) does this occur on?
Android, iOS
### SDK Version (managed workflow only)
45
### Environment
expo-env-info 1.0.3 environment info:
System:
OS: macOS 12.3.1
Shell: 5.8 - /bin/zsh
Binaries:
Node: 16.15.0 - ~/.nvm/versions/node/v16.15.0/bin/node
Yarn: 1.22.19 - /opt/homebrew/bin/yarn
npm: 8.5.5 - ~/.nvm/versions/node/v16.15.0/bin/npm
Watchman: 4.9.0 - /opt/local/bin/watchman
Managers:
CocoaPods: 1.11.3 - /opt/homebrew/lib/ruby/gems/3.0.0/bin/pod
SDKs:
iOS SDK:
Platforms: DriverKit 21.4, iOS 15.4, macOS 12.3, tvOS 15.4, watchOS 8.5
Android SDK:
Android NDK: 21.3.6528147
IDEs:
Xcode: 13.3.1/13E500a - /usr/bin/xcodebuild
npmPackages:
expo: ~45.0.0 => 45.0.5
react: 17.0.2 => 17.0.2
react-dom: 17.0.2 => 17.0.2
react-native: 0.68.2 => 0.68.2
react-native-web: 0.17.7 => 0.17.7
npmGlobalPackages:
eas-cli: 0.52.0
expo-cli: 5.4.12
Expo Workflow: bare
### Reproducible demo
Fresh project created with expo-cli that has `"orientation": "landscape"` in `app.json`
|
1.0
|
[expo-dev-menu] is broken for landscape orientation - ### Summary
`expo-dev-menu` is broken for projects that have `"orientation": "landscape"` in `app.json`.
`expo-dev-menu` is hardcoded for `portrait` orientation and is not usable in landscape(button on initial screen is not pressable)
Affects both Expo Go and app with `expo-dev-client`
### Managed or bare workflow? If you have `ios/` or `android/` directories in your project, the answer is bare!
both bare and managed
### What platform(s) does this occur on?
Android, iOS
### SDK Version (managed workflow only)
45
### Environment
expo-env-info 1.0.3 environment info:
System:
OS: macOS 12.3.1
Shell: 5.8 - /bin/zsh
Binaries:
Node: 16.15.0 - ~/.nvm/versions/node/v16.15.0/bin/node
Yarn: 1.22.19 - /opt/homebrew/bin/yarn
npm: 8.5.5 - ~/.nvm/versions/node/v16.15.0/bin/npm
Watchman: 4.9.0 - /opt/local/bin/watchman
Managers:
CocoaPods: 1.11.3 - /opt/homebrew/lib/ruby/gems/3.0.0/bin/pod
SDKs:
iOS SDK:
Platforms: DriverKit 21.4, iOS 15.4, macOS 12.3, tvOS 15.4, watchOS 8.5
Android SDK:
Android NDK: 21.3.6528147
IDEs:
Xcode: 13.3.1/13E500a - /usr/bin/xcodebuild
npmPackages:
expo: ~45.0.0 => 45.0.5
react: 17.0.2 => 17.0.2
react-dom: 17.0.2 => 17.0.2
react-native: 0.68.2 => 0.68.2
react-native-web: 0.17.7 => 0.17.7
npmGlobalPackages:
eas-cli: 0.52.0
expo-cli: 5.4.12
Expo Workflow: bare
### Reproducible demo
Fresh project created with expo-cli that has `"orientation": "landscape"` in `app.json`
|
non_process
|
is broken for landscape orientation summary expo dev menu is broken for projects that have orientation landscape in app json expo dev menu is hardcoded for portrait orientation and is not usable in landscape button on initial screen is not pressable affects both expo go and app with expo dev client managed or bare workflow if you have ios or android directories in your project the answer is bare both bare and managed what platform s does this occur on android ios sdk version managed workflow only environment expo env info environment info system os macos shell bin zsh binaries node nvm versions node bin node yarn opt homebrew bin yarn npm nvm versions node bin npm watchman opt local bin watchman managers cocoapods opt homebrew lib ruby gems bin pod sdks ios sdk platforms driverkit ios macos tvos watchos android sdk android ndk ides xcode usr bin xcodebuild npmpackages expo react react dom react native react native web npmglobalpackages eas cli expo cli expo workflow bare reproducible demo fresh project created with expo cli that has orientation landscape in app json
| 0
|
6,364
| 9,416,767,728
|
IssuesEvent
|
2019-04-10 15:18:25
|
material-components/material-components-ios
|
https://api.github.com/repos/material-components/material-components-ios
|
closed
|
[Buttons] Promote Theming Extensions to Ready
|
[Buttons] type:Process
|
This was filed as an internal issue. If you are a Googler, please visit [b/124516065](http://b/124516065) for more details.
<!-- Auto-generated content below, do not modify -->
---
#### Internal data
- Associated internal bug: [b/124516065](http://b/124516065)
- Blocked by: https://github.com/material-components/material-components-ios/issues/6656
|
1.0
|
[Buttons] Promote Theming Extensions to Ready - This was filed as an internal issue. If you are a Googler, please visit [b/124516065](http://b/124516065) for more details.
<!-- Auto-generated content below, do not modify -->
---
#### Internal data
- Associated internal bug: [b/124516065](http://b/124516065)
- Blocked by: https://github.com/material-components/material-components-ios/issues/6656
|
process
|
promote theming extensions to ready this was filed as an internal issue if you are a googler please visit for more details internal data associated internal bug blocked by
| 1
|
365,487
| 25,538,393,343
|
IssuesEvent
|
2022-11-29 13:41:59
|
cwinland/FastMoq
|
https://api.github.com/repos/cwinland/FastMoq
|
closed
|
Stack overflow when creating a mock
|
bug documentation Fixed
|
The mock system will attempt to create mocks for all inner mock constructors.
When InnerMockResolution is true (default), this can cause an endless loop.
To fix this, constructors with the same Type are removed if a valid default value is not provided. Meaning, if the mock is created and parameters are passed to it, then constructor with the same Type in the parameter can be used. However, if a default value is required, then instead of creating a mock, it will exclude that constructor.
#14
Also, to prevent a mock from being creating automatically on a constructor:
Mocks.AddType<DelegatingHandler, DelegatingHandler>(_ => null);
|
1.0
|
Stack overflow when creating a mock - The mock system will attempt to create mocks for all inner mock constructors.
When InnerMockResolution is true (default), this can cause an endless loop.
To fix this, constructors with the same Type are removed if a valid default value is not provided. Meaning, if the mock is created and parameters are passed to it, then constructor with the same Type in the parameter can be used. However, if a default value is required, then instead of creating a mock, it will exclude that constructor.
#14
Also, to prevent a mock from being creating automatically on a constructor:
Mocks.AddType<DelegatingHandler, DelegatingHandler>(_ => null);
|
non_process
|
stack overflow when creating a mock the mock system will attempt to create mocks for all inner mock constructors when innermockresolution is true default this can cause an endless loop to fix this constructors with the same type are removed if a valid default value is not provided meaning if the mock is created and parameters are passed to it then constructor with the same type in the parameter can be used however if a default value is required then instead of creating a mock it will exclude that constructor also to prevent a mock from being creating automatically on a constructor mocks addtype null
| 0
|
8,637
| 11,787,245,207
|
IssuesEvent
|
2020-03-17 13:44:05
|
cncf/sig-security
|
https://api.github.com/repos/cncf/sig-security
|
closed
|
[Suggestion] Add recommendations for tooling in security assessments
|
assessment-process suggestion
|
Description: During the security assessments, there are some steps, like taking the self-assessment (that usually comes in a google doc form for collaboration), and make it into a markdown format for the repo.
Impact: This will help lead security reviewers in performing security assessments
Scope: After discussions, should be a short 10 min task to add a line to include tooling.
|
1.0
|
[Suggestion] Add recommendations for tooling in security assessments - Description: During the security assessments, there are some steps, like taking the self-assessment (that usually comes in a google doc form for collaboration), and make it into a markdown format for the repo.
Impact: This will help lead security reviewers in performing security assessments
Scope: After discussions, should be a short 10 min task to add a line to include tooling.
|
process
|
add recommendations for tooling in security assessments description during the security assessments there are some steps like taking the self assessment that usually comes in a google doc form for collaboration and make it into a markdown format for the repo impact this will help lead security reviewers in performing security assessments scope after discussions should be a short min task to add a line to include tooling
| 1
|
8,445
| 11,614,077,314
|
IssuesEvent
|
2020-02-26 11:56:02
|
MHRA/products
|
https://api.github.com/repos/MHRA/products
|
closed
|
Add Health Endpoint to Stub Document Manager API
|
EPIC - Auto Batch Process :oncoming_automobile: HIGH PRIORITY :arrow_double_up:
|
Further to conversations with Dilu, we need to add a health endpoint.
**Acceptance criteria:**
- [x] Should be accessible at `/health`
- [x] Should require authentication
- [x] Should return OK with no contents
|
1.0
|
Add Health Endpoint to Stub Document Manager API - Further to conversations with Dilu, we need to add a health endpoint.
**Acceptance criteria:**
- [x] Should be accessible at `/health`
- [x] Should require authentication
- [x] Should return OK with no contents
|
process
|
add health endpoint to stub document manager api further to conversations with dilu we need to add a health endpoint acceptance criteria should be accessible at health should require authentication should return ok with no contents
| 1
|
67,297
| 16,869,946,743
|
IssuesEvent
|
2021-06-22 02:09:47
|
opencv/opencv
|
https://api.github.com/repos/opencv/opencv
|
closed
|
OpenCV 4.5.2 videoio not compatible with ffmpeg 4.4
|
category: build/install incomplete
|
<!--
If you have a question rather than reporting a bug please go to https://forum.opencv.org where you get much faster responses.
If you need further assistance please read [How To Contribute](https://github.com/opencv/opencv/wiki/How_to_contribute).
This is a template helping you to create an issue which can be processed as quickly as possible. This is the bug reporting section for the OpenCV library.
-->
##### System information (version)
- OpenCV => 4.5.2
- Operating System / Platform => Linux
- Compiler => gcc 4.8.5
Build fails when building opencv 4.5.2 with FFmpeg 4.4:
```
In file included from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg_impl.hpp:100:0,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:50:
/opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg_hw.hpp: In constructor 'HWAccelIterator::HWAccelIterator(cv::VideoAccelerationType, bool, AVDictionary*)':
/opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg_hw.hpp:501:23: error: use of deleted function 'std::basic_istringstream<char>& std::basic_istringstream<char>::operator=(const std::basic_istringstream<char>&)'
s_stream_ = std::istringstream(accel_list);
^
In file included from /usr/include/c++/4.8.2/complex:45:0,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core/cvstd.inl.hpp:47,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core.hpp:3306,
from /opt/build/opencv-4.5.2/modules/videoio/include/opencv2/videoio.hpp:46,
from /opt/build/opencv-4.5.2/modules/videoio/src/precomp.hpp:57,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:42:
/usr/include/c++/4.8.2/sstream:272:11: note: 'std::basic_istringstream<char>& std::basic_istringstream<char>::operator=(const std::basic_istringstream<char>&)' is implicitly deleted because the default definition would be ill-formed:
class basic_istringstream : public basic_istream<_CharT, _Traits>
^
/usr/include/c++/4.8.2/sstream:272:11: error: use of deleted function 'std::basic_istream<char>& std::basic_istream<char>::operator=(const std::basic_istream<char>&)'
In file included from /usr/include/c++/4.8.2/sstream:38:0,
from /usr/include/c++/4.8.2/complex:45,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core/cvstd.inl.hpp:47,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core.hpp:3306,
from /opt/build/opencv-4.5.2/modules/videoio/include/opencv2/videoio.hpp:46,
from /opt/build/opencv-4.5.2/modules/videoio/src/precomp.hpp:57,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:42:
/usr/include/c++/4.8.2/istream:58:11: note: 'std::basic_istream<char>& std::basic_istream<char>::operator=(const std::basic_istream<char>&)' is implicitly deleted because the default definition would be ill-formed:
class basic_istream : virtual public basic_ios<_CharT, _Traits>
^
/usr/include/c++/4.8.2/istream:58:11: error: use of deleted function 'std::basic_ios<char>& std::basic_ios<char>::operator=(const std::basic_ios<char>&)'
In file included from /usr/include/c++/4.8.2/ios:44:0,
from /usr/include/c++/4.8.2/istream:38,
from /usr/include/c++/4.8.2/sstream:38,
from /usr/include/c++/4.8.2/complex:45,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core/cvstd.inl.hpp:47,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core.hpp:3306,
from /opt/build/opencv-4.5.2/modules/videoio/include/opencv2/videoio.hpp:46,
from /opt/build/opencv-4.5.2/modules/videoio/src/precomp.hpp:57,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:42:
/usr/include/c++/4.8.2/bits/basic_ios.h:66:11: note: 'std::basic_ios<char>& std::basic_ios<char>::operator=(const std::basic_ios<char>&)' is implicitly deleted because the default definition would be ill-formed:
class basic_ios : public ios_base
^
In file included from /usr/include/c++/4.8.2/ios:42:0,
from /usr/include/c++/4.8.2/istream:38,
from /usr/include/c++/4.8.2/sstream:38,
from /usr/include/c++/4.8.2/complex:45,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core/cvstd.inl.hpp:47,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core.hpp:3306,
from /opt/build/opencv-4.5.2/modules/videoio/include/opencv2/videoio.hpp:46,
from /opt/build/opencv-4.5.2/modules/videoio/src/precomp.hpp:57,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:42:
/usr/include/c++/4.8.2/bits/ios_base.h:789:5: error: 'std::ios_base& std::ios_base::operator=(const std::ios_base&)' is private
operator=(const ios_base&);
^
In file included from /usr/include/c++/4.8.2/ios:44:0,
from /usr/include/c++/4.8.2/istream:38,
from /usr/include/c++/4.8.2/sstream:38,
from /usr/include/c++/4.8.2/complex:45,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core/cvstd.inl.hpp:47,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core.hpp:3306,
from /opt/build/opencv-4.5.2/modules/videoio/include/opencv2/videoio.hpp:46,
from /opt/build/opencv-4.5.2/modules/videoio/src/precomp.hpp:57,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:42:
/usr/include/c++/4.8.2/bits/basic_ios.h:66:11: error: within this context
class basic_ios : public ios_base
^
In file included from /usr/include/c++/4.8.2/complex:45:0,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core/cvstd.inl.hpp:47,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core.hpp:3306,
from /opt/build/opencv-4.5.2/modules/videoio/include/opencv2/videoio.hpp:46,
from /opt/build/opencv-4.5.2/modules/videoio/src/precomp.hpp:57,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:42:
/usr/include/c++/4.8.2/sstream:272:11: error: use of deleted function 'std::basic_stringbuf<char>& std::basic_stringbuf<char>::operator=(const std::basic_stringbuf<char>&)'
class basic_istringstream : public basic_istream<_CharT, _Traits>
^
/usr/include/c++/4.8.2/sstream:64:11: note: 'std::basic_stringbuf<char>& std::basic_stringbuf<char>::operator=(const std::basic_stringbuf<char>&)' is implicitly deleted because the default definition would be ill-formed:
class basic_stringbuf : public basic_streambuf<_CharT, _Traits>
^
In file included from /usr/include/c++/4.8.2/ios:43:0,
from /usr/include/c++/4.8.2/istream:38,
from /usr/include/c++/4.8.2/sstream:38,
from /usr/include/c++/4.8.2/complex:45,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core/cvstd.inl.hpp:47,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core.hpp:3306,
from /opt/build/opencv-4.5.2/modules/videoio/include/opencv2/videoio.hpp:46,
from /opt/build/opencv-4.5.2/modules/videoio/src/precomp.hpp:57,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:42:
/usr/include/c++/4.8.2/streambuf:810:7: error: 'std::basic_streambuf<_CharT, _Traits>& std::basic_streambuf<_CharT, _Traits>::operator=(const std::basic_streambuf<_CharT, _Traits>&) [with _CharT = char; _Traits = std::char_traits<char>]' is private
operator=(const basic_streambuf& __sb)
^
In file included from /usr/include/c++/4.8.2/complex:45:0,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core/cvstd.inl.hpp:47,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core.hpp:3306,
from /opt/build/opencv-4.5.2/modules/videoio/include/opencv2/videoio.hpp:46,
from /opt/build/opencv-4.5.2/modules/videoio/src/precomp.hpp:57,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:42:
/usr/include/c++/4.8.2/sstream:64:11: error: within this context
class basic_stringbuf : public basic_streambuf<_CharT, _Traits>
^
make[2]: *** [modules/videoio/CMakeFiles/opencv_videoio.dir/src/cap_ffmpeg.cpp.o] Error 1
make[2]: *** Waiting for unfinished jobs....
make[1]: *** [modules/videoio/CMakeFiles/opencv_videoio.dir/all] Error 2
make: *** [all] Error 2
```
##### Steps to reproduce
- Use this [Dockerfile](https://github.com/OpenVisualCloud/Dockerfiles/blob/master/Xeon/ubuntu-20.04/analytics/dev/Dockerfile)
- Make sure Dockerfile is using 4.5.2 (ARG OPENCV_REPO=https://github.com/opencv/opencv/archive/4.5.2.tar.gz)
- The build fails when re-building opencv 4.5.2 with FFmpeg 4.4
##### Issue submission checklist
- [ ] I report the issue, it's not a question
<!--
OpenCV team works with forum.opencv.org, Stack Overflow and other communities
to discuss problems. Tickets with question without real issue statement will be
closed.
-->
- [ ] I checked the problem with documentation, FAQ, open issues,
forum.opencv.org, Stack Overflow, etc and have not found solution
<!--
Places to check:
* OpenCV documentation: https://docs.opencv.org
* FAQ page: https://github.com/opencv/opencv/wiki/FAQ
* OpenCV forum: https://forum.opencv.org
* OpenCV issue tracker: https://github.com/opencv/opencv/issues?q=is%3Aissue
* Stack Overflow branch: https://stackoverflow.com/questions/tagged/opencv
-->
- [ ] I updated to latest OpenCV version and the issue is still there
<!--
master branch for OpenCV 4.x and 3.4 branch for OpenCV 3.x releases.
OpenCV team supports only latest release for each branch.
The ticket is closed, if the problem is not reproduced with modern version.
-->
- [ ] There is reproducer code and related data files: videos, images, onnx, etc
<!--
The best reproducer -- test case for OpenCV that we can add to the library.
Recommendations for media files and binary files:
* Try to reproduce the issue with images and videos in opencv_extra repository
to reduce attachment size
* Use PNG for images, if you report some CV related bug, but not image reader
issue
* Attach the image as archive to the ticket, if you report some reader issue.
Image hosting services compress images and it breaks the repro code.
* Provide ONNX file for some public model or ONNX file with with random weights,
if you report ONNX parsing or handling issue. Architecture details diagram
from netron tool can be very useful too. See https://lutzroeder.github.io/netron/
-->
|
1.0
|
OpenCV 4.5.2 videoio not compatible with ffmpeg 4.4 - <!--
If you have a question rather than reporting a bug please go to https://forum.opencv.org where you get much faster responses.
If you need further assistance please read [How To Contribute](https://github.com/opencv/opencv/wiki/How_to_contribute).
This is a template helping you to create an issue which can be processed as quickly as possible. This is the bug reporting section for the OpenCV library.
-->
##### System information (version)
- OpenCV => 4.5.2
- Operating System / Platform => Linux
- Compiler => gcc 4.8.5
Build fails when building opencv 4.5.2 with FFmpeg 4.4:
```
In file included from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg_impl.hpp:100:0,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:50:
/opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg_hw.hpp: In constructor 'HWAccelIterator::HWAccelIterator(cv::VideoAccelerationType, bool, AVDictionary*)':
/opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg_hw.hpp:501:23: error: use of deleted function 'std::basic_istringstream<char>& std::basic_istringstream<char>::operator=(const std::basic_istringstream<char>&)'
s_stream_ = std::istringstream(accel_list);
^
In file included from /usr/include/c++/4.8.2/complex:45:0,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core/cvstd.inl.hpp:47,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core.hpp:3306,
from /opt/build/opencv-4.5.2/modules/videoio/include/opencv2/videoio.hpp:46,
from /opt/build/opencv-4.5.2/modules/videoio/src/precomp.hpp:57,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:42:
/usr/include/c++/4.8.2/sstream:272:11: note: 'std::basic_istringstream<char>& std::basic_istringstream<char>::operator=(const std::basic_istringstream<char>&)' is implicitly deleted because the default definition would be ill-formed:
class basic_istringstream : public basic_istream<_CharT, _Traits>
^
/usr/include/c++/4.8.2/sstream:272:11: error: use of deleted function 'std::basic_istream<char>& std::basic_istream<char>::operator=(const std::basic_istream<char>&)'
In file included from /usr/include/c++/4.8.2/sstream:38:0,
from /usr/include/c++/4.8.2/complex:45,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core/cvstd.inl.hpp:47,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core.hpp:3306,
from /opt/build/opencv-4.5.2/modules/videoio/include/opencv2/videoio.hpp:46,
from /opt/build/opencv-4.5.2/modules/videoio/src/precomp.hpp:57,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:42:
/usr/include/c++/4.8.2/istream:58:11: note: 'std::basic_istream<char>& std::basic_istream<char>::operator=(const std::basic_istream<char>&)' is implicitly deleted because the default definition would be ill-formed:
class basic_istream : virtual public basic_ios<_CharT, _Traits>
^
/usr/include/c++/4.8.2/istream:58:11: error: use of deleted function 'std::basic_ios<char>& std::basic_ios<char>::operator=(const std::basic_ios<char>&)'
In file included from /usr/include/c++/4.8.2/ios:44:0,
from /usr/include/c++/4.8.2/istream:38,
from /usr/include/c++/4.8.2/sstream:38,
from /usr/include/c++/4.8.2/complex:45,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core/cvstd.inl.hpp:47,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core.hpp:3306,
from /opt/build/opencv-4.5.2/modules/videoio/include/opencv2/videoio.hpp:46,
from /opt/build/opencv-4.5.2/modules/videoio/src/precomp.hpp:57,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:42:
/usr/include/c++/4.8.2/bits/basic_ios.h:66:11: note: 'std::basic_ios<char>& std::basic_ios<char>::operator=(const std::basic_ios<char>&)' is implicitly deleted because the default definition would be ill-formed:
class basic_ios : public ios_base
^
In file included from /usr/include/c++/4.8.2/ios:42:0,
from /usr/include/c++/4.8.2/istream:38,
from /usr/include/c++/4.8.2/sstream:38,
from /usr/include/c++/4.8.2/complex:45,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core/cvstd.inl.hpp:47,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core.hpp:3306,
from /opt/build/opencv-4.5.2/modules/videoio/include/opencv2/videoio.hpp:46,
from /opt/build/opencv-4.5.2/modules/videoio/src/precomp.hpp:57,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:42:
/usr/include/c++/4.8.2/bits/ios_base.h:789:5: error: 'std::ios_base& std::ios_base::operator=(const std::ios_base&)' is private
operator=(const ios_base&);
^
In file included from /usr/include/c++/4.8.2/ios:44:0,
from /usr/include/c++/4.8.2/istream:38,
from /usr/include/c++/4.8.2/sstream:38,
from /usr/include/c++/4.8.2/complex:45,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core/cvstd.inl.hpp:47,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core.hpp:3306,
from /opt/build/opencv-4.5.2/modules/videoio/include/opencv2/videoio.hpp:46,
from /opt/build/opencv-4.5.2/modules/videoio/src/precomp.hpp:57,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:42:
/usr/include/c++/4.8.2/bits/basic_ios.h:66:11: error: within this context
class basic_ios : public ios_base
^
In file included from /usr/include/c++/4.8.2/complex:45:0,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core/cvstd.inl.hpp:47,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core.hpp:3306,
from /opt/build/opencv-4.5.2/modules/videoio/include/opencv2/videoio.hpp:46,
from /opt/build/opencv-4.5.2/modules/videoio/src/precomp.hpp:57,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:42:
/usr/include/c++/4.8.2/sstream:272:11: error: use of deleted function 'std::basic_stringbuf<char>& std::basic_stringbuf<char>::operator=(const std::basic_stringbuf<char>&)'
class basic_istringstream : public basic_istream<_CharT, _Traits>
^
/usr/include/c++/4.8.2/sstream:64:11: note: 'std::basic_stringbuf<char>& std::basic_stringbuf<char>::operator=(const std::basic_stringbuf<char>&)' is implicitly deleted because the default definition would be ill-formed:
class basic_stringbuf : public basic_streambuf<_CharT, _Traits>
^
In file included from /usr/include/c++/4.8.2/ios:43:0,
from /usr/include/c++/4.8.2/istream:38,
from /usr/include/c++/4.8.2/sstream:38,
from /usr/include/c++/4.8.2/complex:45,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core/cvstd.inl.hpp:47,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core.hpp:3306,
from /opt/build/opencv-4.5.2/modules/videoio/include/opencv2/videoio.hpp:46,
from /opt/build/opencv-4.5.2/modules/videoio/src/precomp.hpp:57,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:42:
/usr/include/c++/4.8.2/streambuf:810:7: error: 'std::basic_streambuf<_CharT, _Traits>& std::basic_streambuf<_CharT, _Traits>::operator=(const std::basic_streambuf<_CharT, _Traits>&) [with _CharT = char; _Traits = std::char_traits<char>]' is private
operator=(const basic_streambuf& __sb)
^
In file included from /usr/include/c++/4.8.2/complex:45:0,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core/cvstd.inl.hpp:47,
from /opt/build/opencv-4.5.2/modules/core/include/opencv2/core.hpp:3306,
from /opt/build/opencv-4.5.2/modules/videoio/include/opencv2/videoio.hpp:46,
from /opt/build/opencv-4.5.2/modules/videoio/src/precomp.hpp:57,
from /opt/build/opencv-4.5.2/modules/videoio/src/cap_ffmpeg.cpp:42:
/usr/include/c++/4.8.2/sstream:64:11: error: within this context
class basic_stringbuf : public basic_streambuf<_CharT, _Traits>
^
make[2]: *** [modules/videoio/CMakeFiles/opencv_videoio.dir/src/cap_ffmpeg.cpp.o] Error 1
make[2]: *** Waiting for unfinished jobs....
make[1]: *** [modules/videoio/CMakeFiles/opencv_videoio.dir/all] Error 2
make: *** [all] Error 2
```
##### Steps to reproduce
- Use this [Dockerfile](https://github.com/OpenVisualCloud/Dockerfiles/blob/master/Xeon/ubuntu-20.04/analytics/dev/Dockerfile)
- Make sure Dockerfile is using 4.5.2 (ARG OPENCV_REPO=https://github.com/opencv/opencv/archive/4.5.2.tar.gz)
- The build fails when re-building opencv 4.5.2 with FFmpeg 4.4
##### Issue submission checklist
- [ ] I report the issue, it's not a question
<!--
OpenCV team works with forum.opencv.org, Stack Overflow and other communities
to discuss problems. Tickets with question without real issue statement will be
closed.
-->
- [ ] I checked the problem with documentation, FAQ, open issues,
forum.opencv.org, Stack Overflow, etc and have not found solution
<!--
Places to check:
* OpenCV documentation: https://docs.opencv.org
* FAQ page: https://github.com/opencv/opencv/wiki/FAQ
* OpenCV forum: https://forum.opencv.org
* OpenCV issue tracker: https://github.com/opencv/opencv/issues?q=is%3Aissue
* Stack Overflow branch: https://stackoverflow.com/questions/tagged/opencv
-->
- [ ] I updated to latest OpenCV version and the issue is still there
<!--
master branch for OpenCV 4.x and 3.4 branch for OpenCV 3.x releases.
OpenCV team supports only latest release for each branch.
The ticket is closed, if the problem is not reproduced with modern version.
-->
- [ ] There is reproducer code and related data files: videos, images, onnx, etc
<!--
The best reproducer -- test case for OpenCV that we can add to the library.
Recommendations for media files and binary files:
* Try to reproduce the issue with images and videos in opencv_extra repository
to reduce attachment size
* Use PNG for images, if you report some CV related bug, but not image reader
issue
* Attach the image as archive to the ticket, if you report some reader issue.
Image hosting services compress images and it breaks the repro code.
* Provide ONNX file for some public model or ONNX file with with random weights,
if you report ONNX parsing or handling issue. Architecture details diagram
from netron tool can be very useful too. See https://lutzroeder.github.io/netron/
-->
|
non_process
|
opencv videoio not compatible with ffmpeg if you have a question rather than reporting a bug please go to where you get much faster responses if you need further assistance please read this is a template helping you to create an issue which can be processed as quickly as possible this is the bug reporting section for the opencv library system information version opencv operating system platform linux compiler gcc build fails when building opencv with ffmpeg in file included from opt build opencv modules videoio src cap ffmpeg impl hpp from opt build opencv modules videoio src cap ffmpeg cpp opt build opencv modules videoio src cap ffmpeg hw hpp in constructor hwacceliterator hwacceliterator cv videoaccelerationtype bool avdictionary opt build opencv modules videoio src cap ffmpeg hw hpp error use of deleted function std basic istringstream std basic istringstream operator const std basic istringstream s stream std istringstream accel list in file included from usr include c complex from opt build opencv modules core include core cvstd inl hpp from opt build opencv modules core include core hpp from opt build opencv modules videoio include videoio hpp from opt build opencv modules videoio src precomp hpp from opt build opencv modules videoio src cap ffmpeg cpp usr include c sstream note std basic istringstream std basic istringstream operator const std basic istringstream is implicitly deleted because the default definition would be ill formed class basic istringstream public basic istream usr include c sstream error use of deleted function std basic istream std basic istream operator const std basic istream in file included from usr include c sstream from usr include c complex from opt build opencv modules core include core cvstd inl hpp from opt build opencv modules core include core hpp from opt build opencv modules videoio include videoio hpp from opt build opencv modules videoio src precomp hpp from opt build opencv modules videoio src cap ffmpeg cpp usr include c istream note std basic istream std basic istream operator const std basic istream is implicitly deleted because the default definition would be ill formed class basic istream virtual public basic ios usr include c istream error use of deleted function std basic ios std basic ios operator const std basic ios in file included from usr include c ios from usr include c istream from usr include c sstream from usr include c complex from opt build opencv modules core include core cvstd inl hpp from opt build opencv modules core include core hpp from opt build opencv modules videoio include videoio hpp from opt build opencv modules videoio src precomp hpp from opt build opencv modules videoio src cap ffmpeg cpp usr include c bits basic ios h note std basic ios std basic ios operator const std basic ios is implicitly deleted because the default definition would be ill formed class basic ios public ios base in file included from usr include c ios from usr include c istream from usr include c sstream from usr include c complex from opt build opencv modules core include core cvstd inl hpp from opt build opencv modules core include core hpp from opt build opencv modules videoio include videoio hpp from opt build opencv modules videoio src precomp hpp from opt build opencv modules videoio src cap ffmpeg cpp usr include c bits ios base h error std ios base std ios base operator const std ios base is private operator const ios base in file included from usr include c ios from usr include c istream from usr include c sstream from usr include c complex from opt build opencv modules core include core cvstd inl hpp from opt build opencv modules core include core hpp from opt build opencv modules videoio include videoio hpp from opt build opencv modules videoio src precomp hpp from opt build opencv modules videoio src cap ffmpeg cpp usr include c bits basic ios h error within this context class basic ios public ios base in file included from usr include c complex from opt build opencv modules core include core cvstd inl hpp from opt build opencv modules core include core hpp from opt build opencv modules videoio include videoio hpp from opt build opencv modules videoio src precomp hpp from opt build opencv modules videoio src cap ffmpeg cpp usr include c sstream error use of deleted function std basic stringbuf std basic stringbuf operator const std basic stringbuf class basic istringstream public basic istream usr include c sstream note std basic stringbuf std basic stringbuf operator const std basic stringbuf is implicitly deleted because the default definition would be ill formed class basic stringbuf public basic streambuf in file included from usr include c ios from usr include c istream from usr include c sstream from usr include c complex from opt build opencv modules core include core cvstd inl hpp from opt build opencv modules core include core hpp from opt build opencv modules videoio include videoio hpp from opt build opencv modules videoio src precomp hpp from opt build opencv modules videoio src cap ffmpeg cpp usr include c streambuf error std basic streambuf std basic streambuf operator const std basic streambuf is private operator const basic streambuf sb in file included from usr include c complex from opt build opencv modules core include core cvstd inl hpp from opt build opencv modules core include core hpp from opt build opencv modules videoio include videoio hpp from opt build opencv modules videoio src precomp hpp from opt build opencv modules videoio src cap ffmpeg cpp usr include c sstream error within this context class basic stringbuf public basic streambuf make error make waiting for unfinished jobs make error make error steps to reproduce use this make sure dockerfile is using arg opencv repo the build fails when re building opencv with ffmpeg issue submission checklist i report the issue it s not a question opencv team works with forum opencv org stack overflow and other communities to discuss problems tickets with question without real issue statement will be closed i checked the problem with documentation faq open issues forum opencv org stack overflow etc and have not found solution places to check opencv documentation faq page opencv forum opencv issue tracker stack overflow branch i updated to latest opencv version and the issue is still there master branch for opencv x and branch for opencv x releases opencv team supports only latest release for each branch the ticket is closed if the problem is not reproduced with modern version there is reproducer code and related data files videos images onnx etc the best reproducer test case for opencv that we can add to the library recommendations for media files and binary files try to reproduce the issue with images and videos in opencv extra repository to reduce attachment size use png for images if you report some cv related bug but not image reader issue attach the image as archive to the ticket if you report some reader issue image hosting services compress images and it breaks the repro code provide onnx file for some public model or onnx file with with random weights if you report onnx parsing or handling issue architecture details diagram from netron tool can be very useful too see
| 0
|
6,259
| 9,218,158,630
|
IssuesEvent
|
2019-03-11 12:43:05
|
Open-EO/openeo-api
|
https://api.github.com/repos/Open-EO/openeo-api
|
opened
|
parameters as array to remove parameter_order
|
process discovery
|
Currently, the parameters are stored as object. As JSON objects have no order, we added parameter_order for programming languages without names parameters. parameter_order on the other hand is only required for 2+ parameters and needs special handling in many implementations. Therefore the idea came up to remove parameter_order and make the parameters object an array. The key could be moved to the object in the array.
## Example:
Before:
```json
{
"id": "product",
"summary": "Multiplication of a sequence of numbers",
"description": "This process is an exact alias for the `multiply` process. See ``multiply()`` for more information.",
"categories": [
"math",
"reducer"
],
"parameter_order": [
"data",
"ignore_nodata"
],
"parameters": {
"data": {
"description": "See ``multiply()`` for more information.",
"schema": {
"type": "array",
"items": {
"type": [
"number",
"null"
]
}
},
"required": true
},
"ignore_nodata": {
"description": "See ``multiply()`` for more information.",
"schema": {
"type": "boolean",
"default": true
}
}
},
"returns": {
"description": "See ``multiply()`` for more information.",
"schema": {
"type": [
"number",
"null"
]
}
}
}
```
After:
```json
{
"id": "product",
"summary": "Multiplication of a sequence of numbers",
"description": "This process is an exact alias for the `multiply` process. See ``multiply()`` for more information.",
"categories": [
"math",
"reducer"
],
"parameters": [
{
"name": "data",
"description": "See ``multiply()`` for more information.",
"schema": {
"type": "array",
"items": {
"type": [
"number",
"null"
]
}
},
"required": true
},
{
"name": "ignore_nodata",
"description": "See ``multiply()`` for more information.",
"schema": {
"type": "boolean",
"default": true
}
}
],
"returns": {
"description": "See ``multiply()`` for more information.",
"schema": {
"type": [
"number",
"null"
]
}
}
}
```
|
1.0
|
parameters as array to remove parameter_order - Currently, the parameters are stored as object. As JSON objects have no order, we added parameter_order for programming languages without names parameters. parameter_order on the other hand is only required for 2+ parameters and needs special handling in many implementations. Therefore the idea came up to remove parameter_order and make the parameters object an array. The key could be moved to the object in the array.
## Example:
Before:
```json
{
"id": "product",
"summary": "Multiplication of a sequence of numbers",
"description": "This process is an exact alias for the `multiply` process. See ``multiply()`` for more information.",
"categories": [
"math",
"reducer"
],
"parameter_order": [
"data",
"ignore_nodata"
],
"parameters": {
"data": {
"description": "See ``multiply()`` for more information.",
"schema": {
"type": "array",
"items": {
"type": [
"number",
"null"
]
}
},
"required": true
},
"ignore_nodata": {
"description": "See ``multiply()`` for more information.",
"schema": {
"type": "boolean",
"default": true
}
}
},
"returns": {
"description": "See ``multiply()`` for more information.",
"schema": {
"type": [
"number",
"null"
]
}
}
}
```
After:
```json
{
"id": "product",
"summary": "Multiplication of a sequence of numbers",
"description": "This process is an exact alias for the `multiply` process. See ``multiply()`` for more information.",
"categories": [
"math",
"reducer"
],
"parameters": [
{
"name": "data",
"description": "See ``multiply()`` for more information.",
"schema": {
"type": "array",
"items": {
"type": [
"number",
"null"
]
}
},
"required": true
},
{
"name": "ignore_nodata",
"description": "See ``multiply()`` for more information.",
"schema": {
"type": "boolean",
"default": true
}
}
],
"returns": {
"description": "See ``multiply()`` for more information.",
"schema": {
"type": [
"number",
"null"
]
}
}
}
```
|
process
|
parameters as array to remove parameter order currently the parameters are stored as object as json objects have no order we added parameter order for programming languages without names parameters parameter order on the other hand is only required for parameters and needs special handling in many implementations therefore the idea came up to remove parameter order and make the parameters object an array the key could be moved to the object in the array example before json id product summary multiplication of a sequence of numbers description this process is an exact alias for the multiply process see multiply for more information categories math reducer parameter order data ignore nodata parameters data description see multiply for more information schema type array items type number null required true ignore nodata description see multiply for more information schema type boolean default true returns description see multiply for more information schema type number null after json id product summary multiplication of a sequence of numbers description this process is an exact alias for the multiply process see multiply for more information categories math reducer parameters name data description see multiply for more information schema type array items type number null required true name ignore nodata description see multiply for more information schema type boolean default true returns description see multiply for more information schema type number null
| 1
|
166,516
| 6,305,932,550
|
IssuesEvent
|
2017-07-21 19:38:27
|
fedarko/MetagenomeScope
|
https://api.github.com/repos/fedarko/MetagenomeScope
|
closed
|
Add AGP file export functionality for finished paths
|
highpriorityfeature
|
This has been on the docket for a while. Gonna take care of it soon.
See [the AGP specification](https://www.ncbi.nlm.nih.gov/assembly/agp/AGP_Specification/) for details. I guess we'd export node labels for GML input files and node IDs for other input files.
The main thing I'm unsure about is whether or not to prefix node IDs with anything (e.g. `NODE_-5` instead of just `-5`). Will look around a bit to find out.
|
1.0
|
Add AGP file export functionality for finished paths - This has been on the docket for a while. Gonna take care of it soon.
See [the AGP specification](https://www.ncbi.nlm.nih.gov/assembly/agp/AGP_Specification/) for details. I guess we'd export node labels for GML input files and node IDs for other input files.
The main thing I'm unsure about is whether or not to prefix node IDs with anything (e.g. `NODE_-5` instead of just `-5`). Will look around a bit to find out.
|
non_process
|
add agp file export functionality for finished paths this has been on the docket for a while gonna take care of it soon see for details i guess we d export node labels for gml input files and node ids for other input files the main thing i m unsure about is whether or not to prefix node ids with anything e g node instead of just will look around a bit to find out
| 0
|
268,401
| 8,406,533,362
|
IssuesEvent
|
2018-10-11 18:13:15
|
GCE-NEIIST/GCE-NEIIST-webapp
|
https://api.github.com/repos/GCE-NEIIST/GCE-NEIIST-webapp
|
closed
|
GCE-Thesis: adding semester field
|
Category: Coding Priority: High
|
A student might want to navigate and check older master theses. Adding a field to the entity thesis called "Semester" might help us differentiate between them.
Example:
title: "A title",
...
semester:"2017-2018-1S"
Given this differentiation, we can check for trends and provide a more valuable service to students.
|
1.0
|
GCE-Thesis: adding semester field - A student might want to navigate and check older master theses. Adding a field to the entity thesis called "Semester" might help us differentiate between them.
Example:
title: "A title",
...
semester:"2017-2018-1S"
Given this differentiation, we can check for trends and provide a more valuable service to students.
|
non_process
|
gce thesis adding semester field a student might want to navigate and check older master theses adding a field to the entity thesis called semester might help us differentiate between them example title a title semester given this differentiation we can check for trends and provide a more valuable service to students
| 0
|
8,065
| 11,235,819,848
|
IssuesEvent
|
2020-01-09 09:12:28
|
Altinn/altinn-studio
|
https://api.github.com/repos/Altinn/altinn-studio
|
closed
|
Create Custom PEP for Process API
|
area/authorization area/process kind/user-story solution/app-backend team/steam
|
## Description
As decided in #2521 we need to use the current action step for a current task to decide if a user is authorized to move a process forward.
A standard PEP (Created in #2554) will not be able to do that so we would need to either add this create this generic in pep or create custom code for it
## Considerations
- We probably need to define what is the task action in some config so that in signing step, signing is the
Input (beyond tasks) on how the user story should be solved can be put here.
## Acceptance criteria
- PEP is able to identify the correct action based on the task type / task configuration
## Tasks
- [x] Decide where to store the action
- [x] Implement PEP for process API that identifies the correct action based on the above configuration
|
1.0
|
Create Custom PEP for Process API - ## Description
As decided in #2521 we need to use the current action step for a current task to decide if a user is authorized to move a process forward.
A standard PEP (Created in #2554) will not be able to do that so we would need to either add this create this generic in pep or create custom code for it
## Considerations
- We probably need to define what is the task action in some config so that in signing step, signing is the
Input (beyond tasks) on how the user story should be solved can be put here.
## Acceptance criteria
- PEP is able to identify the correct action based on the task type / task configuration
## Tasks
- [x] Decide where to store the action
- [x] Implement PEP for process API that identifies the correct action based on the above configuration
|
process
|
create custom pep for process api description as decided in we need to use the current action step for a current task to decide if a user is authorized to move a process forward a standard pep created in will not be able to do that so we would need to either add this create this generic in pep or create custom code for it considerations we probably need to define what is the task action in some config so that in signing step signing is the input beyond tasks on how the user story should be solved can be put here acceptance criteria pep is able to identify the correct action based on the task type task configuration tasks decide where to store the action implement pep for process api that identifies the correct action based on the above configuration
| 1
|
5,484
| 8,358,679,619
|
IssuesEvent
|
2018-10-03 04:30:05
|
googleapis/nodejs-spanner
|
https://api.github.com/repos/googleapis/nodejs-spanner
|
closed
|
deep-extend@0.4.2 security vulnerability
|
status: blocked type: process
|
#### Environment details
- OS: macOS High Sierra v10.13.3
- Node.js version: v8.9.4
- npm version: v5.6.0
- `@google-cloud/spanner` version: 1.5.0
#### Steps to reproduce
https://nodesecurity.io/advisories/612
path to package:
`@google-cloud/spanner@1.5.0 > google-gax@0.16.1 > grpc@1.12.4 > node-pre-gyp@0.10.0 > rc@1.2.6 > deep-extend@0.4.2`
#### Suggested fix
- upgrade to deep-extend@0.5.1 (rc@1.2.8 seems to have upgraded deep-extend to 0.6.0)
- use nsp to proactively stub security vulnerabilities out
|
1.0
|
deep-extend@0.4.2 security vulnerability - #### Environment details
- OS: macOS High Sierra v10.13.3
- Node.js version: v8.9.4
- npm version: v5.6.0
- `@google-cloud/spanner` version: 1.5.0
#### Steps to reproduce
https://nodesecurity.io/advisories/612
path to package:
`@google-cloud/spanner@1.5.0 > google-gax@0.16.1 > grpc@1.12.4 > node-pre-gyp@0.10.0 > rc@1.2.6 > deep-extend@0.4.2`
#### Suggested fix
- upgrade to deep-extend@0.5.1 (rc@1.2.8 seems to have upgraded deep-extend to 0.6.0)
- use nsp to proactively stub security vulnerabilities out
|
process
|
deep extend security vulnerability environment details os macos high sierra node js version npm version google cloud spanner version steps to reproduce path to package google cloud spanner google gax grpc node pre gyp rc deep extend suggested fix upgrade to deep extend rc seems to have upgraded deep extend to use nsp to proactively stub security vulnerabilities out
| 1
|
765,287
| 26,840,623,472
|
IssuesEvent
|
2023-02-03 00:00:11
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
closed
|
gRPC python (version 1.51.1) lead to exit django process with exit code 1
|
kind/bug lang/Python priority/P2 disposition/requires reporter action
|
<!--
PLEASE DO NOT POST A QUESTION HERE.
This form is for bug reports and feature requests ONLY!
For general questions and troubleshooting, please ask/look for answers at StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc
For questions that specifically need to be answered by gRPC team members, please ask/look for answers at grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io
Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new)
-->
### What version of gRPC and what language are you using?
I'm using versions 1.51.1 and 1.46.3
### What operating system (Linux, Windows,...) and version?
windows and linux
### What runtime / compiler are you using (e.g. python version or version of gcc)
python 11.1 and python 10.9.1
### What did you do?
I have used grpcio version 1.46.3 which have worked correctly without any bud in django, but after I upgraded my grpcio to 1.51.1 it's ran out from django by exit code -1.
I found this error came from insecure_channel function in grpc module
### What did you expect to see?
### What did you see instead?
Make sure you include information that can help us debug (full error message, exception listing, stack trace, logs).
See [TROUBLESHOOTING.md](https://github.com/grpc/grpc/blob/master/TROUBLESHOOTING.md) for how to diagnose problems better.
### Anything else we should know about your project / environment?
|
1.0
|
gRPC python (version 1.51.1) lead to exit django process with exit code 1 - <!--
PLEASE DO NOT POST A QUESTION HERE.
This form is for bug reports and feature requests ONLY!
For general questions and troubleshooting, please ask/look for answers at StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc
For questions that specifically need to be answered by gRPC team members, please ask/look for answers at grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io
Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new)
-->
### What version of gRPC and what language are you using?
I'm using versions 1.51.1 and 1.46.3
### What operating system (Linux, Windows,...) and version?
windows and linux
### What runtime / compiler are you using (e.g. python version or version of gcc)
python 11.1 and python 10.9.1
### What did you do?
I have used grpcio version 1.46.3 which have worked correctly without any bud in django, but after I upgraded my grpcio to 1.51.1 it's ran out from django by exit code -1.
I found this error came from insecure_channel function in grpc module
### What did you expect to see?
### What did you see instead?
Make sure you include information that can help us debug (full error message, exception listing, stack trace, logs).
See [TROUBLESHOOTING.md](https://github.com/grpc/grpc/blob/master/TROUBLESHOOTING.md) for how to diagnose problems better.
### Anything else we should know about your project / environment?
|
non_process
|
grpc python version lead to exit django process with exit code please do not post a question here this form is for bug reports and feature requests only for general questions and troubleshooting please ask look for answers at stackoverflow with grpc tag for questions that specifically need to be answered by grpc team members please ask look for answers at grpc io mailing list issues specific to grpc java grpc go grpc node grpc dart grpc web should be created in the repository they belong to e g what version of grpc and what language are you using i m using versions and what operating system linux windows and version windows and linux what runtime compiler are you using e g python version or version of gcc python and python what did you do i have used grpcio version which have worked correctly without any bud in django but after i upgraded my grpcio to it s ran out from django by exit code i found this error came from insecure channel function in grpc module what did you expect to see what did you see instead make sure you include information that can help us debug full error message exception listing stack trace logs see for how to diagnose problems better anything else we should know about your project environment
| 0
|
19,130
| 25,184,385,065
|
IssuesEvent
|
2022-11-11 16:29:04
|
bradmartin333/batter
|
https://api.github.com/repos/bradmartin333/batter
|
closed
|
focus score
|
processing
|
- [ ] rewrite RedFocus algorithm
- [ ] add score to hue display
- [ ] add diagnostic into to ROI render
|
1.0
|
focus score - - [ ] rewrite RedFocus algorithm
- [ ] add score to hue display
- [ ] add diagnostic into to ROI render
|
process
|
focus score rewrite redfocus algorithm add score to hue display add diagnostic into to roi render
| 1
|
505,354
| 14,632,033,476
|
IssuesEvent
|
2020-12-23 21:18:48
|
cBioPortal/datahub
|
https://api.github.com/repos/cBioPortal/datahub
|
reopened
|
TRACERx - McGranahan et al. Clonal neoantigens elicit T cell immunoreactivity and sensitivity to immune checkpoint blockade. Science. 2016.
|
immunotherapy missing data/access limitation new public study not curatable priority wontfix
|
https://www.ncbi.nlm.nih.gov/pubmed/26940869
- [ ] [create a issue on datahub](https://github.com/cBioPortal/datahub/issues/new) before curating a study (one issue per study) and copy this checklist to the issue tracker
- [ ] List information of the dataset/paper in the issue, e.g. pmid, paper link, suppl file link
- [ ] Document the curation process, e.g. how and by whom the data was transformed
- [ ] Follow [the data checklist](https://github.com/cBioPortal/datahub/issues/248)
- [ ] Create a pull request to datahub once the data is curated
- [ ] Push to triage portal
- [ ] Import into msk and public portal database
- [ ] Update [cBioPortal news](https://github.com/cBioPortal/cbioportal/blob/master/docs/News.md)
|
1.0
|
TRACERx - McGranahan et al. Clonal neoantigens elicit T cell immunoreactivity and sensitivity to immune checkpoint blockade. Science. 2016. - https://www.ncbi.nlm.nih.gov/pubmed/26940869
- [ ] [create a issue on datahub](https://github.com/cBioPortal/datahub/issues/new) before curating a study (one issue per study) and copy this checklist to the issue tracker
- [ ] List information of the dataset/paper in the issue, e.g. pmid, paper link, suppl file link
- [ ] Document the curation process, e.g. how and by whom the data was transformed
- [ ] Follow [the data checklist](https://github.com/cBioPortal/datahub/issues/248)
- [ ] Create a pull request to datahub once the data is curated
- [ ] Push to triage portal
- [ ] Import into msk and public portal database
- [ ] Update [cBioPortal news](https://github.com/cBioPortal/cbioportal/blob/master/docs/News.md)
|
non_process
|
tracerx mcgranahan et al clonal neoantigens elicit t cell immunoreactivity and sensitivity to immune checkpoint blockade science before curating a study one issue per study and copy this checklist to the issue tracker list information of the dataset paper in the issue e g pmid paper link suppl file link document the curation process e g how and by whom the data was transformed follow create a pull request to datahub once the data is curated push to triage portal import into msk and public portal database update
| 0
|
10,743
| 13,538,554,707
|
IssuesEvent
|
2020-09-16 12:16:46
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Implement a `findFirst` API
|
kind/feature process/candidate team/engines team/product team/typescript
|
## Problem
`findOne` is strict in that it requires a unique constraint (@unique, @id) to find a unique entry. That, however, doesn't allow for more flexible kinds of queries which should also usually return a single result.
## Suggested solution
`findFirst`, doing essentially a `findMany` + `take: 1` under the hood.
|
1.0
|
Implement a `findFirst` API - ## Problem
`findOne` is strict in that it requires a unique constraint (@unique, @id) to find a unique entry. That, however, doesn't allow for more flexible kinds of queries which should also usually return a single result.
## Suggested solution
`findFirst`, doing essentially a `findMany` + `take: 1` under the hood.
|
process
|
implement a findfirst api problem findone is strict in that it requires a unique constraint unique id to find a unique entry that however doesn t allow for more flexible kinds of queries which should also usually return a single result suggested solution findfirst doing essentially a findmany take under the hood
| 1
|
501,101
| 14,521,195,831
|
IssuesEvent
|
2020-12-14 06:57:28
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
accounts.firefox.com - design is broken
|
browser-fenix engine-gecko ml-needsdiagnosis-false ml-probability-high priority-normal
|
<!-- @browser: Firefox Mobile 84.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:84.0) Gecko/84.0 Firefox/84.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/63580 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://accounts.firefox.com/oauth/?action=email&response_type=code&entrypoint=android-components&client_id=a2270f727f45f648&scope=profile%2Bhttps%3A%2F%2Fidentity.mozilla.com%2Fapps%2Foldsync%2Bhttps%3A%2F%2Fidentity.mozilla.com%2Ftokens%2Fsession&state=cUbcmARyyDN0sRIyCH4x5w&code_challenge_method=S256&code_challenge=g626zmwt30CY37agnuTWcIqLaQjeYtDuQ5l15CNAIQg&access_type=offline&keys_jwk=eyJrdHkiOiJFQyIsImNydiI6IlAtMjU2IiwieCI6IkVNRDZnSk8zZERqMkV6T3AyUzFmWWFyd1NTejhObDBXbVBSMk1rX3AteGsiLCJ5IjoiUGJySmpkTlFTeWVmMXhyeTFUTU1yQXFHT1I5cTRScWJfSGdoWTdLT0RkNCJ9&context=oauth_webchannel_v1
**Browser / Version**: Firefox Mobile 84.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Design is broken
**Description**: Items not fully visible
**Steps to Reproduce**:
Need to zoom in and zoom out to see forms
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/12/a7d5b28e-029e-4d2e-bce3-437acc4e10f8.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201206192040</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/12/a7f14ebe-78a8-4ad3-94f2-4b67e11caec9)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
accounts.firefox.com - design is broken - <!-- @browser: Firefox Mobile 84.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:84.0) Gecko/84.0 Firefox/84.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/63580 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://accounts.firefox.com/oauth/?action=email&response_type=code&entrypoint=android-components&client_id=a2270f727f45f648&scope=profile%2Bhttps%3A%2F%2Fidentity.mozilla.com%2Fapps%2Foldsync%2Bhttps%3A%2F%2Fidentity.mozilla.com%2Ftokens%2Fsession&state=cUbcmARyyDN0sRIyCH4x5w&code_challenge_method=S256&code_challenge=g626zmwt30CY37agnuTWcIqLaQjeYtDuQ5l15CNAIQg&access_type=offline&keys_jwk=eyJrdHkiOiJFQyIsImNydiI6IlAtMjU2IiwieCI6IkVNRDZnSk8zZERqMkV6T3AyUzFmWWFyd1NTejhObDBXbVBSMk1rX3AteGsiLCJ5IjoiUGJySmpkTlFTeWVmMXhyeTFUTU1yQXFHT1I5cTRScWJfSGdoWTdLT0RkNCJ9&context=oauth_webchannel_v1
**Browser / Version**: Firefox Mobile 84.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Design is broken
**Description**: Items not fully visible
**Steps to Reproduce**:
Need to zoom in and zoom out to see forms
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/12/a7d5b28e-029e-4d2e-bce3-437acc4e10f8.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201206192040</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/12/a7f14ebe-78a8-4ad3-94f2-4b67e11caec9)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
accounts firefox com design is broken url browser version firefox mobile operating system android tested another browser no problem type design is broken description items not fully visible steps to reproduce need to zoom in and zoom out to see forms view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
299,689
| 9,205,762,516
|
IssuesEvent
|
2019-03-08 11:34:15
|
qissue-bot/QGIS
|
https://api.github.com/repos/qissue-bot/QGIS
|
closed
|
For PostGIS layers save style files in the database
|
Category: Vectors Component: Easy fix? Component: Pull Request or Patch supplied Component: Resolution Priority: Low Project: QGIS Application Status: Closed Tracker: Feature request
|
---
Author Name: **Redmine Admin** (Redmine Admin)
Original Redmine Issue: 1059, https://issues.qgis.org/issues/1059
Original Assignee: nobody -
---
It would be, I think, a good idea to have the option to store the style data in the database. I reckon the comment field/area would be a good place to store it because every table can have a comment and this would negate the need to add any special fields to the table.
|
1.0
|
For PostGIS layers save style files in the database - ---
Author Name: **Redmine Admin** (Redmine Admin)
Original Redmine Issue: 1059, https://issues.qgis.org/issues/1059
Original Assignee: nobody -
---
It would be, I think, a good idea to have the option to store the style data in the database. I reckon the comment field/area would be a good place to store it because every table can have a comment and this would negate the need to add any special fields to the table.
|
non_process
|
for postgis layers save style files in the database author name redmine admin redmine admin original redmine issue original assignee nobody it would be i think a good idea to have the option to store the style data in the database i reckon the comment field area would be a good place to store it because every table can have a comment and this would negate the need to add any special fields to the table
| 0
|
14,331
| 17,364,448,468
|
IssuesEvent
|
2021-07-30 04:16:55
|
googleapis/python-spanner
|
https://api.github.com/repos/googleapis/python-spanner
|
closed
|
samples.samples.quickstart_test: test_quickstart failed
|
api: spanner flakybot: flaky flakybot: issue samples type: process
|
This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 3132587453f7bd0be72ebc393626b5c8b1bab982
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/7d78f9ad-4b52-4472-9847-94be332f8559), [Sponge](http://sponge2/7d78f9ad-4b52-4472-9847-94be332f8559)
status: failed
<details><summary>Test output</summary><br><pre>args = (parent: "projects/python-docs-samples-tests/instances/test-instance"
create_statement: "CREATE DATABASE `my-database-id`"
,)
kwargs = {'metadata': [('google-cloud-resource-prefix', 'projects/python-docs-samples-tests/instances/test-instance/databases/m...-samples-tests/instances/test-instance'), ('x-goog-api-client', 'gl-python/3.6.13 grpc/1.38.1 gax/1.30.0 gccl/3.6.0')]}
@six.wraps(callable_)
def error_remapped_callable(*args, **kwargs):
try:
> return callable_(*args, **kwargs)
.nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:67:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <grpc._channel._UnaryUnaryMultiCallable object at 0x7f43237f3cc0>
request = parent: "projects/python-docs-samples-tests/instances/test-instance"
create_statement: "CREATE DATABASE `my-database-id`"
timeout = None
metadata = [('google-cloud-resource-prefix', 'projects/python-docs-samples-tests/instances/test-instance/databases/my-database-id...s-samples-tests/instances/test-instance'), ('x-goog-api-client', 'gl-python/3.6.13 grpc/1.38.1 gax/1.30.0 gccl/3.6.0')]
credentials = None, wait_for_ready = None, compression = None
def __call__(self,
request,
timeout=None,
metadata=None,
credentials=None,
wait_for_ready=None,
compression=None):
state, call, = self._blocking(request, timeout, metadata, credentials,
wait_for_ready, compression)
> return _end_unary_response_blocking(state, call, False, None)
.nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py:946:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
state = <grpc._channel._RPCState object at 0x7f43237f52e8>
call = <grpc._cython.cygrpc.SegregatedCall object at 0x7f43237a2f88>
with_call = False, deadline = None
def _end_unary_response_blocking(state, call, with_call, deadline):
if state.code is grpc.StatusCode.OK:
if with_call:
rendezvous = _MultiThreadedRendezvous(state, call, None, deadline)
return state.response, rendezvous
else:
return state.response
else:
> raise _InactiveRpcError(state)
E grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
E status = StatusCode.NOT_FOUND
E details = "Instance not found: projects/python-docs-samples-tests/instances/test-instance"
E debug_error_string = "{"created":"@1624700026.817055527","description":"Error received from peer ipv4:74.125.135.95:443","file":"src/core/lib/surface/call.cc","file_line":1066,"grpc_message":"Instance not found: projects/python-docs-samples-tests/instances/test-instance","grpc_status":5}"
E >
.nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py:849: _InactiveRpcError
The above exception was the direct cause of the following exception:
@pytest.fixture
def example_database():
spanner_client = spanner.Client()
instance = spanner_client.instance(SPANNER_INSTANCE)
database = instance.database("my-database-id")
if not database.exists():
> database.create()
quickstart_test.py:48:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
../../google/cloud/spanner_v1/database.py:374: in create
future = api.create_database(request=request, metadata=metadata)
../../google/cloud/spanner_admin_database_v1/services/database_admin/client.py:615: in create_database
response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,)
.nox/py-3-6/lib/python3.6/site-packages/google/api_core/gapic_v1/method.py:145: in __call__
return wrapped_func(*args, **kwargs)
.nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:69: in error_remapped_callable
six.raise_from(exceptions.from_grpc_error(exc), exc)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
value = None
from_value = <_InactiveRpcError of RPC that terminated with:
status = StatusCode.NOT_FOUND
details = "Instance not found: project...066,"grpc_message":"Instance not found: projects/python-docs-samples-tests/instances/test-instance","grpc_status":5}"
>
> ???
E google.api_core.exceptions.NotFound: 404 Instance not found: projects/python-docs-samples-tests/instances/test-instance
<string>:3: NotFound</pre></details>
|
1.0
|
samples.samples.quickstart_test: test_quickstart failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 3132587453f7bd0be72ebc393626b5c8b1bab982
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/7d78f9ad-4b52-4472-9847-94be332f8559), [Sponge](http://sponge2/7d78f9ad-4b52-4472-9847-94be332f8559)
status: failed
<details><summary>Test output</summary><br><pre>args = (parent: "projects/python-docs-samples-tests/instances/test-instance"
create_statement: "CREATE DATABASE `my-database-id`"
,)
kwargs = {'metadata': [('google-cloud-resource-prefix', 'projects/python-docs-samples-tests/instances/test-instance/databases/m...-samples-tests/instances/test-instance'), ('x-goog-api-client', 'gl-python/3.6.13 grpc/1.38.1 gax/1.30.0 gccl/3.6.0')]}
@six.wraps(callable_)
def error_remapped_callable(*args, **kwargs):
try:
> return callable_(*args, **kwargs)
.nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:67:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <grpc._channel._UnaryUnaryMultiCallable object at 0x7f43237f3cc0>
request = parent: "projects/python-docs-samples-tests/instances/test-instance"
create_statement: "CREATE DATABASE `my-database-id`"
timeout = None
metadata = [('google-cloud-resource-prefix', 'projects/python-docs-samples-tests/instances/test-instance/databases/my-database-id...s-samples-tests/instances/test-instance'), ('x-goog-api-client', 'gl-python/3.6.13 grpc/1.38.1 gax/1.30.0 gccl/3.6.0')]
credentials = None, wait_for_ready = None, compression = None
def __call__(self,
request,
timeout=None,
metadata=None,
credentials=None,
wait_for_ready=None,
compression=None):
state, call, = self._blocking(request, timeout, metadata, credentials,
wait_for_ready, compression)
> return _end_unary_response_blocking(state, call, False, None)
.nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py:946:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
state = <grpc._channel._RPCState object at 0x7f43237f52e8>
call = <grpc._cython.cygrpc.SegregatedCall object at 0x7f43237a2f88>
with_call = False, deadline = None
def _end_unary_response_blocking(state, call, with_call, deadline):
if state.code is grpc.StatusCode.OK:
if with_call:
rendezvous = _MultiThreadedRendezvous(state, call, None, deadline)
return state.response, rendezvous
else:
return state.response
else:
> raise _InactiveRpcError(state)
E grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
E status = StatusCode.NOT_FOUND
E details = "Instance not found: projects/python-docs-samples-tests/instances/test-instance"
E debug_error_string = "{"created":"@1624700026.817055527","description":"Error received from peer ipv4:74.125.135.95:443","file":"src/core/lib/surface/call.cc","file_line":1066,"grpc_message":"Instance not found: projects/python-docs-samples-tests/instances/test-instance","grpc_status":5}"
E >
.nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py:849: _InactiveRpcError
The above exception was the direct cause of the following exception:
@pytest.fixture
def example_database():
spanner_client = spanner.Client()
instance = spanner_client.instance(SPANNER_INSTANCE)
database = instance.database("my-database-id")
if not database.exists():
> database.create()
quickstart_test.py:48:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
../../google/cloud/spanner_v1/database.py:374: in create
future = api.create_database(request=request, metadata=metadata)
../../google/cloud/spanner_admin_database_v1/services/database_admin/client.py:615: in create_database
response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,)
.nox/py-3-6/lib/python3.6/site-packages/google/api_core/gapic_v1/method.py:145: in __call__
return wrapped_func(*args, **kwargs)
.nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:69: in error_remapped_callable
six.raise_from(exceptions.from_grpc_error(exc), exc)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
value = None
from_value = <_InactiveRpcError of RPC that terminated with:
status = StatusCode.NOT_FOUND
details = "Instance not found: project...066,"grpc_message":"Instance not found: projects/python-docs-samples-tests/instances/test-instance","grpc_status":5}"
>
> ???
E google.api_core.exceptions.NotFound: 404 Instance not found: projects/python-docs-samples-tests/instances/test-instance
<string>:3: NotFound</pre></details>
|
process
|
samples samples quickstart test test quickstart failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output args parent projects python docs samples tests instances test instance create statement create database my database id kwargs metadata six wraps callable def error remapped callable args kwargs try return callable args kwargs nox py lib site packages google api core grpc helpers py self request parent projects python docs samples tests instances test instance create statement create database my database id timeout none metadata credentials none wait for ready none compression none def call self request timeout none metadata none credentials none wait for ready none compression none state call self blocking request timeout metadata credentials wait for ready compression return end unary response blocking state call false none nox py lib site packages grpc channel py state call with call false deadline none def end unary response blocking state call with call deadline if state code is grpc statuscode ok if with call rendezvous multithreadedrendezvous state call none deadline return state response rendezvous else return state response else raise inactiverpcerror state e grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with e status statuscode not found e details instance not found projects python docs samples tests instances test instance e debug error string created description error received from peer file src core lib surface call cc file line grpc message instance not found projects python docs samples tests instances test instance grpc status e nox py lib site packages grpc channel py inactiverpcerror the above exception was the direct cause of the following exception pytest fixture def example database spanner client spanner client instance spanner client instance spanner instance database instance database my database id if not database exists database create quickstart test py google cloud spanner database py in create future api create database request request metadata metadata google cloud spanner admin database services database admin client py in create database response rpc request retry retry timeout timeout metadata metadata nox py lib site packages google api core gapic method py in call return wrapped func args kwargs nox py lib site packages google api core grpc helpers py in error remapped callable six raise from exceptions from grpc error exc exc value none from value inactiverpcerror of rpc that terminated with status statuscode not found details instance not found project grpc message instance not found projects python docs samples tests instances test instance grpc status e google api core exceptions notfound instance not found projects python docs samples tests instances test instance notfound
| 1
|
241,276
| 18,440,695,042
|
IssuesEvent
|
2021-10-14 17:42:54
|
AlgebraStudentCollab/START-HERE
|
https://api.github.com/repos/AlgebraStudentCollab/START-HERE
|
opened
|
Improve START-HERE and all READMEs
|
documentation
|
Goals
-Help newcomers figure out what this repo is About and how to use it
-Read me with: etiquette, tutorials (git, setup of things...etc)
-How to fork
-How to ask questions (issues etc..)
But for every repo:
-What is this repo about? (ex. Java 2 Swing bla bla)
-How to fork this exact repo?
-Where to find materials
-Deadlines
|
1.0
|
Improve START-HERE and all READMEs - Goals
-Help newcomers figure out what this repo is About and how to use it
-Read me with: etiquette, tutorials (git, setup of things...etc)
-How to fork
-How to ask questions (issues etc..)
But for every repo:
-What is this repo about? (ex. Java 2 Swing bla bla)
-How to fork this exact repo?
-Where to find materials
-Deadlines
|
non_process
|
improve start here and all readmes goals help newcomers figure out what this repo is about and how to use it read me with etiquette tutorials git setup of things etc how to fork how to ask questions issues etc but for every repo what is this repo about ex java swing bla bla how to fork this exact repo where to find materials deadlines
| 0
|
2,665
| 5,438,842,691
|
IssuesEvent
|
2017-03-06 11:39:15
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
An interesting feature process.hrtime
|
process question
|
Version: v7.7.1
Platform: Linux admin 4.9.0-0.bpo.1-amd64 #1 SMP Debian 4.9.2-2~bpo8+1 (2017-01-26) x86_64 GNU/Linux
Subsystem
Catching manual profiling functions with the help of designs
function profile(func) {
var wrapper = function () {
var start = process.hrtime(), diff;
var result = func.apply(null, arguments);
diff = process.hrtime(start);
console.log(func.name, diff[0] * 1e9 + diff[1], "ns");
return result;
};
return wrapper;
}
remarked feature: when the profiler is disabled, some functions
have started to run slower and the overall code has become slower
by about a third. Experimentally established that in some cases
the code
process.hrtime(process.hrtime());
accelerates nodejs.
|
1.0
|
An interesting feature process.hrtime - Version: v7.7.1
Platform: Linux admin 4.9.0-0.bpo.1-amd64 #1 SMP Debian 4.9.2-2~bpo8+1 (2017-01-26) x86_64 GNU/Linux
Subsystem
Catching manual profiling functions with the help of designs
function profile(func) {
var wrapper = function () {
var start = process.hrtime(), diff;
var result = func.apply(null, arguments);
diff = process.hrtime(start);
console.log(func.name, diff[0] * 1e9 + diff[1], "ns");
return result;
};
return wrapper;
}
remarked feature: when the profiler is disabled, some functions
have started to run slower and the overall code has become slower
by about a third. Experimentally established that in some cases
the code
process.hrtime(process.hrtime());
accelerates nodejs.
|
process
|
an interesting feature process hrtime version platform linux admin bpo smp debian gnu linux subsystem catching manual profiling functions with the help of designs function profile func var wrapper function var start process hrtime diff var result func apply null arguments diff process hrtime start console log func name diff diff ns return result return wrapper remarked feature when the profiler is disabled some functions have started to run slower and the overall code has become slower by about a third experimentally established that in some cases the code process hrtime process hrtime accelerates nodejs
| 1
|
3,722
| 6,732,899,033
|
IssuesEvent
|
2017-10-18 13:13:55
|
lockedata/rcms
|
https://api.github.com/repos/lockedata/rcms
|
opened
|
Register
|
attendee odoo processes
|
## Detailed task
Buy a ticket for the conference
## Assessing the task
Try to perform the task. Use google and the system documentation to help - part of what we're trying to assess how easy it is for people to work out how to do tasks.
Use a 👍 (`:+1:`) reaction to this task if you were able to perform the task. Use a 👎 (`:-1:`) reaction to the task if you could not complete it. Add a reply with any comments or feedback.
## Extra Info
- Site: [odoo](//http://188.166.159.192:8069)
- System documentation: [odoo docs](https://www.odoo.com/page/docs)
- Role: Attendee
- Area: Processes
|
1.0
|
Register - ## Detailed task
Buy a ticket for the conference
## Assessing the task
Try to perform the task. Use google and the system documentation to help - part of what we're trying to assess how easy it is for people to work out how to do tasks.
Use a 👍 (`:+1:`) reaction to this task if you were able to perform the task. Use a 👎 (`:-1:`) reaction to the task if you could not complete it. Add a reply with any comments or feedback.
## Extra Info
- Site: [odoo](//http://188.166.159.192:8069)
- System documentation: [odoo docs](https://www.odoo.com/page/docs)
- Role: Attendee
- Area: Processes
|
process
|
register detailed task buy a ticket for the conference assessing the task try to perform the task use google and the system documentation to help part of what we re trying to assess how easy it is for people to work out how to do tasks use a 👍 reaction to this task if you were able to perform the task use a 👎 reaction to the task if you could not complete it add a reply with any comments or feedback extra info site system documentation role attendee area processes
| 1
|
124,406
| 10,310,761,379
|
IssuesEvent
|
2019-08-29 15:47:34
|
ansible/awx
|
https://api.github.com/repos/ansible/awx
|
closed
|
CLI: awx users create does not indicate correct required arguments
|
component:cli priority:medium state:needs_test type:bug
|
##### ISSUE TYPE
- Bug Report
##### SUMMARY
<!-- Briefly describe the problem. -->
Help text only indicates need username, also some strange formatting
##### STEPS TO REPRODUCE
<!-- Please describe exactly how to reproduce the problem. -->
1) Have valid token or username/password in environment
2) `awx users create --help`
##### EXPECTED RESULTS
<!-- What did you expect to happen when running the steps above? -->
```
usage: awx users create [-h] --username TEXT [--first_name TEXT]
[--last_name TEXT] [--email TEXT]
[--is_superuser BOOLEAN] [--is_system_auditor BOOLEAN]
--password TEXT [--last_login LAST_LOGIN]
arguments:
-h, --help show this help message and exit
--username TEXT [REQUIRED] Required. 150 characters or fewer. Letters,
digits and @/./+/-/_ only.
--first_name TEXT
--last_name TEXT
--email TEXT
--is_superuser BOOLEAN
Designates that this user has all permissions without
explicitly assigning them.
--is_system_auditor BOOLEAN
--password TEXT [REQUIRED] Write-only field used to change the password.
--last_login LAST_LOGIN
awx users create: the following arguments are required: --username --password
```
##### ACTUAL RESULTS
<!-- What actually happened? -->
```
awx users create --help
usage: awx users create [-h] --username TEXT [--first_name TEXT]
[--last_name TEXT] [--email TEXT]
[--is_superuser BOOLEAN] [--is_system_auditor BOOLEAN]
[--password TEXT] [--last_login LAST_LOGIN]
optional arguments:
-h, --help show this help message and exit
--username TEXT [REQUIRED] Required. 150 characters or fewer. Letters,
digits and @/./+/-/_ only.
--first_name TEXT
--last_name TEXT
--email TEXT
--is_superuser BOOLEAN
Designates that this user has all permissions without
explicitly assigning them.
--is_system_auditor BOOLEAN
--password TEXT Write-only field used to change the password.
--last_login LAST_LOGIN
awx users create: the following arguments are required: --username
```
diff is:
1) that all arguments are labeled as `optional arguments`, we could fix this by just saying `arguments`
2) Last line `the following arguments are required: --username` should include `--password`
3) In list at top of help `--password` should NOT be in brackets (is not optional)
##### ADDITIONAL INFORMATION
<!-- Include any links to sosreport, database dumps, screenshots or other
information. -->
the `optional arguments` thing is a problem across the board, should be fixed in general way I think
|
1.0
|
CLI: awx users create does not indicate correct required arguments - ##### ISSUE TYPE
- Bug Report
##### SUMMARY
<!-- Briefly describe the problem. -->
Help text only indicates need username, also some strange formatting
##### STEPS TO REPRODUCE
<!-- Please describe exactly how to reproduce the problem. -->
1) Have valid token or username/password in environment
2) `awx users create --help`
##### EXPECTED RESULTS
<!-- What did you expect to happen when running the steps above? -->
```
usage: awx users create [-h] --username TEXT [--first_name TEXT]
[--last_name TEXT] [--email TEXT]
[--is_superuser BOOLEAN] [--is_system_auditor BOOLEAN]
--password TEXT [--last_login LAST_LOGIN]
arguments:
-h, --help show this help message and exit
--username TEXT [REQUIRED] Required. 150 characters or fewer. Letters,
digits and @/./+/-/_ only.
--first_name TEXT
--last_name TEXT
--email TEXT
--is_superuser BOOLEAN
Designates that this user has all permissions without
explicitly assigning them.
--is_system_auditor BOOLEAN
--password TEXT [REQUIRED] Write-only field used to change the password.
--last_login LAST_LOGIN
awx users create: the following arguments are required: --username --password
```
##### ACTUAL RESULTS
<!-- What actually happened? -->
```
awx users create --help
usage: awx users create [-h] --username TEXT [--first_name TEXT]
[--last_name TEXT] [--email TEXT]
[--is_superuser BOOLEAN] [--is_system_auditor BOOLEAN]
[--password TEXT] [--last_login LAST_LOGIN]
optional arguments:
-h, --help show this help message and exit
--username TEXT [REQUIRED] Required. 150 characters or fewer. Letters,
digits and @/./+/-/_ only.
--first_name TEXT
--last_name TEXT
--email TEXT
--is_superuser BOOLEAN
Designates that this user has all permissions without
explicitly assigning them.
--is_system_auditor BOOLEAN
--password TEXT Write-only field used to change the password.
--last_login LAST_LOGIN
awx users create: the following arguments are required: --username
```
diff is:
1) that all arguments are labeled as `optional arguments`, we could fix this by just saying `arguments`
2) Last line `the following arguments are required: --username` should include `--password`
3) In list at top of help `--password` should NOT be in brackets (is not optional)
##### ADDITIONAL INFORMATION
<!-- Include any links to sosreport, database dumps, screenshots or other
information. -->
the `optional arguments` thing is a problem across the board, should be fixed in general way I think
|
non_process
|
cli awx users create does not indicate correct required arguments issue type bug report summary help text only indicates need username also some strange formatting steps to reproduce have valid token or username password in environment awx users create help expected results usage awx users create username text password text arguments h help show this help message and exit username text required characters or fewer letters digits and only first name text last name text email text is superuser boolean designates that this user has all permissions without explicitly assigning them is system auditor boolean password text write only field used to change the password last login last login awx users create the following arguments are required username password actual results awx users create help usage awx users create username text optional arguments h help show this help message and exit username text required characters or fewer letters digits and only first name text last name text email text is superuser boolean designates that this user has all permissions without explicitly assigning them is system auditor boolean password text write only field used to change the password last login last login awx users create the following arguments are required username diff is that all arguments are labeled as optional arguments we could fix this by just saying arguments last line the following arguments are required username should include password in list at top of help password should not be in brackets is not optional additional information include any links to sosreport database dumps screenshots or other information the optional arguments thing is a problem across the board should be fixed in general way i think
| 0
|
123,035
| 10,244,935,793
|
IssuesEvent
|
2019-08-20 11:41:33
|
kyma-project/kyma
|
https://api.github.com/repos/kyma-project/kyma
|
closed
|
Runtimes and central connector service integration tests
|
area/application-connector enhancement quality/testability
|
Provide an integration test for the runtimes and central connector service. Current approach allows running tests during the deployment in the desired environment together with runtime provisioner tests. Extend existing tests to cover the following scenario:
- runtime is provisioned
- connectivity-certs-controller is configured to contact central connector
- verify whether connectivity-certs-controller successfully applied certificate on the runtime
- register application
- hit runtime with the application certificate
|
1.0
|
Runtimes and central connector service integration tests - Provide an integration test for the runtimes and central connector service. Current approach allows running tests during the deployment in the desired environment together with runtime provisioner tests. Extend existing tests to cover the following scenario:
- runtime is provisioned
- connectivity-certs-controller is configured to contact central connector
- verify whether connectivity-certs-controller successfully applied certificate on the runtime
- register application
- hit runtime with the application certificate
|
non_process
|
runtimes and central connector service integration tests provide an integration test for the runtimes and central connector service current approach allows running tests during the deployment in the desired environment together with runtime provisioner tests extend existing tests to cover the following scenario runtime is provisioned connectivity certs controller is configured to contact central connector verify whether connectivity certs controller successfully applied certificate on the runtime register application hit runtime with the application certificate
| 0
|
325,871
| 9,937,040,905
|
IssuesEvent
|
2019-07-02 20:45:36
|
mlr-org/paradox
|
https://api.github.com/repos/mlr-org/paradox
|
closed
|
Function to print x values
|
Priority: Medium Type: Enhancement
|
Function values are stored in named lists.
To transform them to a single string you could use `paste(names(x), x, sep = "=" ,collapse=",")`
This is problematic for
* Long values
* x values that can not be transferred to a character. These should not exists, because complex types are just created by transformation. But we have a untyped param class.
* Real valued numbers with many decimal places.
because they can mess up the output.
So we want to shorten and format some of them.
Formatting and shortening should be configurable.
Each ParamNode should be able to transform a named list to a character.
I propose `Param(Set/Real/...)$value_to_string(x)`.
|
1.0
|
Function to print x values - Function values are stored in named lists.
To transform them to a single string you could use `paste(names(x), x, sep = "=" ,collapse=",")`
This is problematic for
* Long values
* x values that can not be transferred to a character. These should not exists, because complex types are just created by transformation. But we have a untyped param class.
* Real valued numbers with many decimal places.
because they can mess up the output.
So we want to shorten and format some of them.
Formatting and shortening should be configurable.
Each ParamNode should be able to transform a named list to a character.
I propose `Param(Set/Real/...)$value_to_string(x)`.
|
non_process
|
function to print x values function values are stored in named lists to transform them to a single string you could use paste names x x sep collapse this is problematic for long values x values that can not be transferred to a character these should not exists because complex types are just created by transformation but we have a untyped param class real valued numbers with many decimal places because they can mess up the output so we want to shorten and format some of them formatting and shortening should be configurable each paramnode should be able to transform a named list to a character i propose param set real value to string x
| 0
|
2,004
| 4,726,832,332
|
IssuesEvent
|
2016-10-18 11:35:44
|
AdguardTeam/AdguardForAndroid
|
https://api.github.com/repos/AdguardTeam/AdguardForAndroid
|
closed
|
Bug with Updato app when Adguard enabled.
|
Bug Compatibility
|
Updato app (https://play.google.com/store/apps/details?id=samsungupdate.com&hl=en) can't update device because of Adguard. After entering device model and choosing update and download options app redirects to page, which says that we can't proceed it because of Adblocker.
Samsung-N910P
Android 6.0.1
Ver. 2.7.220
Logs
<details>[adguard.zip](https://github.com/AdguardTeam/AdguardForAndroid/files/520125/adguard.zip)</details>
Video guide
<details>[SCR_20161006_175304 (1).zip](https://github.com/AdguardTeam/AdguardForAndroid/files/520126/SCR_20161006_175304.1.zip)</details>
Ticket code 1138191
|
True
|
Bug with Updato app when Adguard enabled. - Updato app (https://play.google.com/store/apps/details?id=samsungupdate.com&hl=en) can't update device because of Adguard. After entering device model and choosing update and download options app redirects to page, which says that we can't proceed it because of Adblocker.
Samsung-N910P
Android 6.0.1
Ver. 2.7.220
Logs
<details>[adguard.zip](https://github.com/AdguardTeam/AdguardForAndroid/files/520125/adguard.zip)</details>
Video guide
<details>[SCR_20161006_175304 (1).zip](https://github.com/AdguardTeam/AdguardForAndroid/files/520126/SCR_20161006_175304.1.zip)</details>
Ticket code 1138191
|
non_process
|
bug with updato app when adguard enabled updato app can t update device because of adguard after entering device model and choosing update and download options app redirects to page which says that we can t proceed it because of adblocker samsung android ver logs video guide ticket code
| 0
|
493,196
| 14,227,955,533
|
IssuesEvent
|
2020-11-18 02:33:47
|
aiidateam/aiida-core
|
https://api.github.com/repos/aiidateam/aiida-core
|
closed
|
Update `verdi import`
|
good first issue priority/quality-of-life topic/verdi type/refactoring
|
As mentioned [here](https://github.com/aiidateam/aiida_core/pull/2820#pullrequestreview-235910419):
Suggestions to update `verdi import` code (by @ltalirz):
- [ ] archives should simply be processed one by one (there is no need to create and validate the full list of archives before you even start. e.g. `rm file1 file2 file3` also will start removing files and only stop when it encounters a file that isn't there)
- [ ] when you look at an archive you figure out whether it's a url or not. if url, validate URL and call function to download. you replace url by path to downloaded file and the following code runs inside a sandbox folder
- [ ] from now on all code is shared...
|
1.0
|
Update `verdi import` - As mentioned [here](https://github.com/aiidateam/aiida_core/pull/2820#pullrequestreview-235910419):
Suggestions to update `verdi import` code (by @ltalirz):
- [ ] archives should simply be processed one by one (there is no need to create and validate the full list of archives before you even start. e.g. `rm file1 file2 file3` also will start removing files and only stop when it encounters a file that isn't there)
- [ ] when you look at an archive you figure out whether it's a url or not. if url, validate URL and call function to download. you replace url by path to downloaded file and the following code runs inside a sandbox folder
- [ ] from now on all code is shared...
|
non_process
|
update verdi import as mentioned suggestions to update verdi import code by ltalirz archives should simply be processed one by one there is no need to create and validate the full list of archives before you even start e g rm also will start removing files and only stop when it encounters a file that isn t there when you look at an archive you figure out whether it s a url or not if url validate url and call function to download you replace url by path to downloaded file and the following code runs inside a sandbox folder from now on all code is shared
| 0
|
8,114
| 11,302,311,845
|
IssuesEvent
|
2020-01-17 17:22:11
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
regulating the host phagocytosis machinery
|
multi-species process
|
<img width="1059" alt="Screenshot 2019-10-09 at 20 09 23" src="https://user-images.githubusercontent.com/7359272/66512395-bad30700-ead0-11e9-868a-f34a9222d3f9.png">
These are the terms related to modulation of host phagocytosis by symbiont
GO:0052190 modulation by symbiont of host phagocytosis
Any process in which an organism modulates the frequency, rate or extent of phagocytosis, the engulfing by phagocytes of external particulate material, in the host organism. The host is defined as the larger of the organisms involved in a symbiotic interaction.
GO:0052191 positive regulation by symbiont of host phagocytosis
This term, in a different place:
GO:0052067 negative regulation by symbiont of entry into host cell via phagocytosis
So the symbiont uses the host phagocytosis machinery for entry into the host.
This is described by the child of GO:0052190
GO:0052191 positive regulation by symbiont of host phagocytosis
Definition Any process in which an organism activates, maintains or increases the frequency, rate or extent of phagocytosis, the engulfing by phagocytes of external particulate material, in the host organism. The host is defined as the larger of the organisms involved in a symbiotic interaction.
i.e if the symbiont is positively regulating the phagocytosis machinery it is using it to enter the host.
Therefore this is the term which should have an "entry into host parentage". I think it would be better to describe this term as a 'hijacking' process.
The GO:0052067 negative regulation by symbiont of entry into host cell via phagocytosis
is describing *avoidence* of engulfment by host phagocytosis (i.e by macrophages).
this term is nothing to do with "entry into host", but this term does have (incorrectly) the entry into host parentage.
|
1.0
|
regulating the host phagocytosis machinery - <img width="1059" alt="Screenshot 2019-10-09 at 20 09 23" src="https://user-images.githubusercontent.com/7359272/66512395-bad30700-ead0-11e9-868a-f34a9222d3f9.png">
These are the terms related to modulation of host phagocytosis by symbiont
GO:0052190 modulation by symbiont of host phagocytosis
Any process in which an organism modulates the frequency, rate or extent of phagocytosis, the engulfing by phagocytes of external particulate material, in the host organism. The host is defined as the larger of the organisms involved in a symbiotic interaction.
GO:0052191 positive regulation by symbiont of host phagocytosis
This term, in a different place:
GO:0052067 negative regulation by symbiont of entry into host cell via phagocytosis
So the symbiont uses the host phagocytosis machinery for entry into the host.
This is described by the child of GO:0052190
GO:0052191 positive regulation by symbiont of host phagocytosis
Definition Any process in which an organism activates, maintains or increases the frequency, rate or extent of phagocytosis, the engulfing by phagocytes of external particulate material, in the host organism. The host is defined as the larger of the organisms involved in a symbiotic interaction.
i.e if the symbiont is positively regulating the phagocytosis machinery it is using it to enter the host.
Therefore this is the term which should have an "entry into host parentage". I think it would be better to describe this term as a 'hijacking' process.
The GO:0052067 negative regulation by symbiont of entry into host cell via phagocytosis
is describing *avoidence* of engulfment by host phagocytosis (i.e by macrophages).
this term is nothing to do with "entry into host", but this term does have (incorrectly) the entry into host parentage.
|
process
|
regulating the host phagocytosis machinery img width alt screenshot at src these are the terms related to modulation of host phagocytosis by symbiont go modulation by symbiont of host phagocytosis any process in which an organism modulates the frequency rate or extent of phagocytosis the engulfing by phagocytes of external particulate material in the host organism the host is defined as the larger of the organisms involved in a symbiotic interaction go positive regulation by symbiont of host phagocytosis this term in a different place go negative regulation by symbiont of entry into host cell via phagocytosis so the symbiont uses the host phagocytosis machinery for entry into the host this is described by the child of go go positive regulation by symbiont of host phagocytosis definition any process in which an organism activates maintains or increases the frequency rate or extent of phagocytosis the engulfing by phagocytes of external particulate material in the host organism the host is defined as the larger of the organisms involved in a symbiotic interaction i e if the symbiont is positively regulating the phagocytosis machinery it is using it to enter the host therefore this is the term which should have an entry into host parentage i think it would be better to describe this term as a hijacking process the go negative regulation by symbiont of entry into host cell via phagocytosis is describing avoidence of engulfment by host phagocytosis i e by macrophages this term is nothing to do with entry into host but this term does have incorrectly the entry into host parentage
| 1
|
18,023
| 24,032,791,877
|
IssuesEvent
|
2022-09-15 16:19:21
|
googleapis/java-gke-multi-cloud
|
https://api.github.com/repos/googleapis/java-gke-multi-cloud
|
opened
|
Your .repo-metadata.json file has a problem 🤒
|
type: process repo-metadata: lint
|
You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'gke-multi-cloud' invalid in .repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'gke-multi-cloud' invalid in .repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname gke multi cloud invalid in repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
| 1
|
501,062
| 14,520,305,187
|
IssuesEvent
|
2020-12-14 05:10:00
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
reopened
|
portal.xero.com - site is not usable
|
browser-firefox engine-gecko priority-important
|
<!-- @browser: Firefox 84.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:84.0) Gecko/20100101 Firefox/84.0 -->
<!-- @reported_with: unknown -->
**URL**: https://portal.xero.com/Agreement/Sign/febb6fe2-46b9-4e0d-93b4-b2f6270734dd
**Browser / Version**: Firefox 84.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
When attempting to go to sign a particular document, I get the page :
---
The cookie functionality is disabled on your browser
Adobe Sign has detected that the cookie functionality is disabled on your browser. To use Adobe Sign, please enable cookies on your browser.
When cookies are enabled, click the "Continue" button below to proceed to Adobe Sign.
Instructions for enabling cookies
Safari for iOS
Safari for Mac
Chrome for phone or tablet
Chrome for desktop OS
Firefox
Click on the Firefox menu.
Select Preferences.
Select the Privacy & Security panel.
Select Standard to enable cookies.
Set Accept third-party cookies to Always to enable third-party cookies.
Edge for Windows
Internet Explorer
When cookies are enabled, click the "Continue" button below to refresh page.
---
Tried with a new profile same thing. Cookies are enabled and tracking protection is set to standard.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/10/ad8d017c-b031-4957-9587-0ca226edb1da.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
portal.xero.com - site is not usable - <!-- @browser: Firefox 84.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:84.0) Gecko/20100101 Firefox/84.0 -->
<!-- @reported_with: unknown -->
**URL**: https://portal.xero.com/Agreement/Sign/febb6fe2-46b9-4e0d-93b4-b2f6270734dd
**Browser / Version**: Firefox 84.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
When attempting to go to sign a particular document, I get the page :
---
The cookie functionality is disabled on your browser
Adobe Sign has detected that the cookie functionality is disabled on your browser. To use Adobe Sign, please enable cookies on your browser.
When cookies are enabled, click the "Continue" button below to proceed to Adobe Sign.
Instructions for enabling cookies
Safari for iOS
Safari for Mac
Chrome for phone or tablet
Chrome for desktop OS
Firefox
Click on the Firefox menu.
Select Preferences.
Select the Privacy & Security panel.
Select Standard to enable cookies.
Set Accept third-party cookies to Always to enable third-party cookies.
Edge for Windows
Internet Explorer
When cookies are enabled, click the "Continue" button below to refresh page.
---
Tried with a new profile same thing. Cookies are enabled and tracking protection is set to standard.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/10/ad8d017c-b031-4957-9587-0ca226edb1da.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
portal xero com site is not usable url browser version firefox operating system windows tested another browser yes chrome problem type site is not usable description page not loading correctly steps to reproduce when attempting to go to sign a particular document i get the page the cookie functionality is disabled on your browser adobe sign has detected that the cookie functionality is disabled on your browser to use adobe sign please enable cookies on your browser when cookies are enabled click the continue button below to proceed to adobe sign instructions for enabling cookies safari for ios safari for mac chrome for phone or tablet chrome for desktop os firefox click on the firefox menu select preferences select the privacy security panel select standard to enable cookies set accept third party cookies to always to enable third party cookies edge for windows internet explorer when cookies are enabled click the continue button below to refresh page tried with a new profile same thing cookies are enabled and tracking protection is set to standard view the screenshot img alt screenshot src browser configuration none from with ❤️
| 0
|
312,306
| 9,545,566,471
|
IssuesEvent
|
2019-05-01 17:26:15
|
knowmetools/km-api
|
https://api.github.com/repos/knowmetools/km-api
|
closed
|
Duplicate Apple receipts may be uploaded
|
Priority: Low Status: In Progress Type: Bug
|
**Environment (production/development)**
Occurs in: production
**Describe the bug**
A user can upload a more recent copy of a receipt as long as the stored copy of the receipt has not been updated yet by a background job.
**To Reproduce**
Steps to reproduce the behavior:
1. Upload a valid Apple receipt
2. Wait for Apple to update the receipt, thus generating a new blob of receipt data. This is easiest to see on the sandbox environment where receipts are refreshed on an order of minutes rather than days or months.
3. From within an iOS client use the "restore purchases" functionality to get back the receipt data which is now updated.
4. Upload that receipt to a different Know Me account before our background job runs to update the receipt.
**Expected behavior**
Attempting to upload the receipt in step 4 should return an error that the receipt is already in use.
**Additional context**
This happens because we use the receipt data itself to ensure that a receipt is unique so if receipt data is provided that we haven't received through the background job to update subscriptions yet we assume it's a new receipt.
To fix this we have to stop relying on the receipt data itself as an indication of uniqueness. We instead have to examine the information Apple provides when we verify the receipt more closely. The `original_transaction_id` field looks promising.
|
1.0
|
Duplicate Apple receipts may be uploaded - **Environment (production/development)**
Occurs in: production
**Describe the bug**
A user can upload a more recent copy of a receipt as long as the stored copy of the receipt has not been updated yet by a background job.
**To Reproduce**
Steps to reproduce the behavior:
1. Upload a valid Apple receipt
2. Wait for Apple to update the receipt, thus generating a new blob of receipt data. This is easiest to see on the sandbox environment where receipts are refreshed on an order of minutes rather than days or months.
3. From within an iOS client use the "restore purchases" functionality to get back the receipt data which is now updated.
4. Upload that receipt to a different Know Me account before our background job runs to update the receipt.
**Expected behavior**
Attempting to upload the receipt in step 4 should return an error that the receipt is already in use.
**Additional context**
This happens because we use the receipt data itself to ensure that a receipt is unique so if receipt data is provided that we haven't received through the background job to update subscriptions yet we assume it's a new receipt.
To fix this we have to stop relying on the receipt data itself as an indication of uniqueness. We instead have to examine the information Apple provides when we verify the receipt more closely. The `original_transaction_id` field looks promising.
|
non_process
|
duplicate apple receipts may be uploaded environment production development occurs in production describe the bug a user can upload a more recent copy of a receipt as long as the stored copy of the receipt has not been updated yet by a background job to reproduce steps to reproduce the behavior upload a valid apple receipt wait for apple to update the receipt thus generating a new blob of receipt data this is easiest to see on the sandbox environment where receipts are refreshed on an order of minutes rather than days or months from within an ios client use the restore purchases functionality to get back the receipt data which is now updated upload that receipt to a different know me account before our background job runs to update the receipt expected behavior attempting to upload the receipt in step should return an error that the receipt is already in use additional context this happens because we use the receipt data itself to ensure that a receipt is unique so if receipt data is provided that we haven t received through the background job to update subscriptions yet we assume it s a new receipt to fix this we have to stop relying on the receipt data itself as an indication of uniqueness we instead have to examine the information apple provides when we verify the receipt more closely the original transaction id field looks promising
| 0
|
35,743
| 9,657,215,267
|
IssuesEvent
|
2019-05-20 08:00:40
|
gradle/gradle
|
https://api.github.com/repos/gradle/gradle
|
closed
|
Introduce BuildCacheService#init/start
|
@build-cache from:contributor
|
The `BuildCacheService` has a `#close` method which will be triggered at the end of the Gradle build to close possible opened resources.
What I is missing a `BuildCacheService#init/start` method, which will be called only once. When the BuildCache is used for the first time.
### Expected Behavior
There should be a `BuildCacheService#init or start` method which will be triggered by the first artifact which can be stored.
### Current Behavior
There is no `init/start` method.
### Context
Assume that I build my own remote `BuildCacheService`.
For whatever reasons my `BuildCacheService#store` took some time (maybe because my server needs a authentication **all the time** (for each stored artifact)).
I could imagine that I create a "session" ID in the `init/start` method which allows the `BuildCacheService#store` method to call a URL with that session ID.
Which means my build cache server does not to authenticate the user again and again - which leads to faster storing...
### Steps to Reproduce (for bugs)
-
### Your Environment
-
|
1.0
|
Introduce BuildCacheService#init/start - The `BuildCacheService` has a `#close` method which will be triggered at the end of the Gradle build to close possible opened resources.
What I is missing a `BuildCacheService#init/start` method, which will be called only once. When the BuildCache is used for the first time.
### Expected Behavior
There should be a `BuildCacheService#init or start` method which will be triggered by the first artifact which can be stored.
### Current Behavior
There is no `init/start` method.
### Context
Assume that I build my own remote `BuildCacheService`.
For whatever reasons my `BuildCacheService#store` took some time (maybe because my server needs a authentication **all the time** (for each stored artifact)).
I could imagine that I create a "session" ID in the `init/start` method which allows the `BuildCacheService#store` method to call a URL with that session ID.
Which means my build cache server does not to authenticate the user again and again - which leads to faster storing...
### Steps to Reproduce (for bugs)
-
### Your Environment
-
|
non_process
|
introduce buildcacheservice init start the buildcacheservice has a close method which will be triggered at the end of the gradle build to close possible opened resources what i is missing a buildcacheservice init start method which will be called only once when the buildcache is used for the first time expected behavior there should be a buildcacheservice init or start method which will be triggered by the first artifact which can be stored current behavior there is no init start method context assume that i build my own remote buildcacheservice for whatever reasons my buildcacheservice store took some time maybe because my server needs a authentication all the time for each stored artifact i could imagine that i create a session id in the init start method which allows the buildcacheservice store method to call a url with that session id which means my build cache server does not to authenticate the user again and again which leads to faster storing steps to reproduce for bugs your environment
| 0
|
22,575
| 31,800,950,712
|
IssuesEvent
|
2023-09-13 11:04:09
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
Performance regression in how we parse query_metadata response leads to notebook being unusable
|
Type:Bug Priority:P1 .Performance Querying/MBQL .Backend .Regression .Blocker .Team/QueryProcessor :hammer_and_wrench: .Escalation
|
### Describe the bug
We introduced a regression in the query_metadata, possibly because of several factors, that lead questions with lots of Metadata to slow down heavily the frontend (Pro tip: use the table metadata and generate a large lorem ipsum on the fields and tables descriptions to make a heavy response along with a "big" question with lots of stages and fields. remember we call the table_metadata for every linked table on a query, so if table 1 has a FK to table 2, we fetch both tables metadata)
Performance in 46:

Profiler


Now let's move to 47:
look at the separation between calls (sizes of responses stayed the same)

profiler:
2X

This will get amplified as more tables are joined in the data model
### To Reproduce
1) go to the data model and try to join a single table to all others. Also try to complete as much metadata as possible
2) then create a GUI question and add that single table, build a question with lots of aggregations, fields, etc
3) compare it between 46 and 47, see the difference in rendering
### Expected behavior
UI should be F A S T
### Logs
NA
### Information about your Metabase installation
```JSON
v47
```
### Severity
P1
### Additional context
_No response_
|
1.0
|
Performance regression in how we parse query_metadata response leads to notebook being unusable - ### Describe the bug
We introduced a regression in the query_metadata, possibly because of several factors, that lead questions with lots of Metadata to slow down heavily the frontend (Pro tip: use the table metadata and generate a large lorem ipsum on the fields and tables descriptions to make a heavy response along with a "big" question with lots of stages and fields. remember we call the table_metadata for every linked table on a query, so if table 1 has a FK to table 2, we fetch both tables metadata)
Performance in 46:

Profiler


Now let's move to 47:
look at the separation between calls (sizes of responses stayed the same)

profiler:
2X

This will get amplified as more tables are joined in the data model
### To Reproduce
1) go to the data model and try to join a single table to all others. Also try to complete as much metadata as possible
2) then create a GUI question and add that single table, build a question with lots of aggregations, fields, etc
3) compare it between 46 and 47, see the difference in rendering
### Expected behavior
UI should be F A S T
### Logs
NA
### Information about your Metabase installation
```JSON
v47
```
### Severity
P1
### Additional context
_No response_
|
process
|
performance regression in how we parse query metadata response leads to notebook being unusable describe the bug we introduced a regression in the query metadata possibly because of several factors that lead questions with lots of metadata to slow down heavily the frontend pro tip use the table metadata and generate a large lorem ipsum on the fields and tables descriptions to make a heavy response along with a big question with lots of stages and fields remember we call the table metadata for every linked table on a query so if table has a fk to table we fetch both tables metadata performance in profiler now let s move to look at the separation between calls sizes of responses stayed the same profiler this will get amplified as more tables are joined in the data model to reproduce go to the data model and try to join a single table to all others also try to complete as much metadata as possible then create a gui question and add that single table build a question with lots of aggregations fields etc compare it between and see the difference in rendering expected behavior ui should be f a s t logs na information about your metabase installation json severity additional context no response
| 1
|
340,202
| 24,645,139,188
|
IssuesEvent
|
2022-10-17 14:23:07
|
music-encoding/music-encoding.github.io
|
https://api.github.com/repos/music-encoding/music-encoding.github.io
|
closed
|
Link Checker Report
|
documentation bug
|
## Summary
| Status | Count |
|---------------|-------|
| 🔍 Total | 432 |
| ✅ Successful | 199 |
| ⏳ Timeouts | 1 |
| 🔀 Redirected | 0 |
| 👻 Excluded | 232 |
| ❓ Unknown | 0 |
| 🚫 Errors | 0 |
## Errors per input
### Errors in ./resources/pedagogy.md
* [https://dhsi.org/](https://dhsi.org/): Timeout (status code: TIMEOUT)
[Full Github Actions output](https://github.com/music-encoding/music-encoding.github.io/actions/runs/3259662048?check_suite_focus=true)
|
1.0
|
Link Checker Report - ## Summary
| Status | Count |
|---------------|-------|
| 🔍 Total | 432 |
| ✅ Successful | 199 |
| ⏳ Timeouts | 1 |
| 🔀 Redirected | 0 |
| 👻 Excluded | 232 |
| ❓ Unknown | 0 |
| 🚫 Errors | 0 |
## Errors per input
### Errors in ./resources/pedagogy.md
* [https://dhsi.org/](https://dhsi.org/): Timeout (status code: TIMEOUT)
[Full Github Actions output](https://github.com/music-encoding/music-encoding.github.io/actions/runs/3259662048?check_suite_focus=true)
|
non_process
|
link checker report summary status count 🔍 total ✅ successful ⏳ timeouts 🔀 redirected 👻 excluded ❓ unknown 🚫 errors errors per input errors in resources pedagogy md timeout status code timeout
| 0
|
90,284
| 10,677,470,109
|
IssuesEvent
|
2019-10-21 15:28:55
|
christianvoigt/argdown
|
https://api.github.com/repos/christianvoigt/argdown
|
closed
|
link to first example broken
|
documentation
|
>In case it's useful I immediately got a dead link https://argdown.org/guide/a-first-example/ from this page https://argdown.org/guide/
>--[pithymaxim](https://news.ycombinator.com/reply?id=20476015&goto=item%3Fid%3D20475865%2320476015)
>FYI to the owner of Argdown.org if around- Your getting started page linked here 404's: https://argdown.org/guide/a-first-example
>-- [pmarreck](https://news.ycombinator.com/reply?id=20481733&goto=item%3Fid%3D20475865%2320481733)
|
1.0
|
link to first example broken - >In case it's useful I immediately got a dead link https://argdown.org/guide/a-first-example/ from this page https://argdown.org/guide/
>--[pithymaxim](https://news.ycombinator.com/reply?id=20476015&goto=item%3Fid%3D20475865%2320476015)
>FYI to the owner of Argdown.org if around- Your getting started page linked here 404's: https://argdown.org/guide/a-first-example
>-- [pmarreck](https://news.ycombinator.com/reply?id=20481733&goto=item%3Fid%3D20475865%2320481733)
|
non_process
|
link to first example broken in case it s useful i immediately got a dead link from this page fyi to the owner of argdown org if around your getting started page linked here s
| 0
|
12,164
| 14,741,546,190
|
IssuesEvent
|
2021-01-07 10:47:18
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Last few error emails - Concern
|
anc-process anp-2.5 ant-bug has attachment
|
In GitLab by @kdjstudios on Jan 24, 2019, 08:37
Hello Team,
Here are the last few error emails we have been receiving. May someone please explain them to me. I have my thoughts on what each one means.
Error: [Cron_root_answernet_binbash_homesabillingmongodb_backupsbackup_mongodb.sh.msg](/uploads/13c68884d888d27e1479e767f2798af8/Cron_root_answernet_binbash_homesabillingmongodb_backupsbackup_mongodb.sh.msg)
* My thought: Judging from the amount "Backup: command not found" it would seem we are not backing things up?
Error: [Cron_sabilling_answernet_binbash_optsabillingdelete_customer_without_account.sh.msg](/uploads/8f6257a1f04e70dd7b6e4028e1c532ae/Cron_sabilling_answernet_binbash_optsabillingdelete_customer_without_account.sh.msg)
* My Thought: Judging from the lack of success or completed message, it would appear this script is possibly failing?
Error: [Cron_sabilling_answernet_homesabillingjobsrun_check_aging.sh.msg](/uploads/d75de2e1c889607e4984cff243a7fb53/Cron_sabilling_answernet_homesabillingjobsrun_check_aging.sh.msg)
* My Thought: Judging from the amount of sites and accounts that appear to show bad AR, it would seem we still have an AR issue?
Error: [Cron_sabilling_answernet_homesabillingjobsrun_daily_ar_snapshot.sh.zip](/uploads/df922790b5dfb51fd32a402cae7ed005/Cron_sabilling_answernet_homesabillingjobsrun_daily_ar_snapshot.sh.zip)
* My Thought: Judging from the length of the email and that I am unable to understand what anything in it means, I need clarification on what to look for. (NOTE: I had to compress this email since it was over 20MB and gitlab would not upload it.)
Error: [SA_Billing_Error_Report_invoices_index__MongoidErrorsDocumentNotFound__nProblemn__Document_s__not_found_for_cl....msg](/uploads/64d8bd2b16996227daaac9f7a18e7653/SA_Billing_Error_Report_invoices_index__MongoidErrorsDocumentNotFound__nProblemn__Document_s__not_found_for_cl....msg)
*My Thought: It looks like an account is missing or some part of an account is missing?
Error: [SA_Billing_Error_Report_accounts_edit__MongoidErrorsDocumentNotFound__nProblemn__Document_s__not_found_for_cla....msg](/uploads/6a4cb7fd69c672d66c7e365b3fe50c0a/SA_Billing_Error_Report_accounts_edit__MongoidErrorsDocumentNotFound__nProblemn__Document_s__not_found_for_cla....msg)
* My Thoughts: Looks like someone tried to edit an account and was unable to as the account is missing? Maybe a cache issue?
Error: [SA_Billing_Error_Report_billing_cycles_upload__ArgumentError__invalid_byte_sequence_in_UTF-8.msg](/uploads/313e1c395608afbb7d1adf66b5cda4a9/SA_Billing_Error_Report_billing_cycles_upload__ArgumentError__invalid_byte_sequence_in_UTF-8.msg)
* My Thought: Looks like someone attempted to upload a invalid formatted usage file. (NOTE: There a like 5 of these emails consecutively.)
Error: [SA_Billing_Error_Report_side_nav_pull_customers__MongoidErrorsDocumentNotFound__nProblemn__Document_s__not_fou....msg](/uploads/b081ef4fc7debd0b34aa2ce468e206b2/SA_Billing_Error_Report_side_nav_pull_customers__MongoidErrorsDocumentNotFound__nProblemn__Document_s__not_fou....msg)
* My Thought: Looks like the drop down is displaying a link to an invalid site?
|
1.0
|
Last few error emails - Concern - In GitLab by @kdjstudios on Jan 24, 2019, 08:37
Hello Team,
Here are the last few error emails we have been receiving. May someone please explain them to me. I have my thoughts on what each one means.
Error: [Cron_root_answernet_binbash_homesabillingmongodb_backupsbackup_mongodb.sh.msg](/uploads/13c68884d888d27e1479e767f2798af8/Cron_root_answernet_binbash_homesabillingmongodb_backupsbackup_mongodb.sh.msg)
* My thought: Judging from the amount "Backup: command not found" it would seem we are not backing things up?
Error: [Cron_sabilling_answernet_binbash_optsabillingdelete_customer_without_account.sh.msg](/uploads/8f6257a1f04e70dd7b6e4028e1c532ae/Cron_sabilling_answernet_binbash_optsabillingdelete_customer_without_account.sh.msg)
* My Thought: Judging from the lack of success or completed message, it would appear this script is possibly failing?
Error: [Cron_sabilling_answernet_homesabillingjobsrun_check_aging.sh.msg](/uploads/d75de2e1c889607e4984cff243a7fb53/Cron_sabilling_answernet_homesabillingjobsrun_check_aging.sh.msg)
* My Thought: Judging from the amount of sites and accounts that appear to show bad AR, it would seem we still have an AR issue?
Error: [Cron_sabilling_answernet_homesabillingjobsrun_daily_ar_snapshot.sh.zip](/uploads/df922790b5dfb51fd32a402cae7ed005/Cron_sabilling_answernet_homesabillingjobsrun_daily_ar_snapshot.sh.zip)
* My Thought: Judging from the length of the email and that I am unable to understand what anything in it means, I need clarification on what to look for. (NOTE: I had to compress this email since it was over 20MB and gitlab would not upload it.)
Error: [SA_Billing_Error_Report_invoices_index__MongoidErrorsDocumentNotFound__nProblemn__Document_s__not_found_for_cl....msg](/uploads/64d8bd2b16996227daaac9f7a18e7653/SA_Billing_Error_Report_invoices_index__MongoidErrorsDocumentNotFound__nProblemn__Document_s__not_found_for_cl....msg)
*My Thought: It looks like an account is missing or some part of an account is missing?
Error: [SA_Billing_Error_Report_accounts_edit__MongoidErrorsDocumentNotFound__nProblemn__Document_s__not_found_for_cla....msg](/uploads/6a4cb7fd69c672d66c7e365b3fe50c0a/SA_Billing_Error_Report_accounts_edit__MongoidErrorsDocumentNotFound__nProblemn__Document_s__not_found_for_cla....msg)
* My Thoughts: Looks like someone tried to edit an account and was unable to as the account is missing? Maybe a cache issue?
Error: [SA_Billing_Error_Report_billing_cycles_upload__ArgumentError__invalid_byte_sequence_in_UTF-8.msg](/uploads/313e1c395608afbb7d1adf66b5cda4a9/SA_Billing_Error_Report_billing_cycles_upload__ArgumentError__invalid_byte_sequence_in_UTF-8.msg)
* My Thought: Looks like someone attempted to upload a invalid formatted usage file. (NOTE: There a like 5 of these emails consecutively.)
Error: [SA_Billing_Error_Report_side_nav_pull_customers__MongoidErrorsDocumentNotFound__nProblemn__Document_s__not_fou....msg](/uploads/b081ef4fc7debd0b34aa2ce468e206b2/SA_Billing_Error_Report_side_nav_pull_customers__MongoidErrorsDocumentNotFound__nProblemn__Document_s__not_fou....msg)
* My Thought: Looks like the drop down is displaying a link to an invalid site?
|
process
|
last few error emails concern in gitlab by kdjstudios on jan hello team here are the last few error emails we have been receiving may someone please explain them to me i have my thoughts on what each one means error uploads cron root answernet binbash homesabillingmongodb backupsbackup mongodb sh msg my thought judging from the amount backup command not found it would seem we are not backing things up error uploads cron sabilling answernet binbash optsabillingdelete customer without account sh msg my thought judging from the lack of success or completed message it would appear this script is possibly failing error uploads cron sabilling answernet homesabillingjobsrun check aging sh msg my thought judging from the amount of sites and accounts that appear to show bad ar it would seem we still have an ar issue error uploads cron sabilling answernet homesabillingjobsrun daily ar snapshot sh zip my thought judging from the length of the email and that i am unable to understand what anything in it means i need clarification on what to look for note i had to compress this email since it was over and gitlab would not upload it error uploads sa billing error report invoices index mongoiderrorsdocumentnotfound nproblemn document s not found for cl msg my thought it looks like an account is missing or some part of an account is missing error uploads sa billing error report accounts edit mongoiderrorsdocumentnotfound nproblemn document s not found for cla msg my thoughts looks like someone tried to edit an account and was unable to as the account is missing maybe a cache issue error uploads sa billing error report billing cycles upload argumenterror invalid byte sequence in utf msg my thought looks like someone attempted to upload a invalid formatted usage file note there a like of these emails consecutively error uploads sa billing error report side nav pull customers mongoiderrorsdocumentnotfound nproblemn document s not fou msg my thought looks like the drop down is displaying a link to an invalid site
| 1
|
21,695
| 30,190,936,253
|
IssuesEvent
|
2023-07-04 15:19:30
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
[MLv2] [Bug] `replace-clause` doesn't handle metrics
|
.Regression/master .metabase-lib .Team/QueryProcessor :hammer_and_wrench:
|
When trying to replace a regular aggregation with a metric, `replace-clause` fails with the following error:
```js
core.cljs:1942 Uncaught (in promise) Error: nth not supported on this type cljs.core/PersistentHashMap
at Function.cljs$core$IFn$_invoke$arity$3 (core.cljs:1942:1)
at convert.cljc:428:1
at core.cljs:4416:1
at Function.G__30280__2 [as cljs$core$IFn$_invoke$arity$2] (core.cljs:4416:1)
at core.cljs:5686:1
at core.cljs:5686:1
at Object.cljs$core$IReduce$_reduce$arity$3 (core.cljs:5690:1)
at Function.cljs$core$IFn$_invoke$arity$3 (core.cljs:2570:1)
at Function.cljs$core$IFn$_invoke$arity$4 (core.cljs:2626:1)
at Function.cljs$core$IFn$_invoke$arity$3 (core.cljs:5275:1)
```
### Steps to reproduce
1. Sign in as admin
2. Create a metric for the sample Orders table at `/admin/datamodel/metrics`
3. Start a question from the Orders table
4. Add an aggregation like Count
5. Try to replace the Count aggregation with the metric from step 2
Code:
```js
import * as Lib from "metabase-lib"
const aggregation = Lib.aggregations(query, 0)[0]
const metric = Lib.availableMetrics(query, 0)[0]
const query2 = Lib.replaceClause(query, 0, aggregation, metric)
// The error shows up when trying to convert this back to MLv1 format:
Lib.toLegacyQuery(query2)
```
|
1.0
|
[MLv2] [Bug] `replace-clause` doesn't handle metrics - When trying to replace a regular aggregation with a metric, `replace-clause` fails with the following error:
```js
core.cljs:1942 Uncaught (in promise) Error: nth not supported on this type cljs.core/PersistentHashMap
at Function.cljs$core$IFn$_invoke$arity$3 (core.cljs:1942:1)
at convert.cljc:428:1
at core.cljs:4416:1
at Function.G__30280__2 [as cljs$core$IFn$_invoke$arity$2] (core.cljs:4416:1)
at core.cljs:5686:1
at core.cljs:5686:1
at Object.cljs$core$IReduce$_reduce$arity$3 (core.cljs:5690:1)
at Function.cljs$core$IFn$_invoke$arity$3 (core.cljs:2570:1)
at Function.cljs$core$IFn$_invoke$arity$4 (core.cljs:2626:1)
at Function.cljs$core$IFn$_invoke$arity$3 (core.cljs:5275:1)
```
### Steps to reproduce
1. Sign in as admin
2. Create a metric for the sample Orders table at `/admin/datamodel/metrics`
3. Start a question from the Orders table
4. Add an aggregation like Count
5. Try to replace the Count aggregation with the metric from step 2
Code:
```js
import * as Lib from "metabase-lib"
const aggregation = Lib.aggregations(query, 0)[0]
const metric = Lib.availableMetrics(query, 0)[0]
const query2 = Lib.replaceClause(query, 0, aggregation, metric)
// The error shows up when trying to convert this back to MLv1 format:
Lib.toLegacyQuery(query2)
```
|
process
|
replace clause doesn t handle metrics when trying to replace a regular aggregation with a metric replace clause fails with the following error js core cljs uncaught in promise error nth not supported on this type cljs core persistenthashmap at function cljs core ifn invoke arity core cljs at convert cljc at core cljs at function g core cljs at core cljs at core cljs at object cljs core ireduce reduce arity core cljs at function cljs core ifn invoke arity core cljs at function cljs core ifn invoke arity core cljs at function cljs core ifn invoke arity core cljs steps to reproduce sign in as admin create a metric for the sample orders table at admin datamodel metrics start a question from the orders table add an aggregation like count try to replace the count aggregation with the metric from step code js import as lib from metabase lib const aggregation lib aggregations query const metric lib availablemetrics query const lib replaceclause query aggregation metric the error shows up when trying to convert this back to format lib tolegacyquery
| 1
|
20,441
| 27,100,573,726
|
IssuesEvent
|
2023-02-15 08:19:02
|
billingran/Newsletter
|
https://api.github.com/repos/billingran/Newsletter
|
closed
|
Validation de l'adresse email d'un fichier CSV
|
processing... Brief 2
|
- [ ] Mettre à jour en conséquence la procédure d'import des données à partir d'un fichier CSV et éviter d'importer un email existe déjà.
|
1.0
|
Validation de l'adresse email d'un fichier CSV - - [ ] Mettre à jour en conséquence la procédure d'import des données à partir d'un fichier CSV et éviter d'importer un email existe déjà.
|
process
|
validation de l adresse email d un fichier csv mettre à jour en conséquence la procédure d import des données à partir d un fichier csv et éviter d importer un email existe déjà
| 1
|
18,069
| 24,084,244,476
|
IssuesEvent
|
2022-09-19 09:28:19
|
Tencent/tdesign-miniprogram
|
https://api.github.com/repos/Tencent/tdesign-miniprogram
|
closed
|
[Upload] Need support wx.chooseMessageFile
|
enhancement in process
|
### 这个功能解决了什么问题
Need support wx.chooseMessageFile
### 你建议的方案是什么
Need support wx.chooseMessageFile
|
1.0
|
[Upload] Need support wx.chooseMessageFile - ### 这个功能解决了什么问题
Need support wx.chooseMessageFile
### 你建议的方案是什么
Need support wx.chooseMessageFile
|
process
|
need support wx choosemessagefile 这个功能解决了什么问题 need support wx choosemessagefile 你建议的方案是什么 need support wx choosemessagefile
| 1
|
179,455
| 14,704,647,528
|
IssuesEvent
|
2021-01-04 16:48:51
|
SketchUp/api-issue-tracker
|
https://api.github.com/repos/SketchUp/api-issue-tracker
|
closed
|
Sketchup::PickHelper class getter by index methods (5) have incorrect iterator code examples
|
Ruby API SketchUp documentation
|
#### SketchUp Ruby API Documentation Issue
**ALL** of the iterator examples using getter methods via index in the [`Sketchup::PickHelper`](https://ruby.sketchup.com/Sketchup/PickHelper.html) class have **erroneous** code examples.
These getters are:
* [`Sketchup::PickHelper#depth_at()`](https://ruby.sketchup.com/Sketchup/PickHelper.html#depth_at-instance_method)
* [`Sketchup::PickHelper#element_at()`](https://ruby.sketchup.com/Sketchup/PickHelper.html#element_at-instance_method)
* [`Sketchup::PickHelper#leaf_at`](https://ruby.sketchup.com/Sketchup/PickHelper.html#leaf_at-instance_method)
* [`Sketchup::PickHelper#path_at()`](https://ruby.sketchup.com/Sketchup/PickHelper.html#path_at-instance_method)
* [`Sketchup::PickHelper#transformation_at`](https://ruby.sketchup.com/Sketchup/PickHelper.html#transformation_at-instance_method)
The error is repeated in each method docstring:
1. The preamble code creates a `ph` reference, but the loop afterward uses a `pickhelper` reference.
Apparently these examples were copy and pasted without realizing the errors.
----
**Incorrect** example (from [`#path_at()`](https://ruby.sketchup.com/Sketchup/PickHelper.html#path_at-instance_method))...
```ruby
ph = view.pick_helper
ph.do_pick(x, y)
# Iterate all pick-routes:
pickhelper.count.times { |pick_path_index|
p pickhelper.path_at(pick_path_index)
}
```
A **correct** example:
```ruby
ph = view.pick_helper
ph.do_pick(x, y)
# Iterate all pick-routes and collect them:
pickpaths = []
ph.count.times do |pick_path_index|
pickpaths << ph.path_at(pick_path_index)
end
```
~
|
1.0
|
Sketchup::PickHelper class getter by index methods (5) have incorrect iterator code examples - #### SketchUp Ruby API Documentation Issue
**ALL** of the iterator examples using getter methods via index in the [`Sketchup::PickHelper`](https://ruby.sketchup.com/Sketchup/PickHelper.html) class have **erroneous** code examples.
These getters are:
* [`Sketchup::PickHelper#depth_at()`](https://ruby.sketchup.com/Sketchup/PickHelper.html#depth_at-instance_method)
* [`Sketchup::PickHelper#element_at()`](https://ruby.sketchup.com/Sketchup/PickHelper.html#element_at-instance_method)
* [`Sketchup::PickHelper#leaf_at`](https://ruby.sketchup.com/Sketchup/PickHelper.html#leaf_at-instance_method)
* [`Sketchup::PickHelper#path_at()`](https://ruby.sketchup.com/Sketchup/PickHelper.html#path_at-instance_method)
* [`Sketchup::PickHelper#transformation_at`](https://ruby.sketchup.com/Sketchup/PickHelper.html#transformation_at-instance_method)
The error is repeated in each method docstring:
1. The preamble code creates a `ph` reference, but the loop afterward uses a `pickhelper` reference.
Apparently these examples were copy and pasted without realizing the errors.
----
**Incorrect** example (from [`#path_at()`](https://ruby.sketchup.com/Sketchup/PickHelper.html#path_at-instance_method))...
```ruby
ph = view.pick_helper
ph.do_pick(x, y)
# Iterate all pick-routes:
pickhelper.count.times { |pick_path_index|
p pickhelper.path_at(pick_path_index)
}
```
A **correct** example:
```ruby
ph = view.pick_helper
ph.do_pick(x, y)
# Iterate all pick-routes and collect them:
pickpaths = []
ph.count.times do |pick_path_index|
pickpaths << ph.path_at(pick_path_index)
end
```
~
|
non_process
|
sketchup pickhelper class getter by index methods have incorrect iterator code examples sketchup ruby api documentation issue all of the iterator examples using getter methods via index in the class have erroneous code examples these getters are the error is repeated in each method docstring the preamble code creates a ph reference but the loop afterward uses a pickhelper reference apparently these examples were copy and pasted without realizing the errors incorrect example from ruby ph view pick helper ph do pick x y iterate all pick routes pickhelper count times pick path index p pickhelper path at pick path index a correct example ruby ph view pick helper ph do pick x y iterate all pick routes and collect them pickpaths ph count times do pick path index pickpaths ph path at pick path index end
| 0
|
714,593
| 24,567,147,774
|
IssuesEvent
|
2022-10-13 04:53:12
|
AY2223S1-CS2103T-T09-4/tp
|
https://api.github.com/repos/AY2223S1-CS2103T-T09-4/tp
|
closed
|
As a tutor who has many students, I can check the students' phone number
|
type.Story priority.HIGH
|
... so that I can contact them.
|
1.0
|
As a tutor who has many students, I can check the students' phone number - ... so that I can contact them.
|
non_process
|
as a tutor who has many students i can check the students phone number so that i can contact them
| 0
|
702,334
| 24,120,889,287
|
IssuesEvent
|
2022-09-20 18:36:03
|
googleapis/nodejs-dns
|
https://api.github.com/repos/googleapis/nodejs-dns
|
closed
|
dns: "before all" hook for "should return 0 or more zones" failed
|
priority: p1 type: bug api: dns flakybot: issue
|
Note: #480 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: b3a77f071df17739c2fcd682f181ac5bdddffae6
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/c06c4077-6671-438f-a7bc-5ddb8ecc02bf), [Sponge](http://sponge2/c06c4077-6671-438f-a7bc-5ddb8ecc02bf)
status: failed
<details><summary>Test output</summary><br><pre>invalid_grant: Invalid JWT Signature.
Error: invalid_grant: Invalid JWT Signature.
at Gaxios._request (node_modules/gaxios/build/src/gaxios.js:130:23)
-> /workspace/node_modules/gaxios/src/gaxios.ts:158:15
at processTicksAndRejections (internal/process/task_queues.js:97:5)
at async GoogleToken.requestToken (node_modules/gtoken/build/src/index.js:238:23)
at async GoogleToken.getTokenAsync (node_modules/gtoken/build/src/index.js:145:20)
at async JWT.refreshTokenNoCache (node_modules/google-auth-library/build/src/auth/jwtclient.js:172:23)
at async JWT.getRequestMetadataAsync (node_modules/google-auth-library/build/src/auth/oauth2client.js:284:17)
at async JWT.getRequestHeaders (node_modules/google-auth-library/build/src/auth/oauth2client.js:247:26)
at async GoogleAuth.authorizeRequest (node_modules/google-auth-library/build/src/auth/googleauth.js:677:25)
at async Promise.all (index 1)
at async prepareRequest (node_modules/@google-cloud/common/build/src/util.js:442:61)</pre></details>
|
1.0
|
dns: "before all" hook for "should return 0 or more zones" failed - Note: #480 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: b3a77f071df17739c2fcd682f181ac5bdddffae6
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/c06c4077-6671-438f-a7bc-5ddb8ecc02bf), [Sponge](http://sponge2/c06c4077-6671-438f-a7bc-5ddb8ecc02bf)
status: failed
<details><summary>Test output</summary><br><pre>invalid_grant: Invalid JWT Signature.
Error: invalid_grant: Invalid JWT Signature.
at Gaxios._request (node_modules/gaxios/build/src/gaxios.js:130:23)
-> /workspace/node_modules/gaxios/src/gaxios.ts:158:15
at processTicksAndRejections (internal/process/task_queues.js:97:5)
at async GoogleToken.requestToken (node_modules/gtoken/build/src/index.js:238:23)
at async GoogleToken.getTokenAsync (node_modules/gtoken/build/src/index.js:145:20)
at async JWT.refreshTokenNoCache (node_modules/google-auth-library/build/src/auth/jwtclient.js:172:23)
at async JWT.getRequestMetadataAsync (node_modules/google-auth-library/build/src/auth/oauth2client.js:284:17)
at async JWT.getRequestHeaders (node_modules/google-auth-library/build/src/auth/oauth2client.js:247:26)
at async GoogleAuth.authorizeRequest (node_modules/google-auth-library/build/src/auth/googleauth.js:677:25)
at async Promise.all (index 1)
at async prepareRequest (node_modules/@google-cloud/common/build/src/util.js:442:61)</pre></details>
|
non_process
|
dns before all hook for should return or more zones failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output invalid grant invalid jwt signature error invalid grant invalid jwt signature at gaxios request node modules gaxios build src gaxios js workspace node modules gaxios src gaxios ts at processticksandrejections internal process task queues js at async googletoken requesttoken node modules gtoken build src index js at async googletoken gettokenasync node modules gtoken build src index js at async jwt refreshtokennocache node modules google auth library build src auth jwtclient js at async jwt getrequestmetadataasync node modules google auth library build src auth js at async jwt getrequestheaders node modules google auth library build src auth js at async googleauth authorizerequest node modules google auth library build src auth googleauth js at async promise all index at async preparerequest node modules google cloud common build src util js
| 0
|
441,551
| 30,789,102,281
|
IssuesEvent
|
2023-07-31 15:00:12
|
software-mansion/scarb
|
https://api.github.com/repos/software-mansion/scarb
|
closed
|
Move roadmap from private google doc to readme of GH Discussion
|
documentation
|
It's problematic for organizational purposes that the roadmap mostly lies in @mkaput head and has only been coded several months ago in a private Google Doc. Let's make it public & also revisit it (make it up to date). See what Foundry did.
|
1.0
|
Move roadmap from private google doc to readme of GH Discussion - It's problematic for organizational purposes that the roadmap mostly lies in @mkaput head and has only been coded several months ago in a private Google Doc. Let's make it public & also revisit it (make it up to date). See what Foundry did.
|
non_process
|
move roadmap from private google doc to readme of gh discussion it s problematic for organizational purposes that the roadmap mostly lies in mkaput head and has only been coded several months ago in a private google doc let s make it public also revisit it make it up to date see what foundry did
| 0
|
16,777
| 4,087,431,401
|
IssuesEvent
|
2016-06-01 10:01:45
|
DevExpress/testcafe
|
https://api.github.com/repos/DevExpress/testcafe
|
closed
|
Make default offsets for mouse actions equal to the center of element
|
!IMPORTANT! API MODIFICATION LEVEL: non-breaking changes AREA: client DOCUMENTATION: not required SYSTEM: automations TYPE: bug
|
Test without offsets
`t.click('#element')`
should click on the center (like in previous API)
|
1.0
|
Make default offsets for mouse actions equal to the center of element - Test without offsets
`t.click('#element')`
should click on the center (like in previous API)
|
non_process
|
make default offsets for mouse actions equal to the center of element test without offsets t click element should click on the center like in previous api
| 0
|
16,973
| 22,335,282,396
|
IssuesEvent
|
2022-06-14 17:57:47
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Enable single file export option for algorithms that export Atlas layouts in processing
|
Processing Feature Request
|
### Feature description
In layout, there is the option to either export layouts as a single output or one per page.

In the processing algorithms to export atlas layout, this is not exposed. It would be nice to add an option for users to be able to select if they need this
#44242
### Additional context
_No response_
|
1.0
|
Enable single file export option for algorithms that export Atlas layouts in processing - ### Feature description
In layout, there is the option to either export layouts as a single output or one per page.

In the processing algorithms to export atlas layout, this is not exposed. It would be nice to add an option for users to be able to select if they need this
#44242
### Additional context
_No response_
|
process
|
enable single file export option for algorithms that export atlas layouts in processing feature description in layout there is the option to either export layouts as a single output or one per page in the processing algorithms to export atlas layout this is not exposed it would be nice to add an option for users to be able to select if they need this additional context no response
| 1
|
18,503
| 24,551,268,133
|
IssuesEvent
|
2022-10-12 12:47:42
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[iOS] Closed study > UI issue
|
Bug P2 iOS Process: Fixed Process: Tested dev
|
AR: Closed study > UI issue > 'Cancel' button is not placed properly
ER: Cancel button should get placed vertically and horizontally center of the white layout

|
2.0
|
[iOS] Closed study > UI issue - AR: Closed study > UI issue > 'Cancel' button is not placed properly
ER: Cancel button should get placed vertically and horizontally center of the white layout

|
process
|
closed study ui issue ar closed study ui issue cancel button is not placed properly er cancel button should get placed vertically and horizontally center of the white layout
| 1
|
16,547
| 21,568,599,019
|
IssuesEvent
|
2022-05-02 04:17:56
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
Add Treasure Chest
|
suggested title in process
|
Title: Treasure Chest
Type (film/tv show): TV show
Film or show in which it appears: Monk
Is the parent film/show streaming anywhere? Amazon Prime, Peacock
About when in the parent film/show does it appear? Season 3, Episode 8, "Mr. Monk and the Game Show"
Actual footage of the film/show can be seen (yes/no)? Yes
Cast:
Roddy Lankman (host)
Lizzie Talvo (Lankman's assistant)
Producer: Dwight Ellison
Known contestants: Val Birch, Adrian Monk
Link to Monk wiki: https://monk.fandom.com/wiki/Mr._Monk_and_the_Game_Show
I'm not sure if game shows are allowed
|
1.0
|
Add Treasure Chest - Title: Treasure Chest
Type (film/tv show): TV show
Film or show in which it appears: Monk
Is the parent film/show streaming anywhere? Amazon Prime, Peacock
About when in the parent film/show does it appear? Season 3, Episode 8, "Mr. Monk and the Game Show"
Actual footage of the film/show can be seen (yes/no)? Yes
Cast:
Roddy Lankman (host)
Lizzie Talvo (Lankman's assistant)
Producer: Dwight Ellison
Known contestants: Val Birch, Adrian Monk
Link to Monk wiki: https://monk.fandom.com/wiki/Mr._Monk_and_the_Game_Show
I'm not sure if game shows are allowed
|
process
|
add treasure chest title treasure chest type film tv show tv show film or show in which it appears monk is the parent film show streaming anywhere amazon prime peacock about when in the parent film show does it appear season episode mr monk and the game show actual footage of the film show can be seen yes no yes cast roddy lankman host lizzie talvo lankman s assistant producer dwight ellison known contestants val birch adrian monk link to monk wiki i m not sure if game shows are allowed
| 1
|
9,531
| 12,501,060,764
|
IssuesEvent
|
2020-06-02 00:05:48
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
[modeler] Any command with "." (e.g. grass') can't be used in expressions
|
Bug Processing
|
**Describe the bug**
In modeler, if a grass algorithm is used, it's output can't be used in expression generating syntax expression error
**How to Reproduce**
1. Chosse a simple grass alg (e.g. r.cost)
2. create a model with inputs to feed the alg
3. add the grass alg and feed it with input from model
4. Add a new alg to the model having as input an expression getting ouput of previous grass alg (e.g. @r.cost_output) => expression syntax error due to "."
**QGIS and OS versions**
QGIS version | 3.13.0-Master | QGIS code revision | 1435076d2f
-- | -- | -- | --
Compiled against Qt | 5.9.5 | Running against Qt | 5.9.5
Compiled against GDAL/OGR | 2.2.3 | Running against GDAL/OGR | 2.2.3
Compiled against GEOS | 3.7.1-CAPI-1.11.1 | Running against GEOS | 3.7.1-CAPI-1.11.1 27a5e771
Compiled against SQLite | 3.22.0 | Running against SQLite | 3.22.0
PostgreSQL Client Version | 12.2 (Ubuntu 12.2-2.pgdg18.04+1) | SpatiaLite Version | 4.3.0a
QWT Version | 6.1.3 | QScintilla2 Version | 2.10.2
PROJ.4 Version | 493
OS Version | Ubuntu 18.04.4 LTS | This copy of QGIS writes debugging output.
Active python plugins | processing_whitebox; plugin_reloader; processing_taudem; quick_map_services; galiciasustentable; SpreadsheetLayers; firstaid; IPyConsole; qconsolidate; excel_sync; db_manager; processing; MetaSearch
|
1.0
|
[modeler] Any command with "." (e.g. grass') can't be used in expressions - **Describe the bug**
In modeler, if a grass algorithm is used, it's output can't be used in expression generating syntax expression error
**How to Reproduce**
1. Chosse a simple grass alg (e.g. r.cost)
2. create a model with inputs to feed the alg
3. add the grass alg and feed it with input from model
4. Add a new alg to the model having as input an expression getting ouput of previous grass alg (e.g. @r.cost_output) => expression syntax error due to "."
**QGIS and OS versions**
QGIS version | 3.13.0-Master | QGIS code revision | 1435076d2f
-- | -- | -- | --
Compiled against Qt | 5.9.5 | Running against Qt | 5.9.5
Compiled against GDAL/OGR | 2.2.3 | Running against GDAL/OGR | 2.2.3
Compiled against GEOS | 3.7.1-CAPI-1.11.1 | Running against GEOS | 3.7.1-CAPI-1.11.1 27a5e771
Compiled against SQLite | 3.22.0 | Running against SQLite | 3.22.0
PostgreSQL Client Version | 12.2 (Ubuntu 12.2-2.pgdg18.04+1) | SpatiaLite Version | 4.3.0a
QWT Version | 6.1.3 | QScintilla2 Version | 2.10.2
PROJ.4 Version | 493
OS Version | Ubuntu 18.04.4 LTS | This copy of QGIS writes debugging output.
Active python plugins | processing_whitebox; plugin_reloader; processing_taudem; quick_map_services; galiciasustentable; SpreadsheetLayers; firstaid; IPyConsole; qconsolidate; excel_sync; db_manager; processing; MetaSearch
|
process
|
any command with e g grass can t be used in expressions describe the bug in modeler if a grass algorithm is used it s output can t be used in expression generating syntax expression error how to reproduce chosse a simple grass alg e g r cost create a model with inputs to feed the alg add the grass alg and feed it with input from model add a new alg to the model having as input an expression getting ouput of previous grass alg e g r cost output expression syntax error due to qgis and os versions qgis version master qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version ubuntu spatialite version qwt version version proj version os version ubuntu lts this copy of qgis writes debugging output active python plugins processing whitebox plugin reloader processing taudem quick map services galiciasustentable spreadsheetlayers firstaid ipyconsole qconsolidate excel sync db manager processing metasearch
| 1
|
99,854
| 4,073,717,268
|
IssuesEvent
|
2016-05-28 00:17:13
|
ampproject/amphtml
|
https://api.github.com/repos/ampproject/amphtml
|
closed
|
Re-attached elements are being built again
|
Priority: High Related to: AMP Core
|
Follow up to #3354
We sometimes re-parent children to wrap them in a container (e.g. Carousel, Lightbox). But those children may have already been built. But since we fire another `attachedCallback`, we'll add them to the unbuilts again, and try to build again. It's cool to add them to the list again, but we should make sure they're really unbuilt before trying to #build them again.
I believe we're already doing this in `ElementProto.build` method.
```javascript
ElementProto.build = function() {
// ...
if (this.isBuilt()) {
return;
}
// ...
}
```
@jridgewell should this be enough? or were you suggesting avoid adding them to the pool all together?
|
1.0
|
Re-attached elements are being built again - Follow up to #3354
We sometimes re-parent children to wrap them in a container (e.g. Carousel, Lightbox). But those children may have already been built. But since we fire another `attachedCallback`, we'll add them to the unbuilts again, and try to build again. It's cool to add them to the list again, but we should make sure they're really unbuilt before trying to #build them again.
I believe we're already doing this in `ElementProto.build` method.
```javascript
ElementProto.build = function() {
// ...
if (this.isBuilt()) {
return;
}
// ...
}
```
@jridgewell should this be enough? or were you suggesting avoid adding them to the pool all together?
|
non_process
|
re attached elements are being built again follow up to we sometimes re parent children to wrap them in a container e g carousel lightbox but those children may have already been built but since we fire another attachedcallback we ll add them to the unbuilts again and try to build again it s cool to add them to the list again but we should make sure they re really unbuilt before trying to build them again i believe we re already doing this in elementproto build method javascript elementproto build function if this isbuilt return jridgewell should this be enough or were you suggesting avoid adding them to the pool all together
| 0
|
135,768
| 30,440,800,239
|
IssuesEvent
|
2023-07-15 03:22:52
|
creativecommons/creativecommons.org
|
https://api.github.com/repos/creativecommons/creativecommons.org
|
opened
|
[Bug] Don’t use hyphens in place of en dashes in legal texts (partially subjective)
|
🟧 priority: high 🚦 status: awaiting triage 🛠 goal: fix 💻 aspect: code
|
## Description
All CC licenses and CC0 1.0 (not CC-PDDL or PDM 1.0) use a hyphen (-) in some places where an en dash (–) might/should be used instead.
The word “lawyer-client” in the “not a law firm” disclaimer should use an en dash, since it mentions two different objects instead of being a single, compound word: “lawyer–client.” (This is not part of the legal code.)
In addition, an en dash is often used to denote a range, e.g. 1–3 or A–D. The legal texts use a hyphen for this instead, e.g. 1-3 or A-D. This one is subjective; the APA and MLA style guides, for example, use an en dash to show ranges, while AMA style uses hyphens. If en dashes were used instead, ranges like “Section 2(b)(1)-(2)” would become “Section 2(b)(1)–(2).” Hyphens might still be preferred in plaintext, in any case.
## Additional context
See [Dash § En dash on English Wikipedia](https://en.wikipedia.org/wiki/Dash#En_dash)
## Resolution
- [ ] I would be interested in resolving this bug. <!-- I can’t fix this -->
|
1.0
|
[Bug] Don’t use hyphens in place of en dashes in legal texts (partially subjective) - ## Description
All CC licenses and CC0 1.0 (not CC-PDDL or PDM 1.0) use a hyphen (-) in some places where an en dash (–) might/should be used instead.
The word “lawyer-client” in the “not a law firm” disclaimer should use an en dash, since it mentions two different objects instead of being a single, compound word: “lawyer–client.” (This is not part of the legal code.)
In addition, an en dash is often used to denote a range, e.g. 1–3 or A–D. The legal texts use a hyphen for this instead, e.g. 1-3 or A-D. This one is subjective; the APA and MLA style guides, for example, use an en dash to show ranges, while AMA style uses hyphens. If en dashes were used instead, ranges like “Section 2(b)(1)-(2)” would become “Section 2(b)(1)–(2).” Hyphens might still be preferred in plaintext, in any case.
## Additional context
See [Dash § En dash on English Wikipedia](https://en.wikipedia.org/wiki/Dash#En_dash)
## Resolution
- [ ] I would be interested in resolving this bug. <!-- I can’t fix this -->
|
non_process
|
don’t use hyphens in place of en dashes in legal texts partially subjective description all cc licenses and not cc pddl or pdm use a hyphen in some places where an en dash – might should be used instead the word “lawyer client” in the “not a law firm” disclaimer should use an en dash since it mentions two different objects instead of being a single compound word “lawyer–client ” this is not part of the legal code in addition an en dash is often used to denote a range e g – or a–d the legal texts use a hyphen for this instead e g or a d this one is subjective the apa and mla style guides for example use an en dash to show ranges while ama style uses hyphens if en dashes were used instead ranges like “section b ” would become “section b – ” hyphens might still be preferred in plaintext in any case additional context see resolution i would be interested in resolving this bug
| 0
|
434,619
| 30,458,503,687
|
IssuesEvent
|
2023-07-17 03:43:16
|
quantified-uncertainty/squiggle
|
https://api.github.com/repos/quantified-uncertainty/squiggle
|
opened
|
Documentation improvement round
|
Documentation
|
- x Concerns documentation
# Description of suggestion or shortcoming:
Spend 2-3 days improving the documentation
|
1.0
|
Documentation improvement round - - x Concerns documentation
# Description of suggestion or shortcoming:
Spend 2-3 days improving the documentation
|
non_process
|
documentation improvement round x concerns documentation description of suggestion or shortcoming spend days improving the documentation
| 0
|
789
| 3,273,292,649
|
IssuesEvent
|
2015-10-26 00:12:30
|
beesmart-it/trend-hrm
|
https://api.github.com/repos/beesmart-it/trend-hrm
|
closed
|
Add actions column to processes table
|
enhancement selection process
|
Add column to processes table with actions:
-View process details
-Access process
-Edit process
-Update process status (stop, start, close, cancel)
|
1.0
|
Add actions column to processes table - Add column to processes table with actions:
-View process details
-Access process
-Edit process
-Update process status (stop, start, close, cancel)
|
process
|
add actions column to processes table add column to processes table with actions view process details access process edit process update process status stop start close cancel
| 1
|
69,970
| 22,772,628,990
|
IssuesEvent
|
2022-07-08 11:32:45
|
primefaces/primereact
|
https://api.github.com/repos/primefaces/primereact
|
closed
|
Hooks: useImperativeHandle is preventing the default behaviour of useRef hook
|
defect 👍 confirmed discussion
|
### Describe the bug
Since the transition from Class components to Hook based components, the access to the methods of a component has been handled with the `useImperativeHandle` hook. [Example 1](https://github.com/primefaces/primereact/blob/63cf7185d2fe2922a30f24c6f7afe8559b1ef5bc/components/lib/datatable/DataTable.js#L1307), [Example 2](https://github.com/primefaces/primereact/blob/63cf7185d2fe2922a30f24c6f7afe8559b1ef5bc/components/lib/autocomplete/AutoComplete.js#L453), etc.
This hook, instead of adding methods to the ref of components, has been replacing the ref completly, blocking the default behaivour of the `useRef` hook, accesing the dom element. In V7 we would also get access to the component props which gives the developer a lot of flexibility and useful state data.
This issue needs two codesandboxes (SEE CONSOLE LOG):
V7: https://codesandbox.io/s/exciting-sound-6xncl2?file=/src/demo/AutoCompleteDemo.js
V8: https://codesandbox.io/s/nervous-shtern-fyhs9h?file=/src/demo/AutoCompleteDemo.js
### Reproducer
https://codesandbox.io/s/nervous-shtern-fyhs9h?file=/src/demo/AutoCompleteDemo.js

### PrimeReact version
8.1.0
### React version
18.x
### Language
ALL
### Build / Runtime
Create React App (CRA)
### Browser(s)
_No response_
### Steps to reproduce the behavior
1. Create an Autocomplete component.
2. Create a const with `useRef` hook.
3. Assign the const to the ref value of the Autocomplete component.
4. Console log the `ref.current` with a `useEffect` hook.
5. See no DOM element nor props.
### Expected behavior
When using a `useRef` hook and assigning the ref to a component, I expect to get access to the dom element (and its props, having used primereact V7).
|
1.0
|
Hooks: useImperativeHandle is preventing the default behaviour of useRef hook - ### Describe the bug
Since the transition from Class components to Hook based components, the access to the methods of a component has been handled with the `useImperativeHandle` hook. [Example 1](https://github.com/primefaces/primereact/blob/63cf7185d2fe2922a30f24c6f7afe8559b1ef5bc/components/lib/datatable/DataTable.js#L1307), [Example 2](https://github.com/primefaces/primereact/blob/63cf7185d2fe2922a30f24c6f7afe8559b1ef5bc/components/lib/autocomplete/AutoComplete.js#L453), etc.
This hook, instead of adding methods to the ref of components, has been replacing the ref completly, blocking the default behaivour of the `useRef` hook, accesing the dom element. In V7 we would also get access to the component props which gives the developer a lot of flexibility and useful state data.
This issue needs two codesandboxes (SEE CONSOLE LOG):
V7: https://codesandbox.io/s/exciting-sound-6xncl2?file=/src/demo/AutoCompleteDemo.js
V8: https://codesandbox.io/s/nervous-shtern-fyhs9h?file=/src/demo/AutoCompleteDemo.js
### Reproducer
https://codesandbox.io/s/nervous-shtern-fyhs9h?file=/src/demo/AutoCompleteDemo.js

### PrimeReact version
8.1.0
### React version
18.x
### Language
ALL
### Build / Runtime
Create React App (CRA)
### Browser(s)
_No response_
### Steps to reproduce the behavior
1. Create an Autocomplete component.
2. Create a const with `useRef` hook.
3. Assign the const to the ref value of the Autocomplete component.
4. Console log the `ref.current` with a `useEffect` hook.
5. See no DOM element nor props.
### Expected behavior
When using a `useRef` hook and assigning the ref to a component, I expect to get access to the dom element (and its props, having used primereact V7).
|
non_process
|
hooks useimperativehandle is preventing the default behaviour of useref hook describe the bug since the transition from class components to hook based components the access to the methods of a component has been handled with the useimperativehandle hook etc this hook instead of adding methods to the ref of components has been replacing the ref completly blocking the default behaivour of the useref hook accesing the dom element in we would also get access to the component props which gives the developer a lot of flexibility and useful state data this issue needs two codesandboxes see console log reproducer primereact version react version x language all build runtime create react app cra browser s no response steps to reproduce the behavior create an autocomplete component create a const with useref hook assign the const to the ref value of the autocomplete component console log the ref current with a useeffect hook see no dom element nor props expected behavior when using a useref hook and assigning the ref to a component i expect to get access to the dom element and its props having used primereact
| 0
|
208,712
| 7,157,646,535
|
IssuesEvent
|
2018-01-26 20:41:39
|
status-im/status-react
|
https://api.github.com/repos/status-im/status-react
|
closed
|
App version for release build is 0.9.14d3 instead of 0.9.13
|
high-priority release
|
### Description
[comment]: # (Feature or Bug? i.e Type: Bug)
*Type*: Bug
[comment]: # (Describe the feature you would like, or briefly summarise the bug and what you did, what you expected to happen, and what actually happens. Sections below)
*Summary*:
On latest release build 70:
iOS app version: 0.9.14d3
Android app version: 0.9.13-3-g848bcaea+

Note that on release build 67 versions were as expected (0.9.13)
#### Expected behavior
[comment]: # (Describe what you expected to happen.)
iOS app version: 0.9.13
Android app version: 0.9.13
#### Actual behavior
[comment]: # (Describe what actually happened.)
iOS app version: 0.9.14d3
Android app version: 0.9.13-3-g848bcaea+
### Reproduction
[comment]: # (Describe how we can replicate the bug step by step.)
- Install Status
- Check app version:
for Android: OS Settings
for iOS: General -> iPhone Storage -> Status
* Status version: release build 70
* Operating System: Android and iOS real devices
|
1.0
|
App version for release build is 0.9.14d3 instead of 0.9.13 -
### Description
[comment]: # (Feature or Bug? i.e Type: Bug)
*Type*: Bug
[comment]: # (Describe the feature you would like, or briefly summarise the bug and what you did, what you expected to happen, and what actually happens. Sections below)
*Summary*:
On latest release build 70:
iOS app version: 0.9.14d3
Android app version: 0.9.13-3-g848bcaea+

Note that on release build 67 versions were as expected (0.9.13)
#### Expected behavior
[comment]: # (Describe what you expected to happen.)
iOS app version: 0.9.13
Android app version: 0.9.13
#### Actual behavior
[comment]: # (Describe what actually happened.)
iOS app version: 0.9.14d3
Android app version: 0.9.13-3-g848bcaea+
### Reproduction
[comment]: # (Describe how we can replicate the bug step by step.)
- Install Status
- Check app version:
for Android: OS Settings
for iOS: General -> iPhone Storage -> Status
* Status version: release build 70
* Operating System: Android and iOS real devices
|
non_process
|
app version for release build is instead of description feature or bug i e type bug type bug describe the feature you would like or briefly summarise the bug and what you did what you expected to happen and what actually happens sections below summary on latest release build ios app version android app version note that on release build versions were as expected expected behavior describe what you expected to happen ios app version android app version actual behavior describe what actually happened ios app version android app version reproduction describe how we can replicate the bug step by step install status check app version for android os settings for ios general iphone storage status status version release build operating system android and ios real devices
| 0
|
93,484
| 10,764,918,533
|
IssuesEvent
|
2019-11-01 09:39:25
|
Xdecosee/ped
|
https://api.github.com/repos/Xdecosee/ped
|
opened
|
Archive does not store questions into pdf
|
severity.Low type.DocumentationBug
|
According to user guide at github, archive will directly create a pdf and store a question to it. But it does not until user uses print command. It would better if the instructions were clearer.

|
1.0
|
Archive does not store questions into pdf - According to user guide at github, archive will directly create a pdf and store a question to it. But it does not until user uses print command. It would better if the instructions were clearer.

|
non_process
|
archive does not store questions into pdf according to user guide at github archive will directly create a pdf and store a question to it but it does not until user uses print command it would better if the instructions were clearer
| 0
|
57,350
| 14,099,391,979
|
IssuesEvent
|
2020-11-06 01:16:12
|
storybookjs/storybook
|
https://api.github.com/repos/storybookjs/storybook
|
closed
|
Module not found error doesn't properly exit during build
|
P1 bug build-storybook has workaround mdx todo tracked
|
**Describe the bug**
When building storybook using `build-storybook` with module resolution errors in an `.mdx` file, a build error occurs but the build process hangs instead of exiting with a non-zero exit code. This is observed when using `@storybook/addons-docs` with a story using `.mdx` with an improper import.
This is problematic for our teams as we run this process during CI/CD pipelines and it hangs our builds if one of our developers makes an error in their `.mdx` files.
```
> storybook-build-hang@0.0.0 build-storybook M:\scratch\storybook-build-hang
> build-storybook
info @storybook/angular v5.3.18
info
info clean outputDir..
info => Copying prebuild dll's..
info => Building manager..
info => Loading manager config..
info => Loading presets
info => Compiling manager..
info => manager built (8.11 s)
info => Building preview..
info => Loading preview config..
info => Loading presets
info => Loading config/preview file in "./.storybook".
info => Adding stories defined in ".storybook\main.js".
info => Found custom tsconfig.json
info => Using default Webpack setup.
info => Using angular project 'storybook-build-hang' for configuring Storybook.
info => Loading angular-cli config.
info => Get angular-cli webpack config.
info => Compiling preview..
Starting type checking service...
Using 1 worker with 2048MB memory limit
70% building 40/90 modules 50 active M:\scratch\storybook-build-hang\node_modules\core-js\internals\set-to-string-tag.jsERR! => Failed to build the preview
ERR! Module not found: Error: Can't resolve './app.componen' in 'M:\scratch\storybook-build-hang\src\app'
(node:39788) UnhandledPromiseRejectionWarning: ModuleNotFoundError: Module not found: Error: Can't resolve './app.componen' in 'M:\scratch\storybook-build-hang\src\app'
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\Compilation.js:925:10
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\NormalModuleFactory.js:401:22
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\NormalModuleFactory.js:130:21
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\NormalModuleFactory.js:224:22
at M:\scratch\storybook-build-hang\node_modules\neo-async\async.js:2830:7
at M:\scratch\storybook-build-hang\node_modules\neo-async\async.js:6877:13
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\NormalModuleFactory.js:214:25
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\Resolver.js:213:14
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\Resolver.js:285:5
at eval (eval at create (M:\scratch\storybook-build-hang\node_modules\tapable\lib\HookCodeFactory.js:33:10), <anonymous>:15:1)
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\UnsafeCachePlugin.js:44:7
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\Resolver.js:285:5
at eval (eval at create (M:\scratch\storybook-build-hang\node_modules\tapable\lib\HookCodeFactory.js:33:10), <anonymous>:15:1)
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\Resolver.js:285:5
at eval (eval at create (M:\scratch\storybook-build-hang\node_modules\tapable\lib\HookCodeFactory.js:33:10), <anonymous>:27:1)
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\DescriptionFilePlugin.js:67:43
(node:39788) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). (rejection id: 1)
(node:39788) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code.
70% building 1553/1553 modules 0 active
```
**To Reproduce**
Steps to reproduce the behavior:
1. Bootstrap a new Angular CLI Project
2. Run `npx -p @storybook/cli sb init` to bootstrap storybook for Angular
3. Install `@storybook/addon-docs` and create a story using MDX
4. Add a typo in a component import path
5. Build the project with `build-storybook`
**Expected behavior**
An error printed to the console and the process exiting with a non-zero exit code.
**Code snippets**
A repository has been created with this issue already reproduced:
https://github.com/esingletary/storybook-build-hang
**System:**
```
Environment Info:
System:
OS: Windows 10 10.0.17134
CPU: (12) x64 Intel(R) Core(TM) i7-8850H CPU @ 2.60GHz
Binaries:
Node: 12.14.1 - C:\Program Files\nodejs\node.EXE
npm: 6.13.7 - C:\Program Files\nodejs\npm.CMD
Browsers:
Edge: 42.17134.1098.0
```
|
1.0
|
Module not found error doesn't properly exit during build - **Describe the bug**
When building storybook using `build-storybook` with module resolution errors in an `.mdx` file, a build error occurs but the build process hangs instead of exiting with a non-zero exit code. This is observed when using `@storybook/addons-docs` with a story using `.mdx` with an improper import.
This is problematic for our teams as we run this process during CI/CD pipelines and it hangs our builds if one of our developers makes an error in their `.mdx` files.
```
> storybook-build-hang@0.0.0 build-storybook M:\scratch\storybook-build-hang
> build-storybook
info @storybook/angular v5.3.18
info
info clean outputDir..
info => Copying prebuild dll's..
info => Building manager..
info => Loading manager config..
info => Loading presets
info => Compiling manager..
info => manager built (8.11 s)
info => Building preview..
info => Loading preview config..
info => Loading presets
info => Loading config/preview file in "./.storybook".
info => Adding stories defined in ".storybook\main.js".
info => Found custom tsconfig.json
info => Using default Webpack setup.
info => Using angular project 'storybook-build-hang' for configuring Storybook.
info => Loading angular-cli config.
info => Get angular-cli webpack config.
info => Compiling preview..
Starting type checking service...
Using 1 worker with 2048MB memory limit
70% building 40/90 modules 50 active M:\scratch\storybook-build-hang\node_modules\core-js\internals\set-to-string-tag.jsERR! => Failed to build the preview
ERR! Module not found: Error: Can't resolve './app.componen' in 'M:\scratch\storybook-build-hang\src\app'
(node:39788) UnhandledPromiseRejectionWarning: ModuleNotFoundError: Module not found: Error: Can't resolve './app.componen' in 'M:\scratch\storybook-build-hang\src\app'
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\Compilation.js:925:10
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\NormalModuleFactory.js:401:22
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\NormalModuleFactory.js:130:21
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\NormalModuleFactory.js:224:22
at M:\scratch\storybook-build-hang\node_modules\neo-async\async.js:2830:7
at M:\scratch\storybook-build-hang\node_modules\neo-async\async.js:6877:13
at M:\scratch\storybook-build-hang\node_modules\webpack\lib\NormalModuleFactory.js:214:25
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\Resolver.js:213:14
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\Resolver.js:285:5
at eval (eval at create (M:\scratch\storybook-build-hang\node_modules\tapable\lib\HookCodeFactory.js:33:10), <anonymous>:15:1)
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\UnsafeCachePlugin.js:44:7
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\Resolver.js:285:5
at eval (eval at create (M:\scratch\storybook-build-hang\node_modules\tapable\lib\HookCodeFactory.js:33:10), <anonymous>:15:1)
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\Resolver.js:285:5
at eval (eval at create (M:\scratch\storybook-build-hang\node_modules\tapable\lib\HookCodeFactory.js:33:10), <anonymous>:27:1)
at M:\scratch\storybook-build-hang\node_modules\enhanced-resolve\lib\DescriptionFilePlugin.js:67:43
(node:39788) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). (rejection id: 1)
(node:39788) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code.
70% building 1553/1553 modules 0 active
```
**To Reproduce**
Steps to reproduce the behavior:
1. Bootstrap a new Angular CLI Project
2. Run `npx -p @storybook/cli sb init` to bootstrap storybook for Angular
3. Install `@storybook/addon-docs` and create a story using MDX
4. Add a typo in a component import path
5. Build the project with `build-storybook`
**Expected behavior**
An error printed to the console and the process exiting with a non-zero exit code.
**Code snippets**
A repository has been created with this issue already reproduced:
https://github.com/esingletary/storybook-build-hang
**System:**
```
Environment Info:
System:
OS: Windows 10 10.0.17134
CPU: (12) x64 Intel(R) Core(TM) i7-8850H CPU @ 2.60GHz
Binaries:
Node: 12.14.1 - C:\Program Files\nodejs\node.EXE
npm: 6.13.7 - C:\Program Files\nodejs\npm.CMD
Browsers:
Edge: 42.17134.1098.0
```
|
non_process
|
module not found error doesn t properly exit during build describe the bug when building storybook using build storybook with module resolution errors in an mdx file a build error occurs but the build process hangs instead of exiting with a non zero exit code this is observed when using storybook addons docs with a story using mdx with an improper import this is problematic for our teams as we run this process during ci cd pipelines and it hangs our builds if one of our developers makes an error in their mdx files storybook build hang build storybook m scratch storybook build hang build storybook info storybook angular info info clean outputdir info copying prebuild dll s info building manager info loading manager config info loading presets info compiling manager info manager built s info building preview info loading preview config info loading presets info loading config preview file in storybook info adding stories defined in storybook main js info found custom tsconfig json info using default webpack setup info using angular project storybook build hang for configuring storybook info loading angular cli config info get angular cli webpack config info compiling preview starting type checking service using worker with memory limit building modules active m scratch storybook build hang node modules core js internals set to string tag jserr failed to build the preview err module not found error can t resolve app componen in m scratch storybook build hang src app node unhandledpromiserejectionwarning modulenotfounderror module not found error can t resolve app componen in m scratch storybook build hang src app at m scratch storybook build hang node modules webpack lib compilation js at m scratch storybook build hang node modules webpack lib normalmodulefactory js at m scratch storybook build hang node modules webpack lib normalmodulefactory js at m scratch storybook build hang node modules webpack lib normalmodulefactory js at m scratch storybook build hang node modules neo async async js at m scratch storybook build hang node modules neo async async js at m scratch storybook build hang node modules webpack lib normalmodulefactory js at m scratch storybook build hang node modules enhanced resolve lib resolver js at m scratch storybook build hang node modules enhanced resolve lib resolver js at eval eval at create m scratch storybook build hang node modules tapable lib hookcodefactory js at m scratch storybook build hang node modules enhanced resolve lib unsafecacheplugin js at m scratch storybook build hang node modules enhanced resolve lib resolver js at eval eval at create m scratch storybook build hang node modules tapable lib hookcodefactory js at m scratch storybook build hang node modules enhanced resolve lib resolver js at eval eval at create m scratch storybook build hang node modules tapable lib hookcodefactory js at m scratch storybook build hang node modules enhanced resolve lib descriptionfileplugin js node unhandledpromiserejectionwarning unhandled promise rejection this error originated either by throwing inside of an async function without a catch block or by rejecting a promise which was not handled with catch rejection id node deprecationwarning unhandled promise rejections are deprecated in the future promise rejections that are not handled will terminate the node js process with a non zero exit code building modules active to reproduce steps to reproduce the behavior bootstrap a new angular cli project run npx p storybook cli sb init to bootstrap storybook for angular install storybook addon docs and create a story using mdx add a typo in a component import path build the project with build storybook expected behavior an error printed to the console and the process exiting with a non zero exit code code snippets a repository has been created with this issue already reproduced system environment info system os windows cpu intel r core tm cpu binaries node c program files nodejs node exe npm c program files nodejs npm cmd browsers edge
| 0
|
13,698
| 16,455,335,978
|
IssuesEvent
|
2021-05-21 11:46:11
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
Release 4.1 - March 2021
|
P1 release team-XProduct type: process
|
# Status of Bazel 4.1
This release will use Bazel 4.0.0 as its baseline and we will apply selected cherry-picks and backports on top of it. Please request cherry-picks that you'd like to get into Bazel 4.1.0 here via a comment.
- Expected release date: March 2021
- [List of release blockers](https://github.com/bazelbuild/bazel/labels/Release%20blocker)
To report a release-blocking bug, please file a bug using the `Release blocker` label, and cc me.
Task list:
- [x] Pick release baseline: https://github.com/bazelbuild/bazel/commit/6b33bdb1e22514304c0e35ce8e067f2175685245
- [x] Create release candidate: https://releases.bazel.build/4.1.0/rc1/
- [x] Check downstream projects: https://buildkite.com/bazel/bazel-at-head-plus-downstream/builds/2040
- [x] [Create draft release announcement](https://docs.google.com/document/d/1wDvulLlj4NAlPZamdlEVFORks3YXJonCjyuQMUQEmB0/edit)
- [x] Send for review the release announcement PR: https://github.com/bazelbuild/bazel-blog/pull/256
- [x] Push the release, notify package maintainers: @vbatts @petemounce @excitoon
- [x] Update the documentation
- [ ] Push the blog post
- [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/tag/4.1.0)
|
1.0
|
Release 4.1 - March 2021 - # Status of Bazel 4.1
This release will use Bazel 4.0.0 as its baseline and we will apply selected cherry-picks and backports on top of it. Please request cherry-picks that you'd like to get into Bazel 4.1.0 here via a comment.
- Expected release date: March 2021
- [List of release blockers](https://github.com/bazelbuild/bazel/labels/Release%20blocker)
To report a release-blocking bug, please file a bug using the `Release blocker` label, and cc me.
Task list:
- [x] Pick release baseline: https://github.com/bazelbuild/bazel/commit/6b33bdb1e22514304c0e35ce8e067f2175685245
- [x] Create release candidate: https://releases.bazel.build/4.1.0/rc1/
- [x] Check downstream projects: https://buildkite.com/bazel/bazel-at-head-plus-downstream/builds/2040
- [x] [Create draft release announcement](https://docs.google.com/document/d/1wDvulLlj4NAlPZamdlEVFORks3YXJonCjyuQMUQEmB0/edit)
- [x] Send for review the release announcement PR: https://github.com/bazelbuild/bazel-blog/pull/256
- [x] Push the release, notify package maintainers: @vbatts @petemounce @excitoon
- [x] Update the documentation
- [ ] Push the blog post
- [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/tag/4.1.0)
|
process
|
release march status of bazel this release will use bazel as its baseline and we will apply selected cherry picks and backports on top of it please request cherry picks that you d like to get into bazel here via a comment expected release date march to report a release blocking bug please file a bug using the release blocker label and cc me task list pick release baseline create release candidate check downstream projects send for review the release announcement pr push the release notify package maintainers vbatts petemounce excitoon update the documentation push the blog post update the
| 1
|
176,846
| 21,443,075,329
|
IssuesEvent
|
2022-04-25 01:05:50
|
jgeraigery/spring-session
|
https://api.github.com/repos/jgeraigery/spring-session
|
closed
|
CVE-2019-9514 (High) detected in netty-codec-http2-4.1.33.Final.jar - autoclosed
|
security vulnerability
|
## CVE-2019-9514 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-codec-http2-4.1.33.Final.jar</b></p></summary>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: spring-session</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/io.netty/netty-codec-http2/4.1.33.Final/4bf1ea831ccfd84aea5cf2e292ec9d0beaa1b2b/netty-codec-http2-4.1.33.Final.jar,/root/.gradle/caches/modules-2/files-2.1/io.netty/netty-codec-http2/4.1.33.Final/4bf1ea831ccfd84aea5cf2e292ec9d0beaa1b2b/netty-codec-http2-4.1.33.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-webflux-2.1.3.RELEASE.jar (Root Library)
- spring-boot-starter-reactor-netty-2.1.3.RELEASE.jar
- reactor-netty-0.8.5.RELEASE.jar
- :x: **netty-codec-http2-4.1.33.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/spring-session/commit/523573994538bfeee4b8160bc4af5bcd4ad95a0d">523573994538bfeee4b8160bc4af5bcd4ad95a0d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Some HTTP/2 implementations are vulnerable to a reset flood, potentially leading to a denial of service. The attacker opens a number of streams and sends an invalid request over each stream that should solicit a stream of RST_STREAM frames from the peer. Depending on how the peer queues the RST_STREAM frames, this can consume excess memory, CPU, or both.
<p>Publish Date: 2019-08-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-9514>CVE-2019-9514</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-9514">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-9514</a></p>
<p>Release Date: 2019-08-13</p>
<p>Fix Resolution: 7.1.7,8.0.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.netty","packageName":"netty-codec-http2","packageVersion":"4.1.33.Final","packageFilePaths":["spring-session"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-webflux:2.1.3.RELEASE;org.springframework.boot:spring-boot-starter-reactor-netty:2.1.3.RELEASE;io.projectreactor.netty:reactor-netty:0.8.5.RELEASE;io.netty:netty-codec-http2:4.1.33.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"7.1.7,8.0.4"}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2019-9514","vulnerabilityDetails":"Some HTTP/2 implementations are vulnerable to a reset flood, potentially leading to a denial of service. The attacker opens a number of streams and sends an invalid request over each stream that should solicit a stream of RST_STREAM frames from the peer. Depending on how the peer queues the RST_STREAM frames, this can consume excess memory, CPU, or both.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-9514","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2019-9514 (High) detected in netty-codec-http2-4.1.33.Final.jar - autoclosed - ## CVE-2019-9514 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-codec-http2-4.1.33.Final.jar</b></p></summary>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: spring-session</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/io.netty/netty-codec-http2/4.1.33.Final/4bf1ea831ccfd84aea5cf2e292ec9d0beaa1b2b/netty-codec-http2-4.1.33.Final.jar,/root/.gradle/caches/modules-2/files-2.1/io.netty/netty-codec-http2/4.1.33.Final/4bf1ea831ccfd84aea5cf2e292ec9d0beaa1b2b/netty-codec-http2-4.1.33.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-webflux-2.1.3.RELEASE.jar (Root Library)
- spring-boot-starter-reactor-netty-2.1.3.RELEASE.jar
- reactor-netty-0.8.5.RELEASE.jar
- :x: **netty-codec-http2-4.1.33.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/spring-session/commit/523573994538bfeee4b8160bc4af5bcd4ad95a0d">523573994538bfeee4b8160bc4af5bcd4ad95a0d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Some HTTP/2 implementations are vulnerable to a reset flood, potentially leading to a denial of service. The attacker opens a number of streams and sends an invalid request over each stream that should solicit a stream of RST_STREAM frames from the peer. Depending on how the peer queues the RST_STREAM frames, this can consume excess memory, CPU, or both.
<p>Publish Date: 2019-08-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-9514>CVE-2019-9514</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-9514">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-9514</a></p>
<p>Release Date: 2019-08-13</p>
<p>Fix Resolution: 7.1.7,8.0.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.netty","packageName":"netty-codec-http2","packageVersion":"4.1.33.Final","packageFilePaths":["spring-session"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-webflux:2.1.3.RELEASE;org.springframework.boot:spring-boot-starter-reactor-netty:2.1.3.RELEASE;io.projectreactor.netty:reactor-netty:0.8.5.RELEASE;io.netty:netty-codec-http2:4.1.33.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"7.1.7,8.0.4"}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2019-9514","vulnerabilityDetails":"Some HTTP/2 implementations are vulnerable to a reset flood, potentially leading to a denial of service. The attacker opens a number of streams and sends an invalid request over each stream that should solicit a stream of RST_STREAM frames from the peer. Depending on how the peer queues the RST_STREAM frames, this can consume excess memory, CPU, or both.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-9514","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in netty codec final jar autoclosed cve high severity vulnerability vulnerable library netty codec final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file spring session path to vulnerable library root gradle caches modules files io netty netty codec final netty codec final jar root gradle caches modules files io netty netty codec final netty codec final jar dependency hierarchy spring boot starter webflux release jar root library spring boot starter reactor netty release jar reactor netty release jar x netty codec final jar vulnerable library found in head commit a href vulnerability details some http implementations are vulnerable to a reset flood potentially leading to a denial of service the attacker opens a number of streams and sends an invalid request over each stream that should solicit a stream of rst stream frames from the peer depending on how the peer queues the rst stream frames this can consume excess memory cpu or both publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org springframework boot spring boot starter webflux release org springframework boot spring boot starter reactor netty release io projectreactor netty reactor netty release io netty netty codec final isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails some http implementations are vulnerable to a reset flood potentially leading to a denial of service the attacker opens a number of streams and sends an invalid request over each stream that should solicit a stream of rst stream frames from the peer depending on how the peer queues the rst stream frames this can consume excess memory cpu or both vulnerabilityurl
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.