Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
93,947
| 8,459,787,280
|
IssuesEvent
|
2018-10-22 16:55:44
|
Microsoft/AzureStorageExplorer
|
https://api.github.com/repos/Microsoft/AzureStorageExplorer
|
closed
|
No 'Toggle Explorer & Manage Accounts & Add account' buttons present
|
:computer: windows testing
|
Storage Explorer Version: 1.5.0
Platform/OS Version: Windows 10
Architecture: ia32
Build Number: 20181022.4
Commit: 492b519a
Regression From: Previous release 1.4.4
#### Steps to Reproduce: ####
1. Launch the Storage Explorer -> Check the top left of the Storage Explorer.
#### Expected Experience: ####
'Toggle Explorer & Manage Accounts & Add account' buttons present at the top of left in storage explorer.
#### Actual Experience: ####
No 'Toggle Explorer & Manage Accounts & Add account' buttons present at the top left of storage explorer.

#### More Info: ####
This issue doesn't reproduce on Mac/Linux.
|
1.0
|
No 'Toggle Explorer & Manage Accounts & Add account' buttons present - Storage Explorer Version: 1.5.0
Platform/OS Version: Windows 10
Architecture: ia32
Build Number: 20181022.4
Commit: 492b519a
Regression From: Previous release 1.4.4
#### Steps to Reproduce: ####
1. Launch the Storage Explorer -> Check the top left of the Storage Explorer.
#### Expected Experience: ####
'Toggle Explorer & Manage Accounts & Add account' buttons present at the top of left in storage explorer.
#### Actual Experience: ####
No 'Toggle Explorer & Manage Accounts & Add account' buttons present at the top left of storage explorer.

#### More Info: ####
This issue doesn't reproduce on Mac/Linux.
|
non_process
|
no toggle explorer manage accounts add account buttons present storage explorer version platform os version windows architecture build number commit regression from previous release steps to reproduce launch the storage explorer check the top left of the storage explorer expected experience toggle explorer manage accounts add account buttons present at the top of left in storage explorer actual experience no toggle explorer manage accounts add account buttons present at the top left of storage explorer more info this issue doesn t reproduce on mac linux
| 0
|
27,830
| 12,709,851,073
|
IssuesEvent
|
2020-06-23 13:02:16
|
LiskHQ/lisk-docs
|
https://api.github.com/repos/LiskHQ/lisk-docs
|
opened
|
Update setup from source
|
service update
|
## Which content is outdated?
- (Source code install) Postgres is also needed
- (Source code install) Freegeoip is outdated and we cannot really recommend any alternative without advertising [Maxmind](https://www.maxmind.com/en/geoip-demo)
|
1.0
|
Update setup from source - ## Which content is outdated?
- (Source code install) Postgres is also needed
- (Source code install) Freegeoip is outdated and we cannot really recommend any alternative without advertising [Maxmind](https://www.maxmind.com/en/geoip-demo)
|
non_process
|
update setup from source which content is outdated source code install postgres is also needed source code install freegeoip is outdated and we cannot really recommend any alternative without advertising
| 0
|
2,164
| 5,011,564,764
|
IssuesEvent
|
2016-12-13 08:23:26
|
CERNDocumentServer/cds
|
https://api.github.com/repos/CERNDocumentServer/cds
|
opened
|
Wrong aspect ratio extracted from video
|
avc_processing enhancement
|
[ff_probe_all](https://github.com/CERNDocumentServer/cds/blob/cdslabs_qa/cds/modules/ffmpeg/ffmpeg.py#L53) should extract the aspect ratio of a video inside `display_aspect_ratio` but most of the time its value is `0:1`.
Maybe a solution could the to add a small check to [ff_probe](https://github.com/CERNDocumentServer/cds/blob/cdslabs_qa/cds/modules/ffmpeg/ffmpeg.py#L34) and [ff_probe_all](https://github.com/CERNDocumentServer/cds/blob/cdslabs_qa/cds/modules/ffmpeg/ffmpeg.py#L53) to calculate the `display_aspect_ratio` from the width and height in case the value of the former is `0:1`.
\* https://trac.ffmpeg.org/ticket/3798
|
1.0
|
Wrong aspect ratio extracted from video - [ff_probe_all](https://github.com/CERNDocumentServer/cds/blob/cdslabs_qa/cds/modules/ffmpeg/ffmpeg.py#L53) should extract the aspect ratio of a video inside `display_aspect_ratio` but most of the time its value is `0:1`.
Maybe a solution could the to add a small check to [ff_probe](https://github.com/CERNDocumentServer/cds/blob/cdslabs_qa/cds/modules/ffmpeg/ffmpeg.py#L34) and [ff_probe_all](https://github.com/CERNDocumentServer/cds/blob/cdslabs_qa/cds/modules/ffmpeg/ffmpeg.py#L53) to calculate the `display_aspect_ratio` from the width and height in case the value of the former is `0:1`.
\* https://trac.ffmpeg.org/ticket/3798
|
process
|
wrong aspect ratio extracted from video should extract the aspect ratio of a video inside display aspect ratio but most of the time its value is maybe a solution could the to add a small check to and to calculate the display aspect ratio from the width and height in case the value of the former is
| 1
|
10,877
| 13,646,018,752
|
IssuesEvent
|
2020-09-25 22:05:50
|
GoogleCloudPlatform/cloud-sql-jdbc-socket-factory
|
https://api.github.com/repos/GoogleCloudPlatform/cloud-sql-jdbc-socket-factory
|
closed
|
Add e2e testing
|
type: process
|
We need to add end to end tests to verify things work against the current SQL Admin version.
We'll probably need to move testing to Kokoro, and possibly separate lint checks out as a separate test.
|
1.0
|
Add e2e testing - We need to add end to end tests to verify things work against the current SQL Admin version.
We'll probably need to move testing to Kokoro, and possibly separate lint checks out as a separate test.
|
process
|
add testing we need to add end to end tests to verify things work against the current sql admin version we ll probably need to move testing to kokoro and possibly separate lint checks out as a separate test
| 1
|
7,356
| 10,490,429,141
|
IssuesEvent
|
2019-09-25 08:59:22
|
Graylog2/graylog2-server
|
https://api.github.com/repos/Graylog2/graylog2-server
|
opened
|
Allow POST in HTTP/JsonPath data adapter
|
feature processing
|
## Expected Behavior
The HTTP/JsonPath data adapter should support the POST method and allow specifying a body template that gets evaluated with the key that is being looked up.
## Current Behavior
The current implementation of the HTTP/JsonPath data adapter only supports the GET method to retrieve data from the remote location.
For third party systems that model their searches as POSTs, such as MISP or even Graylog itself, only support GET is not enough.
## Possible Solution
Add a radio button or drop down for the HTTP method (radio if all we want to support is GET and POST, I think the others don't make sense right now).
Allow pasting a body template for POST and evaluate it the same way we do it for the URL right now.
## Context
When trying to read data from external systems like MISP, you are faced with the problem that it requires POSTing a search, which either forces you to write middleware or implement an entirely new plugin just to be able to send a POST request, which seems excessive.
## Your Environment
* Graylog Version: 3.1
* Elasticsearch Version: not applicable
* MongoDB Version: not applicable
* Operating System: not applicable
* Browser version: not applicable
|
1.0
|
Allow POST in HTTP/JsonPath data adapter - ## Expected Behavior
The HTTP/JsonPath data adapter should support the POST method and allow specifying a body template that gets evaluated with the key that is being looked up.
## Current Behavior
The current implementation of the HTTP/JsonPath data adapter only supports the GET method to retrieve data from the remote location.
For third party systems that model their searches as POSTs, such as MISP or even Graylog itself, only support GET is not enough.
## Possible Solution
Add a radio button or drop down for the HTTP method (radio if all we want to support is GET and POST, I think the others don't make sense right now).
Allow pasting a body template for POST and evaluate it the same way we do it for the URL right now.
## Context
When trying to read data from external systems like MISP, you are faced with the problem that it requires POSTing a search, which either forces you to write middleware or implement an entirely new plugin just to be able to send a POST request, which seems excessive.
## Your Environment
* Graylog Version: 3.1
* Elasticsearch Version: not applicable
* MongoDB Version: not applicable
* Operating System: not applicable
* Browser version: not applicable
|
process
|
allow post in http jsonpath data adapter expected behavior the http jsonpath data adapter should support the post method and allow specifying a body template that gets evaluated with the key that is being looked up current behavior the current implementation of the http jsonpath data adapter only supports the get method to retrieve data from the remote location for third party systems that model their searches as posts such as misp or even graylog itself only support get is not enough possible solution add a radio button or drop down for the http method radio if all we want to support is get and post i think the others don t make sense right now allow pasting a body template for post and evaluate it the same way we do it for the url right now context when trying to read data from external systems like misp you are faced with the problem that it requires posting a search which either forces you to write middleware or implement an entirely new plugin just to be able to send a post request which seems excessive your environment graylog version elasticsearch version not applicable mongodb version not applicable operating system not applicable browser version not applicable
| 1
|
9,991
| 13,039,413,256
|
IssuesEvent
|
2020-07-28 16:42:37
|
Motionlessness/eCommerceSite
|
https://api.github.com/repos/Motionlessness/eCommerceSite
|
closed
|
Add CI Pipeline
|
Dev process
|
Add continuous pipeline that will check to make sure code in a pull request compiles successfully.
|
1.0
|
Add CI Pipeline - Add continuous pipeline that will check to make sure code in a pull request compiles successfully.
|
process
|
add ci pipeline add continuous pipeline that will check to make sure code in a pull request compiles successfully
| 1
|
8,001
| 11,194,504,321
|
IssuesEvent
|
2020-01-03 01:22:51
|
AlmuraDev/SGCraft
|
https://api.github.com/repos/AlmuraDev/SGCraft
|
closed
|
[Suggestion] Add config option to restrict only OPs to see gates' addresses
|
enhancement in process
|
Basically, if you're a regular user and right click a stargate, you get nothing except that facade thing for the bottom blocks. If you're an OP, you get what regular users get currently.
This would be useful to me for my exploration themed modpack, for me and some friends.
Basically, each of us spawn on our own planet. I'll be throwing addresses in a ton of places, but I won't be allowed to look at the addresses after that.
When we actually start playing, we have to find addresses, even the addresses just to go to each others' worlds. That's the whole purpose of this- we can't see our own address, so we have to be careful traveling.
Also, it would be nice if you could have another config to make it so you could still read the address if you have The One Probe, following the idea of hovering over with the proper tool to get extra info.
|
1.0
|
[Suggestion] Add config option to restrict only OPs to see gates' addresses - Basically, if you're a regular user and right click a stargate, you get nothing except that facade thing for the bottom blocks. If you're an OP, you get what regular users get currently.
This would be useful to me for my exploration themed modpack, for me and some friends.
Basically, each of us spawn on our own planet. I'll be throwing addresses in a ton of places, but I won't be allowed to look at the addresses after that.
When we actually start playing, we have to find addresses, even the addresses just to go to each others' worlds. That's the whole purpose of this- we can't see our own address, so we have to be careful traveling.
Also, it would be nice if you could have another config to make it so you could still read the address if you have The One Probe, following the idea of hovering over with the proper tool to get extra info.
|
process
|
add config option to restrict only ops to see gates addresses basically if you re a regular user and right click a stargate you get nothing except that facade thing for the bottom blocks if you re an op you get what regular users get currently this would be useful to me for my exploration themed modpack for me and some friends basically each of us spawn on our own planet i ll be throwing addresses in a ton of places but i won t be allowed to look at the addresses after that when we actually start playing we have to find addresses even the addresses just to go to each others worlds that s the whole purpose of this we can t see our own address so we have to be careful traveling also it would be nice if you could have another config to make it so you could still read the address if you have the one probe following the idea of hovering over with the proper tool to get extra info
| 1
|
1,522
| 4,113,607,254
|
IssuesEvent
|
2016-06-07 14:36:36
|
matz-e/lobster
|
https://api.github.com/repos/matz-e/lobster
|
closed
|
Rework the task wrapper script to quit when one step fails.
|
enhancement fix-ready processing
|
I'm currently looking into why my merge tasks fail (it's xrootd access to the storage element, I think). Currently, out tasks script just proceeds through the steps as if nothing happened. I think we should rework this to stop at the first error. My suggested changes:
1. The main flow of the script should be a bunch of function calls, no real logic (maybe beyond the CMS code distinctions for now), like:
config = json.load(…)
data = {…}
copy_inputs(config, data, …)
execute_prologue(config, data, …)
…
2. Instead of using the context to set the error state, wrap the above functions, and update mona-lisa accordingly, if needed:
@quit_on_failure(exit_code=210)
def copy_outputs(config, data, …):
…
|
1.0
|
Rework the task wrapper script to quit when one step fails. - I'm currently looking into why my merge tasks fail (it's xrootd access to the storage element, I think). Currently, out tasks script just proceeds through the steps as if nothing happened. I think we should rework this to stop at the first error. My suggested changes:
1. The main flow of the script should be a bunch of function calls, no real logic (maybe beyond the CMS code distinctions for now), like:
config = json.load(…)
data = {…}
copy_inputs(config, data, …)
execute_prologue(config, data, …)
…
2. Instead of using the context to set the error state, wrap the above functions, and update mona-lisa accordingly, if needed:
@quit_on_failure(exit_code=210)
def copy_outputs(config, data, …):
…
|
process
|
rework the task wrapper script to quit when one step fails i m currently looking into why my merge tasks fail it s xrootd access to the storage element i think currently out tasks script just proceeds through the steps as if nothing happened i think we should rework this to stop at the first error my suggested changes the main flow of the script should be a bunch of function calls no real logic maybe beyond the cms code distinctions for now like config json load … data … copy inputs config data … execute prologue config data … … instead of using the context to set the error state wrap the above functions and update mona lisa accordingly if needed quit on failure exit code def copy outputs config data … …
| 1
|
97,759
| 4,005,886,993
|
IssuesEvent
|
2016-05-12 13:16:46
|
coreos/rkt
|
https://api.github.com/repos/coreos/rkt
|
closed
|
Godeps: bump appc/spec for the next release
|
area/security dependency/appc spec priority/P0
|
For the next release of rkt-v1.6.0, we should wait for a new appc/spec release and then bump the version in Godeps. It will help with:
- [ ] capabilities #2589
- [x] depends on https://github.com/appc/spec/pull/601 (merged, but not in an appc/spec release yet)
- [ ] seccomp #1614
- [ ] depends on https://github.com/appc/spec/issues/529 (not implemented, might be postponed if not ready)
According to [](https://semaphoreci.com/alban/godep-rkt-test), rkt will need to be fixed with the following (https://github.com/appc/spec/pull/561).
```
rkt/pubkey/pubkey.go:160: undefined: discovery.InsecureHttp
rkt/pubkey/pubkey.go:163: undefined: discovery.InsecureTls
rkt/pubkey/pubkey.go:176: ep.Keys undefined (type discovery.PublicKeys has no field or method Keys)
```
|
1.0
|
Godeps: bump appc/spec for the next release - For the next release of rkt-v1.6.0, we should wait for a new appc/spec release and then bump the version in Godeps. It will help with:
- [ ] capabilities #2589
- [x] depends on https://github.com/appc/spec/pull/601 (merged, but not in an appc/spec release yet)
- [ ] seccomp #1614
- [ ] depends on https://github.com/appc/spec/issues/529 (not implemented, might be postponed if not ready)
According to [](https://semaphoreci.com/alban/godep-rkt-test), rkt will need to be fixed with the following (https://github.com/appc/spec/pull/561).
```
rkt/pubkey/pubkey.go:160: undefined: discovery.InsecureHttp
rkt/pubkey/pubkey.go:163: undefined: discovery.InsecureTls
rkt/pubkey/pubkey.go:176: ep.Keys undefined (type discovery.PublicKeys has no field or method Keys)
```
|
non_process
|
godeps bump appc spec for the next release for the next release of rkt we should wait for a new appc spec release and then bump the version in godeps it will help with capabilities depends on merged but not in an appc spec release yet seccomp depends on not implemented might be postponed if not ready according to rkt will need to be fixed with the following rkt pubkey pubkey go undefined discovery insecurehttp rkt pubkey pubkey go undefined discovery insecuretls rkt pubkey pubkey go ep keys undefined type discovery publickeys has no field or method keys
| 0
|
283,189
| 24,530,993,061
|
IssuesEvent
|
2022-10-11 16:24:47
|
void-linux/void-packages
|
https://api.github.com/repos/void-linux/void-packages
|
closed
|
python3-matplotlib-3.6.0_1: ModuleNotFoundError: No module named 'contourpy'
|
bug needs-testing
|
### Is this a new report?
Yes
### System Info
Void 5.19.10_1 x86_64
### Package(s) Affected
python3-matplotlib-3.6.0_1
### Does a report exist for this bug with the project's home (upstream) and/or another distro?
Not a bug, upstream says:
```
For packagers, this release contains some changes to dependencies:
* Python 3.8 is now required.
* NumPy 1.19 is now required.
* The ContourPy package is now required for contouring.
```
(https://discourse.matplotlib.org/t/matplotlib-announce-ann-matplotlib-3-6-0/23129)
So this should be fixed just by packaging `python3-contourpy`, presumably without the need to change matplotlib itself.
### Expected behaviour
A contour object is created and can be shown:
```
...
>>> ax.contourf(X, Y, Z, levels=levels)
<matplotlib.contour.QuadContourSet object at 0x7f8e831ebd60>
>>>
>>> plt.show()
```
This actually works after downgrading to `python3-matplotlib-3.5.3_1`
### Actual behaviour
Example from https://matplotlib.org/stable/plot_types/arrays/contourf.html#sphx-glr-plot-types-arrays-contourf-py
```
$ python
Python 3.10.7 (main, Sep 7 2022, 15:23:13) [GCC 10.2.1 20201203] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> import matplotlib.pyplot as plt
>>> import numpy as np
>>>
>>> plt.style.use('_mpl-gallery-nogrid')
>>>
>>> # make data
>>> X, Y = np.meshgrid(np.linspace(-3, 3, 256), np.linspace(-3, 3, 256))
>>> Z = (1 - X/2 + X**5 + Y**3) * np.exp(-X**2 - Y**2)
>>> levels = np.linspace(Z.min(), Z.max(), 7)
>>>
>>> # plot
>>> fig, ax = plt.subplots()
>>>
>>> ax.contourf(X, Y, Z, levels=levels)
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/usr/lib/python3.10/site-packages/matplotlib/__init__.py", line 1423, in inner
return func(ax, *map(sanitize_sequence, args), **kwargs)
File "/usr/lib/python3.10/site-packages/matplotlib/axes/_axes.py", line 6382, in contourf
contours = mcontour.QuadContourSet(self, *args, **kwargs)
File "/usr/lib/python3.10/site-packages/matplotlib/contour.py", line 794, in __init__
kwargs = self._process_args(*args, **kwargs)
File "/usr/lib/python3.10/site-packages/matplotlib/contour.py", line 1419, in _process_args
import contourpy
ModuleNotFoundError: No module named 'contourpy'
```
### Steps to reproduce
Try the example from https://matplotlib.org/stable/plot_types/arrays/contourf.html#sphx-glr-plot-types-arrays-contourf-py
|
1.0
|
python3-matplotlib-3.6.0_1: ModuleNotFoundError: No module named 'contourpy' - ### Is this a new report?
Yes
### System Info
Void 5.19.10_1 x86_64
### Package(s) Affected
python3-matplotlib-3.6.0_1
### Does a report exist for this bug with the project's home (upstream) and/or another distro?
Not a bug, upstream says:
```
For packagers, this release contains some changes to dependencies:
* Python 3.8 is now required.
* NumPy 1.19 is now required.
* The ContourPy package is now required for contouring.
```
(https://discourse.matplotlib.org/t/matplotlib-announce-ann-matplotlib-3-6-0/23129)
So this should be fixed just by packaging `python3-contourpy`, presumably without the need to change matplotlib itself.
### Expected behaviour
A contour object is created and can be shown:
```
...
>>> ax.contourf(X, Y, Z, levels=levels)
<matplotlib.contour.QuadContourSet object at 0x7f8e831ebd60>
>>>
>>> plt.show()
```
This actually works after downgrading to `python3-matplotlib-3.5.3_1`
### Actual behaviour
Example from https://matplotlib.org/stable/plot_types/arrays/contourf.html#sphx-glr-plot-types-arrays-contourf-py
```
$ python
Python 3.10.7 (main, Sep 7 2022, 15:23:13) [GCC 10.2.1 20201203] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> import matplotlib.pyplot as plt
>>> import numpy as np
>>>
>>> plt.style.use('_mpl-gallery-nogrid')
>>>
>>> # make data
>>> X, Y = np.meshgrid(np.linspace(-3, 3, 256), np.linspace(-3, 3, 256))
>>> Z = (1 - X/2 + X**5 + Y**3) * np.exp(-X**2 - Y**2)
>>> levels = np.linspace(Z.min(), Z.max(), 7)
>>>
>>> # plot
>>> fig, ax = plt.subplots()
>>>
>>> ax.contourf(X, Y, Z, levels=levels)
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/usr/lib/python3.10/site-packages/matplotlib/__init__.py", line 1423, in inner
return func(ax, *map(sanitize_sequence, args), **kwargs)
File "/usr/lib/python3.10/site-packages/matplotlib/axes/_axes.py", line 6382, in contourf
contours = mcontour.QuadContourSet(self, *args, **kwargs)
File "/usr/lib/python3.10/site-packages/matplotlib/contour.py", line 794, in __init__
kwargs = self._process_args(*args, **kwargs)
File "/usr/lib/python3.10/site-packages/matplotlib/contour.py", line 1419, in _process_args
import contourpy
ModuleNotFoundError: No module named 'contourpy'
```
### Steps to reproduce
Try the example from https://matplotlib.org/stable/plot_types/arrays/contourf.html#sphx-glr-plot-types-arrays-contourf-py
|
non_process
|
matplotlib modulenotfounderror no module named contourpy is this a new report yes system info void package s affected matplotlib does a report exist for this bug with the project s home upstream and or another distro not a bug upstream says for packagers this release contains some changes to dependencies python is now required numpy is now required the contourpy package is now required for contouring so this should be fixed just by packaging contourpy presumably without the need to change matplotlib itself expected behaviour a contour object is created and can be shown ax contourf x y z levels levels plt show this actually works after downgrading to matplotlib actual behaviour example from python python main sep on linux type help copyright credits or license for more information import matplotlib pyplot as plt import numpy as np plt style use mpl gallery nogrid make data x y np meshgrid np linspace np linspace z x x y np exp x y levels np linspace z min z max plot fig ax plt subplots ax contourf x y z levels levels traceback most recent call last file line in file usr lib site packages matplotlib init py line in inner return func ax map sanitize sequence args kwargs file usr lib site packages matplotlib axes axes py line in contourf contours mcontour quadcontourset self args kwargs file usr lib site packages matplotlib contour py line in init kwargs self process args args kwargs file usr lib site packages matplotlib contour py line in process args import contourpy modulenotfounderror no module named contourpy steps to reproduce try the example from
| 0
|
276,036
| 20,965,982,866
|
IssuesEvent
|
2022-03-28 06:44:45
|
jon-harper/clock-3
|
https://api.github.com/repos/jon-harper/clock-3
|
closed
|
Fusion 360 script to visualize t-nuts is needed
|
documentation enhancement Fusion Model Fusion Script
|
It would be extraordinarily helpful to be able to isolate *just* the t-nuts and frame to assist counting and identifying location
|
1.0
|
Fusion 360 script to visualize t-nuts is needed - It would be extraordinarily helpful to be able to isolate *just* the t-nuts and frame to assist counting and identifying location
|
non_process
|
fusion script to visualize t nuts is needed it would be extraordinarily helpful to be able to isolate just the t nuts and frame to assist counting and identifying location
| 0
|
246,027
| 7,893,117,845
|
IssuesEvent
|
2018-06-28 16:59:04
|
visit-dav/issues-test
|
https://api.github.com/repos/visit-dav/issues-test
|
closed
|
PFLOTRAN reader segv's reading vector var
|
Bug Likelihood: 3 - Occasional OS: All Priority: Normal Severity: 4 - Crash / Wrong Results Support Group: Any version: 2.8.2
|
Glenn Hammond at Sandia reported this and provided sample data.
He said 2.7.3 works, but 2.8.x and 2.9 do not.
gdb shows segv line 1403 avtPFLOTRANFileFormat.C (2.9RC version), so looks like an indexing error.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Kathleen Biagas
Original creation: 03/26/2015 12:30 pm
Original update: 03/26/2015 01:22 pm
Ticket number: 2190
|
1.0
|
PFLOTRAN reader segv's reading vector var - Glenn Hammond at Sandia reported this and provided sample data.
He said 2.7.3 works, but 2.8.x and 2.9 do not.
gdb shows segv line 1403 avtPFLOTRANFileFormat.C (2.9RC version), so looks like an indexing error.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Kathleen Biagas
Original creation: 03/26/2015 12:30 pm
Original update: 03/26/2015 01:22 pm
Ticket number: 2190
|
non_process
|
pflotran reader segv s reading vector var glenn hammond at sandia reported this and provided sample data he said works but x and do not gdb shows segv line avtpflotranfileformat c version so looks like an indexing error redmine migration this ticket was migrated from redmine the following information could not be accurately captured in the new ticket original author kathleen biagas original creation pm original update pm ticket number
| 0
|
19,040
| 25,042,669,557
|
IssuesEvent
|
2022-11-04 23:09:28
|
USGS-WiM/StreamStats
|
https://api.github.com/repos/USGS-WiM/StreamStats
|
opened
|
BP: Allow other geometry uploads
|
Batch Processor
|
Part of #1455
- [ ] Rename the "Upload Shapefile (.zip):" file uploader to "Upload Point Geometry: shapefile (.zip), GeoJSON (.json, .geojson), CSV (.csv)"
- [ ] Allow the user to upload a geoJSON
- [ ] Allow the user to upload a CSV
- [ ] Validate the file formats when the user clicks "Submit Batch"
|
1.0
|
BP: Allow other geometry uploads - Part of #1455
- [ ] Rename the "Upload Shapefile (.zip):" file uploader to "Upload Point Geometry: shapefile (.zip), GeoJSON (.json, .geojson), CSV (.csv)"
- [ ] Allow the user to upload a geoJSON
- [ ] Allow the user to upload a CSV
- [ ] Validate the file formats when the user clicks "Submit Batch"
|
process
|
bp allow other geometry uploads part of rename the upload shapefile zip file uploader to upload point geometry shapefile zip geojson json geojson csv csv allow the user to upload a geojson allow the user to upload a csv validate the file formats when the user clicks submit batch
| 1
|
23,538
| 16,384,132,056
|
IssuesEvent
|
2021-05-17 08:16:47
|
google/web-stories-wp
|
https://api.github.com/repos/google/web-stories-wp
|
opened
|
E2E Tests: take screenshots on failures
|
Package: E2E Tests Pod: WP & Infra Type: Infrastructure
|
<!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ -->
## Task Description
<!-- A clear and concise description of what this task is about. -->
Taking screenshots when an e2e test fails on CI would make it much easier for us to debug.
Some prior art:
https://github.com/WordPress/gutenberg/pull/26664
https://github.com/WordPress/gutenberg/pull/28449
Just like for #7549, it would mean using `jest-circus` as the test runner due to the events it emits to make this easier.
We could try using Percy for the screenshots, but storing them as GitHub Actions artifacts is cheaper.
|
1.0
|
E2E Tests: take screenshots on failures - <!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ -->
## Task Description
<!-- A clear and concise description of what this task is about. -->
Taking screenshots when an e2e test fails on CI would make it much easier for us to debug.
Some prior art:
https://github.com/WordPress/gutenberg/pull/26664
https://github.com/WordPress/gutenberg/pull/28449
Just like for #7549, it would mean using `jest-circus` as the test runner due to the events it emits to make this easier.
We could try using Percy for the screenshots, but storing them as GitHub Actions artifacts is cheaper.
|
non_process
|
tests take screenshots on failures task description taking screenshots when an test fails on ci would make it much easier for us to debug some prior art just like for it would mean using jest circus as the test runner due to the events it emits to make this easier we could try using percy for the screenshots but storing them as github actions artifacts is cheaper
| 0
|
19,153
| 25,234,629,765
|
IssuesEvent
|
2022-11-14 23:10:48
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
nociception is not the same as sensory perception of pain
|
organism-level process
|
**GO term**
GO:0019233 'sensory perception of pain'
**Issue description**
'sensory perception of pain' `has_exact_synonym` nociception
According this [reference](https://www.physio-pedia.com/Nociception):
> The terms nociception and [pain](https://www.physio-pedia.com/Pain_Behaviours) should not be used synonymously, because each can occur without the other.[[1]](https://www.physio-pedia.com/Nociception#cite_note-1.-1). [Pain](https://www.physio-pedia.com/Pain_Mechanisms) arising from activation of the nociceptors is called nociceptive pain.
|
1.0
|
nociception is not the same as sensory perception of pain - **GO term**
GO:0019233 'sensory perception of pain'
**Issue description**
'sensory perception of pain' `has_exact_synonym` nociception
According this [reference](https://www.physio-pedia.com/Nociception):
> The terms nociception and [pain](https://www.physio-pedia.com/Pain_Behaviours) should not be used synonymously, because each can occur without the other.[[1]](https://www.physio-pedia.com/Nociception#cite_note-1.-1). [Pain](https://www.physio-pedia.com/Pain_Mechanisms) arising from activation of the nociceptors is called nociceptive pain.
|
process
|
nociception is not the same as sensory perception of pain go term go sensory perception of pain issue description sensory perception of pain has exact synonym nociception according this the terms nociception and should not be used synonymously because each can occur without the other arising from activation of the nociceptors is called nociceptive pain
| 1
|
2,127
| 4,969,547,023
|
IssuesEvent
|
2016-12-05 13:44:38
|
AnalyticalGraphicsInc/cesium
|
https://api.github.com/repos/AnalyticalGraphicsInc/cesium
|
closed
|
Less greenkeeper noise
|
beginner dev process
|
Is it possible to improve the signal-to-noise ratio of this repo's pull requests by configuring greenkeeper to only do pull requests at some interval, perhaps biweekly? Since the updates are rarely significant, it has created an incredible amount of noise in our pull request history.
|
1.0
|
Less greenkeeper noise - Is it possible to improve the signal-to-noise ratio of this repo's pull requests by configuring greenkeeper to only do pull requests at some interval, perhaps biweekly? Since the updates are rarely significant, it has created an incredible amount of noise in our pull request history.
|
process
|
less greenkeeper noise is it possible to improve the signal to noise ratio of this repo s pull requests by configuring greenkeeper to only do pull requests at some interval perhaps biweekly since the updates are rarely significant it has created an incredible amount of noise in our pull request history
| 1
|
192,833
| 14,630,514,824
|
IssuesEvent
|
2020-12-23 17:52:21
|
deathlyrage/pot-demo-bugs
|
https://api.github.com/repos/deathlyrage/pot-demo-bugs
|
closed
|
Sucho Returning to the Surface Not Working
|
fixed needs testing
|
## Expected Behavior
If you dive underwater as a suchomimus, you should be able to return and stay at the surface.
## Current Behavior
If you sprint into the water and dive, you are unable to stay at the surface. You continously sink underwater every time until your stamina runs out or until you turn off sprint.
## Reproduction Steps
- Play as a sucho
- Sprint into a lake
- Dive underwater
- Attempt to stay at the surface
## Evidence (Screenshots & Videos)
https://youtu.be/aubpAXE8T38
## Crash Logs
(If Applicable. No logs needed if there is no crash.)
## Build Information
**Game Version:** 0.0.0.10071
**Operating System & Version:** iPhone 6s, iOS 14.2
**Alderon Games ID:** 246-116-163
|
1.0
|
Sucho Returning to the Surface Not Working - ## Expected Behavior
If you dive underwater as a suchomimus, you should be able to return and stay at the surface.
## Current Behavior
If you sprint into the water and dive, you are unable to stay at the surface. You continously sink underwater every time until your stamina runs out or until you turn off sprint.
## Reproduction Steps
- Play as a sucho
- Sprint into a lake
- Dive underwater
- Attempt to stay at the surface
## Evidence (Screenshots & Videos)
https://youtu.be/aubpAXE8T38
## Crash Logs
(If Applicable. No logs needed if there is no crash.)
## Build Information
**Game Version:** 0.0.0.10071
**Operating System & Version:** iPhone 6s, iOS 14.2
**Alderon Games ID:** 246-116-163
|
non_process
|
sucho returning to the surface not working expected behavior if you dive underwater as a suchomimus you should be able to return and stay at the surface current behavior if you sprint into the water and dive you are unable to stay at the surface you continously sink underwater every time until your stamina runs out or until you turn off sprint reproduction steps play as a sucho sprint into a lake dive underwater attempt to stay at the surface evidence screenshots videos crash logs if applicable no logs needed if there is no crash build information game version operating system version iphone ios alderon games id
| 0
|
14,435
| 17,484,522,805
|
IssuesEvent
|
2021-08-09 09:13:50
|
googleapis/python-spanner
|
https://api.github.com/repos/googleapis/python-spanner
|
closed
|
samples.samples.snippets_test: test_create_client_with_query_options failed
|
api: spanner type: process samples flakybot: issue flakybot: flaky
|
This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 7b8d8563651a9732631b22bef38cb2030eaadfa7
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/74f0a66c-c356-4d06-9f52-85f92729e353), [Sponge](http://sponge2/74f0a66c-c356-4d06-9f52-85f92729e353)
status: failed
<details><summary>Test output</summary><br><pre>args = (session: "projects/python-docs-samples-tests/instances/test-instance-03175cbad1/databases/test-db-78e4b8da2e/sessions...s"
params {
}
query_options {
optimizer_version: "1"
optimizer_statistics_package: "auto_20191128_14_47_22UTC"
}
,)
kwargs = {'metadata': [('google-cloud-resource-prefix', 'projects/python-docs-samples-tests/instances/test-instance-03175cbad1/...0j9Vuyc7VWPLqdjZfVQ'), ('x-goog-api-client', 'gl-python/3.6.13 grpc/1.38.1 gax/1.31.0 gccl/3.6.0')], 'timeout': 3600.0}
result = <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.INVALID_ARGUMENT
details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}"
>
prefetch_first = True
@general_helpers.wraps(callable_)
def error_remapped_callable(*args, **kwargs):
try:
result = callable_(*args, **kwargs)
# Auto-fetching the first result causes PubSub client's streaming pull
# to hang when re-opening the stream, thus we need examine the hacky
# hidden flag to see if pre-fetching is disabled.
# https://github.com/googleapis/python-pubsub/issues/93#issuecomment-630762257
prefetch_first = getattr(callable_, "_prefetch_first_result_", True)
return _StreamingResponseIterator(
> result, prefetch_first_result=prefetch_first
)
.nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:161:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.api_core.grpc_helpers._StreamingResponseIterator object at 0x7f2e0595d710>
wrapped = <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.INVALID_ARGUMENT
details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}"
>
prefetch_first_result = True
def __init__(self, wrapped, prefetch_first_result=True):
self._wrapped = wrapped
# This iterator is used in a retry context, and returned outside after init.
# gRPC will not throw an exception until the stream is consumed, so we need
# to retrieve the first result, in order to fail, in order to trigger a retry.
try:
if prefetch_first_result:
> self._stored_first_result = six.next(self._wrapped)
.nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:83:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.INVALID_ARGUMENT
details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}"
>
def __next__(self):
> return self._next()
.nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py:426:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.INVALID_ARGUMENT
details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}"
>
def _next(self):
with self._state.condition:
if self._state.code is None:
event_handler = _event_handler(self._state,
self._response_deserializer)
self._state.due.add(cygrpc.OperationType.receive_message)
operating = self._call.operate(
(cygrpc.ReceiveMessageOperation(_EMPTY_FLAGS),),
event_handler)
if not operating:
self._state.due.remove(cygrpc.OperationType.receive_message)
elif self._state.code is grpc.StatusCode.OK:
raise StopIteration()
else:
raise self
def _response_ready():
return (self._state.response is not None or
(cygrpc.OperationType.receive_message
not in self._state.due and
self._state.code is not None))
_common.wait(self._state.condition.wait, _response_ready)
if self._state.response is not None:
response = self._state.response
self._state.response = None
return response
elif cygrpc.OperationType.receive_message not in self._state.due:
if self._state.code is grpc.StatusCode.OK:
raise StopIteration()
elif self._state.code is not None:
> raise self
E grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
E status = StatusCode.INVALID_ARGUMENT
E details = "Requested optimizer statistics package not found: auto_20191128_14_47_22UTC."
E debug_error_string = "{"created":"@1625909969.479153083","description":"Error received from peer ipv4:74.125.195.95:443","file":"src/core/lib/surface/call.cc","file_line":1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}"
E >
.nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py:826: _MultiThreadedRendezvous
The above exception was the direct cause of the following exception:
capsys = <_pytest.capture.CaptureFixture object at 0x7f2e0470a940>
def test_create_client_with_query_options(capsys):
> snippets.create_client_with_query_options(INSTANCE_ID, DATABASE_ID)
snippets_test.py:417:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
snippets.py:1735: in create_client_with_query_options
for row in results:
../../google/cloud/spanner_v1/streamed.py:149: in __iter__
self._consume_next()
../../google/cloud/spanner_v1/streamed.py:121: in _consume_next
response = six.next(self._response_iterator)
../../google/cloud/spanner_v1/snapshot.py:59: in _restart_on_unavailable
iterator = method(request=request)
../../google/cloud/spanner_v1/services/spanner/client.py:884: in execute_streaming_sql
response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,)
.nox/py-3-6/lib/python3.6/site-packages/google/api_core/gapic_v1/method.py:145: in __call__
return wrapped_func(*args, **kwargs)
.nox/py-3-6/lib/python3.6/site-packages/google/api_core/timeout.py:102: in func_with_timeout
return func(*args, **kwargs)
.nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:164: in error_remapped_callable
six.raise_from(exceptions.from_grpc_error(exc), exc)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
value = None
from_value = <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.INVALID_ARGUMENT
details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}"
>
> ???
E google.api_core.exceptions.InvalidArgument: 400 Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.
<string>:3: InvalidArgument</pre></details>
|
1.0
|
samples.samples.snippets_test: test_create_client_with_query_options failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 7b8d8563651a9732631b22bef38cb2030eaadfa7
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/74f0a66c-c356-4d06-9f52-85f92729e353), [Sponge](http://sponge2/74f0a66c-c356-4d06-9f52-85f92729e353)
status: failed
<details><summary>Test output</summary><br><pre>args = (session: "projects/python-docs-samples-tests/instances/test-instance-03175cbad1/databases/test-db-78e4b8da2e/sessions...s"
params {
}
query_options {
optimizer_version: "1"
optimizer_statistics_package: "auto_20191128_14_47_22UTC"
}
,)
kwargs = {'metadata': [('google-cloud-resource-prefix', 'projects/python-docs-samples-tests/instances/test-instance-03175cbad1/...0j9Vuyc7VWPLqdjZfVQ'), ('x-goog-api-client', 'gl-python/3.6.13 grpc/1.38.1 gax/1.31.0 gccl/3.6.0')], 'timeout': 3600.0}
result = <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.INVALID_ARGUMENT
details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}"
>
prefetch_first = True
@general_helpers.wraps(callable_)
def error_remapped_callable(*args, **kwargs):
try:
result = callable_(*args, **kwargs)
# Auto-fetching the first result causes PubSub client's streaming pull
# to hang when re-opening the stream, thus we need examine the hacky
# hidden flag to see if pre-fetching is disabled.
# https://github.com/googleapis/python-pubsub/issues/93#issuecomment-630762257
prefetch_first = getattr(callable_, "_prefetch_first_result_", True)
return _StreamingResponseIterator(
> result, prefetch_first_result=prefetch_first
)
.nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:161:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.api_core.grpc_helpers._StreamingResponseIterator object at 0x7f2e0595d710>
wrapped = <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.INVALID_ARGUMENT
details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}"
>
prefetch_first_result = True
def __init__(self, wrapped, prefetch_first_result=True):
self._wrapped = wrapped
# This iterator is used in a retry context, and returned outside after init.
# gRPC will not throw an exception until the stream is consumed, so we need
# to retrieve the first result, in order to fail, in order to trigger a retry.
try:
if prefetch_first_result:
> self._stored_first_result = six.next(self._wrapped)
.nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:83:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.INVALID_ARGUMENT
details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}"
>
def __next__(self):
> return self._next()
.nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py:426:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.INVALID_ARGUMENT
details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}"
>
def _next(self):
with self._state.condition:
if self._state.code is None:
event_handler = _event_handler(self._state,
self._response_deserializer)
self._state.due.add(cygrpc.OperationType.receive_message)
operating = self._call.operate(
(cygrpc.ReceiveMessageOperation(_EMPTY_FLAGS),),
event_handler)
if not operating:
self._state.due.remove(cygrpc.OperationType.receive_message)
elif self._state.code is grpc.StatusCode.OK:
raise StopIteration()
else:
raise self
def _response_ready():
return (self._state.response is not None or
(cygrpc.OperationType.receive_message
not in self._state.due and
self._state.code is not None))
_common.wait(self._state.condition.wait, _response_ready)
if self._state.response is not None:
response = self._state.response
self._state.response = None
return response
elif cygrpc.OperationType.receive_message not in self._state.due:
if self._state.code is grpc.StatusCode.OK:
raise StopIteration()
elif self._state.code is not None:
> raise self
E grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with:
E status = StatusCode.INVALID_ARGUMENT
E details = "Requested optimizer statistics package not found: auto_20191128_14_47_22UTC."
E debug_error_string = "{"created":"@1625909969.479153083","description":"Error received from peer ipv4:74.125.195.95:443","file":"src/core/lib/surface/call.cc","file_line":1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}"
E >
.nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py:826: _MultiThreadedRendezvous
The above exception was the direct cause of the following exception:
capsys = <_pytest.capture.CaptureFixture object at 0x7f2e0470a940>
def test_create_client_with_query_options(capsys):
> snippets.create_client_with_query_options(INSTANCE_ID, DATABASE_ID)
snippets_test.py:417:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
snippets.py:1735: in create_client_with_query_options
for row in results:
../../google/cloud/spanner_v1/streamed.py:149: in __iter__
self._consume_next()
../../google/cloud/spanner_v1/streamed.py:121: in _consume_next
response = six.next(self._response_iterator)
../../google/cloud/spanner_v1/snapshot.py:59: in _restart_on_unavailable
iterator = method(request=request)
../../google/cloud/spanner_v1/services/spanner/client.py:884: in execute_streaming_sql
response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,)
.nox/py-3-6/lib/python3.6/site-packages/google/api_core/gapic_v1/method.py:145: in __call__
return wrapped_func(*args, **kwargs)
.nox/py-3-6/lib/python3.6/site-packages/google/api_core/timeout.py:102: in func_with_timeout
return func(*args, **kwargs)
.nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:164: in error_remapped_callable
six.raise_from(exceptions.from_grpc_error(exc), exc)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
value = None
from_value = <_MultiThreadedRendezvous of RPC that terminated with:
status = StatusCode.INVALID_ARGUMENT
details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}"
>
> ???
E google.api_core.exceptions.InvalidArgument: 400 Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.
<string>:3: InvalidArgument</pre></details>
|
process
|
samples samples snippets test test create client with query options failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output args session projects python docs samples tests instances test instance databases test db sessions s params query options optimizer version optimizer statistics package auto kwargs metadata timeout result multithreadedrendezvous of rpc that terminated with status statuscode invalid argument details requested opt grpc message requested optimizer statistics package not found auto grpc status prefetch first true general helpers wraps callable def error remapped callable args kwargs try result callable args kwargs auto fetching the first result causes pubsub client s streaming pull to hang when re opening the stream thus we need examine the hacky hidden flag to see if pre fetching is disabled prefetch first getattr callable prefetch first result true return streamingresponseiterator result prefetch first result prefetch first nox py lib site packages google api core grpc helpers py self wrapped multithreadedrendezvous of rpc that terminated with status statuscode invalid argument details requested opt grpc message requested optimizer statistics package not found auto grpc status prefetch first result true def init self wrapped prefetch first result true self wrapped wrapped this iterator is used in a retry context and returned outside after init grpc will not throw an exception until the stream is consumed so we need to retrieve the first result in order to fail in order to trigger a retry try if prefetch first result self stored first result six next self wrapped nox py lib site packages google api core grpc helpers py self multithreadedrendezvous of rpc that terminated with status statuscode invalid argument details requested opt grpc message requested optimizer statistics package not found auto grpc status def next self return self next nox py lib site packages grpc channel py self multithreadedrendezvous of rpc that terminated with status statuscode invalid argument details requested opt grpc message requested optimizer statistics package not found auto grpc status def next self with self state condition if self state code is none event handler event handler self state self response deserializer self state due add cygrpc operationtype receive message operating self call operate cygrpc receivemessageoperation empty flags event handler if not operating self state due remove cygrpc operationtype receive message elif self state code is grpc statuscode ok raise stopiteration else raise self def response ready return self state response is not none or cygrpc operationtype receive message not in self state due and self state code is not none common wait self state condition wait response ready if self state response is not none response self state response self state response none return response elif cygrpc operationtype receive message not in self state due if self state code is grpc statuscode ok raise stopiteration elif self state code is not none raise self e grpc channel multithreadedrendezvous multithreadedrendezvous of rpc that terminated with e status statuscode invalid argument e details requested optimizer statistics package not found auto e debug error string created description error received from peer file src core lib surface call cc file line grpc message requested optimizer statistics package not found auto grpc status e nox py lib site packages grpc channel py multithreadedrendezvous the above exception was the direct cause of the following exception capsys def test create client with query options capsys snippets create client with query options instance id database id snippets test py snippets py in create client with query options for row in results google cloud spanner streamed py in iter self consume next google cloud spanner streamed py in consume next response six next self response iterator google cloud spanner snapshot py in restart on unavailable iterator method request request google cloud spanner services spanner client py in execute streaming sql response rpc request retry retry timeout timeout metadata metadata nox py lib site packages google api core gapic method py in call return wrapped func args kwargs nox py lib site packages google api core timeout py in func with timeout return func args kwargs nox py lib site packages google api core grpc helpers py in error remapped callable six raise from exceptions from grpc error exc exc value none from value multithreadedrendezvous of rpc that terminated with status statuscode invalid argument details requested opt grpc message requested optimizer statistics package not found auto grpc status e google api core exceptions invalidargument requested optimizer statistics package not found auto invalidargument
| 1
|
5,154
| 7,933,323,593
|
IssuesEvent
|
2018-07-08 04:05:13
|
Great-Hill-Corporation/quickBlocks
|
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
|
closed
|
If a class is upgraded to a new schema, it does not save its data
|
libs-etherlib status-inprocess type-enhancement
|
Do this:
1. Save ${CACHE}/prices/poloniex.bin
2. remove ${CACHE}/prices/poloniex.bin
3. Run test for ethprice
There will be a discrepancy in the schema from 0.2.0 to 0.3.0. This happens because even though we do upgrade the data on read, we don't re-write that data to hard drive. Each time we read, we should keep track of 'dirty' bit on the archive, and write the data if 'dirty' is true.
|
1.0
|
If a class is upgraded to a new schema, it does not save its data - Do this:
1. Save ${CACHE}/prices/poloniex.bin
2. remove ${CACHE}/prices/poloniex.bin
3. Run test for ethprice
There will be a discrepancy in the schema from 0.2.0 to 0.3.0. This happens because even though we do upgrade the data on read, we don't re-write that data to hard drive. Each time we read, we should keep track of 'dirty' bit on the archive, and write the data if 'dirty' is true.
|
process
|
if a class is upgraded to a new schema it does not save its data do this save cache prices poloniex bin remove cache prices poloniex bin run test for ethprice there will be a discrepancy in the schema from to this happens because even though we do upgrade the data on read we don t re write that data to hard drive each time we read we should keep track of dirty bit on the archive and write the data if dirty is true
| 1
|
155,653
| 24,496,222,620
|
IssuesEvent
|
2022-10-10 08:54:56
|
anubhab1710/FLOSS-Graphics-Templates
|
https://api.github.com/repos/anubhab1710/FLOSS-Graphics-Templates
|
opened
|
Create Social Media Templates using GIMP
|
good first issue hacktoberfest design beginner-friendly
|
Use GIMP to design visually attractive Social Media Templates!
|
1.0
|
Create Social Media Templates using GIMP - Use GIMP to design visually attractive Social Media Templates!
|
non_process
|
create social media templates using gimp use gimp to design visually attractive social media templates
| 0
|
446,815
| 12,879,129,569
|
IssuesEvent
|
2020-07-11 20:19:30
|
clinwiki-org/clinwiki
|
https://api.github.com/repos/clinwiki-org/clinwiki
|
closed
|
restore navigation back to search results from study page
|
Priority 3
|
used to be able to get back to search results from within study page and maintain filter/sort - need this from anywhere on study pages
|
1.0
|
restore navigation back to search results from study page - used to be able to get back to search results from within study page and maintain filter/sort - need this from anywhere on study pages
|
non_process
|
restore navigation back to search results from study page used to be able to get back to search results from within study page and maintain filter sort need this from anywhere on study pages
| 0
|
318,293
| 27,297,075,801
|
IssuesEvent
|
2023-02-23 21:21:23
|
nucleus-security/Test-repo
|
https://api.github.com/repos/nucleus-security/Test-repo
|
closed
|
Nucleus - [High] - 440057
|
Test
|
Source: QUALYS
Finding Description: CentOS has released security update for kernel to fix the vulnerabilities. Affected Products: centos 6
Impact: Successful exploitation allows attacker to compromise the system.
Target(s): Asset name: 192.168.56.103
Solution: To resolve this issue, upgrade to the latest packages which contain a patch. Refer to CentOS advisory centos 6 (https://lists.centos.org/pipermail/centos-announce/2016-July/021977.html) for updates and patch information.
Patch:
Following are links for downloading patches to fix the vulnerabilities:
CESA-2016:1406: centos 6 (https://lists.centos.org/pipermail/centos-announce/2016-July/021977.html)
References:
QID:440057
CVE:CVE-2016-4565
Category:CentOS
PCI Flagged:yes
Vendor References:CESA-2016:1406 centos 6
Bugtraq IDs:90301
Severity: High
Date Discovered: 2022-11-12 08:04:44
Nucleus Notification Rules Triggered: Rule GitHub
Project Name: 6716
Please see Nucleus for more information on these vulnerabilities:https://192.168.56.101/nucleus/public/app/index.html#vuln/201000007/NDQwMDU3/UVVBTFlT/VnVsbg--/false/MjAxMDAwMDA3/c3VtbWFyeQ--/false
|
1.0
|
Nucleus - [High] - 440057 - Source: QUALYS
Finding Description: CentOS has released security update for kernel to fix the vulnerabilities. Affected Products: centos 6
Impact: Successful exploitation allows attacker to compromise the system.
Target(s): Asset name: 192.168.56.103
Solution: To resolve this issue, upgrade to the latest packages which contain a patch. Refer to CentOS advisory centos 6 (https://lists.centos.org/pipermail/centos-announce/2016-July/021977.html) for updates and patch information.
Patch:
Following are links for downloading patches to fix the vulnerabilities:
CESA-2016:1406: centos 6 (https://lists.centos.org/pipermail/centos-announce/2016-July/021977.html)
References:
QID:440057
CVE:CVE-2016-4565
Category:CentOS
PCI Flagged:yes
Vendor References:CESA-2016:1406 centos 6
Bugtraq IDs:90301
Severity: High
Date Discovered: 2022-11-12 08:04:44
Nucleus Notification Rules Triggered: Rule GitHub
Project Name: 6716
Please see Nucleus for more information on these vulnerabilities:https://192.168.56.101/nucleus/public/app/index.html#vuln/201000007/NDQwMDU3/UVVBTFlT/VnVsbg--/false/MjAxMDAwMDA3/c3VtbWFyeQ--/false
|
non_process
|
nucleus source qualys finding description centos has released security update for kernel to fix the vulnerabilities affected products centos impact successful exploitation allows attacker to compromise the system target s asset name solution to resolve this issue upgrade to the latest packages which contain a patch refer to centos advisory centos for updates and patch information patch following are links for downloading patches to fix the vulnerabilities cesa centos references qid cve cve category centos pci flagged yes vendor references cesa centos bugtraq ids severity high date discovered nucleus notification rules triggered rule github project name please see nucleus for more information on these vulnerabilities
| 0
|
19,575
| 25,895,131,180
|
IssuesEvent
|
2022-12-14 21:39:30
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
DISABLED test_success_first_then_exception (__main__.SpawnTest)
|
high priority triage review module: multiprocessing triaged module: flaky-tests skipped
|
Platforms: linux
This test was disabled because it is failing on trunk. See [recent examples](http://torch-ci.com/failure/test_success_first_then_exception%2C%20SpawnTest) and the most recent
[workflow logs](https://github.com/pytorch/pytorch/actions/runs/1820236472).
Over the past 6 hours, it has been determined flaky in 1 workflow(s) with
1 red and 3 green.
cc @ezyang @gchanan @zou3519 @VitalyFedyunin
|
1.0
|
DISABLED test_success_first_then_exception (__main__.SpawnTest) - Platforms: linux
This test was disabled because it is failing on trunk. See [recent examples](http://torch-ci.com/failure/test_success_first_then_exception%2C%20SpawnTest) and the most recent
[workflow logs](https://github.com/pytorch/pytorch/actions/runs/1820236472).
Over the past 6 hours, it has been determined flaky in 1 workflow(s) with
1 red and 3 green.
cc @ezyang @gchanan @zou3519 @VitalyFedyunin
|
process
|
disabled test success first then exception main spawntest platforms linux this test was disabled because it is failing on trunk see and the most recent over the past hours it has been determined flaky in workflow s with red and green cc ezyang gchanan vitalyfedyunin
| 1
|
298,271
| 22,473,595,190
|
IssuesEvent
|
2022-06-22 10:13:29
|
ices-tools-dev/RDBES
|
https://api.github.com/repos/ices-tools-dev/RDBES
|
reopened
|
out of frame ports in a strata
|
Documentation selectionMethod Estimation issue Need plenary
|
Hello,
I have read in the documentation that out of frame strata should not be declared. And I have doubts about how to report my data. I will try to explain the situation.
In my sampling design, I have considered different fleets as strata (each of them defined with non-overlapping vessel lists).
Some strata (fleets) are out of my sampling frame. They are never sampled. I understand that these are not reported in the RDBES. No problems with this.
But in every sampled strata, I have some ports which are sampled, and some other ports that are not included in my sampling frame (due to practical constrains) and therefore will never be sampled. I thought that these should be reported with sampling = 0, because they will be need in order to know the total PSU in the strata. But somehow I also think that it would be needed to indicate that they are out of the sampling frame and therefore their sampling probability is zero.
Shall include information of these ports out of my sampling frame?
Thank you!
|
1.0
|
out of frame ports in a strata - Hello,
I have read in the documentation that out of frame strata should not be declared. And I have doubts about how to report my data. I will try to explain the situation.
In my sampling design, I have considered different fleets as strata (each of them defined with non-overlapping vessel lists).
Some strata (fleets) are out of my sampling frame. They are never sampled. I understand that these are not reported in the RDBES. No problems with this.
But in every sampled strata, I have some ports which are sampled, and some other ports that are not included in my sampling frame (due to practical constrains) and therefore will never be sampled. I thought that these should be reported with sampling = 0, because they will be need in order to know the total PSU in the strata. But somehow I also think that it would be needed to indicate that they are out of the sampling frame and therefore their sampling probability is zero.
Shall include information of these ports out of my sampling frame?
Thank you!
|
non_process
|
out of frame ports in a strata hello i have read in the documentation that out of frame strata should not be declared and i have doubts about how to report my data i will try to explain the situation in my sampling design i have considered different fleets as strata each of them defined with non overlapping vessel lists some strata fleets are out of my sampling frame they are never sampled i understand that these are not reported in the rdbes no problems with this but in every sampled strata i have some ports which are sampled and some other ports that are not included in my sampling frame due to practical constrains and therefore will never be sampled i thought that these should be reported with sampling because they will be need in order to know the total psu in the strata but somehow i also think that it would be needed to indicate that they are out of the sampling frame and therefore their sampling probability is zero shall include information of these ports out of my sampling frame thank you
| 0
|
14,472
| 9,321,790,968
|
IssuesEvent
|
2019-03-27 05:45:17
|
danielwii/react-exercises
|
https://api.github.com/repos/danielwii/react-exercises
|
opened
|
CVE-2018-1000620 High Severity Vulnerability detected by WhiteSource
|
security vulnerability
|
## CVE-2018-1000620 - High Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cryptiles-2.0.5.tgz</b></p></summary>
<p>General purpose crypto utilities</p>
<p>path: null</p>
<p>
<p>Library home page: <a href=http://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz>http://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz</a></p>
Dependency Hierarchy:
- react-scripts-0.8.4.tgz (Root Library)
- fsevents-1.0.14.tgz
- node-pre-gyp-0.6.29.tgz
- request-2.73.0.tgz
- hawk-3.1.3.tgz
- :x: **cryptiles-2.0.5.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2.
<p>Publish Date: 2018-07-09
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000620>CVE-2018-1000620</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-1000620 High Severity Vulnerability detected by WhiteSource - ## CVE-2018-1000620 - High Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cryptiles-2.0.5.tgz</b></p></summary>
<p>General purpose crypto utilities</p>
<p>path: null</p>
<p>
<p>Library home page: <a href=http://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz>http://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz</a></p>
Dependency Hierarchy:
- react-scripts-0.8.4.tgz (Root Library)
- fsevents-1.0.14.tgz
- node-pre-gyp-0.6.29.tgz
- request-2.73.0.tgz
- hawk-3.1.3.tgz
- :x: **cryptiles-2.0.5.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2.
<p>Publish Date: 2018-07-09
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000620>CVE-2018-1000620</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high severity vulnerability detected by whitesource cve high severity vulnerability vulnerable library cryptiles tgz general purpose crypto utilities path null library home page a href dependency hierarchy react scripts tgz root library fsevents tgz node pre gyp tgz request tgz hawk tgz x cryptiles tgz vulnerable library vulnerability details eran hammer cryptiles version earlier contains a cwe insufficient entropy vulnerability in randomdigits method that can result in an attacker is more likely to be able to brute force something that was supposed to be random this attack appear to be exploitable via depends upon the calling application this vulnerability appears to have been fixed in publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href step up your open source security game with whitesource
| 0
|
9,685
| 12,685,673,065
|
IssuesEvent
|
2020-06-20 06:05:33
|
ClickHouse/ClickHouse
|
https://api.github.com/repos/ClickHouse/ClickHouse
|
closed
|
Master: pipeline stuck
|
bug comp-processors prio-major v20.5-affected
|
```
CREATE TABLE pipeline_stuck ( `id` String, `val` SimpleAggregateFunction(anyLast, String), INDEX ind val TYPE bloom_filter() GRANULARITY 1 ) ENGINE = AggregatingMergeTree() ORDER BY id ;
insert into pipeline_stuck select toString(number%100), toString(number) from numbers(10000000);
select * from pipeline_stuck final limit 10;
select * from pipeline_stuck final where val = '9999900';
alter table pipeline_stuck drop index ind;
select * from pipeline_stuck final where val = '9999900';
```
|
1.0
|
Master: pipeline stuck - ```
CREATE TABLE pipeline_stuck ( `id` String, `val` SimpleAggregateFunction(anyLast, String), INDEX ind val TYPE bloom_filter() GRANULARITY 1 ) ENGINE = AggregatingMergeTree() ORDER BY id ;
insert into pipeline_stuck select toString(number%100), toString(number) from numbers(10000000);
select * from pipeline_stuck final limit 10;
select * from pipeline_stuck final where val = '9999900';
alter table pipeline_stuck drop index ind;
select * from pipeline_stuck final where val = '9999900';
```
|
process
|
master pipeline stuck create table pipeline stuck id string val simpleaggregatefunction anylast string index ind val type bloom filter granularity engine aggregatingmergetree order by id insert into pipeline stuck select tostring number tostring number from numbers select from pipeline stuck final limit select from pipeline stuck final where val alter table pipeline stuck drop index ind select from pipeline stuck final where val
| 1
|
20,130
| 26,666,079,141
|
IssuesEvent
|
2023-01-26 04:02:31
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
Flaky test TestProcessorConsumeMetricsErrors
|
bug processor/spanmetrics flaky test
|
### Component(s)
processor/spanmetrics
### What happened?
See https://github.com/open-telemetry/opentelemetry-collector-contrib/actions/runs/3993460821/jobs/6850258904
```
=== RUN TestProcessorConsumeMetricsErrors
processor_test.go:297:
Error Trace: /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/spanmetricsprocessor/processor_test.go:297
Error: Should NOT be empty, but was []
Test: TestProcessorConsumeMetricsErrors
--- FAIL: TestProcessorConsumeMetricsErrors (0.01s)
```
### Collector version
0.70.0
### Environment information
_No response_
### OpenTelemetry Collector configuration
_No response_
### Log output
_No response_
### Additional context
_No response_
|
1.0
|
Flaky test TestProcessorConsumeMetricsErrors - ### Component(s)
processor/spanmetrics
### What happened?
See https://github.com/open-telemetry/opentelemetry-collector-contrib/actions/runs/3993460821/jobs/6850258904
```
=== RUN TestProcessorConsumeMetricsErrors
processor_test.go:297:
Error Trace: /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/spanmetricsprocessor/processor_test.go:297
Error: Should NOT be empty, but was []
Test: TestProcessorConsumeMetricsErrors
--- FAIL: TestProcessorConsumeMetricsErrors (0.01s)
```
### Collector version
0.70.0
### Environment information
_No response_
### OpenTelemetry Collector configuration
_No response_
### Log output
_No response_
### Additional context
_No response_
|
process
|
flaky test testprocessorconsumemetricserrors component s processor spanmetrics what happened see run testprocessorconsumemetricserrors processor test go error trace home runner work opentelemetry collector contrib opentelemetry collector contrib processor spanmetricsprocessor processor test go error should not be empty but was test testprocessorconsumemetricserrors fail testprocessorconsumemetricserrors collector version environment information no response opentelemetry collector configuration no response log output no response additional context no response
| 1
|
23,347
| 6,419,936,737
|
IssuesEvent
|
2017-08-08 22:29:49
|
xhqiao89/HydroDesktop_test
|
https://api.github.com/repos/xhqiao89/HydroDesktop_test
|
closed
|
Add option 'save WaterML files' to search download settings
|
CodePlex
|
<b>jirikadlec2[CodePlex]</b> <br />It seems that HydroDesktop is saving all the GetValues WaterML files to the C:\users\user name\appdata\local\temp\HydroDesktop folder. This uses large hard drive space on the user's computer. For example I found that the folder size with the WaterML files
was 2 Gigabytes after I downloaded all data from the 'Little Bear River experimental watershed' web service.
Improvement suggestion:
Add a 'save WaterML files' checkbox to the search download settings. If the checkbox is unchecked then don't save the downloaded WaterML files to the temporary folder.
Show the name of the folder where the WaterML files are saved (usually it's C:\users\user name\appdata\local\temp\HydroDesktop)
|
1.0
|
Add option 'save WaterML files' to search download settings - <b>jirikadlec2[CodePlex]</b> <br />It seems that HydroDesktop is saving all the GetValues WaterML files to the C:\users\user name\appdata\local\temp\HydroDesktop folder. This uses large hard drive space on the user's computer. For example I found that the folder size with the WaterML files
was 2 Gigabytes after I downloaded all data from the 'Little Bear River experimental watershed' web service.
Improvement suggestion:
Add a 'save WaterML files' checkbox to the search download settings. If the checkbox is unchecked then don't save the downloaded WaterML files to the temporary folder.
Show the name of the folder where the WaterML files are saved (usually it's C:\users\user name\appdata\local\temp\HydroDesktop)
|
non_process
|
add option save waterml files to search download settings it seems that hydrodesktop is saving all the getvalues waterml files to the c users user name appdata local temp hydrodesktop folder this uses large hard drive space on the user s computer for example i found that the folder size with the waterml files was gigabytes after i downloaded all data from the little bear river experimental watershed web service improvement suggestion add a save waterml files checkbox to the search download settings if the checkbox is unchecked then don t save the downloaded waterml files to the temporary folder show the name of the folder where the waterml files are saved usually it s c users user name appdata local temp hydrodesktop
| 0
|
554,111
| 16,389,382,480
|
IssuesEvent
|
2021-05-17 14:23:27
|
Bibliome/alvisnlp
|
https://api.github.com/repos/Bibliome/alvisnlp
|
closed
|
Function after bug
|
Priority: High Status: Done Topic: Factory Type: Bug
|
Function `after:LAYER` should return annotations that start at the same offset than the context annotation.
Currently `after` skips the adjacent annotation after the context.
|
1.0
|
Function after bug - Function `after:LAYER` should return annotations that start at the same offset than the context annotation.
Currently `after` skips the adjacent annotation after the context.
|
non_process
|
function after bug function after layer should return annotations that start at the same offset than the context annotation currently after skips the adjacent annotation after the context
| 0
|
747,202
| 26,077,171,352
|
IssuesEvent
|
2022-12-24 18:21:32
|
pystardust/ani-cli
|
https://api.github.com/repos/pystardust/ani-cli
|
opened
|
Episodes not released yet!
|
type: bug priority 2: medium
|
Version: 3.4.7
OS: Windows 10
Shell: git bash
Anime: all i searched for
every time i search for an anime and select the title it gives the error "Episodes not released yet!"
**Steps To Reproduce**
1. Run `ani-cli `
2. search for an anime
3. Choose any one
**Expected behavior**
it should ask for the episode number
**Screenshots (if applicable; you can just drag the image onto github)**

|
1.0
|
Episodes not released yet! -
Version: 3.4.7
OS: Windows 10
Shell: git bash
Anime: all i searched for
every time i search for an anime and select the title it gives the error "Episodes not released yet!"
**Steps To Reproduce**
1. Run `ani-cli `
2. search for an anime
3. Choose any one
**Expected behavior**
it should ask for the episode number
**Screenshots (if applicable; you can just drag the image onto github)**

|
non_process
|
episodes not released yet version os windows shell git bash anime all i searched for every time i search for an anime and select the title it gives the error episodes not released yet steps to reproduce run ani cli search for an anime choose any one expected behavior it should ask for the episode number screenshots if applicable you can just drag the image onto github
| 0
|
8,454
| 11,628,002,550
|
IssuesEvent
|
2020-02-27 17:29:32
|
qri-io/qri
|
https://api.github.com/repos/qri-io/qri
|
closed
|
executable bit not set on binary releases, date is wrong as well
|
bug release process
|
https://github.com/qri-io/qri/releases/download/v0.9.4/qri_linux_amd64.zip
Executable bit is not set.
Modified date is 12/31/1979.
I did not check any other architectures.
|
1.0
|
executable bit not set on binary releases, date is wrong as well - https://github.com/qri-io/qri/releases/download/v0.9.4/qri_linux_amd64.zip
Executable bit is not set.
Modified date is 12/31/1979.
I did not check any other architectures.
|
process
|
executable bit not set on binary releases date is wrong as well executable bit is not set modified date is i did not check any other architectures
| 1
|
8,635
| 6,599,860,081
|
IssuesEvent
|
2017-09-17 02:33:23
|
thrust/thrust
|
https://api.github.com/repos/thrust/thrust
|
closed
|
slow compilation with CUDA 6.5
|
performance bug wontfix
|
I just switched from CUDA 6 to CUDA 6.5 and noticed that compilation of my thrust based code is slowed down **a lot**.
Because my own, more complex code did not yet finish compiling I created a simple demo program:
https://gist.github.com/dachziegel/c7bc0226fe2f16196f3e
When I use the thrust version which is delivered with CUDA 6.5 this takes around 6 seconds:
```
$ time nvcc -arch sm_30 compile_time.cu
real 0m6.469s
user 0m6.010s
sys 0m0.452s
```
When I use the thrust development version from github it takes around 30 (!) seconds:
```
$time nvcc -arch sm_30 -I/tmp/thrust-github/ compile_time.cu
real 0m31.051s
user 0m30.237s
sys 0m0.782s
```
The difference is even higher if the program is a bit more complex, e.g. this one from another bug report: https://gist.github.com/dachziegel/11288269
when using CUDA 6.5:
thrust from CUDA 6.5 takes around **7 seconds**, thrust development version from github took **7.44 minutes**(!).
when using CUDA 6:
thrust development version from github takes **51 seconds**.
when using CUDA 7:
thrust development version from github takes **41 seconds**.
My system is a 64bit Ubuntu 14.04.
|
True
|
slow compilation with CUDA 6.5 - I just switched from CUDA 6 to CUDA 6.5 and noticed that compilation of my thrust based code is slowed down **a lot**.
Because my own, more complex code did not yet finish compiling I created a simple demo program:
https://gist.github.com/dachziegel/c7bc0226fe2f16196f3e
When I use the thrust version which is delivered with CUDA 6.5 this takes around 6 seconds:
```
$ time nvcc -arch sm_30 compile_time.cu
real 0m6.469s
user 0m6.010s
sys 0m0.452s
```
When I use the thrust development version from github it takes around 30 (!) seconds:
```
$time nvcc -arch sm_30 -I/tmp/thrust-github/ compile_time.cu
real 0m31.051s
user 0m30.237s
sys 0m0.782s
```
The difference is even higher if the program is a bit more complex, e.g. this one from another bug report: https://gist.github.com/dachziegel/11288269
when using CUDA 6.5:
thrust from CUDA 6.5 takes around **7 seconds**, thrust development version from github took **7.44 minutes**(!).
when using CUDA 6:
thrust development version from github takes **51 seconds**.
when using CUDA 7:
thrust development version from github takes **41 seconds**.
My system is a 64bit Ubuntu 14.04.
|
non_process
|
slow compilation with cuda i just switched from cuda to cuda and noticed that compilation of my thrust based code is slowed down a lot because my own more complex code did not yet finish compiling i created a simple demo program when i use the thrust version which is delivered with cuda this takes around seconds time nvcc arch sm compile time cu real user sys when i use the thrust development version from github it takes around seconds time nvcc arch sm i tmp thrust github compile time cu real user sys the difference is even higher if the program is a bit more complex e g this one from another bug report when using cuda thrust from cuda takes around seconds thrust development version from github took minutes when using cuda thrust development version from github takes seconds when using cuda thrust development version from github takes seconds my system is a ubuntu
| 0
|
14,058
| 16,870,223,295
|
IssuesEvent
|
2021-06-22 02:49:17
|
Leviatan-Analytics/LA-data-processing
|
https://api.github.com/repos/Leviatan-Analytics/LA-data-processing
|
closed
|
Test YoloV3 model [2]
|
Data Processing Sprint 2 Week 4
|
Estimated time: 1 hs per assignee
Test the accuracy of the trained model with different inputs.
Output: Document with the result of the different inputs test.
|
1.0
|
Test YoloV3 model [2] - Estimated time: 1 hs per assignee
Test the accuracy of the trained model with different inputs.
Output: Document with the result of the different inputs test.
|
process
|
test model estimated time hs per assignee test the accuracy of the trained model with different inputs output document with the result of the different inputs test
| 1
|
19,046
| 25,048,270,715
|
IssuesEvent
|
2022-11-05 14:51:54
|
fluent/fluent-bit
|
https://api.github.com/repos/fluent/fluent-bit
|
closed
|
Changes in v2.0.0+ break static compilation on armhf
|
work-in-process
|
## Bug Report
**Describe the bug**
Changes in v2.0.0+ break static compilation on `armhf`.
I have been building statically-compiled Fluent-Bit on the 4 supported platforms (i386, amd64, armhf, arm64) for a while now in [this project](https://github.com/brian-maloney/fluent-bit-static). I use Docker with a Debian bullseye base to build my binaries.
With the release of v2.0.0 all my builds broke.
After some research, I tracked down that changes in `fb_hash` to directly use OpenSSL instead of Monkey required me to add `libssl-dev` to my build image, which caused further downstream issues requiring me to then disable WASM and LUAJIT. All of these are acceptable for my use case, but while I was able to solve the issues in most general cases, I am stuck on a linking issue with `armhf`.
**To Reproduce**
- Link error:
```
/usr/bin/ld: /usr/lib/gcc/arm-linux-gnueabihf/10/../../../arm-linux-gnueabihf/libc.a(getcontext.o): in function `__getcontext':
(.text+0x74): undefined reference to `_rtld_global_ro'
/usr/bin/ld: /usr/lib/gcc/arm-linux-gnueabihf/10/../../../arm-linux-gnueabihf/libc.a(setcontext.o): in function `__startcontext':
(.text+0x80): undefined reference to `_rtld_global_ro'
```
- Steps to reproduce the problem:
```
git clone https://github.com/brian-maloney/fluent-bit-static.git
cd fluent-bit-static
docker build --progress plain --build-arg "TAG=v2.0.0" --build-arg "BUILDARCH=armhf" --platform 'linux/arm/v7' -t fluent-bit-static .
```
**Expected behavior**
Successful build and link
**Your Environment**
<!--- Include as many relevant details about the environment you experienced the bug in -->
* Version used: 2.0.0
* Configuration: Build in Docker
**Additional context**
<!--- How has this issue affected you? What are you trying to accomplish? -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
I understand this isn't a directly supported use case, and I'm looking for some unusual assistance, but I'd love to avoid having to drop `armhf` support on these builds that I supply to the community.
|
1.0
|
Changes in v2.0.0+ break static compilation on armhf - ## Bug Report
**Describe the bug**
Changes in v2.0.0+ break static compilation on `armhf`.
I have been building statically-compiled Fluent-Bit on the 4 supported platforms (i386, amd64, armhf, arm64) for a while now in [this project](https://github.com/brian-maloney/fluent-bit-static). I use Docker with a Debian bullseye base to build my binaries.
With the release of v2.0.0 all my builds broke.
After some research, I tracked down that changes in `fb_hash` to directly use OpenSSL instead of Monkey required me to add `libssl-dev` to my build image, which caused further downstream issues requiring me to then disable WASM and LUAJIT. All of these are acceptable for my use case, but while I was able to solve the issues in most general cases, I am stuck on a linking issue with `armhf`.
**To Reproduce**
- Link error:
```
/usr/bin/ld: /usr/lib/gcc/arm-linux-gnueabihf/10/../../../arm-linux-gnueabihf/libc.a(getcontext.o): in function `__getcontext':
(.text+0x74): undefined reference to `_rtld_global_ro'
/usr/bin/ld: /usr/lib/gcc/arm-linux-gnueabihf/10/../../../arm-linux-gnueabihf/libc.a(setcontext.o): in function `__startcontext':
(.text+0x80): undefined reference to `_rtld_global_ro'
```
- Steps to reproduce the problem:
```
git clone https://github.com/brian-maloney/fluent-bit-static.git
cd fluent-bit-static
docker build --progress plain --build-arg "TAG=v2.0.0" --build-arg "BUILDARCH=armhf" --platform 'linux/arm/v7' -t fluent-bit-static .
```
**Expected behavior**
Successful build and link
**Your Environment**
<!--- Include as many relevant details about the environment you experienced the bug in -->
* Version used: 2.0.0
* Configuration: Build in Docker
**Additional context**
<!--- How has this issue affected you? What are you trying to accomplish? -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
I understand this isn't a directly supported use case, and I'm looking for some unusual assistance, but I'd love to avoid having to drop `armhf` support on these builds that I supply to the community.
|
process
|
changes in break static compilation on armhf bug report describe the bug changes in break static compilation on armhf i have been building statically compiled fluent bit on the supported platforms armhf for a while now in i use docker with a debian bullseye base to build my binaries with the release of all my builds broke after some research i tracked down that changes in fb hash to directly use openssl instead of monkey required me to add libssl dev to my build image which caused further downstream issues requiring me to then disable wasm and luajit all of these are acceptable for my use case but while i was able to solve the issues in most general cases i am stuck on a linking issue with armhf to reproduce link error usr bin ld usr lib gcc arm linux gnueabihf arm linux gnueabihf libc a getcontext o in function getcontext text undefined reference to rtld global ro usr bin ld usr lib gcc arm linux gnueabihf arm linux gnueabihf libc a setcontext o in function startcontext text undefined reference to rtld global ro steps to reproduce the problem git clone cd fluent bit static docker build progress plain build arg tag build arg buildarch armhf platform linux arm t fluent bit static expected behavior successful build and link your environment version used configuration build in docker additional context i understand this isn t a directly supported use case and i m looking for some unusual assistance but i d love to avoid having to drop armhf support on these builds that i supply to the community
| 1
|
6,729
| 9,842,157,502
|
IssuesEvent
|
2019-06-18 08:40:35
|
syndesisio/syndesis
|
https://api.github.com/repos/syndesisio/syndesis
|
opened
|
Continuous Integration improvement: Improve Elephant carpaccio
|
cat/process cat/research
|
We should find ways to cut features in smaller end-to-end deliverables. It would allow us to integrate our software faster and recover feedback sooner.
|
1.0
|
Continuous Integration improvement: Improve Elephant carpaccio - We should find ways to cut features in smaller end-to-end deliverables. It would allow us to integrate our software faster and recover feedback sooner.
|
process
|
continuous integration improvement improve elephant carpaccio we should find ways to cut features in smaller end to end deliverables it would allow us to integrate our software faster and recover feedback sooner
| 1
|
18,251
| 24,334,011,045
|
IssuesEvent
|
2022-09-30 23:08:18
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
Log flooded with `string to non-string with op EQ`
|
bug priority:p2 processor/transform
|
### What happened?
## Description
Upgraded to 0.60.0.
Using a transform processor with a clause such as:
```
transform:
metrics:
queries:
- set(attributes["foo"], "bar") where attributes["foo"] == "foobar"
```
## Steps to Reproduce
Set up a pipeline using a transform processor that uses an eq operator.
## Expected Result
The pipeline should perform correctly.
## Actual Result
Logs are flooded with:
```
string to non-string with op EQ
string to non-string with op EQ
string to non-string with op EQ
string to non-string with op EQ
string to non-string with op EQ
string to non-string with op EQ
string to non-string with op EQ
string to non-string with op EQ
```
### Collector version
0.60.0
### Environment information
## Environment
OS: (e.g., "Ubuntu 20.04")
Compiler(if manually compiled): (e.g., "go 14.2")
### OpenTelemetry Collector configuration
```yaml
receivers:
splunk_hec:
exporters:
splunk_hec/metrics:
# Splunk HTTP Event Collector token.
token: "00000000-0000-0000-0000-0000000000000"
# URL to a Splunk instance to send data to.
endpoint: "https://splunk:8088/services/collector"
# Optional Splunk source: https://docs.splunk.com/Splexicon:Source
source: "app"
# Optional Splunk source type: https://docs.splunk.com/Splexicon:Sourcetype
sourcetype: "telegraf"
# Splunk index, optional name of the Splunk index targeted.
index: "metrics"
# Maximum HTTP connections to use simultaneously when sending data. Defaults to 100.
max_connections: 20
# Whether to disable gzip compression over HTTP. Defaults to false.
disable_compression: false
# HTTP timeout when sending data. Defaults to 10s.
timeout: 10s
# Whether to skip checking the certificate of the HEC endpoint when sending data over HTTPS. Defaults to false.
# For this demo, we use a self-signed certificate on the Splunk docker instance, so this flag is set to true.
insecure_skip_verify: true
processors:
batch:
transform:
metrics:
queries:
- set(attributes["foo"], "bar") where attributes["foo"] == "foobar"
extensions:
health_check:
endpoint: 0.0.0.0:13133
pprof:
endpoint: :1888
zpages:
endpoint: :55679
service:
telemetry:
logs:
level: "debug"
extensions: [pprof, zpages, health_check]
pipelines:
metrics:
receivers: [splunk_hec]
processors: [batch, transform]
exporters: [splunk_hec/metrics]
```
### Log output
_No response_
### Additional context
_No response_
|
1.0
|
Log flooded with `string to non-string with op EQ` - ### What happened?
## Description
Upgraded to 0.60.0.
Using a transform processor with a clause such as:
```
transform:
metrics:
queries:
- set(attributes["foo"], "bar") where attributes["foo"] == "foobar"
```
## Steps to Reproduce
Set up a pipeline using a transform processor that uses an eq operator.
## Expected Result
The pipeline should perform correctly.
## Actual Result
Logs are flooded with:
```
string to non-string with op EQ
string to non-string with op EQ
string to non-string with op EQ
string to non-string with op EQ
string to non-string with op EQ
string to non-string with op EQ
string to non-string with op EQ
string to non-string with op EQ
```
### Collector version
0.60.0
### Environment information
## Environment
OS: (e.g., "Ubuntu 20.04")
Compiler(if manually compiled): (e.g., "go 14.2")
### OpenTelemetry Collector configuration
```yaml
receivers:
splunk_hec:
exporters:
splunk_hec/metrics:
# Splunk HTTP Event Collector token.
token: "00000000-0000-0000-0000-0000000000000"
# URL to a Splunk instance to send data to.
endpoint: "https://splunk:8088/services/collector"
# Optional Splunk source: https://docs.splunk.com/Splexicon:Source
source: "app"
# Optional Splunk source type: https://docs.splunk.com/Splexicon:Sourcetype
sourcetype: "telegraf"
# Splunk index, optional name of the Splunk index targeted.
index: "metrics"
# Maximum HTTP connections to use simultaneously when sending data. Defaults to 100.
max_connections: 20
# Whether to disable gzip compression over HTTP. Defaults to false.
disable_compression: false
# HTTP timeout when sending data. Defaults to 10s.
timeout: 10s
# Whether to skip checking the certificate of the HEC endpoint when sending data over HTTPS. Defaults to false.
# For this demo, we use a self-signed certificate on the Splunk docker instance, so this flag is set to true.
insecure_skip_verify: true
processors:
batch:
transform:
metrics:
queries:
- set(attributes["foo"], "bar") where attributes["foo"] == "foobar"
extensions:
health_check:
endpoint: 0.0.0.0:13133
pprof:
endpoint: :1888
zpages:
endpoint: :55679
service:
telemetry:
logs:
level: "debug"
extensions: [pprof, zpages, health_check]
pipelines:
metrics:
receivers: [splunk_hec]
processors: [batch, transform]
exporters: [splunk_hec/metrics]
```
### Log output
_No response_
### Additional context
_No response_
|
process
|
log flooded with string to non string with op eq what happened description upgraded to using a transform processor with a clause such as transform metrics queries set attributes bar where attributes foobar steps to reproduce set up a pipeline using a transform processor that uses an eq operator expected result the pipeline should perform correctly actual result logs are flooded with string to non string with op eq string to non string with op eq string to non string with op eq string to non string with op eq string to non string with op eq string to non string with op eq string to non string with op eq string to non string with op eq collector version environment information environment os e g ubuntu compiler if manually compiled e g go opentelemetry collector configuration yaml receivers splunk hec exporters splunk hec metrics splunk http event collector token token url to a splunk instance to send data to endpoint optional splunk source source app optional splunk source type sourcetype telegraf splunk index optional name of the splunk index targeted index metrics maximum http connections to use simultaneously when sending data defaults to max connections whether to disable gzip compression over http defaults to false disable compression false http timeout when sending data defaults to timeout whether to skip checking the certificate of the hec endpoint when sending data over https defaults to false for this demo we use a self signed certificate on the splunk docker instance so this flag is set to true insecure skip verify true processors batch transform metrics queries set attributes bar where attributes foobar extensions health check endpoint pprof endpoint zpages endpoint service telemetry logs level debug extensions pipelines metrics receivers processors exporters log output no response additional context no response
| 1
|
1,230
| 3,769,103,470
|
IssuesEvent
|
2016-03-16 09:16:19
|
e-government-ua/i
|
https://api.github.com/repos/e-government-ua/i
|
closed
|
Поправить футер
|
active In process of testing test _central-js
|
1. под последней услугой больше интервал.
2. левый край услуг должен быть по одной линии с левым краем конктента в футере.
|
1.0
|
Поправить футер - 1. под последней услугой больше интервал.
2. левый край услуг должен быть по одной линии с левым краем конктента в футере.
|
process
|
поправить футер под последней услугой больше интервал левый край услуг должен быть по одной линии с левым краем конктента в футере
| 1
|
16,800
| 5,290,798,002
|
IssuesEvent
|
2017-02-08 20:49:58
|
dotnet/coreclr
|
https://api.github.com/repos/dotnet/coreclr
|
opened
|
Optimize default(T) == null at compile time
|
area-CodeGen optimization
|
`default(T) == null` in generic code is a popular* pattern to determine if `T` is a nullable or a reference type.
Example:
```csharp
using System;
using System.Runtime.CompilerServices;
class Program
{
[MethodImpl(MethodImplOptions.NoInlining)]
static bool IsNotNullableValueType<T>()
{
return default(T) != null;
}
private static void Main(string[] args)
{
Console.WriteLine(IsNotNullableValueType<string>());
Console.WriteLine(IsNotNullableValueType<int>());
Console.WriteLine(IsNotNullableValueType<int?>());
}
}
```
Will print `False True False`.
While the codegen for the "it's a reference type" case is pretty efficient:
```nasm
xor eax,eax
test rax,rax
setb al
movzx eax,al
ret
```
For the "it's a valuetype" and "it's a nullable" case we get a useless allocation:
```nasm
sub rsp,28h
call __NewHelper_System_Private_CoreLib_System_Int32
xor edx,edx
mov dword ptr [rax+8],edx
test rax,rax
seta al
movzx eax,al
add rsp,28h
ret
```
The places that use this pattern would likely benefit from being able to statically evaluate this because it could lead to elimination of entire branches and smaller code (aside from removing a useless allocation).
\* There are 17 hits for the `default\(.*\) == null` regular expression in the CoreFX repo and a couple more for the "not equal to null" case.
|
1.0
|
Optimize default(T) == null at compile time - `default(T) == null` in generic code is a popular* pattern to determine if `T` is a nullable or a reference type.
Example:
```csharp
using System;
using System.Runtime.CompilerServices;
class Program
{
[MethodImpl(MethodImplOptions.NoInlining)]
static bool IsNotNullableValueType<T>()
{
return default(T) != null;
}
private static void Main(string[] args)
{
Console.WriteLine(IsNotNullableValueType<string>());
Console.WriteLine(IsNotNullableValueType<int>());
Console.WriteLine(IsNotNullableValueType<int?>());
}
}
```
Will print `False True False`.
While the codegen for the "it's a reference type" case is pretty efficient:
```nasm
xor eax,eax
test rax,rax
setb al
movzx eax,al
ret
```
For the "it's a valuetype" and "it's a nullable" case we get a useless allocation:
```nasm
sub rsp,28h
call __NewHelper_System_Private_CoreLib_System_Int32
xor edx,edx
mov dword ptr [rax+8],edx
test rax,rax
seta al
movzx eax,al
add rsp,28h
ret
```
The places that use this pattern would likely benefit from being able to statically evaluate this because it could lead to elimination of entire branches and smaller code (aside from removing a useless allocation).
\* There are 17 hits for the `default\(.*\) == null` regular expression in the CoreFX repo and a couple more for the "not equal to null" case.
|
non_process
|
optimize default t null at compile time default t null in generic code is a popular pattern to determine if t is a nullable or a reference type example csharp using system using system runtime compilerservices class program static bool isnotnullablevaluetype return default t null private static void main string args console writeline isnotnullablevaluetype console writeline isnotnullablevaluetype console writeline isnotnullablevaluetype will print false true false while the codegen for the it s a reference type case is pretty efficient nasm xor eax eax test rax rax setb al movzx eax al ret for the it s a valuetype and it s a nullable case we get a useless allocation nasm sub rsp call newhelper system private corelib system xor edx edx mov dword ptr edx test rax rax seta al movzx eax al add rsp ret the places that use this pattern would likely benefit from being able to statically evaluate this because it could lead to elimination of entire branches and smaller code aside from removing a useless allocation there are hits for the default null regular expression in the corefx repo and a couple more for the not equal to null case
| 0
|
21,229
| 28,320,587,790
|
IssuesEvent
|
2023-04-11 00:31:38
|
nephio-project/sig-release
|
https://api.github.com/repos/nephio-project/sig-release
|
closed
|
Establish Roles and Responsibilities in the community for helping with Nephio Releases.
|
area/process-mgmt sig/release
|
We need to establish clear roles and responsibilities in the community for helping with the Nephio release process. We can start by having a Release Lead (RL) role which will be a role that will be rotated across SIG release members for every minor release.
Release lead will be responsible for the following,
Coordinate with other SIG(s) on finalizing the issues that are being targeted for the release.
Lead all the phases of the release process defined earlier.
Lead the release postmortem for the release.
Identifying the RL for the next release by working with SIG release.
We can define additional roles as required for the release process as we learn from our experience with doing a few releases.
|
1.0
|
Establish Roles and Responsibilities in the community for helping with Nephio Releases. - We need to establish clear roles and responsibilities in the community for helping with the Nephio release process. We can start by having a Release Lead (RL) role which will be a role that will be rotated across SIG release members for every minor release.
Release lead will be responsible for the following,
Coordinate with other SIG(s) on finalizing the issues that are being targeted for the release.
Lead all the phases of the release process defined earlier.
Lead the release postmortem for the release.
Identifying the RL for the next release by working with SIG release.
We can define additional roles as required for the release process as we learn from our experience with doing a few releases.
|
process
|
establish roles and responsibilities in the community for helping with nephio releases we need to establish clear roles and responsibilities in the community for helping with the nephio release process we can start by having a release lead rl role which will be a role that will be rotated across sig release members for every minor release release lead will be responsible for the following coordinate with other sig s on finalizing the issues that are being targeted for the release lead all the phases of the release process defined earlier lead the release postmortem for the release identifying the rl for the next release by working with sig release we can define additional roles as required for the release process as we learn from our experience with doing a few releases
| 1
|
21,191
| 28,209,461,989
|
IssuesEvent
|
2023-04-05 02:00:10
|
lizhihao6/get-daily-arxiv-noti
|
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
|
opened
|
New submissions for Wed, 5 Apr 23
|
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
|
## Keyword: events
### Fully Convolutional Networks for Dense Water Flow Intensity Prediction in Swedish Catchment Areas
- **Authors:** Aleksis Pirinen, Olof Mogren, Mårten Västerdal
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2304.01658
- **Pdf link:** https://arxiv.org/pdf/2304.01658
- **Abstract**
Intensifying climate change will lead to more extreme weather events, including heavy rainfall and drought. Accurate stream flow prediction models which are adaptable and robust to new circumstances in a changing climate will be an important source of information for decisions on climate adaptation efforts, especially regarding mitigation of the risks of and damages associated with flooding. In this work we propose a machine learning-based approach for predicting water flow intensities in inland watercourses based on the physical characteristics of the catchment areas, obtained from geospatial data (including elevation and soil maps, as well as satellite imagery), in addition to temporal information about past rainfall quantities and temperature variations. We target the one-day-ahead regime, where a fully convolutional neural network model receives spatio-temporal inputs and predicts the water flow intensity in every coordinate of the spatial input for the subsequent day. To the best of our knowledge, we are the first to tackle the task of dense water flow intensity prediction; earlier works have considered predicting flow intensities at a sparse set of locations at a time. An extensive set of model evaluations and ablations are performed, which empirically justify our various design choices. Code and preprocessed data have been made publicly available at https://github.com/aleksispi/fcn-water-flow.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
### IterativePFN: True Iterative Point Cloud Filtering
- **Authors:** Dasith de Silva Edirimuni, Xuequan Lu, Zhiwen Shao, Gang Li, Antonio Robles-Kelly, Ying He
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2304.01529
- **Pdf link:** https://arxiv.org/pdf/2304.01529
- **Abstract**
The quality of point clouds is often limited by noise introduced during their capture process. Consequently, a fundamental 3D vision task is the removal of noise, known as point cloud filtering or denoising. State-of-the-art learning based methods focus on training neural networks to infer filtered displacements and directly shift noisy points onto the underlying clean surfaces. In high noise conditions, they iterate the filtering process. However, this iterative filtering is only done at test time and is less effective at ensuring points converge quickly onto the clean surfaces. We propose IterativePFN (iterative point cloud filtering network), which consists of multiple IterationModules that model the true iterative filtering process internally, within a single network. We train our IterativePFN network using a novel loss function that utilizes an adaptive ground truth target at each iteration to capture the relationship between intermediate filtering results during training. This ensures that the filtered results converge faster to the clean surfaces. Our method is able to obtain better performance compared to state-of-the-art methods. The source code can be found at: https://github.com/ddsediri/IterativePFN.
### Fully Convolutional Networks for Dense Water Flow Intensity Prediction in Swedish Catchment Areas
- **Authors:** Aleksis Pirinen, Olof Mogren, Mårten Västerdal
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2304.01658
- **Pdf link:** https://arxiv.org/pdf/2304.01658
- **Abstract**
Intensifying climate change will lead to more extreme weather events, including heavy rainfall and drought. Accurate stream flow prediction models which are adaptable and robust to new circumstances in a changing climate will be an important source of information for decisions on climate adaptation efforts, especially regarding mitigation of the risks of and damages associated with flooding. In this work we propose a machine learning-based approach for predicting water flow intensities in inland watercourses based on the physical characteristics of the catchment areas, obtained from geospatial data (including elevation and soil maps, as well as satellite imagery), in addition to temporal information about past rainfall quantities and temperature variations. We target the one-day-ahead regime, where a fully convolutional neural network model receives spatio-temporal inputs and predicts the water flow intensity in every coordinate of the spatial input for the subsequent day. To the best of our knowledge, we are the first to tackle the task of dense water flow intensity prediction; earlier works have considered predicting flow intensities at a sparse set of locations at a time. An extensive set of model evaluations and ablations are performed, which empirically justify our various design choices. Code and preprocessed data have been made publicly available at https://github.com/aleksispi/fcn-water-flow.
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
There is no result
## Keyword: RAW
### A Scale-Invariant Trajectory Simplification Method for Efficient Data Collection in Videos
- **Authors:** Yang Liu, Luiz Gustavo Hafemann
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2304.01340
- **Pdf link:** https://arxiv.org/pdf/2304.01340
- **Abstract**
Training data is a critical requirement for machine learning tasks, and labeled training data can be expensive to acquire, often requiring manual or semi-automated data collection pipelines. For tracking applications, the data collection involves drawing bounding boxes around the classes of interest on each frame, and associate detections of the same "instance" over frames. In a semi-automated data collection pipeline, this can be achieved by running a baseline detection and tracking algorithm, and relying on manual correction to add/remove/change bounding boxes on each frame, as well as resolving errors in the associations over frames (track switches). In this paper, we propose a data correction pipeline to generate ground-truth data more efficiently in this semi-automated scenario. Our method simplifies the trajectories from the tracking systems and let the annotator verify and correct the objects in the sampled keyframes. Once the objects in the keyframes are corrected, the bounding boxes in the other frames are obtained by interpolation. Our method achieves substantial reduction in the number of frames requiring manual correction. In the MOT dataset, it reduces the number of frames by 30x while maintaining a HOTA score of 89.61% . Moreover, it reduces the number of frames by a factor of 10x while achieving a HOTA score of 79.24% in the SoccerNet dataset, and 85.79% in the DanceTrack dataset. The project code and data are publicly released at https://github.com/foreverYoungGitHub/trajectory-simplify-benchmark.
### Self-Supervised Image Denoising for Real-World Images with Context-aware Transformer
- **Authors:** Dan Zhang, Fangfang Zhou
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2304.01627
- **Pdf link:** https://arxiv.org/pdf/2304.01627
- **Abstract**
In recent years, the development of deep learning has been pushing image denoising to a new level. Among them, self-supervised denoising is increasingly popular because it does not require any prior knowledge. Most of the existing self-supervised methods are based on convolutional neural networks (CNN), which are restricted by the locality of the receptive field and would cause color shifts or textures loss. In this paper, we propose a novel Denoise Transformer for real-world image denoising, which is mainly constructed with Context-aware Denoise Transformer (CADT) units and Secondary Noise Extractor (SNE) block. CADT is designed as a dual-branch structure, where the global branch uses a window-based Transformer encoder to extract the global information, while the local branch focuses on the extraction of local features with small receptive field. By incorporating CADT as basic components, we build a hierarchical network to directly learn the noise distribution information through residual learning and obtain the first stage denoised output. Then, we design SNE in low computation for secondary global noise extraction. Finally the blind spots are collected from the Denoise Transformer output and reconstructed, forming the final denoised image. Extensive experiments on the real-world SIDD benchmark achieve 50.62/0.990 for PSNR/SSIM, which is competitive with the current state-of-the-art method and only 0.17/0.001 lower. Visual comparisons on public sRGB, Raw-RGB and greyscale datasets prove that our proposed Denoise Transformer has a competitive performance, especially on blurred textures and low-light images, without using additional knowledge, e.g., noise level or noise type, regarding the underlying unknown noise.
### Trace and Pace: Controllable Pedestrian Animation via Guided Trajectory Diffusion
- **Authors:** Davis Rempe, Zhengyi Luo, Xue Bin Peng, Ye Yuan, Kris Kitani, Karsten Kreis, Sanja Fidler, Or Litany
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR); Machine Learning (cs.LG)
- **Arxiv link:** https://arxiv.org/abs/2304.01893
- **Pdf link:** https://arxiv.org/pdf/2304.01893
- **Abstract**
We introduce a method for generating realistic pedestrian trajectories and full-body animations that can be controlled to meet user-defined goals. We draw on recent advances in guided diffusion modeling to achieve test-time controllability of trajectories, which is normally only associated with rule-based systems. Our guided diffusion model allows users to constrain trajectories through target waypoints, speed, and specified social groups while accounting for the surrounding environment context. This trajectory diffusion model is integrated with a novel physics-based humanoid controller to form a closed-loop, full-body pedestrian animation system capable of placing large crowds in a simulated environment with varying terrains. We further propose utilizing the value function learned during RL training of the animation controller to guide diffusion to produce trajectories better suited for particular scenarios such as collision avoidance and traversing uneven terrain. Video results are available on the project page at https://nv-tlabs.github.io/trace-pace .
## Keyword: raw image
There is no result
|
2.0
|
New submissions for Wed, 5 Apr 23 - ## Keyword: events
### Fully Convolutional Networks for Dense Water Flow Intensity Prediction in Swedish Catchment Areas
- **Authors:** Aleksis Pirinen, Olof Mogren, Mårten Västerdal
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2304.01658
- **Pdf link:** https://arxiv.org/pdf/2304.01658
- **Abstract**
Intensifying climate change will lead to more extreme weather events, including heavy rainfall and drought. Accurate stream flow prediction models which are adaptable and robust to new circumstances in a changing climate will be an important source of information for decisions on climate adaptation efforts, especially regarding mitigation of the risks of and damages associated with flooding. In this work we propose a machine learning-based approach for predicting water flow intensities in inland watercourses based on the physical characteristics of the catchment areas, obtained from geospatial data (including elevation and soil maps, as well as satellite imagery), in addition to temporal information about past rainfall quantities and temperature variations. We target the one-day-ahead regime, where a fully convolutional neural network model receives spatio-temporal inputs and predicts the water flow intensity in every coordinate of the spatial input for the subsequent day. To the best of our knowledge, we are the first to tackle the task of dense water flow intensity prediction; earlier works have considered predicting flow intensities at a sparse set of locations at a time. An extensive set of model evaluations and ablations are performed, which empirically justify our various design choices. Code and preprocessed data have been made publicly available at https://github.com/aleksispi/fcn-water-flow.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
### IterativePFN: True Iterative Point Cloud Filtering
- **Authors:** Dasith de Silva Edirimuni, Xuequan Lu, Zhiwen Shao, Gang Li, Antonio Robles-Kelly, Ying He
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2304.01529
- **Pdf link:** https://arxiv.org/pdf/2304.01529
- **Abstract**
The quality of point clouds is often limited by noise introduced during their capture process. Consequently, a fundamental 3D vision task is the removal of noise, known as point cloud filtering or denoising. State-of-the-art learning based methods focus on training neural networks to infer filtered displacements and directly shift noisy points onto the underlying clean surfaces. In high noise conditions, they iterate the filtering process. However, this iterative filtering is only done at test time and is less effective at ensuring points converge quickly onto the clean surfaces. We propose IterativePFN (iterative point cloud filtering network), which consists of multiple IterationModules that model the true iterative filtering process internally, within a single network. We train our IterativePFN network using a novel loss function that utilizes an adaptive ground truth target at each iteration to capture the relationship between intermediate filtering results during training. This ensures that the filtered results converge faster to the clean surfaces. Our method is able to obtain better performance compared to state-of-the-art methods. The source code can be found at: https://github.com/ddsediri/IterativePFN.
### Fully Convolutional Networks for Dense Water Flow Intensity Prediction in Swedish Catchment Areas
- **Authors:** Aleksis Pirinen, Olof Mogren, Mårten Västerdal
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2304.01658
- **Pdf link:** https://arxiv.org/pdf/2304.01658
- **Abstract**
Intensifying climate change will lead to more extreme weather events, including heavy rainfall and drought. Accurate stream flow prediction models which are adaptable and robust to new circumstances in a changing climate will be an important source of information for decisions on climate adaptation efforts, especially regarding mitigation of the risks of and damages associated with flooding. In this work we propose a machine learning-based approach for predicting water flow intensities in inland watercourses based on the physical characteristics of the catchment areas, obtained from geospatial data (including elevation and soil maps, as well as satellite imagery), in addition to temporal information about past rainfall quantities and temperature variations. We target the one-day-ahead regime, where a fully convolutional neural network model receives spatio-temporal inputs and predicts the water flow intensity in every coordinate of the spatial input for the subsequent day. To the best of our knowledge, we are the first to tackle the task of dense water flow intensity prediction; earlier works have considered predicting flow intensities at a sparse set of locations at a time. An extensive set of model evaluations and ablations are performed, which empirically justify our various design choices. Code and preprocessed data have been made publicly available at https://github.com/aleksispi/fcn-water-flow.
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
There is no result
## Keyword: RAW
### A Scale-Invariant Trajectory Simplification Method for Efficient Data Collection in Videos
- **Authors:** Yang Liu, Luiz Gustavo Hafemann
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2304.01340
- **Pdf link:** https://arxiv.org/pdf/2304.01340
- **Abstract**
Training data is a critical requirement for machine learning tasks, and labeled training data can be expensive to acquire, often requiring manual or semi-automated data collection pipelines. For tracking applications, the data collection involves drawing bounding boxes around the classes of interest on each frame, and associate detections of the same "instance" over frames. In a semi-automated data collection pipeline, this can be achieved by running a baseline detection and tracking algorithm, and relying on manual correction to add/remove/change bounding boxes on each frame, as well as resolving errors in the associations over frames (track switches). In this paper, we propose a data correction pipeline to generate ground-truth data more efficiently in this semi-automated scenario. Our method simplifies the trajectories from the tracking systems and let the annotator verify and correct the objects in the sampled keyframes. Once the objects in the keyframes are corrected, the bounding boxes in the other frames are obtained by interpolation. Our method achieves substantial reduction in the number of frames requiring manual correction. In the MOT dataset, it reduces the number of frames by 30x while maintaining a HOTA score of 89.61% . Moreover, it reduces the number of frames by a factor of 10x while achieving a HOTA score of 79.24% in the SoccerNet dataset, and 85.79% in the DanceTrack dataset. The project code and data are publicly released at https://github.com/foreverYoungGitHub/trajectory-simplify-benchmark.
### Self-Supervised Image Denoising for Real-World Images with Context-aware Transformer
- **Authors:** Dan Zhang, Fangfang Zhou
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2304.01627
- **Pdf link:** https://arxiv.org/pdf/2304.01627
- **Abstract**
In recent years, the development of deep learning has been pushing image denoising to a new level. Among them, self-supervised denoising is increasingly popular because it does not require any prior knowledge. Most of the existing self-supervised methods are based on convolutional neural networks (CNN), which are restricted by the locality of the receptive field and would cause color shifts or textures loss. In this paper, we propose a novel Denoise Transformer for real-world image denoising, which is mainly constructed with Context-aware Denoise Transformer (CADT) units and Secondary Noise Extractor (SNE) block. CADT is designed as a dual-branch structure, where the global branch uses a window-based Transformer encoder to extract the global information, while the local branch focuses on the extraction of local features with small receptive field. By incorporating CADT as basic components, we build a hierarchical network to directly learn the noise distribution information through residual learning and obtain the first stage denoised output. Then, we design SNE in low computation for secondary global noise extraction. Finally the blind spots are collected from the Denoise Transformer output and reconstructed, forming the final denoised image. Extensive experiments on the real-world SIDD benchmark achieve 50.62/0.990 for PSNR/SSIM, which is competitive with the current state-of-the-art method and only 0.17/0.001 lower. Visual comparisons on public sRGB, Raw-RGB and greyscale datasets prove that our proposed Denoise Transformer has a competitive performance, especially on blurred textures and low-light images, without using additional knowledge, e.g., noise level or noise type, regarding the underlying unknown noise.
### Trace and Pace: Controllable Pedestrian Animation via Guided Trajectory Diffusion
- **Authors:** Davis Rempe, Zhengyi Luo, Xue Bin Peng, Ye Yuan, Kris Kitani, Karsten Kreis, Sanja Fidler, Or Litany
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR); Machine Learning (cs.LG)
- **Arxiv link:** https://arxiv.org/abs/2304.01893
- **Pdf link:** https://arxiv.org/pdf/2304.01893
- **Abstract**
We introduce a method for generating realistic pedestrian trajectories and full-body animations that can be controlled to meet user-defined goals. We draw on recent advances in guided diffusion modeling to achieve test-time controllability of trajectories, which is normally only associated with rule-based systems. Our guided diffusion model allows users to constrain trajectories through target waypoints, speed, and specified social groups while accounting for the surrounding environment context. This trajectory diffusion model is integrated with a novel physics-based humanoid controller to form a closed-loop, full-body pedestrian animation system capable of placing large crowds in a simulated environment with varying terrains. We further propose utilizing the value function learned during RL training of the animation controller to guide diffusion to produce trajectories better suited for particular scenarios such as collision avoidance and traversing uneven terrain. Video results are available on the project page at https://nv-tlabs.github.io/trace-pace .
## Keyword: raw image
There is no result
|
process
|
new submissions for wed apr keyword events fully convolutional networks for dense water flow intensity prediction in swedish catchment areas authors aleksis pirinen olof mogren mårten västerdal subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract intensifying climate change will lead to more extreme weather events including heavy rainfall and drought accurate stream flow prediction models which are adaptable and robust to new circumstances in a changing climate will be an important source of information for decisions on climate adaptation efforts especially regarding mitigation of the risks of and damages associated with flooding in this work we propose a machine learning based approach for predicting water flow intensities in inland watercourses based on the physical characteristics of the catchment areas obtained from geospatial data including elevation and soil maps as well as satellite imagery in addition to temporal information about past rainfall quantities and temperature variations we target the one day ahead regime where a fully convolutional neural network model receives spatio temporal inputs and predicts the water flow intensity in every coordinate of the spatial input for the subsequent day to the best of our knowledge we are the first to tackle the task of dense water flow intensity prediction earlier works have considered predicting flow intensities at a sparse set of locations at a time an extensive set of model evaluations and ablations are performed which empirically justify our various design choices code and preprocessed data have been made publicly available at keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp iterativepfn true iterative point cloud filtering authors dasith de silva edirimuni xuequan lu zhiwen shao gang li antonio robles kelly ying he subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract the quality of point clouds is often limited by noise introduced during their capture process consequently a fundamental vision task is the removal of noise known as point cloud filtering or denoising state of the art learning based methods focus on training neural networks to infer filtered displacements and directly shift noisy points onto the underlying clean surfaces in high noise conditions they iterate the filtering process however this iterative filtering is only done at test time and is less effective at ensuring points converge quickly onto the clean surfaces we propose iterativepfn iterative point cloud filtering network which consists of multiple iterationmodules that model the true iterative filtering process internally within a single network we train our iterativepfn network using a novel loss function that utilizes an adaptive ground truth target at each iteration to capture the relationship between intermediate filtering results during training this ensures that the filtered results converge faster to the clean surfaces our method is able to obtain better performance compared to state of the art methods the source code can be found at fully convolutional networks for dense water flow intensity prediction in swedish catchment areas authors aleksis pirinen olof mogren mårten västerdal subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract intensifying climate change will lead to more extreme weather events including heavy rainfall and drought accurate stream flow prediction models which are adaptable and robust to new circumstances in a changing climate will be an important source of information for decisions on climate adaptation efforts especially regarding mitigation of the risks of and damages associated with flooding in this work we propose a machine learning based approach for predicting water flow intensities in inland watercourses based on the physical characteristics of the catchment areas obtained from geospatial data including elevation and soil maps as well as satellite imagery in addition to temporal information about past rainfall quantities and temperature variations we target the one day ahead regime where a fully convolutional neural network model receives spatio temporal inputs and predicts the water flow intensity in every coordinate of the spatial input for the subsequent day to the best of our knowledge we are the first to tackle the task of dense water flow intensity prediction earlier works have considered predicting flow intensities at a sparse set of locations at a time an extensive set of model evaluations and ablations are performed which empirically justify our various design choices code and preprocessed data have been made publicly available at keyword image signal processing there is no result keyword image signal process there is no result keyword compression there is no result keyword raw a scale invariant trajectory simplification method for efficient data collection in videos authors yang liu luiz gustavo hafemann subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract training data is a critical requirement for machine learning tasks and labeled training data can be expensive to acquire often requiring manual or semi automated data collection pipelines for tracking applications the data collection involves drawing bounding boxes around the classes of interest on each frame and associate detections of the same instance over frames in a semi automated data collection pipeline this can be achieved by running a baseline detection and tracking algorithm and relying on manual correction to add remove change bounding boxes on each frame as well as resolving errors in the associations over frames track switches in this paper we propose a data correction pipeline to generate ground truth data more efficiently in this semi automated scenario our method simplifies the trajectories from the tracking systems and let the annotator verify and correct the objects in the sampled keyframes once the objects in the keyframes are corrected the bounding boxes in the other frames are obtained by interpolation our method achieves substantial reduction in the number of frames requiring manual correction in the mot dataset it reduces the number of frames by while maintaining a hota score of moreover it reduces the number of frames by a factor of while achieving a hota score of in the soccernet dataset and in the dancetrack dataset the project code and data are publicly released at self supervised image denoising for real world images with context aware transformer authors dan zhang fangfang zhou subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract in recent years the development of deep learning has been pushing image denoising to a new level among them self supervised denoising is increasingly popular because it does not require any prior knowledge most of the existing self supervised methods are based on convolutional neural networks cnn which are restricted by the locality of the receptive field and would cause color shifts or textures loss in this paper we propose a novel denoise transformer for real world image denoising which is mainly constructed with context aware denoise transformer cadt units and secondary noise extractor sne block cadt is designed as a dual branch structure where the global branch uses a window based transformer encoder to extract the global information while the local branch focuses on the extraction of local features with small receptive field by incorporating cadt as basic components we build a hierarchical network to directly learn the noise distribution information through residual learning and obtain the first stage denoised output then we design sne in low computation for secondary global noise extraction finally the blind spots are collected from the denoise transformer output and reconstructed forming the final denoised image extensive experiments on the real world sidd benchmark achieve for psnr ssim which is competitive with the current state of the art method and only lower visual comparisons on public srgb raw rgb and greyscale datasets prove that our proposed denoise transformer has a competitive performance especially on blurred textures and low light images without using additional knowledge e g noise level or noise type regarding the underlying unknown noise trace and pace controllable pedestrian animation via guided trajectory diffusion authors davis rempe zhengyi luo xue bin peng ye yuan kris kitani karsten kreis sanja fidler or litany subjects computer vision and pattern recognition cs cv graphics cs gr machine learning cs lg arxiv link pdf link abstract we introduce a method for generating realistic pedestrian trajectories and full body animations that can be controlled to meet user defined goals we draw on recent advances in guided diffusion modeling to achieve test time controllability of trajectories which is normally only associated with rule based systems our guided diffusion model allows users to constrain trajectories through target waypoints speed and specified social groups while accounting for the surrounding environment context this trajectory diffusion model is integrated with a novel physics based humanoid controller to form a closed loop full body pedestrian animation system capable of placing large crowds in a simulated environment with varying terrains we further propose utilizing the value function learned during rl training of the animation controller to guide diffusion to produce trajectories better suited for particular scenarios such as collision avoidance and traversing uneven terrain video results are available on the project page at keyword raw image there is no result
| 1
|
4,053
| 6,988,245,602
|
IssuesEvent
|
2017-12-14 12:12:04
|
nlbdev/pipeline
|
https://api.github.com/repos/nlbdev/pipeline
|
closed
|
move colophon, hide titlepage and print toc
|
enhancement pre-processing Priority:3 - High
|
Maybe delete most of it and keep only dedications and forewords?
|
1.0
|
move colophon, hide titlepage and print toc - Maybe delete most of it and keep only dedications and forewords?
|
process
|
move colophon hide titlepage and print toc maybe delete most of it and keep only dedications and forewords
| 1
|
22,528
| 31,627,016,475
|
IssuesEvent
|
2023-09-06 06:25:37
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
Backwards Compatibility: Combine always_sample with a not
|
question processor/tailsampling
|
### Component(s)
processor/tailsampling
### Is your feature request related to a problem? Please describe.
We want to start implementing tail based sampling in our organization in a backwards compatible way.
Some services will use tail sampling, for the others we want to start by using the always_sample policy.
Is there a straightforward way to implement this policy:
```
if service.name not in ['list', 'of', 'services', 'already', 'using', 'tail', 'sampling']
then always_sample
```
Here are a few things I'm trying
### Option 1
```
{
name: backwards-compatibility-policy-v1,
type: and,
and: {
and_sub_policy:
[
{
name: services-using-tail_sampling-policy,
type: string_attribute,
string_attribute:
{
key: service.name,
values:
[
list,
of,
services,
already,
using,
tail_sampling
],
invert_match: true,
},
},
{ name: sample-all-policy, type: always_sample },
],
},
}
```
However, this might not work, because this will lead to an inverted not sample decision - meaning the actual policies for these services will have no impact. Did I understand this correctly ?
-----------------------------------------
### Option 2
```
{
name: backwards-compatibility-policy-v2,
type: and,
and: {
and_sub_policy:
[
{
name: services-not-using-tail_sampling-policy,
type: string_attribute,
string_attribute:
{
key: service.name,
values:
[
list,
of,
services,
not,
using,
tail_sampling
],
},
},
{ name: sample-all-policy, type: always_sample },
],
},
}
```
This might work - however, this means that the entire list of legacy services would need to be added here, and any new service that is not using tail sampling would also need to be added here. This unfortunately means we need to maintain a redundant list of services here, thus shifting the burden on teams that are not ready to use tail sampling, and onto those that are just trying out Otel in the first place.
---------------------
### Option 3
```
{
name: backwards-compatibility-policy-v3,
type: and,
and: {
and_sub_policy:
[
{
name: services-using-tail_sampling-policy,
type: ottl_condition,
ottl_condition:
{
span: [
"attributes[\"service.name\"] != \"list\" ",
"attributes[\"service.name\"] != \"of\" ",
"attributes[\"service.name\"] != \"services\" ",
"attributes[\"service.name\"] != \"already\" ",
"attributes[\"service.name\"] != \"using\" ",
"attributes[\"service.name\"] != \"tail_sampling\" ",
],
spanevent: [
"attributes[\"service.name\"] != \"list\" ",
"attributes[\"service.name\"] != \"of\" ",
"attributes[\"service.name\"] != \"services\" ",
"attributes[\"service.name\"] != \"already\" ",
"attributes[\"service.name\"] != \"using\" ",
"attributes[\"service.name\"] != \"tail_sampling\" ",
],
}
},
{ name: sample-all-policy, type: always_sample },
],
},
}
```
Does this approach work ? Where can I find reference to the syntax of OTTL conditionals. I'm hoping this will create an OR between the list of services - Is that correct ?
----------------------------
Could someone help out with this problem. Is there a more straightforward way of solving this ?
Thanks!
### Describe the solution you'd like
Either some help would the above 3 approaches would be appreciated.
Or it would also be useful, if we could use a simple `not` statement with `string_attribute` policy - that leads to a sample decision for string values that do not match - rather than leading to an explicit inverted not sampled decision.
Something like
```
not: {
string_attribute: {
key: http.route,
values: [/v1/providers/.+],
},
},
```
### Describe alternatives you've considered
Described above.
### Additional context
_No response_
|
1.0
|
Backwards Compatibility: Combine always_sample with a not - ### Component(s)
processor/tailsampling
### Is your feature request related to a problem? Please describe.
We want to start implementing tail based sampling in our organization in a backwards compatible way.
Some services will use tail sampling, for the others we want to start by using the always_sample policy.
Is there a straightforward way to implement this policy:
```
if service.name not in ['list', 'of', 'services', 'already', 'using', 'tail', 'sampling']
then always_sample
```
Here are a few things I'm trying
### Option 1
```
{
name: backwards-compatibility-policy-v1,
type: and,
and: {
and_sub_policy:
[
{
name: services-using-tail_sampling-policy,
type: string_attribute,
string_attribute:
{
key: service.name,
values:
[
list,
of,
services,
already,
using,
tail_sampling
],
invert_match: true,
},
},
{ name: sample-all-policy, type: always_sample },
],
},
}
```
However, this might not work, because this will lead to an inverted not sample decision - meaning the actual policies for these services will have no impact. Did I understand this correctly ?
-----------------------------------------
### Option 2
```
{
name: backwards-compatibility-policy-v2,
type: and,
and: {
and_sub_policy:
[
{
name: services-not-using-tail_sampling-policy,
type: string_attribute,
string_attribute:
{
key: service.name,
values:
[
list,
of,
services,
not,
using,
tail_sampling
],
},
},
{ name: sample-all-policy, type: always_sample },
],
},
}
```
This might work - however, this means that the entire list of legacy services would need to be added here, and any new service that is not using tail sampling would also need to be added here. This unfortunately means we need to maintain a redundant list of services here, thus shifting the burden on teams that are not ready to use tail sampling, and onto those that are just trying out Otel in the first place.
---------------------
### Option 3
```
{
name: backwards-compatibility-policy-v3,
type: and,
and: {
and_sub_policy:
[
{
name: services-using-tail_sampling-policy,
type: ottl_condition,
ottl_condition:
{
span: [
"attributes[\"service.name\"] != \"list\" ",
"attributes[\"service.name\"] != \"of\" ",
"attributes[\"service.name\"] != \"services\" ",
"attributes[\"service.name\"] != \"already\" ",
"attributes[\"service.name\"] != \"using\" ",
"attributes[\"service.name\"] != \"tail_sampling\" ",
],
spanevent: [
"attributes[\"service.name\"] != \"list\" ",
"attributes[\"service.name\"] != \"of\" ",
"attributes[\"service.name\"] != \"services\" ",
"attributes[\"service.name\"] != \"already\" ",
"attributes[\"service.name\"] != \"using\" ",
"attributes[\"service.name\"] != \"tail_sampling\" ",
],
}
},
{ name: sample-all-policy, type: always_sample },
],
},
}
```
Does this approach work ? Where can I find reference to the syntax of OTTL conditionals. I'm hoping this will create an OR between the list of services - Is that correct ?
----------------------------
Could someone help out with this problem. Is there a more straightforward way of solving this ?
Thanks!
### Describe the solution you'd like
Either some help would the above 3 approaches would be appreciated.
Or it would also be useful, if we could use a simple `not` statement with `string_attribute` policy - that leads to a sample decision for string values that do not match - rather than leading to an explicit inverted not sampled decision.
Something like
```
not: {
string_attribute: {
key: http.route,
values: [/v1/providers/.+],
},
},
```
### Describe alternatives you've considered
Described above.
### Additional context
_No response_
|
process
|
backwards compatibility combine always sample with a not component s processor tailsampling is your feature request related to a problem please describe we want to start implementing tail based sampling in our organization in a backwards compatible way some services will use tail sampling for the others we want to start by using the always sample policy is there a straightforward way to implement this policy if service name not in then always sample here are a few things i m trying option name backwards compatibility policy type and and and sub policy name services using tail sampling policy type string attribute string attribute key service name values list of services already using tail sampling invert match true name sample all policy type always sample however this might not work because this will lead to an inverted not sample decision meaning the actual policies for these services will have no impact did i understand this correctly option name backwards compatibility policy type and and and sub policy name services not using tail sampling policy type string attribute string attribute key service name values list of services not using tail sampling name sample all policy type always sample this might work however this means that the entire list of legacy services would need to be added here and any new service that is not using tail sampling would also need to be added here this unfortunately means we need to maintain a redundant list of services here thus shifting the burden on teams that are not ready to use tail sampling and onto those that are just trying out otel in the first place option name backwards compatibility policy type and and and sub policy name services using tail sampling policy type ottl condition ottl condition span attributes list attributes of attributes services attributes already attributes using attributes tail sampling spanevent attributes list attributes of attributes services attributes already attributes using attributes tail sampling name sample all policy type always sample does this approach work where can i find reference to the syntax of ottl conditionals i m hoping this will create an or between the list of services is that correct could someone help out with this problem is there a more straightforward way of solving this thanks describe the solution you d like either some help would the above approaches would be appreciated or it would also be useful if we could use a simple not statement with string attribute policy that leads to a sample decision for string values that do not match rather than leading to an explicit inverted not sampled decision something like not string attribute key http route values describe alternatives you ve considered described above additional context no response
| 1
|
399,517
| 11,756,416,501
|
IssuesEvent
|
2020-03-13 11:31:22
|
TrinityCore/TrinityCore
|
https://api.github.com/repos/TrinityCore/TrinityCore
|
closed
|
DB/Creature: Kirin Tor Mage in Ulduar
|
Comp-Database Priority-Cosmetic Sub-Spawns
|
Description:
Due to the error of NPC spawn point, the characters are overlapped.
.go xyz -814.109985 -201.125 429.924988 603
NPC ID: 33672
Kirin Tor Mage in Ulduar Raid must be summoned via an event
https://www.youtube.com/watch?v=UuPPW5jJaP4&feature=youtu.be
Branch(es):
3.3.5
TC rev. hash/commit:
961adea3727e32c068245735e76db2445c2fcd2d
Operating system:
CentOS
|
1.0
|
DB/Creature: Kirin Tor Mage in Ulduar - Description:
Due to the error of NPC spawn point, the characters are overlapped.
.go xyz -814.109985 -201.125 429.924988 603
NPC ID: 33672
Kirin Tor Mage in Ulduar Raid must be summoned via an event
https://www.youtube.com/watch?v=UuPPW5jJaP4&feature=youtu.be
Branch(es):
3.3.5
TC rev. hash/commit:
961adea3727e32c068245735e76db2445c2fcd2d
Operating system:
CentOS
|
non_process
|
db creature kirin tor mage in ulduar description due to the error of npc spawn point the characters are overlapped go xyz npc id kirin tor mage in ulduar raid must be summoned via an event branch es tc rev hash commit operating system centos
| 0
|
7,235
| 10,384,414,931
|
IssuesEvent
|
2019-09-10 11:57:45
|
RIOT-OS/RIOT
|
https://api.github.com/repos/RIOT-OS/RIOT
|
closed
|
periph/timer: API inconstency
|
Area: drivers Process: API change State: stale Type: enhancement
|
Reviewing #4089 I realized that `periph/timer` API seems to be inconsistent. `timer_init()` returns 0 on success, while all other functions return 1 on success. I think all periph functions which just return a boolean value should return the same one. Ideally this would be rather a macro like `PERIPH_SUCCESS` than a literal.
|
1.0
|
periph/timer: API inconstency - Reviewing #4089 I realized that `periph/timer` API seems to be inconsistent. `timer_init()` returns 0 on success, while all other functions return 1 on success. I think all periph functions which just return a boolean value should return the same one. Ideally this would be rather a macro like `PERIPH_SUCCESS` than a literal.
|
process
|
periph timer api inconstency reviewing i realized that periph timer api seems to be inconsistent timer init returns on success while all other functions return on success i think all periph functions which just return a boolean value should return the same one ideally this would be rather a macro like periph success than a literal
| 1
|
79,143
| 22,623,271,115
|
IssuesEvent
|
2022-06-30 08:28:22
|
adoptium/temurin-build
|
https://api.github.com/repos/adoptium/temurin-build
|
closed
|
JDK11 / Win 32 Build Broken
|
windows buildbreak
|
The JDK11 build has been failing since 21st June.
https://ci.adoptopenjdk.net/view/Failing%20Temurin%20jobs/job/build-scripts/job/jobs/job/jdk11u/job/jdk11u-windows-x86-32-temurin/
Believe this may be due to the change to VS2019, currently investigating
New build : https://ci.adoptopenjdk.net/view/Failing%20Temurin%20jobs/job/build-scripts/job/jobs/job/jdk11u/job/jdk11u-windows-x86-32-temurin/77/
|
1.0
|
JDK11 / Win 32 Build Broken - The JDK11 build has been failing since 21st June.
https://ci.adoptopenjdk.net/view/Failing%20Temurin%20jobs/job/build-scripts/job/jobs/job/jdk11u/job/jdk11u-windows-x86-32-temurin/
Believe this may be due to the change to VS2019, currently investigating
New build : https://ci.adoptopenjdk.net/view/Failing%20Temurin%20jobs/job/build-scripts/job/jobs/job/jdk11u/job/jdk11u-windows-x86-32-temurin/77/
|
non_process
|
win build broken the build has been failing since june believe this may be due to the change to currently investigating new build
| 0
|
5,174
| 7,959,437,504
|
IssuesEvent
|
2018-07-13 00:47:51
|
GoogleCloudPlatform/google-cloud-cpp
|
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-cpp
|
closed
|
Document the CurlRequest API as a concept.
|
status: will not fix storage type: process
|
This is two parts: one is to write a comment or brief document describing the requirements for `CurlRequest`. The second is to write a series of compile-time assertions that check this concept, so we can enforce the requirements in `storage::DefaultClient<>` and generate good error messages when we make a mistake.
|
1.0
|
Document the CurlRequest API as a concept. - This is two parts: one is to write a comment or brief document describing the requirements for `CurlRequest`. The second is to write a series of compile-time assertions that check this concept, so we can enforce the requirements in `storage::DefaultClient<>` and generate good error messages when we make a mistake.
|
process
|
document the curlrequest api as a concept this is two parts one is to write a comment or brief document describing the requirements for curlrequest the second is to write a series of compile time assertions that check this concept so we can enforce the requirements in storage defaultclient and generate good error messages when we make a mistake
| 1
|
13,852
| 16,614,007,117
|
IssuesEvent
|
2021-06-02 14:40:47
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
In raster calculator, the AND and OR buttons insert the translated words
|
Bug Processing
|
When using Spanish language and raster calculator the AND and OR buttons insert the translated words Y (=AND) and O (=OR) into the expression instead of the correct, untranslated ones, resulting in a syntax error.
1. Set language to Spanish
2. Open raster calculator
3. Hit AND or OR buttons (labelled Y or O)
4. The spanish words Y / O are inserted, instead of AND / OR, the expression is incorrect
**QGIS and OS versions**
Versión de QGIS | 3.16.2-Hannover | Revisión del código de QGIS | f1660f9da5
-- | -- | -- | --
Compilado contra Qt | 5.11.2 | Ejecutándose contra Qt | 5.11.2
Compilado contra GDAL/OGR | 3.1.4 | Ejecutándose contra GDAL/OGR | 3.1.4
Compilado contra GEOS | 3.8.1-CAPI-1.13.3 | Ejecutándose contra GEOS | 3.8.1-CAPI-1.13.3
Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0
Versión del cliente PostgreSQL | 11.5 | Versión de SpatiaLite | 4.3.0
Versión de QWT | 6.1.3 | Versión de QScintilla2 | 2.10.8
Compilado contra PROJ | 6.3.2 | Ejecutándose contra PROJ | Rel. 6.3.2, May 1st, 2020
Versión del SO | Windows 10 (10.0)
Active python plugins | gdb2gpkg; mapswipetool_plugin; openlayers_plugin; db_manager; MetaSearch; processing
|
1.0
|
In raster calculator, the AND and OR buttons insert the translated words - When using Spanish language and raster calculator the AND and OR buttons insert the translated words Y (=AND) and O (=OR) into the expression instead of the correct, untranslated ones, resulting in a syntax error.
1. Set language to Spanish
2. Open raster calculator
3. Hit AND or OR buttons (labelled Y or O)
4. The spanish words Y / O are inserted, instead of AND / OR, the expression is incorrect
**QGIS and OS versions**
Versión de QGIS | 3.16.2-Hannover | Revisión del código de QGIS | f1660f9da5
-- | -- | -- | --
Compilado contra Qt | 5.11.2 | Ejecutándose contra Qt | 5.11.2
Compilado contra GDAL/OGR | 3.1.4 | Ejecutándose contra GDAL/OGR | 3.1.4
Compilado contra GEOS | 3.8.1-CAPI-1.13.3 | Ejecutándose contra GEOS | 3.8.1-CAPI-1.13.3
Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0
Versión del cliente PostgreSQL | 11.5 | Versión de SpatiaLite | 4.3.0
Versión de QWT | 6.1.3 | Versión de QScintilla2 | 2.10.8
Compilado contra PROJ | 6.3.2 | Ejecutándose contra PROJ | Rel. 6.3.2, May 1st, 2020
Versión del SO | Windows 10 (10.0)
Active python plugins | gdb2gpkg; mapswipetool_plugin; openlayers_plugin; db_manager; MetaSearch; processing
|
process
|
in raster calculator the and and or buttons insert the translated words when using spanish language and raster calculator the and and or buttons insert the translated words y and and o or into the expression instead of the correct untranslated ones resulting in a syntax error set language to spanish open raster calculator hit and or or buttons labelled y or o the spanish words y o are inserted instead of and or the expression is incorrect qgis and os versions versión de qgis hannover revisión del código de qgis compilado contra qt ejecutándose contra qt compilado contra gdal ogr ejecutándose contra gdal ogr compilado contra geos capi ejecutándose contra geos capi compiled against sqlite running against sqlite versión del cliente postgresql versión de spatialite versión de qwt versión de compilado contra proj ejecutándose contra proj rel may versión del so windows active python plugins mapswipetool plugin openlayers plugin db manager metasearch processing
| 1
|
25,504
| 6,668,088,604
|
IssuesEvent
|
2017-10-03 14:44:26
|
rust-lang/rust
|
https://api.github.com/repos/rust-lang/rust
|
closed
|
enable mir inlining across trait methods
|
A-codegen A-mir C-enhancement T-compiler
|
As the fix to https://github.com/rust-lang/rust/issues/40473, @qmx disabled MIR inlining of trait calls altogether. This issue tracks the task of re-enabling the inlining, where possible! This is a slightly more involved task.
Here is the example text:
```rust
pub trait Foo {
fn bar(&self) {}
}
impl Foo for () {
fn bar(&self) { println!("Hello, World!"); }
}
pub fn main() {
().bar();
}
```
The goal would be to have the call to `bar()` inlined. But we should draw up some more interesting tasks showing edge-cases: for example, generic functions and so forth (we should be able to use `RevealMode::All` to "see through" specialization, since inlining executes quite late).
I think roughly speaking the steps to solve this issue "properly" are to:
- Refactor `Instance` and `InstanceDef` out of `librustc_trans` and (probably) into `librustc`
- Rework inlining to use those mechanisms to resolve method calls (including trait ones)
|
1.0
|
enable mir inlining across trait methods - As the fix to https://github.com/rust-lang/rust/issues/40473, @qmx disabled MIR inlining of trait calls altogether. This issue tracks the task of re-enabling the inlining, where possible! This is a slightly more involved task.
Here is the example text:
```rust
pub trait Foo {
fn bar(&self) {}
}
impl Foo for () {
fn bar(&self) { println!("Hello, World!"); }
}
pub fn main() {
().bar();
}
```
The goal would be to have the call to `bar()` inlined. But we should draw up some more interesting tasks showing edge-cases: for example, generic functions and so forth (we should be able to use `RevealMode::All` to "see through" specialization, since inlining executes quite late).
I think roughly speaking the steps to solve this issue "properly" are to:
- Refactor `Instance` and `InstanceDef` out of `librustc_trans` and (probably) into `librustc`
- Rework inlining to use those mechanisms to resolve method calls (including trait ones)
|
non_process
|
enable mir inlining across trait methods as the fix to qmx disabled mir inlining of trait calls altogether this issue tracks the task of re enabling the inlining where possible this is a slightly more involved task here is the example text rust pub trait foo fn bar self impl foo for fn bar self println hello world pub fn main bar the goal would be to have the call to bar inlined but we should draw up some more interesting tasks showing edge cases for example generic functions and so forth we should be able to use revealmode all to see through specialization since inlining executes quite late i think roughly speaking the steps to solve this issue properly are to refactor instance and instancedef out of librustc trans and probably into librustc rework inlining to use those mechanisms to resolve method calls including trait ones
| 0
|
71,388
| 3,356,379,343
|
IssuesEvent
|
2015-11-18 20:14:48
|
TechReborn/TechReborn
|
https://api.github.com/repos/TechReborn/TechReborn
|
closed
|
Missing texture with standard machine casing.
|
bug Medium priority
|
Techreborn: 0.5.6.1004
reborncore:1.0.0.9
forge:10.13.4.1558
ic2/3:2.2.2.791
# Enable Connected textures
B:"Enable Connected textures"=false

|
1.0
|
Missing texture with standard machine casing. - Techreborn: 0.5.6.1004
reborncore:1.0.0.9
forge:10.13.4.1558
ic2/3:2.2.2.791
# Enable Connected textures
B:"Enable Connected textures"=false

|
non_process
|
missing texture with standard machine casing techreborn reborncore forge enable connected textures b enable connected textures false
| 0
|
16,278
| 20,884,553,965
|
IssuesEvent
|
2022-03-23 02:34:50
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
Add Da Derp Dee Derp Da Teetley Derpee Derpee Dumb
|
suggested title in process
|
Please add as much of the following info as you can:
Title: Da Derp Dee Derp Da Teetley Derpee Derpee Dumb
Type (film/tv show): Film
Film or show in which it appears: South Park (https://www.imdb.com/title/tt0705968/ Season 06 Episode 15)
Is the parent film/show streaming anywhere? Amazon Prime in the UK.
About when in the parent film/show does it appear? 15 minute
Actual footage of the film/show can be seen (yes/no)? https://www.youtube.com/watch?v=ZxYA6duF-9E
|
1.0
|
Add Da Derp Dee Derp Da Teetley Derpee Derpee Dumb - Please add as much of the following info as you can:
Title: Da Derp Dee Derp Da Teetley Derpee Derpee Dumb
Type (film/tv show): Film
Film or show in which it appears: South Park (https://www.imdb.com/title/tt0705968/ Season 06 Episode 15)
Is the parent film/show streaming anywhere? Amazon Prime in the UK.
About when in the parent film/show does it appear? 15 minute
Actual footage of the film/show can be seen (yes/no)? https://www.youtube.com/watch?v=ZxYA6duF-9E
|
process
|
add da derp dee derp da teetley derpee derpee dumb please add as much of the following info as you can title da derp dee derp da teetley derpee derpee dumb type film tv show film film or show in which it appears south park season episode is the parent film show streaming anywhere amazon prime in the uk about when in the parent film show does it appear minute actual footage of the film show can be seen yes no
| 1
|
303,338
| 22,971,499,485
|
IssuesEvent
|
2022-07-20 03:53:59
|
a-happy-dev-day/yanawa-server
|
https://api.github.com/repos/a-happy-dev-day/yanawa-server
|
closed
|
[Feature] Search Service
|
documentation feature
|
### 용어 사전
용어 | 영어 | 설명
-- | -- | --
지역 | location | 참여 가능한 위치
시작 일자 | startDate | 참여 가능한 매칭 시작 일자
종료 일자 | endDate | 참여 가능한 매칭 마지막 일자
시작 시간 | startTime | 참여 가능한 매칭 시작 시간
종료 시간 | endTime | 참여 가능한 매칭 마지막 시간
구력 | annual | 원하는 매칭의 구력
### 행위
- 회원은 지역 정보와 예정 날짜, 예정 시간, 구력을 이용해 매칭을 검색할 수 있다.
- 지역 정보는 코트장의 지역 정보이다.
- 예정 날짜는 찾으려는 날짜의 기간이 입력된다.
- 예정 시간은 시작 시간과 종료 시간이 입력된다.
- 구력 정보는 각 연차별 이상 및 미만의 정보가 입력된다.
- **값이 들어가지 않으면 해당 조건은 검색하지 않는다.**
### 다이어그램
```mermaid
sequenceDiagram
Server -->> Search : request search
Search -->> Matching : search matching
Matching ->> Matching : search
Matching ->> Search : return matching json value
Search ->> Server : response matching
```
### 이미지
<img width="904" alt="스크린샷 2022-07-06 오후 12 40 51" src="https://user-images.githubusercontent.com/92219795/177463244-abcc7625-b93c-4999-affc-f2280998ebb7.png">
|
1.0
|
[Feature] Search Service - ### 용어 사전
용어 | 영어 | 설명
-- | -- | --
지역 | location | 참여 가능한 위치
시작 일자 | startDate | 참여 가능한 매칭 시작 일자
종료 일자 | endDate | 참여 가능한 매칭 마지막 일자
시작 시간 | startTime | 참여 가능한 매칭 시작 시간
종료 시간 | endTime | 참여 가능한 매칭 마지막 시간
구력 | annual | 원하는 매칭의 구력
### 행위
- 회원은 지역 정보와 예정 날짜, 예정 시간, 구력을 이용해 매칭을 검색할 수 있다.
- 지역 정보는 코트장의 지역 정보이다.
- 예정 날짜는 찾으려는 날짜의 기간이 입력된다.
- 예정 시간은 시작 시간과 종료 시간이 입력된다.
- 구력 정보는 각 연차별 이상 및 미만의 정보가 입력된다.
- **값이 들어가지 않으면 해당 조건은 검색하지 않는다.**
### 다이어그램
```mermaid
sequenceDiagram
Server -->> Search : request search
Search -->> Matching : search matching
Matching ->> Matching : search
Matching ->> Search : return matching json value
Search ->> Server : response matching
```
### 이미지
<img width="904" alt="스크린샷 2022-07-06 오후 12 40 51" src="https://user-images.githubusercontent.com/92219795/177463244-abcc7625-b93c-4999-affc-f2280998ebb7.png">
|
non_process
|
search service 용어 사전 용어 영어 설명 지역 location 참여 가능한 위치 시작 일자 startdate 참여 가능한 매칭 시작 일자 종료 일자 enddate 참여 가능한 매칭 마지막 일자 시작 시간 starttime 참여 가능한 매칭 시작 시간 종료 시간 endtime 참여 가능한 매칭 마지막 시간 구력 annual 원하는 매칭의 구력 행위 회원은 지역 정보와 예정 날짜 예정 시간 구력을 이용해 매칭을 검색할 수 있다 지역 정보는 코트장의 지역 정보이다 예정 날짜는 찾으려는 날짜의 기간이 입력된다 예정 시간은 시작 시간과 종료 시간이 입력된다 구력 정보는 각 연차별 이상 및 미만의 정보가 입력된다 값이 들어가지 않으면 해당 조건은 검색하지 않는다 다이어그램 mermaid sequencediagram server search request search search matching search matching matching matching search matching search return matching json value search server response matching 이미지 img width alt 스크린샷 오후 src
| 0
|
13,921
| 16,677,496,367
|
IssuesEvent
|
2021-06-07 18:09:57
|
ORNL-AMO/AMO-Tools-Suite
|
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Suite
|
closed
|
Heat Cascade Error
|
Needs Verification Process Heating bug
|
Issue overview
--------------
Fix error from the Algo doc...
currently: line 21
` const double energySavings = eqEnergySupply * effOpHours;`
Should be: line 21
` const double energySavings = std::min(eqEnergySupply, secFiringRate) * effOpHours;`
Looks like secFiringRate may need bindings
Also pls make a binding for "hourlySavings" which = std::min(eqEnergySupply, secFiringRate)
If you want to make that tidier (add a line for hourlySavings and use that in energySavings), that is fine
|
1.0
|
Heat Cascade Error - Issue overview
--------------
Fix error from the Algo doc...
currently: line 21
` const double energySavings = eqEnergySupply * effOpHours;`
Should be: line 21
` const double energySavings = std::min(eqEnergySupply, secFiringRate) * effOpHours;`
Looks like secFiringRate may need bindings
Also pls make a binding for "hourlySavings" which = std::min(eqEnergySupply, secFiringRate)
If you want to make that tidier (add a line for hourlySavings and use that in energySavings), that is fine
|
process
|
heat cascade error issue overview fix error from the algo doc currently line const double energysavings eqenergysupply effophours should be line const double energysavings std min eqenergysupply secfiringrate effophours looks like secfiringrate may need bindings also pls make a binding for hourlysavings which std min eqenergysupply secfiringrate if you want to make that tidier add a line for hourlysavings and use that in energysavings that is fine
| 1
|
203,984
| 15,396,544,827
|
IssuesEvent
|
2021-03-03 20:47:49
|
projectcontour/contour
|
https://api.github.com/repos/projectcontour/contour
|
closed
|
SNI integration test is flaky
|
area/testing
|
See https://travis-ci.com/github/projectcontour/contour/jobs/365814535
Some relevant output:
```
# reading document from ./_integration/testsuite/httpproxy/002-header-condition-match.yaml
...
ok 46 - running object update check
# Fatal - internal/check/grnqoUyHmVtJ:5: eval_conflict_error: complete rules must not produce multiple outputs
not ok 47 - running Rego check lines 160-172
---
- Message: 'internal/check/grnqoUyHmVtJ:5: eval_conflict_error: complete rules must
not produce multiple outputs'
Severity: Fatal
Timestamp: "2020-07-28T18:28:52.766611607Z"
...
# skipping
ok 48 - deleting test objects
1..48
...
# reading document from ./_integration/testsuite/httpproxy/003-path-condition-match.yaml
...
ok 31 - running Rego check lines 98-106
# Error - raised predicate "error_proxy_is_not_valid"
# Error - HTTP proxy for 'conditions.projectcontour.io' is not valid
# Error - status: invalid
# Error - desc: fqdn "conditions.projectcontour.io" is used in multiple HTTPProxies: default/header-conditions, default/path-conditions
not ok 32 - running Rego check lines 108-121
---
- Message: |-
raised predicate "error_proxy_is_not_valid"
HTTP proxy for 'conditions.projectcontour.io' is not valid
status: invalid
desc: fqdn "conditions.projectcontour.io" is used in multiple HTTPProxies: default/header-conditions, default/path-conditions
Severity: Error
Timestamp: "0001-01-01T00:00:00Z"
...
# Error - internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45072->127.0.0.1:9080: read: connection reset by peer
# Error - internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45076->127.0.0.1:9080: read: connection reset by peer
# Error - internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45080->127.0.0.1:9080: read: connection reset by peer
not ok 33 - running Rego check lines 123-182
---
- Message: 'internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/":
read tcp 127.0.0.1:45072->127.0.0.1:9080: read: connection reset by peer'
Severity: Error
Timestamp: "0001-01-01T00:00:00Z"
- Message: 'internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/":
read tcp 127.0.0.1:45076->127.0.0.1:9080: read: connection reset by peer'
Severity: Error
Timestamp: "0001-01-01T00:00:00Z"
- Message: 'internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/":
read tcp 127.0.0.1:45080->127.0.0.1:9080: read: connection reset by peer'
Severity: Error
Timestamp: "0001-01-01T00:00:00Z"
...
ok 34 - deleting test objects
1..34
```
|
1.0
|
SNI integration test is flaky - See https://travis-ci.com/github/projectcontour/contour/jobs/365814535
Some relevant output:
```
# reading document from ./_integration/testsuite/httpproxy/002-header-condition-match.yaml
...
ok 46 - running object update check
# Fatal - internal/check/grnqoUyHmVtJ:5: eval_conflict_error: complete rules must not produce multiple outputs
not ok 47 - running Rego check lines 160-172
---
- Message: 'internal/check/grnqoUyHmVtJ:5: eval_conflict_error: complete rules must
not produce multiple outputs'
Severity: Fatal
Timestamp: "2020-07-28T18:28:52.766611607Z"
...
# skipping
ok 48 - deleting test objects
1..48
...
# reading document from ./_integration/testsuite/httpproxy/003-path-condition-match.yaml
...
ok 31 - running Rego check lines 98-106
# Error - raised predicate "error_proxy_is_not_valid"
# Error - HTTP proxy for 'conditions.projectcontour.io' is not valid
# Error - status: invalid
# Error - desc: fqdn "conditions.projectcontour.io" is used in multiple HTTPProxies: default/header-conditions, default/path-conditions
not ok 32 - running Rego check lines 108-121
---
- Message: |-
raised predicate "error_proxy_is_not_valid"
HTTP proxy for 'conditions.projectcontour.io' is not valid
status: invalid
desc: fqdn "conditions.projectcontour.io" is used in multiple HTTPProxies: default/header-conditions, default/path-conditions
Severity: Error
Timestamp: "0001-01-01T00:00:00Z"
...
# Error - internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45072->127.0.0.1:9080: read: connection reset by peer
# Error - internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45076->127.0.0.1:9080: read: connection reset by peer
# Error - internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45080->127.0.0.1:9080: read: connection reset by peer
not ok 33 - running Rego check lines 123-182
---
- Message: 'internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/":
read tcp 127.0.0.1:45072->127.0.0.1:9080: read: connection reset by peer'
Severity: Error
Timestamp: "0001-01-01T00:00:00Z"
- Message: 'internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/":
read tcp 127.0.0.1:45076->127.0.0.1:9080: read: connection reset by peer'
Severity: Error
Timestamp: "0001-01-01T00:00:00Z"
- Message: 'internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/":
read tcp 127.0.0.1:45080->127.0.0.1:9080: read: connection reset by peer'
Severity: Error
Timestamp: "0001-01-01T00:00:00Z"
...
ok 34 - deleting test objects
1..34
```
|
non_process
|
sni integration test is flaky see some relevant output reading document from integration testsuite httpproxy header condition match yaml ok running object update check fatal internal check grnqouyhmvtj eval conflict error complete rules must not produce multiple outputs not ok running rego check lines message internal check grnqouyhmvtj eval conflict error complete rules must not produce multiple outputs severity fatal timestamp skipping ok deleting test objects reading document from integration testsuite httpproxy path condition match yaml ok running rego check lines error raised predicate error proxy is not valid error http proxy for conditions projectcontour io is not valid error status invalid error desc fqdn conditions projectcontour io is used in multiple httpproxies default header conditions default path conditions not ok running rego check lines message raised predicate error proxy is not valid http proxy for conditions projectcontour io is not valid status invalid desc fqdn conditions projectcontour io is used in multiple httpproxies default header conditions default path conditions severity error timestamp error internal check raybffqqegji eval builtin error http send get read tcp read connection reset by peer error internal check raybffqqegji eval builtin error http send get read tcp read connection reset by peer error internal check raybffqqegji eval builtin error http send get read tcp read connection reset by peer not ok running rego check lines message internal check raybffqqegji eval builtin error http send get read tcp read connection reset by peer severity error timestamp message internal check raybffqqegji eval builtin error http send get read tcp read connection reset by peer severity error timestamp message internal check raybffqqegji eval builtin error http send get read tcp read connection reset by peer severity error timestamp ok deleting test objects
| 0
|
22,142
| 11,694,116,267
|
IssuesEvent
|
2020-03-06 02:52:59
|
letpoc/faq
|
https://api.github.com/repos/letpoc/faq
|
closed
|
Restful Service - User login
|
P01-S01 PI-01 Restful Service
|
As a developer, develop user login functionalities with security feature with secret token, authorization code
Acceptance Criteria:
Only Valid user can login in to the application
|
1.0
|
Restful Service - User login - As a developer, develop user login functionalities with security feature with secret token, authorization code
Acceptance Criteria:
Only Valid user can login in to the application
|
non_process
|
restful service user login as a developer develop user login functionalities with security feature with secret token authorization code acceptance criteria only valid user can login in to the application
| 0
|
17,197
| 22,773,833,309
|
IssuesEvent
|
2022-07-08 12:41:41
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Missing "Networking" tab
|
automation/svc triaged assigned-to-author doc-bug process-automation/subsvc Pri2
|
Hello, the document doesn't include details on the "Networking" tab:
<img width="732" alt="Screen Shot 2022-05-24 at 10 49 39" src="https://user-images.githubusercontent.com/1591516/170065395-e791d6c3-33c5-4e9a-94b5-658eea018e96.png">
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 87f330cc-8989-9721-7d02-965dce0e988d
* Version Independent ID: 4242498b-4503-2f16-9922-3c401e155b11
* Content: [Quickstart - Create an Azure Automation account using the portal](https://docs.microsoft.com/en-us/azure/automation/quickstarts/create-account-portal)
* Content Source: [articles/automation/quickstarts/create-account-portal.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/quickstarts/create-account-portal.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SGSneha
* Microsoft Alias: **sudhirsneha**
|
1.0
|
Missing "Networking" tab - Hello, the document doesn't include details on the "Networking" tab:
<img width="732" alt="Screen Shot 2022-05-24 at 10 49 39" src="https://user-images.githubusercontent.com/1591516/170065395-e791d6c3-33c5-4e9a-94b5-658eea018e96.png">
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 87f330cc-8989-9721-7d02-965dce0e988d
* Version Independent ID: 4242498b-4503-2f16-9922-3c401e155b11
* Content: [Quickstart - Create an Azure Automation account using the portal](https://docs.microsoft.com/en-us/azure/automation/quickstarts/create-account-portal)
* Content Source: [articles/automation/quickstarts/create-account-portal.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/quickstarts/create-account-portal.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SGSneha
* Microsoft Alias: **sudhirsneha**
|
process
|
missing networking tab hello the document doesn t include details on the networking tab img width alt screen shot at src document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login sgsneha microsoft alias sudhirsneha
| 1
|
604
| 2,577,791,304
|
IssuesEvent
|
2015-02-12 19:09:33
|
chrsmith/quake2-gwt-port
|
https://api.github.com/repos/chrsmith/quake2-gwt-port
|
opened
|
can't connect to multiplayer from a computer other than the server
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. start the server
2. hit the url from another machine
3. try to join the multiplayer game
What is the expected output? What do you see instead?
Being in the game.
Connecting <hostname> repeatedly.
What version of the product are you using? On what operating system?
Current Webkit nightly and Chromium nightly.
```
-----
Original issue reported on code.google.com by b...@getdown.org on 2 Apr 2010 at 2:09
|
1.0
|
can't connect to multiplayer from a computer other than the server - ```
What steps will reproduce the problem?
1. start the server
2. hit the url from another machine
3. try to join the multiplayer game
What is the expected output? What do you see instead?
Being in the game.
Connecting <hostname> repeatedly.
What version of the product are you using? On what operating system?
Current Webkit nightly and Chromium nightly.
```
-----
Original issue reported on code.google.com by b...@getdown.org on 2 Apr 2010 at 2:09
|
non_process
|
can t connect to multiplayer from a computer other than the server what steps will reproduce the problem start the server hit the url from another machine try to join the multiplayer game what is the expected output what do you see instead being in the game connecting repeatedly what version of the product are you using on what operating system current webkit nightly and chromium nightly original issue reported on code google com by b getdown org on apr at
| 0
|
11,519
| 14,400,980,262
|
IssuesEvent
|
2020-12-03 13:08:07
|
gain-wyj/gain-wyj.github.io
|
https://api.github.com/repos/gain-wyj/gain-wyj.github.io
|
opened
|
数学分析模型(一):数据的无量纲处理方法及示例(附完整代码) | 梦魇回生的博客
|
2020/12/01/dimensionless-processing-of-data/ Gitalk
|
https://gain-wyj.cn/2020/12/01/dimensionless-processing-of-data/
在对实际问题建模过程中,特别是在建立指标评价体系时,常常会面临不同类型的数据处理及融合。而各个指标之间由于计量单位和数量级的不尽相同,从而使得各指标间不具有可比性。在数据分析之前,通常需要先将数据标准化,利用标准化后的数据进行分析。数据标准
|
1.0
|
数学分析模型(一):数据的无量纲处理方法及示例(附完整代码) | 梦魇回生的博客 - https://gain-wyj.cn/2020/12/01/dimensionless-processing-of-data/
在对实际问题建模过程中,特别是在建立指标评价体系时,常常会面临不同类型的数据处理及融合。而各个指标之间由于计量单位和数量级的不尽相同,从而使得各指标间不具有可比性。在数据分析之前,通常需要先将数据标准化,利用标准化后的数据进行分析。数据标准
|
process
|
数学分析模型(一):数据的无量纲处理方法及示例(附完整代码) 梦魇回生的博客 在对实际问题建模过程中,特别是在建立指标评价体系时,常常会面临不同类型的数据处理及融合。而各个指标之间由于计量单位和数量级的不尽相同,从而使得各指标间不具有可比性。在数据分析之前,通常需要先将数据标准化,利用标准化后的数据进行分析。数据标准
| 1
|
10,038
| 13,044,161,595
|
IssuesEvent
|
2020-07-29 03:47:24
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `AddDateDatetimeReal` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `AddDateDatetimeReal` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @lonng
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `AddDateDatetimeReal` from TiDB -
## Description
Port the scalar function `AddDateDatetimeReal` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @lonng
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function adddatedatetimereal from tidb description port the scalar function adddatedatetimereal from tidb to coprocessor score mentor s lonng recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
121,523
| 25,982,875,242
|
IssuesEvent
|
2022-12-19 20:35:02
|
Clueless-Community/seamless-ui
|
https://api.github.com/repos/Clueless-Community/seamless-ui
|
closed
|
account options dropdown
|
codepeak 22
|
One need to make this component using `HTML` and `Tailwind CSS`. I would suggest to use [Tailwind Playgrounds](https://play.tailwindcss.com/) to make things faster and quicker.
Here is a reference to the component.
<img width="235" alt="Screenshot 2022-12-20 at 12 00 15 AM" src="https://user-images.githubusercontent.com/107763774/208495181-1181c129-dc48-43df-9db4-3040cdf17c5d.png">
After building the component please raise a PR with a screenshot of the component and add the component in `path of the file`.
If you need to use any icon please use it from [Hero Icons](https://heroicons.com/)
Good luck.
|
1.0
|
account options dropdown - One need to make this component using `HTML` and `Tailwind CSS`. I would suggest to use [Tailwind Playgrounds](https://play.tailwindcss.com/) to make things faster and quicker.
Here is a reference to the component.
<img width="235" alt="Screenshot 2022-12-20 at 12 00 15 AM" src="https://user-images.githubusercontent.com/107763774/208495181-1181c129-dc48-43df-9db4-3040cdf17c5d.png">
After building the component please raise a PR with a screenshot of the component and add the component in `path of the file`.
If you need to use any icon please use it from [Hero Icons](https://heroicons.com/)
Good luck.
|
non_process
|
account options dropdown one need to make this component using html and tailwind css i would suggest to use to make things faster and quicker here is a reference to the component img width alt screenshot at am src after building the component please raise a pr with a screenshot of the component and add the component in path of the file if you need to use any icon please use it from good luck
| 0
|
21,146
| 28,125,460,482
|
IssuesEvent
|
2023-03-31 17:18:57
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
NTR: maternal contribution terms
|
New term request RNA processes development organism-level process Needed for GO-CAM
|
general reference for all of these https://www.ncbi.nlm.nih.gov/books/NBK53192/
maternal contribution of gene product to offspring
part_of oocyte development
is_a developmental process involved in reproduction
def: The process whose specific outcome is the contribution of maternal gene products to offspring for use until the offspring's genome is activated.
maternally deposition of RNA in oocyte
is_a maternal contribution of gene product to offspring
def: The process whose specific outcome is the deposition of maternal RNA in the developing oocyte.
maternally deposition of protein in oocyte
is_amaternal contribution of gene product to offspring
def: The process whose specific outcome is the deposition of maternal protein in the developing oocyte.
regulation of maternally deposited RNA stability
is_a regulation of mRNA stability
def: Any process that modulates the propensity of maternally deposited RNA molecules to degradation. Includes processes that both stabilize and destabilize RNAs
regulation of maternally deposited protein stability
is_a regulation of protein stability
def: Any process that affects the structure and integrity of a maternally deposited protein, altering the likelihood of its degradation or aggregation. Includes processes that both stabilize and destabilize proteins.
|
2.0
|
NTR: maternal contribution terms - general reference for all of these https://www.ncbi.nlm.nih.gov/books/NBK53192/
maternal contribution of gene product to offspring
part_of oocyte development
is_a developmental process involved in reproduction
def: The process whose specific outcome is the contribution of maternal gene products to offspring for use until the offspring's genome is activated.
maternally deposition of RNA in oocyte
is_a maternal contribution of gene product to offspring
def: The process whose specific outcome is the deposition of maternal RNA in the developing oocyte.
maternally deposition of protein in oocyte
is_amaternal contribution of gene product to offspring
def: The process whose specific outcome is the deposition of maternal protein in the developing oocyte.
regulation of maternally deposited RNA stability
is_a regulation of mRNA stability
def: Any process that modulates the propensity of maternally deposited RNA molecules to degradation. Includes processes that both stabilize and destabilize RNAs
regulation of maternally deposited protein stability
is_a regulation of protein stability
def: Any process that affects the structure and integrity of a maternally deposited protein, altering the likelihood of its degradation or aggregation. Includes processes that both stabilize and destabilize proteins.
|
process
|
ntr maternal contribution terms general reference for all of these maternal contribution of gene product to offspring part of oocyte development is a developmental process involved in reproduction def the process whose specific outcome is the contribution of maternal gene products to offspring for use until the offspring s genome is activated maternally deposition of rna in oocyte is a maternal contribution of gene product to offspring def the process whose specific outcome is the deposition of maternal rna in the developing oocyte maternally deposition of protein in oocyte is amaternal contribution of gene product to offspring def the process whose specific outcome is the deposition of maternal protein in the developing oocyte regulation of maternally deposited rna stability is a regulation of mrna stability def any process that modulates the propensity of maternally deposited rna molecules to degradation includes processes that both stabilize and destabilize rnas regulation of maternally deposited protein stability is a regulation of protein stability def any process that affects the structure and integrity of a maternally deposited protein altering the likelihood of its degradation or aggregation includes processes that both stabilize and destabilize proteins
| 1
|
398,760
| 11,742,308,471
|
IssuesEvent
|
2020-03-12 00:18:56
|
thaliawww/concrexit
|
https://api.github.com/repos/thaliawww/concrexit
|
closed
|
'become a member' forms not present on website
|
bug priority: medium
|
In GitLab by @thomwiggers on Dec 21, 2017, 18:07
### One-sentence description
They're not on https://thalia.nu/admin/members/becomeamemberdocument/ / https://thalia.nu/become-a-member/
|
1.0
|
'become a member' forms not present on website - In GitLab by @thomwiggers on Dec 21, 2017, 18:07
### One-sentence description
They're not on https://thalia.nu/admin/members/becomeamemberdocument/ / https://thalia.nu/become-a-member/
|
non_process
|
become a member forms not present on website in gitlab by thomwiggers on dec one sentence description they re not on
| 0
|
115,737
| 11,886,450,348
|
IssuesEvent
|
2020-03-27 21:58:14
|
generaliinformatik/mycoradar
|
https://api.github.com/repos/generaliinformatik/mycoradar
|
closed
|
Translate README to english
|
documentation
|
Translate the README to english to address english speaking contributors.
This is one of the tasks from the team vote of 25th March 2020. With the development of the website mycoradar.eu, which is also to be written in English, English-speaking interested parties may be increasingly addressed
|
1.0
|
Translate README to english - Translate the README to english to address english speaking contributors.
This is one of the tasks from the team vote of 25th March 2020. With the development of the website mycoradar.eu, which is also to be written in English, English-speaking interested parties may be increasingly addressed
|
non_process
|
translate readme to english translate the readme to english to address english speaking contributors this is one of the tasks from the team vote of march with the development of the website mycoradar eu which is also to be written in english english speaking interested parties may be increasingly addressed
| 0
|
475,500
| 13,711,607,676
|
IssuesEvent
|
2020-10-02 05:05:00
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
cse.google.com - see bug description
|
browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-critical
|
<!-- @browser: Firefox 82.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:82.0) Gecko/20100101 Firefox/82.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/59149 -->
**URL**: https://cse.google.com/cse?q=gjkfshsg&sa=Search&ie=UTF-8&cx=partner-pub%2D6638247779433690:3873384991#%9C&gsc.tab=0&gsc.q=youtube
**Browser / Version**: Firefox 82.0
**Operating System**: Windows 7
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: programable scarch engine started how to fix
**Steps to Reproduce**:
progamable screch engine
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/10/2267ed51-92aa-4513-bf6c-d1169c122ee0.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200926073307</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/10/2d30d28d-2dde-43b9-941c-0e58574e0704)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
cse.google.com - see bug description - <!-- @browser: Firefox 82.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:82.0) Gecko/20100101 Firefox/82.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/59149 -->
**URL**: https://cse.google.com/cse?q=gjkfshsg&sa=Search&ie=UTF-8&cx=partner-pub%2D6638247779433690:3873384991#%9C&gsc.tab=0&gsc.q=youtube
**Browser / Version**: Firefox 82.0
**Operating System**: Windows 7
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: programable scarch engine started how to fix
**Steps to Reproduce**:
progamable screch engine
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/10/2267ed51-92aa-4513-bf6c-d1169c122ee0.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200926073307</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/10/2d30d28d-2dde-43b9-941c-0e58574e0704)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
cse google com see bug description url browser version firefox operating system windows tested another browser yes chrome problem type something else description programable scarch engine started how to fix steps to reproduce progamable screch engine view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
18,681
| 24,594,565,706
|
IssuesEvent
|
2022-10-14 07:08:40
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[FHIR] Questionnaire resource > JSON > Branching logic data is not getting mapped into the JSON file for text choice + Other option
|
Bug P1 Response datastore Process: Fixed Process: Tested QA Process: Tested dev
|
**AR:** Questionnaire resource > JSON > Branching logic data is not getting mapped into the JSON file for text choice + Other option
**ER:** Branching logic data should get mapped into the JSON file as configured in the SB
|
3.0
|
[FHIR] Questionnaire resource > JSON > Branching logic data is not getting mapped into the JSON file for text choice + Other option - **AR:** Questionnaire resource > JSON > Branching logic data is not getting mapped into the JSON file for text choice + Other option
**ER:** Branching logic data should get mapped into the JSON file as configured in the SB
|
process
|
questionnaire resource json branching logic data is not getting mapped into the json file for text choice other option ar questionnaire resource json branching logic data is not getting mapped into the json file for text choice other option er branching logic data should get mapped into the json file as configured in the sb
| 1
|
17,412
| 23,228,628,463
|
IssuesEvent
|
2022-08-03 04:44:38
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
Add more advanced actions to processing dialog, including copy as qgis_process command, copy/paste as json (Request in QGIS)
|
Processing 3.24
|
### Request for documentation
From pull request QGIS/qgis#46514
Author: @nyalldawson
QGIS version: 3.24
**Add more advanced actions to processing dialog, including copy as qgis_process command, copy/paste as json**
### PR Description:
Together these actions:
- Allow for easy generation of qgis_process commands via the QGIS gui, including any tricky parameter values like gpkg outputs with specific layer (refs https://github.com/qgis/QGIS/issues/38272)
- Provide a way for users to copy the settings defined in the dialog in a text format, so they can store these easily and then
restore later by pasting the value
- Provides an easy way for users to copy the settings into the json format consumed by qgis_process (following #46497), so that it is easy for users to see the expected format even for complex parameters (like tin interpolation parameters)
Refs paleolimbot/qgisprocess#56
Refs paleolimbot/qgisprocess#44
Sponsored by the Research Institute for Nature and Forest, Flemish Govt
### Commits tagged with [need-docs] or [FEATURE]
"[feature] Add action in advanced menu for processing algorithm to copy\nequivalent qgis_process command\n\nAllows for easy generation of qgis_process commands via the QGIS gui" "[feature] Add action to processing algorithm advanced menu to copy the\ncurrent settings as JSON and paste json settings from clipboard\n\nThe utility of this is two fold:\n\n1. It provides a way for users to copy the settings defined in the\ndialog in a text format, so they can store these easily and then\nrestore later by pasting the value\n\n2. It provides an easy way for users to copy the settings into\nthe json format consumed by qgis_process (following\nhttps://github.com/qgis/QGIS/pull/46497), so that it is easy\nfor users to see the expected format even for complex parameters\n(like tin interpolation parameters)"
|
1.0
|
Add more advanced actions to processing dialog, including copy as qgis_process command, copy/paste as json (Request in QGIS) - ### Request for documentation
From pull request QGIS/qgis#46514
Author: @nyalldawson
QGIS version: 3.24
**Add more advanced actions to processing dialog, including copy as qgis_process command, copy/paste as json**
### PR Description:
Together these actions:
- Allow for easy generation of qgis_process commands via the QGIS gui, including any tricky parameter values like gpkg outputs with specific layer (refs https://github.com/qgis/QGIS/issues/38272)
- Provide a way for users to copy the settings defined in the dialog in a text format, so they can store these easily and then
restore later by pasting the value
- Provides an easy way for users to copy the settings into the json format consumed by qgis_process (following #46497), so that it is easy for users to see the expected format even for complex parameters (like tin interpolation parameters)
Refs paleolimbot/qgisprocess#56
Refs paleolimbot/qgisprocess#44
Sponsored by the Research Institute for Nature and Forest, Flemish Govt
### Commits tagged with [need-docs] or [FEATURE]
"[feature] Add action in advanced menu for processing algorithm to copy\nequivalent qgis_process command\n\nAllows for easy generation of qgis_process commands via the QGIS gui" "[feature] Add action to processing algorithm advanced menu to copy the\ncurrent settings as JSON and paste json settings from clipboard\n\nThe utility of this is two fold:\n\n1. It provides a way for users to copy the settings defined in the\ndialog in a text format, so they can store these easily and then\nrestore later by pasting the value\n\n2. It provides an easy way for users to copy the settings into\nthe json format consumed by qgis_process (following\nhttps://github.com/qgis/QGIS/pull/46497), so that it is easy\nfor users to see the expected format even for complex parameters\n(like tin interpolation parameters)"
|
process
|
add more advanced actions to processing dialog including copy as qgis process command copy paste as json request in qgis request for documentation from pull request qgis qgis author nyalldawson qgis version add more advanced actions to processing dialog including copy as qgis process command copy paste as json pr description together these actions allow for easy generation of qgis process commands via the qgis gui including any tricky parameter values like gpkg outputs with specific layer refs provide a way for users to copy the settings defined in the dialog in a text format so they can store these easily and then restore later by pasting the value provides an easy way for users to copy the settings into the json format consumed by qgis process following so that it is easy for users to see the expected format even for complex parameters like tin interpolation parameters refs paleolimbot qgisprocess refs paleolimbot qgisprocess sponsored by the research institute for nature and forest flemish govt commits tagged with or add action in advanced menu for processing algorithm to copy nequivalent qgis process command n nallows for easy generation of qgis process commands via the qgis gui add action to processing algorithm advanced menu to copy the ncurrent settings as json and paste json settings from clipboard n nthe utility of this is two fold n it provides a way for users to copy the settings defined in the ndialog in a text format so they can store these easily and then nrestore later by pasting the value n it provides an easy way for users to copy the settings into nthe json format consumed by qgis process following n so that it is easy nfor users to see the expected format even for complex parameters n like tin interpolation parameters
| 1
|
8,287
| 11,453,542,964
|
IssuesEvent
|
2020-02-06 15:34:33
|
ION28/BLUESPAWN
|
https://api.github.com/repos/ION28/BLUESPAWN
|
closed
|
T1055, T1093, T1186 - Implement Process Related Hunt Techniques
|
difficulty/hard lang/c++ module/processes priority/low type/enhancement
|
Will implement after @jnmcd finishes up process analytics testing
|
1.0
|
T1055, T1093, T1186 - Implement Process Related Hunt Techniques - Will implement after @jnmcd finishes up process analytics testing
|
process
|
implement process related hunt techniques will implement after jnmcd finishes up process analytics testing
| 1
|
22,604
| 31,825,171,562
|
IssuesEvent
|
2023-09-14 07:00:02
|
0xPolygonMiden/miden-vm
|
https://api.github.com/repos/0xPolygonMiden/miden-vm
|
opened
|
Expand capabilities of the `debug` instruction
|
assembly processor
|
In #1069 we introduced the basic `debug` instruction which prints out the current state of the stack. We should expand the capabilities of this instruction to also enable printing of memory and procedure locals (as described in https://github.com/0xPolygonMiden/miden-vm/issues/988#issuecomment-1711099264).
|
1.0
|
Expand capabilities of the `debug` instruction - In #1069 we introduced the basic `debug` instruction which prints out the current state of the stack. We should expand the capabilities of this instruction to also enable printing of memory and procedure locals (as described in https://github.com/0xPolygonMiden/miden-vm/issues/988#issuecomment-1711099264).
|
process
|
expand capabilities of the debug instruction in we introduced the basic debug instruction which prints out the current state of the stack we should expand the capabilities of this instruction to also enable printing of memory and procedure locals as described in
| 1
|
11,657
| 14,519,299,281
|
IssuesEvent
|
2020-12-14 02:28:20
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Pseudohyphal growth definition is S. cerevisie-specific
|
cellular processes textual definition
|
This is a side issue brought up by Karen in #13157 so it should be a separate ticket.
The current definition for GO:0007124 'pseudohyphal growth' specifically describes the situation in S. cerevisiae diploid cells. Other fungi do not necessarily require nitrogen limitation and an abundant fermentable carbon source to grow pseudohyphally. These conditions are specific to S. cerevisiae diploid strains. In S. cerevisiae haploids, these same conditions stimulate "(haploid) invasive growth."' I'm not sure if S. pombe requires the same conditions/limitations. Where it states that cells become elongated, this implies a non-elongated (yeast) switches but some fungi (Candida albicans) can switch from true hyphal growth to pdeudohyphal growth. In this case, "becomes elongated" is not correct.
Both SGD and CGD have many annotations to this term and it is appropriate but the definition is too narrow and specific to S. cerevisiae. See PMID: 19347739.
This definition should be broadened to remove the S. cerevisiae-specific facts.
"The ability of cells to grow as a chain of physically attached, elongated cells in response to an environmental stimulus or stimuli."
This broader definition is preferable because orthologs in Sc and Ca are often involved in this process and the definition should be correct for more than just S. cerevisiae. This or a similar definition for GO:0007124 and related will not cause any need for reannotation.
@mah11
@srengel
|
1.0
|
Pseudohyphal growth definition is S. cerevisie-specific - This is a side issue brought up by Karen in #13157 so it should be a separate ticket.
The current definition for GO:0007124 'pseudohyphal growth' specifically describes the situation in S. cerevisiae diploid cells. Other fungi do not necessarily require nitrogen limitation and an abundant fermentable carbon source to grow pseudohyphally. These conditions are specific to S. cerevisiae diploid strains. In S. cerevisiae haploids, these same conditions stimulate "(haploid) invasive growth."' I'm not sure if S. pombe requires the same conditions/limitations. Where it states that cells become elongated, this implies a non-elongated (yeast) switches but some fungi (Candida albicans) can switch from true hyphal growth to pdeudohyphal growth. In this case, "becomes elongated" is not correct.
Both SGD and CGD have many annotations to this term and it is appropriate but the definition is too narrow and specific to S. cerevisiae. See PMID: 19347739.
This definition should be broadened to remove the S. cerevisiae-specific facts.
"The ability of cells to grow as a chain of physically attached, elongated cells in response to an environmental stimulus or stimuli."
This broader definition is preferable because orthologs in Sc and Ca are often involved in this process and the definition should be correct for more than just S. cerevisiae. This or a similar definition for GO:0007124 and related will not cause any need for reannotation.
@mah11
@srengel
|
process
|
pseudohyphal growth definition is s cerevisie specific this is a side issue brought up by karen in so it should be a separate ticket the current definition for go pseudohyphal growth specifically describes the situation in s cerevisiae diploid cells other fungi do not necessarily require nitrogen limitation and an abundant fermentable carbon source to grow pseudohyphally these conditions are specific to s cerevisiae diploid strains in s cerevisiae haploids these same conditions stimulate haploid invasive growth i m not sure if s pombe requires the same conditions limitations where it states that cells become elongated this implies a non elongated yeast switches but some fungi candida albicans can switch from true hyphal growth to pdeudohyphal growth in this case becomes elongated is not correct both sgd and cgd have many annotations to this term and it is appropriate but the definition is too narrow and specific to s cerevisiae see pmid this definition should be broadened to remove the s cerevisiae specific facts the ability of cells to grow as a chain of physically attached elongated cells in response to an environmental stimulus or stimuli this broader definition is preferable because orthologs in sc and ca are often involved in this process and the definition should be correct for more than just s cerevisiae this or a similar definition for go and related will not cause any need for reannotation srengel
| 1
|
568,183
| 16,961,668,191
|
IssuesEvent
|
2021-06-29 05:21:24
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
closed
|
Speedreader and reader mode icon is too small
|
OS/Desktop QA/No feature/speedreader priority/P3 release-notes/exclude
|
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
The Speedreader and reader mode icon is too small.
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Go to any readable page
2. See that the icon is too small
## Actual result:
<!--Please add screenshots if needed-->


## Expected result:
<img width="201" alt="Screen Shot 2021-06-28 at 2 08 35 PM" src="https://user-images.githubusercontent.com/2388823/123704236-56168d80-d81a-11eb-83f1-a9a5be2d7c74.png">
<img width="257" alt="Screen Shot 2021-06-28 at 2 08 44 PM" src="https://user-images.githubusercontent.com/2388823/123704264-5adb4180-d81a-11eb-8ebe-4249df61d6ae.png">
<img width="300" alt="Screen Shot 2021-06-28 at 2 08 56 PM" src="https://user-images.githubusercontent.com/2388823/123704312-62024f80-d81a-11eb-961e-306c13b29c4f.png">
<img width="337" alt="Screen Shot 2021-06-28 at 2 09 03 PM" src="https://user-images.githubusercontent.com/2388823/123704334-6595d680-d81a-11eb-8154-f12c0fa3cea7.png">
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easily
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? n/a
- Can you reproduce this issue with the beta channel? n/a
- Can you reproduce this issue with the nightly channel? Yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? n/a
- Does the issue resolve itself when disabling Brave Rewards? n/a
- Is the issue reproducible on the latest version of Chrome? n/a
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
|
1.0
|
Speedreader and reader mode icon is too small - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
The Speedreader and reader mode icon is too small.
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Go to any readable page
2. See that the icon is too small
## Actual result:
<!--Please add screenshots if needed-->


## Expected result:
<img width="201" alt="Screen Shot 2021-06-28 at 2 08 35 PM" src="https://user-images.githubusercontent.com/2388823/123704236-56168d80-d81a-11eb-83f1-a9a5be2d7c74.png">
<img width="257" alt="Screen Shot 2021-06-28 at 2 08 44 PM" src="https://user-images.githubusercontent.com/2388823/123704264-5adb4180-d81a-11eb-8ebe-4249df61d6ae.png">
<img width="300" alt="Screen Shot 2021-06-28 at 2 08 56 PM" src="https://user-images.githubusercontent.com/2388823/123704312-62024f80-d81a-11eb-961e-306c13b29c4f.png">
<img width="337" alt="Screen Shot 2021-06-28 at 2 09 03 PM" src="https://user-images.githubusercontent.com/2388823/123704334-6595d680-d81a-11eb-8154-f12c0fa3cea7.png">
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easily
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? n/a
- Can you reproduce this issue with the beta channel? n/a
- Can you reproduce this issue with the nightly channel? Yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? n/a
- Does the issue resolve itself when disabling Brave Rewards? n/a
- Is the issue reproducible on the latest version of Chrome? n/a
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
|
non_process
|
speedreader and reader mode icon is too small have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description the speedreader and reader mode icon is too small steps to reproduce go to any readable page see that the icon is too small actual result expected result img width alt screen shot at pm src img width alt screen shot at pm src img width alt screen shot at pm src img width alt screen shot at pm src reproduces how often easily brave version brave version info version channel information can you reproduce this issue with the current release n a can you reproduce this issue with the beta channel n a can you reproduce this issue with the nightly channel yes other additional information does the issue resolve itself when disabling brave shields n a does the issue resolve itself when disabling brave rewards n a is the issue reproducible on the latest version of chrome n a miscellaneous information
| 0
|
22,544
| 31,717,828,087
|
IssuesEvent
|
2023-09-10 03:49:28
|
Flow-Glow/Code-Jam-2023-Async-Aggregators
|
https://api.github.com/repos/Flow-Glow/Code-Jam-2023-Async-Aggregators
|
closed
|
Add real time image updating to UI
|
enhancement image processing
|
When performing edits to an image in the UI, we want the image to change in real time inside of the Pyqt window.
|
1.0
|
Add real time image updating to UI - When performing edits to an image in the UI, we want the image to change in real time inside of the Pyqt window.
|
process
|
add real time image updating to ui when performing edits to an image in the ui we want the image to change in real time inside of the pyqt window
| 1
|
1,885
| 4,712,358,396
|
IssuesEvent
|
2016-10-14 16:30:07
|
material-motion/material-motion-family-pop-swift
|
https://api.github.com/repos/material-motion/material-motion-family-pop-swift
|
closed
|
Publish the v1.0.0 release
|
Process
|
Must be ran by a member of the @material-motion/core-team
`mdm release publish 1.0.0`
|
1.0
|
Publish the v1.0.0 release - Must be ran by a member of the @material-motion/core-team
`mdm release publish 1.0.0`
|
process
|
publish the release must be ran by a member of the material motion core team mdm release publish
| 1
|
775,692
| 27,235,410,346
|
IssuesEvent
|
2023-02-21 15:59:13
|
ascheid/itsg33-pbmm-issue-gen
|
https://api.github.com/repos/ascheid/itsg33-pbmm-issue-gen
|
closed
|
AC-2(-1): Account Management
|
Priority: P2 Class: Technical ITSG-33 Control: AC-2 Suggested Assignment: IT Operations Group
|
#Control Definition
ACCOUNT MANAGEMENT | AUTOMATED SYSTEM ACCOUNT MANAGEMENT
The organization employs automated mechanisms to support the management of information system accounts.
#Class
Technical
#Supplemental Guidance
The use of automated mechanisms can include, for example: using email or text messaging to automatically notify account managers when users are terminated or transferred; using the information system to monitor account usage; and using telephonic notification to report atypical system account usage.
#General Guide
This security control/enhancement can be met using readily available Commercial-Off-The-Shelf (COTS) components. Consequently, inclusion of this security control/enhancement is strongly encouraged in most cases.
#Suggested Placeholder Values
X
#Suggested Assignment
IT Operations Group
#Support Teams
IT Projects
|
1.0
|
AC-2(-1): Account Management - #Control Definition
ACCOUNT MANAGEMENT | AUTOMATED SYSTEM ACCOUNT MANAGEMENT
The organization employs automated mechanisms to support the management of information system accounts.
#Class
Technical
#Supplemental Guidance
The use of automated mechanisms can include, for example: using email or text messaging to automatically notify account managers when users are terminated or transferred; using the information system to monitor account usage; and using telephonic notification to report atypical system account usage.
#General Guide
This security control/enhancement can be met using readily available Commercial-Off-The-Shelf (COTS) components. Consequently, inclusion of this security control/enhancement is strongly encouraged in most cases.
#Suggested Placeholder Values
X
#Suggested Assignment
IT Operations Group
#Support Teams
IT Projects
|
non_process
|
ac account management control definition account management automated system account management the organization employs automated mechanisms to support the management of information system accounts class technical supplemental guidance the use of automated mechanisms can include for example using email or text messaging to automatically notify account managers when users are terminated or transferred using the information system to monitor account usage and using telephonic notification to report atypical system account usage general guide this security control enhancement can be met using readily available commercial off the shelf cots components consequently inclusion of this security control enhancement is strongly encouraged in most cases suggested placeholder values x suggested assignment it operations group support teams it projects
| 0
|
630,704
| 20,116,707,105
|
IssuesEvent
|
2022-02-07 20:20:57
|
googleapis/python-api-core
|
https://api.github.com/repos/googleapis/python-api-core
|
closed
|
Field mask helper ignores optional fields explicitly set to default values
|
type: bug priority: p2 :rotating_light:
|
#### Environment details
- OS type and version: glinux
- Python version: `3.7.0`
- pip version: `10.0.1`
- `google-api-core` version: `1.31.0`
#### Steps to reproduce
Run the below code example.
#### Code example
```python
import proto
from google.api_core.protobuf_helpers import field_mask
class TestMessage(proto.Message):
description = proto.Field(proto.STRING, number=1, optional=True)
optimization_score = proto.Field(proto.INT64, number=2, optional=True)
test = TestMessage()
# Possibly related to Issue #230. When an optional field is explicitly set to its default value,
# i.e. empty string or zero, the field mask helper does not recognize that it's been set.
# Users must manually add the field path in these cases.
test.description = ""
test.optimization_score = 0
fm = field_mask(None, test._pb)
assert fm.paths == ["description", "optimization_score"]
```
#### Stack trace
```
Traceback (most recent call last):
File "meow.py", line 15, in <module>
assert fm.paths == ["description", "optimization_score"]
AssertionError
```
|
1.0
|
Field mask helper ignores optional fields explicitly set to default values - #### Environment details
- OS type and version: glinux
- Python version: `3.7.0`
- pip version: `10.0.1`
- `google-api-core` version: `1.31.0`
#### Steps to reproduce
Run the below code example.
#### Code example
```python
import proto
from google.api_core.protobuf_helpers import field_mask
class TestMessage(proto.Message):
description = proto.Field(proto.STRING, number=1, optional=True)
optimization_score = proto.Field(proto.INT64, number=2, optional=True)
test = TestMessage()
# Possibly related to Issue #230. When an optional field is explicitly set to its default value,
# i.e. empty string or zero, the field mask helper does not recognize that it's been set.
# Users must manually add the field path in these cases.
test.description = ""
test.optimization_score = 0
fm = field_mask(None, test._pb)
assert fm.paths == ["description", "optimization_score"]
```
#### Stack trace
```
Traceback (most recent call last):
File "meow.py", line 15, in <module>
assert fm.paths == ["description", "optimization_score"]
AssertionError
```
|
non_process
|
field mask helper ignores optional fields explicitly set to default values environment details os type and version glinux python version pip version google api core version steps to reproduce run the below code example code example python import proto from google api core protobuf helpers import field mask class testmessage proto message description proto field proto string number optional true optimization score proto field proto number optional true test testmessage possibly related to issue when an optional field is explicitly set to its default value i e empty string or zero the field mask helper does not recognize that it s been set users must manually add the field path in these cases test description test optimization score fm field mask none test pb assert fm paths stack trace traceback most recent call last file meow py line in assert fm paths assertionerror
| 0
|
17,401
| 23,218,983,949
|
IssuesEvent
|
2022-08-02 16:19:53
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
opened
|
testdata/runfiles_test/WORKSPACE.mock pins to a very old rules_cc
|
type: process
|
### Description of the bug:
src/test/py/bazel/testdata/runfiles_test/WORKSPACE.mock uses an ancient commit from rules_cc.
https://github.com/bazelbuild/rules_cc/commit/8bd6cd75d03c01bb82561a96d9c1f9f7157b13d0
We should generate this file from distdir_deps.bzl using gen_workspace_stanza. That will align the tests using this declaration with the other integration tests that use the rules_cc we build with.
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
No need to reproduce. This is a build file change.
### Which operating system are you running Bazel on?
n/a
### What is the output of `bazel info release`?
n/a
### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel.
n/a
### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ?
```text
n/a
```
### Have you found anything relevant by searching the web?
_No response_
### Any other information, logs, or outputs that you want to share?
_No response_
|
1.0
|
testdata/runfiles_test/WORKSPACE.mock pins to a very old rules_cc - ### Description of the bug:
src/test/py/bazel/testdata/runfiles_test/WORKSPACE.mock uses an ancient commit from rules_cc.
https://github.com/bazelbuild/rules_cc/commit/8bd6cd75d03c01bb82561a96d9c1f9f7157b13d0
We should generate this file from distdir_deps.bzl using gen_workspace_stanza. That will align the tests using this declaration with the other integration tests that use the rules_cc we build with.
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
No need to reproduce. This is a build file change.
### Which operating system are you running Bazel on?
n/a
### What is the output of `bazel info release`?
n/a
### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel.
n/a
### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ?
```text
n/a
```
### Have you found anything relevant by searching the web?
_No response_
### Any other information, logs, or outputs that you want to share?
_No response_
|
process
|
testdata runfiles test workspace mock pins to a very old rules cc description of the bug src test py bazel testdata runfiles test workspace mock uses an ancient commit from rules cc we should generate this file from distdir deps bzl using gen workspace stanza that will align the tests using this declaration with the other integration tests that use the rules cc we build with what s the simplest easiest way to reproduce this bug please provide a minimal example if possible no need to reproduce this is a build file change which operating system are you running bazel on n a what is the output of bazel info release n a if bazel info release returns development version or non git tell us how you built bazel n a what s the output of git remote get url origin git rev parse master git rev parse head text n a have you found anything relevant by searching the web no response any other information logs or outputs that you want to share no response
| 1
|
1,528
| 4,118,762,877
|
IssuesEvent
|
2016-06-08 12:48:44
|
World4Fly/Interface-for-Arduino
|
https://api.github.com/repos/World4Fly/Interface-for-Arduino
|
closed
|
Design complex system to process commands
|
process
|
Create the concept on how the firmware handles messages from the interface and sends them to it by considering verification and so on...
|
1.0
|
Design complex system to process commands - Create the concept on how the firmware handles messages from the interface and sends them to it by considering verification and so on...
|
process
|
design complex system to process commands create the concept on how the firmware handles messages from the interface and sends them to it by considering verification and so on
| 1
|
4,890
| 7,763,757,039
|
IssuesEvent
|
2018-06-01 17:43:53
|
StrikeNP/trac_test
|
https://api.github.com/repos/StrikeNP/trac_test
|
closed
|
Add a function to convert.m to changed a pressure profile into altitude (Trac #4)
|
Migrated from Trac enhancement fasching@uwm.edu post_processing
|
Add a function to convert.m to changed a pressure profile into altitude. This would be useful for cases that do not specify things in terms of altitude.
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/4
```json
{
"status": "closed",
"changetime": "2009-05-16T10:07:24",
"description": "Add a function to convert.m to changed a pressure profile into altitude. This would be useful for cases that do not specify things in terms of altitude.",
"reporter": "fasching@uwm.edu",
"cc": "",
"resolution": "Verified by V. Larson",
"_ts": "1242468444000000",
"component": "post_processing",
"summary": "Add a function to convert.m to changed a pressure profile into altitude",
"priority": "minor",
"keywords": "conversion, MATLAB",
"time": "2009-05-01T21:20:08",
"milestone": "",
"owner": "fasching@uwm.edu",
"type": "enhancement"
}
```
|
1.0
|
Add a function to convert.m to changed a pressure profile into altitude (Trac #4) - Add a function to convert.m to changed a pressure profile into altitude. This would be useful for cases that do not specify things in terms of altitude.
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/4
```json
{
"status": "closed",
"changetime": "2009-05-16T10:07:24",
"description": "Add a function to convert.m to changed a pressure profile into altitude. This would be useful for cases that do not specify things in terms of altitude.",
"reporter": "fasching@uwm.edu",
"cc": "",
"resolution": "Verified by V. Larson",
"_ts": "1242468444000000",
"component": "post_processing",
"summary": "Add a function to convert.m to changed a pressure profile into altitude",
"priority": "minor",
"keywords": "conversion, MATLAB",
"time": "2009-05-01T21:20:08",
"milestone": "",
"owner": "fasching@uwm.edu",
"type": "enhancement"
}
```
|
process
|
add a function to convert m to changed a pressure profile into altitude trac add a function to convert m to changed a pressure profile into altitude this would be useful for cases that do not specify things in terms of altitude migrated from json status closed changetime description add a function to convert m to changed a pressure profile into altitude this would be useful for cases that do not specify things in terms of altitude reporter fasching uwm edu cc resolution verified by v larson ts component post processing summary add a function to convert m to changed a pressure profile into altitude priority minor keywords conversion matlab time milestone owner fasching uwm edu type enhancement
| 1
|
172,199
| 27,243,694,636
|
IssuesEvent
|
2023-02-21 23:05:15
|
SAP/fundamental
|
https://api.github.com/repos/SAP/fundamental
|
closed
|
Website Improvement: Foundation Landing Page
|
Improvement Design Website Documentation UX Visual
|
Create some kind of information and visual treatment for this landing page.
|
1.0
|
Website Improvement: Foundation Landing Page - Create some kind of information and visual treatment for this landing page.
|
non_process
|
website improvement foundation landing page create some kind of information and visual treatment for this landing page
| 0
|
5,782
| 8,632,439,914
|
IssuesEvent
|
2018-11-22 10:47:56
|
kiwicom/orbit-components
|
https://api.github.com/repos/kiwicom/orbit-components
|
closed
|
<TripSector>'s 100% height breaks layout
|
Bug Processing
|
`<TripSector>` breaks layout as it has rule of 100% height.
## Expected Behavior
This component shouldn't break layout
<img width="622" alt="screen shot 2018-11-13 at 13 44 19" src="https://user-images.githubusercontent.com/3975660/48416306-16c1ce00-e750-11e8-954d-f6f33c2ce4b3.png">
|
1.0
|
<TripSector>'s 100% height breaks layout - `<TripSector>` breaks layout as it has rule of 100% height.
## Expected Behavior
This component shouldn't break layout
<img width="622" alt="screen shot 2018-11-13 at 13 44 19" src="https://user-images.githubusercontent.com/3975660/48416306-16c1ce00-e750-11e8-954d-f6f33c2ce4b3.png">
|
process
|
s height breaks layout breaks layout as it has rule of height expected behavior this component shouldn t break layout img width alt screen shot at src
| 1
|
15,187
| 18,956,009,498
|
IssuesEvent
|
2021-11-18 20:20:59
|
NixOS/nixpkgs
|
https://api.github.com/repos/NixOS/nixpkgs
|
closed
|
Add `pkgconfig` aliasing to release notes
|
2.status: stale 6.topic: release process
|
Follow up item to #109887
Add in the release notes that it's been properly included into `aliases.nix` and will not be available if `config.allowAliases` is set to `false`.
|
1.0
|
Add `pkgconfig` aliasing to release notes - Follow up item to #109887
Add in the release notes that it's been properly included into `aliases.nix` and will not be available if `config.allowAliases` is set to `false`.
|
process
|
add pkgconfig aliasing to release notes follow up item to add in the release notes that it s been properly included into aliases nix and will not be available if config allowaliases is set to false
| 1
|
147,238
| 5,635,838,877
|
IssuesEvent
|
2017-04-06 02:34:17
|
Polymer/polymer-analyzer
|
https://api.github.com/repos/Polymer/polymer-analyzer
|
closed
|
Analyze @namespace jsdoc annotations
|
Priority: Critical Status: Accepted Type: Enhancement
|
We need to find variable declarations with a `@namespace` annotation and create ScannedNamespace and Namespace features.
Examples:
```javascript
/**
* @namespace
*/
const Polymer = function() { … }
/**
* @namespace
* @memberof Polymer
*/
Polymer.dom = function() { … }
/**
* @memberof Polymer.dom
*/
Polymer.dom.foo = …;
/**
* @memberof Polymer.dom
*/
Polymer.dom.bar = …;
/**
* @namespace
*/
Polymer.Async = {
run() {...},
};
```
`@namespace` can take a name to the right of it, otherwise we can get the name from the left side of the assignment.
`@memberof` should be handled as a separate issue.
|
1.0
|
Analyze @namespace jsdoc annotations - We need to find variable declarations with a `@namespace` annotation and create ScannedNamespace and Namespace features.
Examples:
```javascript
/**
* @namespace
*/
const Polymer = function() { … }
/**
* @namespace
* @memberof Polymer
*/
Polymer.dom = function() { … }
/**
* @memberof Polymer.dom
*/
Polymer.dom.foo = …;
/**
* @memberof Polymer.dom
*/
Polymer.dom.bar = …;
/**
* @namespace
*/
Polymer.Async = {
run() {...},
};
```
`@namespace` can take a name to the right of it, otherwise we can get the name from the left side of the assignment.
`@memberof` should be handled as a separate issue.
|
non_process
|
analyze namespace jsdoc annotations we need to find variable declarations with a namespace annotation and create scannednamespace and namespace features examples javascript namespace const polymer function … namespace memberof polymer polymer dom function … memberof polymer dom polymer dom foo … memberof polymer dom polymer dom bar … namespace polymer async run namespace can take a name to the right of it otherwise we can get the name from the left side of the assignment memberof should be handled as a separate issue
| 0
|
6,669
| 9,784,009,318
|
IssuesEvent
|
2019-06-08 15:30:11
|
initc3/HoneyBadgerMPC
|
https://api.github.com/repos/initc3/HoneyBadgerMPC
|
closed
|
Blockchain integration
|
applications avss secret sharing feature request hyperledger integration preprocessing
|
hbMPC should ship with three blockchain integrations:
- [x] the built-in `commonsubset` from hbBFT
- [ ] the hyperledger fabric integration
- [x] web3 / solidity integration
A blockchain integration consists of essentially four components:
1. Coordination of client input (through hbAVSS or preprocess-AVSS)
2. Initiation of MPC online phase
3. Managing offline phase
4. Making results available to the blockchain
|
1.0
|
Blockchain integration - hbMPC should ship with three blockchain integrations:
- [x] the built-in `commonsubset` from hbBFT
- [ ] the hyperledger fabric integration
- [x] web3 / solidity integration
A blockchain integration consists of essentially four components:
1. Coordination of client input (through hbAVSS or preprocess-AVSS)
2. Initiation of MPC online phase
3. Managing offline phase
4. Making results available to the blockchain
|
process
|
blockchain integration hbmpc should ship with three blockchain integrations the built in commonsubset from hbbft the hyperledger fabric integration solidity integration a blockchain integration consists of essentially four components coordination of client input through hbavss or preprocess avss initiation of mpc online phase managing offline phase making results available to the blockchain
| 1
|
372,788
| 11,028,168,950
|
IssuesEvent
|
2019-12-06 11:02:17
|
jdi-testing/jdi-light
|
https://api.github.com/repos/jdi-testing/jdi-light
|
closed
|
Attach screenshots to Allure
|
priority:high
|
While option screenshot.strategy=on fail in test.properties Attach screenshots to Allure report by default
|
1.0
|
Attach screenshots to Allure - While option screenshot.strategy=on fail in test.properties Attach screenshots to Allure report by default
|
non_process
|
attach screenshots to allure while option screenshot strategy on fail in test properties attach screenshots to allure report by default
| 0
|
19,217
| 25,352,684,285
|
IssuesEvent
|
2022-11-20 00:21:25
|
ethereum/EIPs
|
https://api.github.com/repos/ethereum/EIPs
|
closed
|
EIP editors should explain that ERC was renamed to EIP
|
w-stale question r-process
|
### Pull Request
_No response_
### What happened?
The EIP editors have apparently decided that ERCs shall be renamed to EIPs. (I can see this in the CI bots' notices.)
That's fine and I respect your control of this naming.
Before I go around telling everybody that ERC-721 is now renamed to EIP-721, it will be helpful if the EIP official documentation does not contradict me. Please do these:
- [ ] Update https://github.com/ethereum/EIPs/blob/master/README.md to use this new wording
- [ ] Make a blog post documenting this significant change and link to it prominently from the README and EIP-1
### Relevant log output
_No response_
|
1.0
|
EIP editors should explain that ERC was renamed to EIP - ### Pull Request
_No response_
### What happened?
The EIP editors have apparently decided that ERCs shall be renamed to EIPs. (I can see this in the CI bots' notices.)
That's fine and I respect your control of this naming.
Before I go around telling everybody that ERC-721 is now renamed to EIP-721, it will be helpful if the EIP official documentation does not contradict me. Please do these:
- [ ] Update https://github.com/ethereum/EIPs/blob/master/README.md to use this new wording
- [ ] Make a blog post documenting this significant change and link to it prominently from the README and EIP-1
### Relevant log output
_No response_
|
process
|
eip editors should explain that erc was renamed to eip pull request no response what happened the eip editors have apparently decided that ercs shall be renamed to eips i can see this in the ci bots notices that s fine and i respect your control of this naming before i go around telling everybody that erc is now renamed to eip it will be helpful if the eip official documentation does not contradict me please do these update to use this new wording make a blog post documenting this significant change and link to it prominently from the readme and eip relevant log output no response
| 1
|
6,346
| 9,389,961,017
|
IssuesEvent
|
2019-04-06 00:19:47
|
natario1/CameraView
|
https://api.github.com/repos/natario1/CameraView
|
closed
|
Throw errors when released frames are accessed
|
about:frame processing is:enhancement
|
### Problem to be addressed
After releasing, `Frame`s are invalid but can still be inspected.
### Describe the solution you'd like
`Frame` access after releasing should throw a readable error.
|
1.0
|
Throw errors when released frames are accessed - ### Problem to be addressed
After releasing, `Frame`s are invalid but can still be inspected.
### Describe the solution you'd like
`Frame` access after releasing should throw a readable error.
|
process
|
throw errors when released frames are accessed problem to be addressed after releasing frame s are invalid but can still be inspected describe the solution you d like frame access after releasing should throw a readable error
| 1
|
16,946
| 5,311,045,660
|
IssuesEvent
|
2017-02-13 01:03:15
|
WayofTime/BloodMagic
|
https://api.github.com/repos/WayofTime/BloodMagic
|
closed
|
PacketNoSpamChat isn't handled on main thread
|
1.10 1.11 bug code complete
|
#### Issue Description:
`PacketNoSpamChat`'s handler doesn't seem to schedule itself on the main thread. Instead, it adds to chat on the current thread.
#### What happens:
Crash with a ConcurrentModificationException when rendering the chat (from tabbychat).
I wrote some code to catch it, so now the stacktrace is more useful.
https://gist.github.com/killjoy1221/efede90225e4511a03a483a602dddfe6
It's been fixed on my side, but it's gonna spam the log with stacktraces each time it happens outside the minecraft thread.
#### Steps to reproduce:
1. Install TabbyChat
2. Hold right click with a sigil on the blood altar
____
#### Affected Versions (Do *not* use "latest"):
- BloodMagic: commit: https://github.com/WayofTime/BloodMagic/commit/11fe41c654678cbf6d96427b82b73ec431081d31 (It's been seen in 1.8)
- Minecraft: 1.11.2
- Forge: 13.20.0.2226
|
1.0
|
PacketNoSpamChat isn't handled on main thread - #### Issue Description:
`PacketNoSpamChat`'s handler doesn't seem to schedule itself on the main thread. Instead, it adds to chat on the current thread.
#### What happens:
Crash with a ConcurrentModificationException when rendering the chat (from tabbychat).
I wrote some code to catch it, so now the stacktrace is more useful.
https://gist.github.com/killjoy1221/efede90225e4511a03a483a602dddfe6
It's been fixed on my side, but it's gonna spam the log with stacktraces each time it happens outside the minecraft thread.
#### Steps to reproduce:
1. Install TabbyChat
2. Hold right click with a sigil on the blood altar
____
#### Affected Versions (Do *not* use "latest"):
- BloodMagic: commit: https://github.com/WayofTime/BloodMagic/commit/11fe41c654678cbf6d96427b82b73ec431081d31 (It's been seen in 1.8)
- Minecraft: 1.11.2
- Forge: 13.20.0.2226
|
non_process
|
packetnospamchat isn t handled on main thread issue description packetnospamchat s handler doesn t seem to schedule itself on the main thread instead it adds to chat on the current thread what happens crash with a concurrentmodificationexception when rendering the chat from tabbychat i wrote some code to catch it so now the stacktrace is more useful it s been fixed on my side but it s gonna spam the log with stacktraces each time it happens outside the minecraft thread steps to reproduce install tabbychat hold right click with a sigil on the blood altar affected versions do not use latest bloodmagic commit it s been seen in minecraft forge
| 0
|
90,188
| 3,812,655,548
|
IssuesEvent
|
2016-03-27 18:58:32
|
HubTurbo/HubTurbo
|
https://api.github.com/repos/HubTurbo/HubTurbo
|
closed
|
Create Issue Picker to choose issue easily
|
aspect-ui priority.high
|
## UI Design
- text field for searching
- suggestion menu that will pop-up when typing
|
1.0
|
Create Issue Picker to choose issue easily - ## UI Design
- text field for searching
- suggestion menu that will pop-up when typing
|
non_process
|
create issue picker to choose issue easily ui design text field for searching suggestion menu that will pop up when typing
| 0
|
9,494
| 12,487,035,615
|
IssuesEvent
|
2020-05-31 06:35:06
|
oppia/oppia-android
|
https://api.github.com/repos/oppia/oppia-android
|
closed
|
Add DragDropAndSort case in StateReceiver
|
Priority: Essential Status: Finished Type: Process Where: Exploration player
|
Add testing exploration file with handling rules for Drag & Drop in StateReciever
Target PR date: 19 May 2020
Target completion date: 20 May 2020
|
1.0
|
Add DragDropAndSort case in StateReceiver - Add testing exploration file with handling rules for Drag & Drop in StateReciever
Target PR date: 19 May 2020
Target completion date: 20 May 2020
|
process
|
add dragdropandsort case in statereceiver add testing exploration file with handling rules for drag drop in statereciever target pr date may target completion date may
| 1
|
501,899
| 14,536,166,260
|
IssuesEvent
|
2020-12-15 07:08:23
|
kubesphere/kubesphere
|
https://api.github.com/repos/kubesphere/kubesphere
|
closed
|
The submission record is not displayed on the pipeline activity details page
|
area/console kind/bug priority/medium stale
|
**Describe the Bug**
The submission record is not displayed on the pipeline activity details page

**Versions Used**
host-k8s: v1.15.12
member-k8s: v1.18.6
/kind bug
/area console
/assign @leoendless
/milestone 3.1.0
/priority medium
|
1.0
|
The submission record is not displayed on the pipeline activity details page -
**Describe the Bug**
The submission record is not displayed on the pipeline activity details page

**Versions Used**
host-k8s: v1.15.12
member-k8s: v1.18.6
/kind bug
/area console
/assign @leoendless
/milestone 3.1.0
/priority medium
|
non_process
|
the submission record is not displayed on the pipeline activity details page describe the bug the submission record is not displayed on the pipeline activity details page versions used host member kind bug area console assign leoendless milestone priority medium
| 0
|
49,706
| 26,268,808,133
|
IssuesEvent
|
2023-01-06 15:04:37
|
NVIDIA/spark-rapids
|
https://api.github.com/repos/NVIDIA/spark-rapids
|
closed
|
[BUG] Parquet Chunked reader causing performance overhead
|
bug performance
|
**Describe the bug**
I have a customer job where the parquet chunked reader is causing the stage to go from 2.1 minutes to 3 minutes.
It looks like with the chunked reader we have about 3 times the number of batches and the shuffle read and/or coalescebatches is taking much longer.
Needs more investigation.
|
True
|
[BUG] Parquet Chunked reader causing performance overhead - **Describe the bug**
I have a customer job where the parquet chunked reader is causing the stage to go from 2.1 minutes to 3 minutes.
It looks like with the chunked reader we have about 3 times the number of batches and the shuffle read and/or coalescebatches is taking much longer.
Needs more investigation.
|
non_process
|
parquet chunked reader causing performance overhead describe the bug i have a customer job where the parquet chunked reader is causing the stage to go from minutes to minutes it looks like with the chunked reader we have about times the number of batches and the shuffle read and or coalescebatches is taking much longer needs more investigation
| 0
|
92,506
| 15,857,087,389
|
IssuesEvent
|
2021-04-08 03:56:11
|
Thezone1975/send
|
https://api.github.com/repos/Thezone1975/send
|
opened
|
CVE-2020-8237 (High) detected in json-bigint-0.3.0.tgz
|
security vulnerability
|
## CVE-2020-8237 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-bigint-0.3.0.tgz</b></p></summary>
<p>JSON.parse with bigints support</p>
<p>Library home page: <a href="https://registry.npmjs.org/json-bigint/-/json-bigint-0.3.0.tgz">https://registry.npmjs.org/json-bigint/-/json-bigint-0.3.0.tgz</a></p>
<p>Path to dependency file: /send/package.json</p>
<p>Path to vulnerable library: send/node_modules/json-bigint/package.json</p>
<p>
Dependency Hierarchy:
- storage-3.0.1.tgz (Root Library)
- common-2.0.1.tgz
- google-auth-library-4.2.1.tgz
- gcp-metadata-2.0.0.tgz
- :x: **json-bigint-0.3.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution in json-bigint npm package < 1.0.0 may lead to a denial-of-service (DoS) attack.
<p>Publish Date: 2020-09-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8237>CVE-2020-8237</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://hackerone.com/reports/916430">https://hackerone.com/reports/916430</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: v1.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-8237 (High) detected in json-bigint-0.3.0.tgz - ## CVE-2020-8237 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-bigint-0.3.0.tgz</b></p></summary>
<p>JSON.parse with bigints support</p>
<p>Library home page: <a href="https://registry.npmjs.org/json-bigint/-/json-bigint-0.3.0.tgz">https://registry.npmjs.org/json-bigint/-/json-bigint-0.3.0.tgz</a></p>
<p>Path to dependency file: /send/package.json</p>
<p>Path to vulnerable library: send/node_modules/json-bigint/package.json</p>
<p>
Dependency Hierarchy:
- storage-3.0.1.tgz (Root Library)
- common-2.0.1.tgz
- google-auth-library-4.2.1.tgz
- gcp-metadata-2.0.0.tgz
- :x: **json-bigint-0.3.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution in json-bigint npm package < 1.0.0 may lead to a denial-of-service (DoS) attack.
<p>Publish Date: 2020-09-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8237>CVE-2020-8237</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://hackerone.com/reports/916430">https://hackerone.com/reports/916430</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: v1.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in json bigint tgz cve high severity vulnerability vulnerable library json bigint tgz json parse with bigints support library home page a href path to dependency file send package json path to vulnerable library send node modules json bigint package json dependency hierarchy storage tgz root library common tgz google auth library tgz gcp metadata tgz x json bigint tgz vulnerable library vulnerability details prototype pollution in json bigint npm package may lead to a denial of service dos attack publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
711,352
| 24,460,116,162
|
IssuesEvent
|
2022-10-07 10:20:12
|
wazuh/wazuh-documentation
|
https://api.github.com/repos/wazuh/wazuh-documentation
|
closed
|
Release notes for 3.13.1/3.13.2 not showing support for Open Distro 1.9
|
priority: high type: bug
|
Hi team!
The release notes for [3.13.1](https://documentation.wazuh.com/3.13/release-notes/release_3_13_1.html) and [3.13.2](https://documentation.wazuh.com/3.13/release-notes/release_3_13_2.html) are not showing that they need Open Distro 1.9 to work as specify [here](https://github.com/wazuh/wazuh-kibana-app#wazuh---kibana---open-distro-version-compatibility-matrix)
Regards.
|
1.0
|
Release notes for 3.13.1/3.13.2 not showing support for Open Distro 1.9 - Hi team!
The release notes for [3.13.1](https://documentation.wazuh.com/3.13/release-notes/release_3_13_1.html) and [3.13.2](https://documentation.wazuh.com/3.13/release-notes/release_3_13_2.html) are not showing that they need Open Distro 1.9 to work as specify [here](https://github.com/wazuh/wazuh-kibana-app#wazuh---kibana---open-distro-version-compatibility-matrix)
Regards.
|
non_process
|
release notes for not showing support for open distro hi team the release notes for and are not showing that they need open distro to work as specify regards
| 0
|
191,881
| 15,306,361,749
|
IssuesEvent
|
2021-02-24 19:22:19
|
ryanheise/just_audio
|
https://api.github.com/repos/ryanheise/just_audio
|
opened
|
Tutorial, written/video
|
1 backlog documentation
|
<!--
PLEASE READ CAREFULLY!
FOR YOUR DOCUMENTATION REQUEST TO BE PROCESSED, YOU WILL NEED
TO FILL IN ALL SECTIONS BELOW. DON'T DELETE THE HEADINGS.
THANK YOU :-D
-->
**To which pages does your suggestion apply?**
- Direct URL 1
- Direct URL 2
- ...
**Quote the sentences(s) from the documentation to be improved (if any)**
> Insert here. (Skip if you are proposing an entirely new section.)
**Describe your suggestion**
I apologize if this is the wrong venue for my question but is there a tutorial, written or video, on how to get started with this audio player?
...
|
1.0
|
Tutorial, written/video - <!--
PLEASE READ CAREFULLY!
FOR YOUR DOCUMENTATION REQUEST TO BE PROCESSED, YOU WILL NEED
TO FILL IN ALL SECTIONS BELOW. DON'T DELETE THE HEADINGS.
THANK YOU :-D
-->
**To which pages does your suggestion apply?**
- Direct URL 1
- Direct URL 2
- ...
**Quote the sentences(s) from the documentation to be improved (if any)**
> Insert here. (Skip if you are proposing an entirely new section.)
**Describe your suggestion**
I apologize if this is the wrong venue for my question but is there a tutorial, written or video, on how to get started with this audio player?
...
|
non_process
|
tutorial written video please read carefully for your documentation request to be processed you will need to fill in all sections below don t delete the headings thank you d to which pages does your suggestion apply direct url direct url quote the sentences s from the documentation to be improved if any insert here skip if you are proposing an entirely new section describe your suggestion i apologize if this is the wrong venue for my question but is there a tutorial written or video on how to get started with this audio player
| 0
|
13,381
| 15,859,241,584
|
IssuesEvent
|
2021-04-08 07:47:14
|
arcus-azure/arcus.messaging
|
https://api.github.com/repos/arcus-azure/arcus.messaging
|
closed
|
Move message handling types to abstractions projects
|
area:message-processing breaking-change enhancement
|
**Is your feature request related to a problem? Please describe.**
The Azure Service Bus message pump project contains several message handling abstractions that is now tightly-coupled to the message pump project. It would be good to extract these interfaces in a separate project.
**Describe the solution you'd like**
* Move the `MessageHandling` types from the `Arcus.Messaging.Pumps.Abstractions` to the `Arcus.Messaging.Abstractions` project/namespace
* Move the `MessageHandling` types from the `Arcus.Messaging.Pumps.ServiceBus` to the `Arcus.Messaging.ServiceBus.Abstractions` project/namespace.
|
1.0
|
Move message handling types to abstractions projects - **Is your feature request related to a problem? Please describe.**
The Azure Service Bus message pump project contains several message handling abstractions that is now tightly-coupled to the message pump project. It would be good to extract these interfaces in a separate project.
**Describe the solution you'd like**
* Move the `MessageHandling` types from the `Arcus.Messaging.Pumps.Abstractions` to the `Arcus.Messaging.Abstractions` project/namespace
* Move the `MessageHandling` types from the `Arcus.Messaging.Pumps.ServiceBus` to the `Arcus.Messaging.ServiceBus.Abstractions` project/namespace.
|
process
|
move message handling types to abstractions projects is your feature request related to a problem please describe the azure service bus message pump project contains several message handling abstractions that is now tightly coupled to the message pump project it would be good to extract these interfaces in a separate project describe the solution you d like move the messagehandling types from the arcus messaging pumps abstractions to the arcus messaging abstractions project namespace move the messagehandling types from the arcus messaging pumps servicebus to the arcus messaging servicebus abstractions project namespace
| 1
|
5,766
| 8,609,080,822
|
IssuesEvent
|
2018-11-18 18:07:12
|
RosenborgSupporterSoftware/RBKweb-ext
|
https://api.github.com/repos/RosenborgSupporterSoftware/RBKweb-ext
|
opened
|
Håndtering av engelsk/norsk språk i foruminnstillinger
|
process
|
Dette vil nok ikke påvirke så mye, men noe kan det være.
Vi bør etterstrebe at utvidelsen fungerer helt likt uavhengig av hvilket språk som er valgt i innstillinger på phpbb. Det er enkelte steder i RFS hvor dette ikke var tilfelle (husker ikke hvilke i farten), men dette er noe å være obs på.
|
1.0
|
Håndtering av engelsk/norsk språk i foruminnstillinger - Dette vil nok ikke påvirke så mye, men noe kan det være.
Vi bør etterstrebe at utvidelsen fungerer helt likt uavhengig av hvilket språk som er valgt i innstillinger på phpbb. Det er enkelte steder i RFS hvor dette ikke var tilfelle (husker ikke hvilke i farten), men dette er noe å være obs på.
|
process
|
håndtering av engelsk norsk språk i foruminnstillinger dette vil nok ikke påvirke så mye men noe kan det være vi bør etterstrebe at utvidelsen fungerer helt likt uavhengig av hvilket språk som er valgt i innstillinger på phpbb det er enkelte steder i rfs hvor dette ikke var tilfelle husker ikke hvilke i farten men dette er noe å være obs på
| 1
|
11,373
| 14,215,317,384
|
IssuesEvent
|
2020-11-17 07:09:04
|
microsoft/react-native-windows
|
https://api.github.com/repos/microsoft/react-native-windows
|
closed
|
Update promote-release Script to Mark Most Packages in Stable Branches as Private
|
Area: Release Process enhancement
|
There are packages like react-native-windows-init that we publish only from our master branch. Publishing outside of our master branch would break both PR and publish pipelines in our master branch, along with creating confusion. The lack of guardrails for this is an especially sharp edge as of 0.63, as bumps to the react-native-windows version will now bump react-native-windows-init, as the latter takes a devDependency on the former to use its typings.
We should prevent the wrong thing from happening by marking most packages as private in stable branches. This can be done pretty easily in the `promote-release` script, to mark any package.json file as private that doesn't add the `promoteRelease: true` flag.
|
1.0
|
Update promote-release Script to Mark Most Packages in Stable Branches as Private - There are packages like react-native-windows-init that we publish only from our master branch. Publishing outside of our master branch would break both PR and publish pipelines in our master branch, along with creating confusion. The lack of guardrails for this is an especially sharp edge as of 0.63, as bumps to the react-native-windows version will now bump react-native-windows-init, as the latter takes a devDependency on the former to use its typings.
We should prevent the wrong thing from happening by marking most packages as private in stable branches. This can be done pretty easily in the `promote-release` script, to mark any package.json file as private that doesn't add the `promoteRelease: true` flag.
|
process
|
update promote release script to mark most packages in stable branches as private there are packages like react native windows init that we publish only from our master branch publishing outside of our master branch would break both pr and publish pipelines in our master branch along with creating confusion the lack of guardrails for this is an especially sharp edge as of as bumps to the react native windows version will now bump react native windows init as the latter takes a devdependency on the former to use its typings we should prevent the wrong thing from happening by marking most packages as private in stable branches this can be done pretty easily in the promote release script to mark any package json file as private that doesn t add the promoterelease true flag
| 1
|
140,037
| 21,002,256,156
|
IssuesEvent
|
2022-03-29 18:39:20
|
zesty-io/nextjs-website
|
https://api.github.com/repos/zesty-io/nextjs-website
|
opened
|
Pricing Page - H1 and H2
|
CSS Design SEO
|
https://www.zesty.io/pricing/
Can we please change the title "Headless CMS Pricing" to an H3 and the sentence below it ("Flexible headless CMS pricing for where your business is at and where you plan to go") to an H1? Keep the same styling.
|
1.0
|
Pricing Page - H1 and H2 - https://www.zesty.io/pricing/
Can we please change the title "Headless CMS Pricing" to an H3 and the sentence below it ("Flexible headless CMS pricing for where your business is at and where you plan to go") to an H1? Keep the same styling.
|
non_process
|
pricing page and can we please change the title headless cms pricing to an and the sentence below it flexible headless cms pricing for where your business is at and where you plan to go to an keep the same styling
| 0
|
9,796
| 12,810,607,752
|
IssuesEvent
|
2020-07-03 19:17:06
|
Maximus5/ConEmu
|
https://api.github.com/repos/Maximus5/ConEmu
|
closed
|
FR: Support variable errorlevel for use in a CMD prompt
|
processes
|
Hi,
I'm trying to add the errorlevel (saved in the internal variable `%errorlevel%`) to my CMD prompt.
The expected result is, that the errorlevel of the last command is displayed if it exits and the prompt appears again.
There's already the possibility to add environment variables with an ANSI sequence:
`$E]9;8;"USERNAME"$E\` adds the variable `USERNAME` to the prompt, which works fine and is possible for any environment variable displayed by the `SET` command.
Anyway `$E]9;8;"ERRORLEVEL"$E\` adds an empty string to the prompt, `ERRORLEVEL` is an internal variable and not displayed by `SET`, but can be accessed in scripts like any environment variable by `%ERRORLEVEL%`
|
1.0
|
FR: Support variable errorlevel for use in a CMD prompt - Hi,
I'm trying to add the errorlevel (saved in the internal variable `%errorlevel%`) to my CMD prompt.
The expected result is, that the errorlevel of the last command is displayed if it exits and the prompt appears again.
There's already the possibility to add environment variables with an ANSI sequence:
`$E]9;8;"USERNAME"$E\` adds the variable `USERNAME` to the prompt, which works fine and is possible for any environment variable displayed by the `SET` command.
Anyway `$E]9;8;"ERRORLEVEL"$E\` adds an empty string to the prompt, `ERRORLEVEL` is an internal variable and not displayed by `SET`, but can be accessed in scripts like any environment variable by `%ERRORLEVEL%`
|
process
|
fr support variable errorlevel for use in a cmd prompt hi i m trying to add the errorlevel saved in the internal variable errorlevel to my cmd prompt the expected result is that the errorlevel of the last command is displayed if it exits and the prompt appears again there s already the possibility to add environment variables with an ansi sequence e username e adds the variable username to the prompt which works fine and is possible for any environment variable displayed by the set command anyway e errorlevel e adds an empty string to the prompt errorlevel is an internal variable and not displayed by set but can be accessed in scripts like any environment variable by errorlevel
| 1
|
14,599
| 17,703,578,276
|
IssuesEvent
|
2021-08-25 03:19:14
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
Change term - country
|
Term - change Class - Location non-normative Process - complete
|
## Change term
* Submitter: John Wieczorek (following issue raised by Ian Engelbrecht @ianengelbrecht)
* Justification (why is this change necessary?): Clarity
* Proponents (who needs this change): Everyone
Current Term definition: https://dwc.tdwg.org/list/#dwc_country
Proposed new attributes of the term:
* Term name (in lowerCamelCase): country
* Organized in Class (e.g. Location, Taxon): Location
* Definition of the term: The name of the country or major administrative unit in which the Location occurs.
* Usage comments (recommendations regarding content, etc.): Recommended best practice is to use a controlled vocabulary such as the Getty Thesaurus of Geographic Names. **Recommended best practice is to leave this field blank if the Location spans multiple entities at this administrative level or if the Location might be in one or another of multiple possible entities at this level. Multiplicity and uncertainty of the geographic entity can be captured either in the term higherGeography or in the term locality, or both.**
* Examples: `Denmark`, `Colombia`, `España`
* Refines (identifier of the broader term this term refines, if applicable): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): http://rs.tdwg.org/dwc/terms/version/country-2017-10-06
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG, if applicable): DataSets/DataSet/Units/Unit/Gathering/Country/Name
This change proposal arises from discussions in Issue #221 and https://github.com/tdwg/dwc-qa/issues/141.
I would like to recommend the same exact amendment to the usage notes for each of the geography terms continent, waterbody, islandGroup, island, stateProvince, country, municipality.
|
1.0
|
Change term - country - ## Change term
* Submitter: John Wieczorek (following issue raised by Ian Engelbrecht @ianengelbrecht)
* Justification (why is this change necessary?): Clarity
* Proponents (who needs this change): Everyone
Current Term definition: https://dwc.tdwg.org/list/#dwc_country
Proposed new attributes of the term:
* Term name (in lowerCamelCase): country
* Organized in Class (e.g. Location, Taxon): Location
* Definition of the term: The name of the country or major administrative unit in which the Location occurs.
* Usage comments (recommendations regarding content, etc.): Recommended best practice is to use a controlled vocabulary such as the Getty Thesaurus of Geographic Names. **Recommended best practice is to leave this field blank if the Location spans multiple entities at this administrative level or if the Location might be in one or another of multiple possible entities at this level. Multiplicity and uncertainty of the geographic entity can be captured either in the term higherGeography or in the term locality, or both.**
* Examples: `Denmark`, `Colombia`, `España`
* Refines (identifier of the broader term this term refines, if applicable): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): http://rs.tdwg.org/dwc/terms/version/country-2017-10-06
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG, if applicable): DataSets/DataSet/Units/Unit/Gathering/Country/Name
This change proposal arises from discussions in Issue #221 and https://github.com/tdwg/dwc-qa/issues/141.
I would like to recommend the same exact amendment to the usage notes for each of the geography terms continent, waterbody, islandGroup, island, stateProvince, country, municipality.
|
process
|
change term country change term submitter john wieczorek following issue raised by ian engelbrecht ianengelbrecht justification why is this change necessary clarity proponents who needs this change everyone current term definition proposed new attributes of the term term name in lowercamelcase country organized in class e g location taxon location definition of the term the name of the country or major administrative unit in which the location occurs usage comments recommendations regarding content etc recommended best practice is to use a controlled vocabulary such as the getty thesaurus of geographic names recommended best practice is to leave this field blank if the location spans multiple entities at this administrative level or if the location might be in one or another of multiple possible entities at this level multiplicity and uncertainty of the geographic entity can be captured either in the term highergeography or in the term locality or both examples denmark colombia españa refines identifier of the broader term this term refines if applicable none replaces identifier of the existing term that would be deprecated and replaced by this term if applicable abcd xpath of the equivalent term in abcd or efg if applicable datasets dataset units unit gathering country name this change proposal arises from discussions in issue and i would like to recommend the same exact amendment to the usage notes for each of the geography terms continent waterbody islandgroup island stateprovince country municipality
| 1
|
175,196
| 6,547,979,260
|
IssuesEvent
|
2017-09-04 17:53:04
|
commonality/archetypes
|
https://api.github.com/repos/commonality/archetypes
|
closed
|
feat(quantity): specify the Quantity archetype pattern with OpenAPI 2.0
|
Points: 8 Priority: Medium Status: Completed Type: Docs Type: Feature
|
## User story
As an API consumer,
I want to consistently describe the amount of something measured according to some standard of measurement,
In order to conduct business unambiguously and clearly.
## Acceptance criteria
- [x] 1. `Metric` represents a standard of measurement for quantities.
- [x] 2. `Unit` represents a type of `Metrics` that is part of a `SystemOfUnits`
- [x] 3. `SystemOfUnits` represents a set of related `Units` defined by a standard such as the International System of Units (SI)
- [x] 4. `SiBaseUnit` provides common properties of SI base units
- [x] 4.1. `Ampere` represents an SI unit of electric current
- [x] 4.2. `Candela` represents an SI unit of luminous intensity
- [x] 4.3. `Kelvin` represents an SI unit of temperature
- [x] 4.4. `Kilogram` represents an SI unit of mass
- [x] 4.5 `Meter` represents an SI unit of length
- [x] 4.6. `Mole` represents an SI unit of amount of substance
- [x] 4.7. `Second` represents an SI unit of time
- [x] 5. `DerivedUnit` represents a combination of one or more base `Units` according to a specific equation
- [x] 5.1. `DerivedUnitTerm` represents a term, i.e., a single `Unit` and its power within a `DerivedUnit`
- [x] 6. `LaborHour` represents the amount of work corresponding to one person working for one hour.
- [x] 7. `Quantity` specifies an amount that is measured in some `Metric` and can be used for
- [x] 7.1. Arithmetic operations
- [x] 7.2. Comparison operations
- [x] 7.3. Rounding operations with a
- [x] 7.3.1. `RoundingPolicy` that defines the mathematical semantics of the rounding operation, and a
- [x] 7.3.2. `RoundingStrategy` that represents the type of rounding to be applied
- [x] 8. Conversion with
- [x] 8.1. `StandardConversion` defines a `conversionFactor` that can be used to convert a source `Quantity` to a `Quantity` in a `targetUnit`
- [x] 8.2. `UnitConverter` represents a conversion process
## Compliance with standards
| Standard | Contents |
|----------|-----------|
| [SI][si-url] | International System of Units (BIPM) |
## Note
All archetypes have been defined, but not all of them have direct RESTful services that can operate on values.
[si-url]: http://www.bipm.org/en/measurement-units/
|
1.0
|
feat(quantity): specify the Quantity archetype pattern with OpenAPI 2.0 - ## User story
As an API consumer,
I want to consistently describe the amount of something measured according to some standard of measurement,
In order to conduct business unambiguously and clearly.
## Acceptance criteria
- [x] 1. `Metric` represents a standard of measurement for quantities.
- [x] 2. `Unit` represents a type of `Metrics` that is part of a `SystemOfUnits`
- [x] 3. `SystemOfUnits` represents a set of related `Units` defined by a standard such as the International System of Units (SI)
- [x] 4. `SiBaseUnit` provides common properties of SI base units
- [x] 4.1. `Ampere` represents an SI unit of electric current
- [x] 4.2. `Candela` represents an SI unit of luminous intensity
- [x] 4.3. `Kelvin` represents an SI unit of temperature
- [x] 4.4. `Kilogram` represents an SI unit of mass
- [x] 4.5 `Meter` represents an SI unit of length
- [x] 4.6. `Mole` represents an SI unit of amount of substance
- [x] 4.7. `Second` represents an SI unit of time
- [x] 5. `DerivedUnit` represents a combination of one or more base `Units` according to a specific equation
- [x] 5.1. `DerivedUnitTerm` represents a term, i.e., a single `Unit` and its power within a `DerivedUnit`
- [x] 6. `LaborHour` represents the amount of work corresponding to one person working for one hour.
- [x] 7. `Quantity` specifies an amount that is measured in some `Metric` and can be used for
- [x] 7.1. Arithmetic operations
- [x] 7.2. Comparison operations
- [x] 7.3. Rounding operations with a
- [x] 7.3.1. `RoundingPolicy` that defines the mathematical semantics of the rounding operation, and a
- [x] 7.3.2. `RoundingStrategy` that represents the type of rounding to be applied
- [x] 8. Conversion with
- [x] 8.1. `StandardConversion` defines a `conversionFactor` that can be used to convert a source `Quantity` to a `Quantity` in a `targetUnit`
- [x] 8.2. `UnitConverter` represents a conversion process
## Compliance with standards
| Standard | Contents |
|----------|-----------|
| [SI][si-url] | International System of Units (BIPM) |
## Note
All archetypes have been defined, but not all of them have direct RESTful services that can operate on values.
[si-url]: http://www.bipm.org/en/measurement-units/
|
non_process
|
feat quantity specify the quantity archetype pattern with openapi user story as an api consumer i want to consistently describe the amount of something measured according to some standard of measurement in order to conduct business unambiguously and clearly acceptance criteria metric represents a standard of measurement for quantities unit represents a type of metrics that is part of a systemofunits systemofunits represents a set of related units defined by a standard such as the international system of units si sibaseunit provides common properties of si base units ampere represents an si unit of electric current candela represents an si unit of luminous intensity kelvin represents an si unit of temperature kilogram represents an si unit of mass meter represents an si unit of length mole represents an si unit of amount of substance second represents an si unit of time derivedunit represents a combination of one or more base units according to a specific equation derivedunitterm represents a term i e a single unit and its power within a derivedunit laborhour represents the amount of work corresponding to one person working for one hour quantity specifies an amount that is measured in some metric and can be used for arithmetic operations comparison operations rounding operations with a roundingpolicy that defines the mathematical semantics of the rounding operation and a roundingstrategy that represents the type of rounding to be applied conversion with standardconversion defines a conversionfactor that can be used to convert a source quantity to a quantity in a targetunit unitconverter represents a conversion process compliance with standards standard contents international system of units bipm note all archetypes have been defined but not all of them have direct restful services that can operate on values
| 0
|
6,060
| 8,887,298,730
|
IssuesEvent
|
2019-01-15 05:00:45
|
mick-warehime/sixth_corp
|
https://api.github.com/repos/mick-warehime/sixth_corp
|
closed
|
remove this comment from character_factory.build
|
development process
|
# This is an intermediate fix as in future these things will be handled
# by the chassis.
chassis = data.chassis_type.build()
in the future create more issues and link them to these comments or use the TODO comment keyword so we dont forget anything
|
1.0
|
remove this comment from character_factory.build - # This is an intermediate fix as in future these things will be handled
# by the chassis.
chassis = data.chassis_type.build()
in the future create more issues and link them to these comments or use the TODO comment keyword so we dont forget anything
|
process
|
remove this comment from character factory build this is an intermediate fix as in future these things will be handled by the chassis chassis data chassis type build in the future create more issues and link them to these comments or use the todo comment keyword so we dont forget anything
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.