Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
93,947
8,459,787,280
IssuesEvent
2018-10-22 16:55:44
Microsoft/AzureStorageExplorer
https://api.github.com/repos/Microsoft/AzureStorageExplorer
closed
No 'Toggle Explorer & Manage Accounts & Add account' buttons present
:computer: windows testing
Storage Explorer Version: 1.5.0 Platform/OS Version: Windows 10 Architecture: ia32 Build Number: 20181022.4 Commit: 492b519a Regression From: Previous release 1.4.4 #### Steps to Reproduce: #### 1. Launch the Storage Explorer -> Check the top left of the Storage Explorer. #### Expected Experience: #### 'Toggle Explorer & Manage Accounts & Add account' buttons present at the top of left in storage explorer. #### Actual Experience: #### No 'Toggle Explorer & Manage Accounts & Add account' buttons present at the top left of storage explorer. ![image](https://user-images.githubusercontent.com/41351993/47286251-78df5580-d620-11e8-95e9-f3e30c9edea7.png) #### More Info: #### This issue doesn't reproduce on Mac/Linux.
1.0
No 'Toggle Explorer & Manage Accounts & Add account' buttons present - Storage Explorer Version: 1.5.0 Platform/OS Version: Windows 10 Architecture: ia32 Build Number: 20181022.4 Commit: 492b519a Regression From: Previous release 1.4.4 #### Steps to Reproduce: #### 1. Launch the Storage Explorer -> Check the top left of the Storage Explorer. #### Expected Experience: #### 'Toggle Explorer & Manage Accounts & Add account' buttons present at the top of left in storage explorer. #### Actual Experience: #### No 'Toggle Explorer & Manage Accounts & Add account' buttons present at the top left of storage explorer. ![image](https://user-images.githubusercontent.com/41351993/47286251-78df5580-d620-11e8-95e9-f3e30c9edea7.png) #### More Info: #### This issue doesn't reproduce on Mac/Linux.
non_process
no toggle explorer manage accounts add account buttons present storage explorer version platform os version windows architecture build number commit regression from previous release steps to reproduce launch the storage explorer check the top left of the storage explorer expected experience toggle explorer manage accounts add account buttons present at the top of left in storage explorer actual experience no toggle explorer manage accounts add account buttons present at the top left of storage explorer more info this issue doesn t reproduce on mac linux
0
27,830
12,709,851,073
IssuesEvent
2020-06-23 13:02:16
LiskHQ/lisk-docs
https://api.github.com/repos/LiskHQ/lisk-docs
opened
Update setup from source
service update
## Which content is outdated? - (Source code install) Postgres is also needed - (Source code install) Freegeoip is outdated and we cannot really recommend any alternative without advertising [Maxmind](https://www.maxmind.com/en/geoip-demo)
1.0
Update setup from source - ## Which content is outdated? - (Source code install) Postgres is also needed - (Source code install) Freegeoip is outdated and we cannot really recommend any alternative without advertising [Maxmind](https://www.maxmind.com/en/geoip-demo)
non_process
update setup from source which content is outdated source code install postgres is also needed source code install freegeoip is outdated and we cannot really recommend any alternative without advertising
0
2,164
5,011,564,764
IssuesEvent
2016-12-13 08:23:26
CERNDocumentServer/cds
https://api.github.com/repos/CERNDocumentServer/cds
opened
Wrong aspect ratio extracted from video
avc_processing enhancement
[ff_probe_all](https://github.com/CERNDocumentServer/cds/blob/cdslabs_qa/cds/modules/ffmpeg/ffmpeg.py#L53) should extract the aspect ratio of a video inside `display_aspect_ratio` but most of the time its value is `0:1`. Maybe a solution could the to add a small check to [ff_probe](https://github.com/CERNDocumentServer/cds/blob/cdslabs_qa/cds/modules/ffmpeg/ffmpeg.py#L34) and [ff_probe_all](https://github.com/CERNDocumentServer/cds/blob/cdslabs_qa/cds/modules/ffmpeg/ffmpeg.py#L53) to calculate the `display_aspect_ratio` from the width and height in case the value of the former is `0:1`. \* https://trac.ffmpeg.org/ticket/3798
1.0
Wrong aspect ratio extracted from video - [ff_probe_all](https://github.com/CERNDocumentServer/cds/blob/cdslabs_qa/cds/modules/ffmpeg/ffmpeg.py#L53) should extract the aspect ratio of a video inside `display_aspect_ratio` but most of the time its value is `0:1`. Maybe a solution could the to add a small check to [ff_probe](https://github.com/CERNDocumentServer/cds/blob/cdslabs_qa/cds/modules/ffmpeg/ffmpeg.py#L34) and [ff_probe_all](https://github.com/CERNDocumentServer/cds/blob/cdslabs_qa/cds/modules/ffmpeg/ffmpeg.py#L53) to calculate the `display_aspect_ratio` from the width and height in case the value of the former is `0:1`. \* https://trac.ffmpeg.org/ticket/3798
process
wrong aspect ratio extracted from video should extract the aspect ratio of a video inside display aspect ratio but most of the time its value is maybe a solution could the to add a small check to and to calculate the display aspect ratio from the width and height in case the value of the former is
1
10,877
13,646,018,752
IssuesEvent
2020-09-25 22:05:50
GoogleCloudPlatform/cloud-sql-jdbc-socket-factory
https://api.github.com/repos/GoogleCloudPlatform/cloud-sql-jdbc-socket-factory
closed
Add e2e testing
type: process
We need to add end to end tests to verify things work against the current SQL Admin version. We'll probably need to move testing to Kokoro, and possibly separate lint checks out as a separate test.
1.0
Add e2e testing - We need to add end to end tests to verify things work against the current SQL Admin version. We'll probably need to move testing to Kokoro, and possibly separate lint checks out as a separate test.
process
add testing we need to add end to end tests to verify things work against the current sql admin version we ll probably need to move testing to kokoro and possibly separate lint checks out as a separate test
1
7,356
10,490,429,141
IssuesEvent
2019-09-25 08:59:22
Graylog2/graylog2-server
https://api.github.com/repos/Graylog2/graylog2-server
opened
Allow POST in HTTP/JsonPath data adapter
feature processing
## Expected Behavior The HTTP/JsonPath data adapter should support the POST method and allow specifying a body template that gets evaluated with the key that is being looked up. ## Current Behavior The current implementation of the HTTP/JsonPath data adapter only supports the GET method to retrieve data from the remote location. For third party systems that model their searches as POSTs, such as MISP or even Graylog itself, only support GET is not enough. ## Possible Solution Add a radio button or drop down for the HTTP method (radio if all we want to support is GET and POST, I think the others don't make sense right now). Allow pasting a body template for POST and evaluate it the same way we do it for the URL right now. ## Context When trying to read data from external systems like MISP, you are faced with the problem that it requires POSTing a search, which either forces you to write middleware or implement an entirely new plugin just to be able to send a POST request, which seems excessive. ## Your Environment * Graylog Version: 3.1 * Elasticsearch Version: not applicable * MongoDB Version: not applicable * Operating System: not applicable * Browser version: not applicable
1.0
Allow POST in HTTP/JsonPath data adapter - ## Expected Behavior The HTTP/JsonPath data adapter should support the POST method and allow specifying a body template that gets evaluated with the key that is being looked up. ## Current Behavior The current implementation of the HTTP/JsonPath data adapter only supports the GET method to retrieve data from the remote location. For third party systems that model their searches as POSTs, such as MISP or even Graylog itself, only support GET is not enough. ## Possible Solution Add a radio button or drop down for the HTTP method (radio if all we want to support is GET and POST, I think the others don't make sense right now). Allow pasting a body template for POST and evaluate it the same way we do it for the URL right now. ## Context When trying to read data from external systems like MISP, you are faced with the problem that it requires POSTing a search, which either forces you to write middleware or implement an entirely new plugin just to be able to send a POST request, which seems excessive. ## Your Environment * Graylog Version: 3.1 * Elasticsearch Version: not applicable * MongoDB Version: not applicable * Operating System: not applicable * Browser version: not applicable
process
allow post in http jsonpath data adapter expected behavior the http jsonpath data adapter should support the post method and allow specifying a body template that gets evaluated with the key that is being looked up current behavior the current implementation of the http jsonpath data adapter only supports the get method to retrieve data from the remote location for third party systems that model their searches as posts such as misp or even graylog itself only support get is not enough possible solution add a radio button or drop down for the http method radio if all we want to support is get and post i think the others don t make sense right now allow pasting a body template for post and evaluate it the same way we do it for the url right now context when trying to read data from external systems like misp you are faced with the problem that it requires posting a search which either forces you to write middleware or implement an entirely new plugin just to be able to send a post request which seems excessive your environment graylog version elasticsearch version not applicable mongodb version not applicable operating system not applicable browser version not applicable
1
9,991
13,039,413,256
IssuesEvent
2020-07-28 16:42:37
Motionlessness/eCommerceSite
https://api.github.com/repos/Motionlessness/eCommerceSite
closed
Add CI Pipeline
Dev process
Add continuous pipeline that will check to make sure code in a pull request compiles successfully.
1.0
Add CI Pipeline - Add continuous pipeline that will check to make sure code in a pull request compiles successfully.
process
add ci pipeline add continuous pipeline that will check to make sure code in a pull request compiles successfully
1
8,001
11,194,504,321
IssuesEvent
2020-01-03 01:22:51
AlmuraDev/SGCraft
https://api.github.com/repos/AlmuraDev/SGCraft
closed
[Suggestion] Add config option to restrict only OPs to see gates' addresses
enhancement in process
Basically, if you're a regular user and right click a stargate, you get nothing except that facade thing for the bottom blocks. If you're an OP, you get what regular users get currently. This would be useful to me for my exploration themed modpack, for me and some friends. Basically, each of us spawn on our own planet. I'll be throwing addresses in a ton of places, but I won't be allowed to look at the addresses after that. When we actually start playing, we have to find addresses, even the addresses just to go to each others' worlds. That's the whole purpose of this- we can't see our own address, so we have to be careful traveling. Also, it would be nice if you could have another config to make it so you could still read the address if you have The One Probe, following the idea of hovering over with the proper tool to get extra info.
1.0
[Suggestion] Add config option to restrict only OPs to see gates' addresses - Basically, if you're a regular user and right click a stargate, you get nothing except that facade thing for the bottom blocks. If you're an OP, you get what regular users get currently. This would be useful to me for my exploration themed modpack, for me and some friends. Basically, each of us spawn on our own planet. I'll be throwing addresses in a ton of places, but I won't be allowed to look at the addresses after that. When we actually start playing, we have to find addresses, even the addresses just to go to each others' worlds. That's the whole purpose of this- we can't see our own address, so we have to be careful traveling. Also, it would be nice if you could have another config to make it so you could still read the address if you have The One Probe, following the idea of hovering over with the proper tool to get extra info.
process
add config option to restrict only ops to see gates addresses basically if you re a regular user and right click a stargate you get nothing except that facade thing for the bottom blocks if you re an op you get what regular users get currently this would be useful to me for my exploration themed modpack for me and some friends basically each of us spawn on our own planet i ll be throwing addresses in a ton of places but i won t be allowed to look at the addresses after that when we actually start playing we have to find addresses even the addresses just to go to each others worlds that s the whole purpose of this we can t see our own address so we have to be careful traveling also it would be nice if you could have another config to make it so you could still read the address if you have the one probe following the idea of hovering over with the proper tool to get extra info
1
1,522
4,113,607,254
IssuesEvent
2016-06-07 14:36:36
matz-e/lobster
https://api.github.com/repos/matz-e/lobster
closed
Rework the task wrapper script to quit when one step fails.
enhancement fix-ready processing
I'm currently looking into why my merge tasks fail (it's xrootd access to the storage element, I think). Currently, out tasks script just proceeds through the steps as if nothing happened. I think we should rework this to stop at the first error. My suggested changes: 1. The main flow of the script should be a bunch of function calls, no real logic (maybe beyond the CMS code distinctions for now), like: config = json.load(…) data = {…} copy_inputs(config, data, …) execute_prologue(config, data, …) … 2. Instead of using the context to set the error state, wrap the above functions, and update mona-lisa accordingly, if needed: @quit_on_failure(exit_code=210) def copy_outputs(config, data, …): …
1.0
Rework the task wrapper script to quit when one step fails. - I'm currently looking into why my merge tasks fail (it's xrootd access to the storage element, I think). Currently, out tasks script just proceeds through the steps as if nothing happened. I think we should rework this to stop at the first error. My suggested changes: 1. The main flow of the script should be a bunch of function calls, no real logic (maybe beyond the CMS code distinctions for now), like: config = json.load(…) data = {…} copy_inputs(config, data, …) execute_prologue(config, data, …) … 2. Instead of using the context to set the error state, wrap the above functions, and update mona-lisa accordingly, if needed: @quit_on_failure(exit_code=210) def copy_outputs(config, data, …): …
process
rework the task wrapper script to quit when one step fails i m currently looking into why my merge tasks fail it s xrootd access to the storage element i think currently out tasks script just proceeds through the steps as if nothing happened i think we should rework this to stop at the first error my suggested changes the main flow of the script should be a bunch of function calls no real logic maybe beyond the cms code distinctions for now like config json load … data … copy inputs config data … execute prologue config data … … instead of using the context to set the error state wrap the above functions and update mona lisa accordingly if needed quit on failure exit code def copy outputs config data … …
1
97,759
4,005,886,993
IssuesEvent
2016-05-12 13:16:46
coreos/rkt
https://api.github.com/repos/coreos/rkt
closed
Godeps: bump appc/spec for the next release
area/security dependency/appc spec priority/P0
For the next release of rkt-v1.6.0, we should wait for a new appc/spec release and then bump the version in Godeps. It will help with: - [ ] capabilities #2589 - [x] depends on https://github.com/appc/spec/pull/601 (merged, but not in an appc/spec release yet) - [ ] seccomp #1614 - [ ] depends on https://github.com/appc/spec/issues/529 (not implemented, might be postponed if not ready) According to [![Build Status](https://semaphoreci.com/api/v1/alban/godep-rkt-test/branches/master/badge.svg)](https://semaphoreci.com/alban/godep-rkt-test), rkt will need to be fixed with the following (https://github.com/appc/spec/pull/561). ``` rkt/pubkey/pubkey.go:160: undefined: discovery.InsecureHttp rkt/pubkey/pubkey.go:163: undefined: discovery.InsecureTls rkt/pubkey/pubkey.go:176: ep.Keys undefined (type discovery.PublicKeys has no field or method Keys) ```
1.0
Godeps: bump appc/spec for the next release - For the next release of rkt-v1.6.0, we should wait for a new appc/spec release and then bump the version in Godeps. It will help with: - [ ] capabilities #2589 - [x] depends on https://github.com/appc/spec/pull/601 (merged, but not in an appc/spec release yet) - [ ] seccomp #1614 - [ ] depends on https://github.com/appc/spec/issues/529 (not implemented, might be postponed if not ready) According to [![Build Status](https://semaphoreci.com/api/v1/alban/godep-rkt-test/branches/master/badge.svg)](https://semaphoreci.com/alban/godep-rkt-test), rkt will need to be fixed with the following (https://github.com/appc/spec/pull/561). ``` rkt/pubkey/pubkey.go:160: undefined: discovery.InsecureHttp rkt/pubkey/pubkey.go:163: undefined: discovery.InsecureTls rkt/pubkey/pubkey.go:176: ep.Keys undefined (type discovery.PublicKeys has no field or method Keys) ```
non_process
godeps bump appc spec for the next release for the next release of rkt we should wait for a new appc spec release and then bump the version in godeps it will help with capabilities depends on merged but not in an appc spec release yet seccomp depends on not implemented might be postponed if not ready according to rkt will need to be fixed with the following rkt pubkey pubkey go undefined discovery insecurehttp rkt pubkey pubkey go undefined discovery insecuretls rkt pubkey pubkey go ep keys undefined type discovery publickeys has no field or method keys
0
283,189
24,530,993,061
IssuesEvent
2022-10-11 16:24:47
void-linux/void-packages
https://api.github.com/repos/void-linux/void-packages
closed
python3-matplotlib-3.6.0_1: ModuleNotFoundError: No module named 'contourpy'
bug needs-testing
### Is this a new report? Yes ### System Info Void 5.19.10_1 x86_64 ### Package(s) Affected python3-matplotlib-3.6.0_1 ### Does a report exist for this bug with the project's home (upstream) and/or another distro? Not a bug, upstream says: ``` For packagers, this release contains some changes to dependencies: * Python 3.8 is now required. * NumPy 1.19 is now required. * The ContourPy package is now required for contouring. ``` (https://discourse.matplotlib.org/t/matplotlib-announce-ann-matplotlib-3-6-0/23129) So this should be fixed just by packaging `python3-contourpy`, presumably without the need to change matplotlib itself. ### Expected behaviour A contour object is created and can be shown: ``` ... >>> ax.contourf(X, Y, Z, levels=levels) <matplotlib.contour.QuadContourSet object at 0x7f8e831ebd60> >>> >>> plt.show() ``` This actually works after downgrading to `python3-matplotlib-3.5.3_1` ### Actual behaviour Example from https://matplotlib.org/stable/plot_types/arrays/contourf.html#sphx-glr-plot-types-arrays-contourf-py ``` $ python Python 3.10.7 (main, Sep 7 2022, 15:23:13) [GCC 10.2.1 20201203] on linux Type "help", "copyright", "credits" or "license" for more information. >>> import matplotlib.pyplot as plt >>> import numpy as np >>> >>> plt.style.use('_mpl-gallery-nogrid') >>> >>> # make data >>> X, Y = np.meshgrid(np.linspace(-3, 3, 256), np.linspace(-3, 3, 256)) >>> Z = (1 - X/2 + X**5 + Y**3) * np.exp(-X**2 - Y**2) >>> levels = np.linspace(Z.min(), Z.max(), 7) >>> >>> # plot >>> fig, ax = plt.subplots() >>> >>> ax.contourf(X, Y, Z, levels=levels) Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/usr/lib/python3.10/site-packages/matplotlib/__init__.py", line 1423, in inner return func(ax, *map(sanitize_sequence, args), **kwargs) File "/usr/lib/python3.10/site-packages/matplotlib/axes/_axes.py", line 6382, in contourf contours = mcontour.QuadContourSet(self, *args, **kwargs) File "/usr/lib/python3.10/site-packages/matplotlib/contour.py", line 794, in __init__ kwargs = self._process_args(*args, **kwargs) File "/usr/lib/python3.10/site-packages/matplotlib/contour.py", line 1419, in _process_args import contourpy ModuleNotFoundError: No module named 'contourpy' ``` ### Steps to reproduce Try the example from https://matplotlib.org/stable/plot_types/arrays/contourf.html#sphx-glr-plot-types-arrays-contourf-py
1.0
python3-matplotlib-3.6.0_1: ModuleNotFoundError: No module named 'contourpy' - ### Is this a new report? Yes ### System Info Void 5.19.10_1 x86_64 ### Package(s) Affected python3-matplotlib-3.6.0_1 ### Does a report exist for this bug with the project's home (upstream) and/or another distro? Not a bug, upstream says: ``` For packagers, this release contains some changes to dependencies: * Python 3.8 is now required. * NumPy 1.19 is now required. * The ContourPy package is now required for contouring. ``` (https://discourse.matplotlib.org/t/matplotlib-announce-ann-matplotlib-3-6-0/23129) So this should be fixed just by packaging `python3-contourpy`, presumably without the need to change matplotlib itself. ### Expected behaviour A contour object is created and can be shown: ``` ... >>> ax.contourf(X, Y, Z, levels=levels) <matplotlib.contour.QuadContourSet object at 0x7f8e831ebd60> >>> >>> plt.show() ``` This actually works after downgrading to `python3-matplotlib-3.5.3_1` ### Actual behaviour Example from https://matplotlib.org/stable/plot_types/arrays/contourf.html#sphx-glr-plot-types-arrays-contourf-py ``` $ python Python 3.10.7 (main, Sep 7 2022, 15:23:13) [GCC 10.2.1 20201203] on linux Type "help", "copyright", "credits" or "license" for more information. >>> import matplotlib.pyplot as plt >>> import numpy as np >>> >>> plt.style.use('_mpl-gallery-nogrid') >>> >>> # make data >>> X, Y = np.meshgrid(np.linspace(-3, 3, 256), np.linspace(-3, 3, 256)) >>> Z = (1 - X/2 + X**5 + Y**3) * np.exp(-X**2 - Y**2) >>> levels = np.linspace(Z.min(), Z.max(), 7) >>> >>> # plot >>> fig, ax = plt.subplots() >>> >>> ax.contourf(X, Y, Z, levels=levels) Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/usr/lib/python3.10/site-packages/matplotlib/__init__.py", line 1423, in inner return func(ax, *map(sanitize_sequence, args), **kwargs) File "/usr/lib/python3.10/site-packages/matplotlib/axes/_axes.py", line 6382, in contourf contours = mcontour.QuadContourSet(self, *args, **kwargs) File "/usr/lib/python3.10/site-packages/matplotlib/contour.py", line 794, in __init__ kwargs = self._process_args(*args, **kwargs) File "/usr/lib/python3.10/site-packages/matplotlib/contour.py", line 1419, in _process_args import contourpy ModuleNotFoundError: No module named 'contourpy' ``` ### Steps to reproduce Try the example from https://matplotlib.org/stable/plot_types/arrays/contourf.html#sphx-glr-plot-types-arrays-contourf-py
non_process
matplotlib modulenotfounderror no module named contourpy is this a new report yes system info void package s affected matplotlib does a report exist for this bug with the project s home upstream and or another distro not a bug upstream says for packagers this release contains some changes to dependencies python is now required numpy is now required the contourpy package is now required for contouring so this should be fixed just by packaging contourpy presumably without the need to change matplotlib itself expected behaviour a contour object is created and can be shown ax contourf x y z levels levels plt show this actually works after downgrading to matplotlib actual behaviour example from python python main sep on linux type help copyright credits or license for more information import matplotlib pyplot as plt import numpy as np plt style use mpl gallery nogrid make data x y np meshgrid np linspace np linspace z x x y np exp x y levels np linspace z min z max plot fig ax plt subplots ax contourf x y z levels levels traceback most recent call last file line in file usr lib site packages matplotlib init py line in inner return func ax map sanitize sequence args kwargs file usr lib site packages matplotlib axes axes py line in contourf contours mcontour quadcontourset self args kwargs file usr lib site packages matplotlib contour py line in init kwargs self process args args kwargs file usr lib site packages matplotlib contour py line in process args import contourpy modulenotfounderror no module named contourpy steps to reproduce try the example from
0
276,036
20,965,982,866
IssuesEvent
2022-03-28 06:44:45
jon-harper/clock-3
https://api.github.com/repos/jon-harper/clock-3
closed
Fusion 360 script to visualize t-nuts is needed
documentation enhancement Fusion Model Fusion Script
It would be extraordinarily helpful to be able to isolate *just* the t-nuts and frame to assist counting and identifying location
1.0
Fusion 360 script to visualize t-nuts is needed - It would be extraordinarily helpful to be able to isolate *just* the t-nuts and frame to assist counting and identifying location
non_process
fusion script to visualize t nuts is needed it would be extraordinarily helpful to be able to isolate just the t nuts and frame to assist counting and identifying location
0
246,027
7,893,117,845
IssuesEvent
2018-06-28 16:59:04
visit-dav/issues-test
https://api.github.com/repos/visit-dav/issues-test
closed
PFLOTRAN reader segv's reading vector var
Bug Likelihood: 3 - Occasional OS: All Priority: Normal Severity: 4 - Crash / Wrong Results Support Group: Any version: 2.8.2
Glenn Hammond at Sandia reported this and provided sample data. He said 2.7.3 works, but 2.8.x and 2.9 do not. gdb shows segv line 1403 avtPFLOTRANFileFormat.C (2.9RC version), so looks like an indexing error. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. The following information could not be accurately captured in the new ticket: Original author: Kathleen Biagas Original creation: 03/26/2015 12:30 pm Original update: 03/26/2015 01:22 pm Ticket number: 2190
1.0
PFLOTRAN reader segv's reading vector var - Glenn Hammond at Sandia reported this and provided sample data. He said 2.7.3 works, but 2.8.x and 2.9 do not. gdb shows segv line 1403 avtPFLOTRANFileFormat.C (2.9RC version), so looks like an indexing error. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. The following information could not be accurately captured in the new ticket: Original author: Kathleen Biagas Original creation: 03/26/2015 12:30 pm Original update: 03/26/2015 01:22 pm Ticket number: 2190
non_process
pflotran reader segv s reading vector var glenn hammond at sandia reported this and provided sample data he said works but x and do not gdb shows segv line avtpflotranfileformat c version so looks like an indexing error redmine migration this ticket was migrated from redmine the following information could not be accurately captured in the new ticket original author kathleen biagas original creation pm original update pm ticket number
0
19,040
25,042,669,557
IssuesEvent
2022-11-04 23:09:28
USGS-WiM/StreamStats
https://api.github.com/repos/USGS-WiM/StreamStats
opened
BP: Allow other geometry uploads
Batch Processor
Part of #1455 - [ ] Rename the "Upload Shapefile (.zip):" file uploader to "Upload Point Geometry: shapefile (.zip), GeoJSON (.json, .geojson), CSV (.csv)" - [ ] Allow the user to upload a geoJSON - [ ] Allow the user to upload a CSV - [ ] Validate the file formats when the user clicks "Submit Batch"
1.0
BP: Allow other geometry uploads - Part of #1455 - [ ] Rename the "Upload Shapefile (.zip):" file uploader to "Upload Point Geometry: shapefile (.zip), GeoJSON (.json, .geojson), CSV (.csv)" - [ ] Allow the user to upload a geoJSON - [ ] Allow the user to upload a CSV - [ ] Validate the file formats when the user clicks "Submit Batch"
process
bp allow other geometry uploads part of rename the upload shapefile zip file uploader to upload point geometry shapefile zip geojson json geojson csv csv allow the user to upload a geojson allow the user to upload a csv validate the file formats when the user clicks submit batch
1
23,538
16,384,132,056
IssuesEvent
2021-05-17 08:16:47
google/web-stories-wp
https://api.github.com/repos/google/web-stories-wp
opened
E2E Tests: take screenshots on failures
Package: E2E Tests Pod: WP & Infra Type: Infrastructure
<!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ --> ## Task Description <!-- A clear and concise description of what this task is about. --> Taking screenshots when an e2e test fails on CI would make it much easier for us to debug. Some prior art: https://github.com/WordPress/gutenberg/pull/26664 https://github.com/WordPress/gutenberg/pull/28449 Just like for #7549, it would mean using `jest-circus` as the test runner due to the events it emits to make this easier. We could try using Percy for the screenshots, but storing them as GitHub Actions artifacts is cheaper.
1.0
E2E Tests: take screenshots on failures - <!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ --> ## Task Description <!-- A clear and concise description of what this task is about. --> Taking screenshots when an e2e test fails on CI would make it much easier for us to debug. Some prior art: https://github.com/WordPress/gutenberg/pull/26664 https://github.com/WordPress/gutenberg/pull/28449 Just like for #7549, it would mean using `jest-circus` as the test runner due to the events it emits to make this easier. We could try using Percy for the screenshots, but storing them as GitHub Actions artifacts is cheaper.
non_process
tests take screenshots on failures task description taking screenshots when an test fails on ci would make it much easier for us to debug some prior art just like for it would mean using jest circus as the test runner due to the events it emits to make this easier we could try using percy for the screenshots but storing them as github actions artifacts is cheaper
0
19,153
25,234,629,765
IssuesEvent
2022-11-14 23:10:48
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
nociception is not the same as sensory perception of pain
organism-level process
**GO term** GO:0019233 'sensory perception of pain' **Issue description** 'sensory perception of pain' `has_exact_synonym` nociception According this [reference](https://www.physio-pedia.com/Nociception): > The terms nociception and [pain](https://www.physio-pedia.com/Pain_Behaviours) should not be used synonymously, because each can occur without the other.[[1]](https://www.physio-pedia.com/Nociception#cite_note-1.-1). [Pain](https://www.physio-pedia.com/Pain_Mechanisms) arising from activation of the nociceptors is called nociceptive pain.
1.0
nociception is not the same as sensory perception of pain - **GO term** GO:0019233 'sensory perception of pain' **Issue description** 'sensory perception of pain' `has_exact_synonym` nociception According this [reference](https://www.physio-pedia.com/Nociception): > The terms nociception and [pain](https://www.physio-pedia.com/Pain_Behaviours) should not be used synonymously, because each can occur without the other.[[1]](https://www.physio-pedia.com/Nociception#cite_note-1.-1). [Pain](https://www.physio-pedia.com/Pain_Mechanisms) arising from activation of the nociceptors is called nociceptive pain.
process
nociception is not the same as sensory perception of pain go term go sensory perception of pain issue description sensory perception of pain has exact synonym nociception according this the terms nociception and should not be used synonymously because each can occur without the other arising from activation of the nociceptors is called nociceptive pain
1
2,127
4,969,547,023
IssuesEvent
2016-12-05 13:44:38
AnalyticalGraphicsInc/cesium
https://api.github.com/repos/AnalyticalGraphicsInc/cesium
closed
Less greenkeeper noise
beginner dev process
Is it possible to improve the signal-to-noise ratio of this repo's pull requests by configuring greenkeeper to only do pull requests at some interval, perhaps biweekly? Since the updates are rarely significant, it has created an incredible amount of noise in our pull request history.
1.0
Less greenkeeper noise - Is it possible to improve the signal-to-noise ratio of this repo's pull requests by configuring greenkeeper to only do pull requests at some interval, perhaps biweekly? Since the updates are rarely significant, it has created an incredible amount of noise in our pull request history.
process
less greenkeeper noise is it possible to improve the signal to noise ratio of this repo s pull requests by configuring greenkeeper to only do pull requests at some interval perhaps biweekly since the updates are rarely significant it has created an incredible amount of noise in our pull request history
1
192,833
14,630,514,824
IssuesEvent
2020-12-23 17:52:21
deathlyrage/pot-demo-bugs
https://api.github.com/repos/deathlyrage/pot-demo-bugs
closed
Sucho Returning to the Surface Not Working
fixed needs testing
## Expected Behavior If you dive underwater as a suchomimus, you should be able to return and stay at the surface. ## Current Behavior If you sprint into the water and dive, you are unable to stay at the surface. You continously sink underwater every time until your stamina runs out or until you turn off sprint. ## Reproduction Steps - Play as a sucho - Sprint into a lake - Dive underwater - Attempt to stay at the surface ## Evidence (Screenshots & Videos) https://youtu.be/aubpAXE8T38 ## Crash Logs (If Applicable. No logs needed if there is no crash.) ## Build Information **Game Version:** 0.0.0.10071 **Operating System & Version:** iPhone 6s, iOS 14.2 **Alderon Games ID:** 246-116-163
1.0
Sucho Returning to the Surface Not Working - ## Expected Behavior If you dive underwater as a suchomimus, you should be able to return and stay at the surface. ## Current Behavior If you sprint into the water and dive, you are unable to stay at the surface. You continously sink underwater every time until your stamina runs out or until you turn off sprint. ## Reproduction Steps - Play as a sucho - Sprint into a lake - Dive underwater - Attempt to stay at the surface ## Evidence (Screenshots & Videos) https://youtu.be/aubpAXE8T38 ## Crash Logs (If Applicable. No logs needed if there is no crash.) ## Build Information **Game Version:** 0.0.0.10071 **Operating System & Version:** iPhone 6s, iOS 14.2 **Alderon Games ID:** 246-116-163
non_process
sucho returning to the surface not working expected behavior if you dive underwater as a suchomimus you should be able to return and stay at the surface current behavior if you sprint into the water and dive you are unable to stay at the surface you continously sink underwater every time until your stamina runs out or until you turn off sprint reproduction steps play as a sucho sprint into a lake dive underwater attempt to stay at the surface evidence screenshots videos crash logs if applicable no logs needed if there is no crash build information game version operating system version iphone ios alderon games id
0
14,435
17,484,522,805
IssuesEvent
2021-08-09 09:13:50
googleapis/python-spanner
https://api.github.com/repos/googleapis/python-spanner
closed
samples.samples.snippets_test: test_create_client_with_query_options failed
api: spanner type: process samples flakybot: issue flakybot: flaky
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 7b8d8563651a9732631b22bef38cb2030eaadfa7 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/74f0a66c-c356-4d06-9f52-85f92729e353), [Sponge](http://sponge2/74f0a66c-c356-4d06-9f52-85f92729e353) status: failed <details><summary>Test output</summary><br><pre>args = (session: "projects/python-docs-samples-tests/instances/test-instance-03175cbad1/databases/test-db-78e4b8da2e/sessions...s" params { } query_options { optimizer_version: "1" optimizer_statistics_package: "auto_20191128_14_47_22UTC" } ,) kwargs = {'metadata': [('google-cloud-resource-prefix', 'projects/python-docs-samples-tests/instances/test-instance-03175cbad1/...0j9Vuyc7VWPLqdjZfVQ'), ('x-goog-api-client', 'gl-python/3.6.13 grpc/1.38.1 gax/1.31.0 gccl/3.6.0')], 'timeout': 3600.0} result = <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.INVALID_ARGUMENT details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}" > prefetch_first = True @general_helpers.wraps(callable_) def error_remapped_callable(*args, **kwargs): try: result = callable_(*args, **kwargs) # Auto-fetching the first result causes PubSub client's streaming pull # to hang when re-opening the stream, thus we need examine the hacky # hidden flag to see if pre-fetching is disabled. # https://github.com/googleapis/python-pubsub/issues/93#issuecomment-630762257 prefetch_first = getattr(callable_, "_prefetch_first_result_", True) return _StreamingResponseIterator( > result, prefetch_first_result=prefetch_first ) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:161: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.api_core.grpc_helpers._StreamingResponseIterator object at 0x7f2e0595d710> wrapped = <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.INVALID_ARGUMENT details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}" > prefetch_first_result = True def __init__(self, wrapped, prefetch_first_result=True): self._wrapped = wrapped # This iterator is used in a retry context, and returned outside after init. # gRPC will not throw an exception until the stream is consumed, so we need # to retrieve the first result, in order to fail, in order to trigger a retry. try: if prefetch_first_result: > self._stored_first_result = six.next(self._wrapped) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:83: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.INVALID_ARGUMENT details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}" > def __next__(self): > return self._next() .nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py:426: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.INVALID_ARGUMENT details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}" > def _next(self): with self._state.condition: if self._state.code is None: event_handler = _event_handler(self._state, self._response_deserializer) self._state.due.add(cygrpc.OperationType.receive_message) operating = self._call.operate( (cygrpc.ReceiveMessageOperation(_EMPTY_FLAGS),), event_handler) if not operating: self._state.due.remove(cygrpc.OperationType.receive_message) elif self._state.code is grpc.StatusCode.OK: raise StopIteration() else: raise self def _response_ready(): return (self._state.response is not None or (cygrpc.OperationType.receive_message not in self._state.due and self._state.code is not None)) _common.wait(self._state.condition.wait, _response_ready) if self._state.response is not None: response = self._state.response self._state.response = None return response elif cygrpc.OperationType.receive_message not in self._state.due: if self._state.code is grpc.StatusCode.OK: raise StopIteration() elif self._state.code is not None: > raise self E grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with: E status = StatusCode.INVALID_ARGUMENT E details = "Requested optimizer statistics package not found: auto_20191128_14_47_22UTC." E debug_error_string = "{"created":"@1625909969.479153083","description":"Error received from peer ipv4:74.125.195.95:443","file":"src/core/lib/surface/call.cc","file_line":1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}" E > .nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py:826: _MultiThreadedRendezvous The above exception was the direct cause of the following exception: capsys = <_pytest.capture.CaptureFixture object at 0x7f2e0470a940> def test_create_client_with_query_options(capsys): > snippets.create_client_with_query_options(INSTANCE_ID, DATABASE_ID) snippets_test.py:417: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ snippets.py:1735: in create_client_with_query_options for row in results: ../../google/cloud/spanner_v1/streamed.py:149: in __iter__ self._consume_next() ../../google/cloud/spanner_v1/streamed.py:121: in _consume_next response = six.next(self._response_iterator) ../../google/cloud/spanner_v1/snapshot.py:59: in _restart_on_unavailable iterator = method(request=request) ../../google/cloud/spanner_v1/services/spanner/client.py:884: in execute_streaming_sql response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/gapic_v1/method.py:145: in __call__ return wrapped_func(*args, **kwargs) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/timeout.py:102: in func_with_timeout return func(*args, **kwargs) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:164: in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ value = None from_value = <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.INVALID_ARGUMENT details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}" > > ??? E google.api_core.exceptions.InvalidArgument: 400 Requested optimizer statistics package not found: auto_20191128_14_47_22UTC. <string>:3: InvalidArgument</pre></details>
1.0
samples.samples.snippets_test: test_create_client_with_query_options failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 7b8d8563651a9732631b22bef38cb2030eaadfa7 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/74f0a66c-c356-4d06-9f52-85f92729e353), [Sponge](http://sponge2/74f0a66c-c356-4d06-9f52-85f92729e353) status: failed <details><summary>Test output</summary><br><pre>args = (session: "projects/python-docs-samples-tests/instances/test-instance-03175cbad1/databases/test-db-78e4b8da2e/sessions...s" params { } query_options { optimizer_version: "1" optimizer_statistics_package: "auto_20191128_14_47_22UTC" } ,) kwargs = {'metadata': [('google-cloud-resource-prefix', 'projects/python-docs-samples-tests/instances/test-instance-03175cbad1/...0j9Vuyc7VWPLqdjZfVQ'), ('x-goog-api-client', 'gl-python/3.6.13 grpc/1.38.1 gax/1.31.0 gccl/3.6.0')], 'timeout': 3600.0} result = <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.INVALID_ARGUMENT details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}" > prefetch_first = True @general_helpers.wraps(callable_) def error_remapped_callable(*args, **kwargs): try: result = callable_(*args, **kwargs) # Auto-fetching the first result causes PubSub client's streaming pull # to hang when re-opening the stream, thus we need examine the hacky # hidden flag to see if pre-fetching is disabled. # https://github.com/googleapis/python-pubsub/issues/93#issuecomment-630762257 prefetch_first = getattr(callable_, "_prefetch_first_result_", True) return _StreamingResponseIterator( > result, prefetch_first_result=prefetch_first ) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:161: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.api_core.grpc_helpers._StreamingResponseIterator object at 0x7f2e0595d710> wrapped = <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.INVALID_ARGUMENT details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}" > prefetch_first_result = True def __init__(self, wrapped, prefetch_first_result=True): self._wrapped = wrapped # This iterator is used in a retry context, and returned outside after init. # gRPC will not throw an exception until the stream is consumed, so we need # to retrieve the first result, in order to fail, in order to trigger a retry. try: if prefetch_first_result: > self._stored_first_result = six.next(self._wrapped) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:83: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.INVALID_ARGUMENT details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}" > def __next__(self): > return self._next() .nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py:426: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.INVALID_ARGUMENT details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}" > def _next(self): with self._state.condition: if self._state.code is None: event_handler = _event_handler(self._state, self._response_deserializer) self._state.due.add(cygrpc.OperationType.receive_message) operating = self._call.operate( (cygrpc.ReceiveMessageOperation(_EMPTY_FLAGS),), event_handler) if not operating: self._state.due.remove(cygrpc.OperationType.receive_message) elif self._state.code is grpc.StatusCode.OK: raise StopIteration() else: raise self def _response_ready(): return (self._state.response is not None or (cygrpc.OperationType.receive_message not in self._state.due and self._state.code is not None)) _common.wait(self._state.condition.wait, _response_ready) if self._state.response is not None: response = self._state.response self._state.response = None return response elif cygrpc.OperationType.receive_message not in self._state.due: if self._state.code is grpc.StatusCode.OK: raise StopIteration() elif self._state.code is not None: > raise self E grpc._channel._MultiThreadedRendezvous: <_MultiThreadedRendezvous of RPC that terminated with: E status = StatusCode.INVALID_ARGUMENT E details = "Requested optimizer statistics package not found: auto_20191128_14_47_22UTC." E debug_error_string = "{"created":"@1625909969.479153083","description":"Error received from peer ipv4:74.125.195.95:443","file":"src/core/lib/surface/call.cc","file_line":1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}" E > .nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py:826: _MultiThreadedRendezvous The above exception was the direct cause of the following exception: capsys = <_pytest.capture.CaptureFixture object at 0x7f2e0470a940> def test_create_client_with_query_options(capsys): > snippets.create_client_with_query_options(INSTANCE_ID, DATABASE_ID) snippets_test.py:417: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ snippets.py:1735: in create_client_with_query_options for row in results: ../../google/cloud/spanner_v1/streamed.py:149: in __iter__ self._consume_next() ../../google/cloud/spanner_v1/streamed.py:121: in _consume_next response = six.next(self._response_iterator) ../../google/cloud/spanner_v1/snapshot.py:59: in _restart_on_unavailable iterator = method(request=request) ../../google/cloud/spanner_v1/services/spanner/client.py:884: in execute_streaming_sql response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/gapic_v1/method.py:145: in __call__ return wrapped_func(*args, **kwargs) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/timeout.py:102: in func_with_timeout return func(*args, **kwargs) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:164: in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ value = None from_value = <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.INVALID_ARGUMENT details = "Requested opt...:1066,"grpc_message":"Requested optimizer statistics package not found: auto_20191128_14_47_22UTC.","grpc_status":3}" > > ??? E google.api_core.exceptions.InvalidArgument: 400 Requested optimizer statistics package not found: auto_20191128_14_47_22UTC. <string>:3: InvalidArgument</pre></details>
process
samples samples snippets test test create client with query options failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output args session projects python docs samples tests instances test instance databases test db sessions s params query options optimizer version optimizer statistics package auto kwargs metadata timeout result multithreadedrendezvous of rpc that terminated with status statuscode invalid argument details requested opt grpc message requested optimizer statistics package not found auto grpc status prefetch first true general helpers wraps callable def error remapped callable args kwargs try result callable args kwargs auto fetching the first result causes pubsub client s streaming pull to hang when re opening the stream thus we need examine the hacky hidden flag to see if pre fetching is disabled prefetch first getattr callable prefetch first result true return streamingresponseiterator result prefetch first result prefetch first nox py lib site packages google api core grpc helpers py self wrapped multithreadedrendezvous of rpc that terminated with status statuscode invalid argument details requested opt grpc message requested optimizer statistics package not found auto grpc status prefetch first result true def init self wrapped prefetch first result true self wrapped wrapped this iterator is used in a retry context and returned outside after init grpc will not throw an exception until the stream is consumed so we need to retrieve the first result in order to fail in order to trigger a retry try if prefetch first result self stored first result six next self wrapped nox py lib site packages google api core grpc helpers py self multithreadedrendezvous of rpc that terminated with status statuscode invalid argument details requested opt grpc message requested optimizer statistics package not found auto grpc status def next self return self next nox py lib site packages grpc channel py self multithreadedrendezvous of rpc that terminated with status statuscode invalid argument details requested opt grpc message requested optimizer statistics package not found auto grpc status def next self with self state condition if self state code is none event handler event handler self state self response deserializer self state due add cygrpc operationtype receive message operating self call operate cygrpc receivemessageoperation empty flags event handler if not operating self state due remove cygrpc operationtype receive message elif self state code is grpc statuscode ok raise stopiteration else raise self def response ready return self state response is not none or cygrpc operationtype receive message not in self state due and self state code is not none common wait self state condition wait response ready if self state response is not none response self state response self state response none return response elif cygrpc operationtype receive message not in self state due if self state code is grpc statuscode ok raise stopiteration elif self state code is not none raise self e grpc channel multithreadedrendezvous multithreadedrendezvous of rpc that terminated with e status statuscode invalid argument e details requested optimizer statistics package not found auto e debug error string created description error received from peer file src core lib surface call cc file line grpc message requested optimizer statistics package not found auto grpc status e nox py lib site packages grpc channel py multithreadedrendezvous the above exception was the direct cause of the following exception capsys def test create client with query options capsys snippets create client with query options instance id database id snippets test py snippets py in create client with query options for row in results google cloud spanner streamed py in iter self consume next google cloud spanner streamed py in consume next response six next self response iterator google cloud spanner snapshot py in restart on unavailable iterator method request request google cloud spanner services spanner client py in execute streaming sql response rpc request retry retry timeout timeout metadata metadata nox py lib site packages google api core gapic method py in call return wrapped func args kwargs nox py lib site packages google api core timeout py in func with timeout return func args kwargs nox py lib site packages google api core grpc helpers py in error remapped callable six raise from exceptions from grpc error exc exc value none from value multithreadedrendezvous of rpc that terminated with status statuscode invalid argument details requested opt grpc message requested optimizer statistics package not found auto grpc status e google api core exceptions invalidargument requested optimizer statistics package not found auto invalidargument
1
5,154
7,933,323,593
IssuesEvent
2018-07-08 04:05:13
Great-Hill-Corporation/quickBlocks
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
closed
If a class is upgraded to a new schema, it does not save its data
libs-etherlib status-inprocess type-enhancement
Do this: 1. Save ${CACHE}/prices/poloniex.bin 2. remove ${CACHE}/prices/poloniex.bin 3. Run test for ethprice There will be a discrepancy in the schema from 0.2.0 to 0.3.0. This happens because even though we do upgrade the data on read, we don't re-write that data to hard drive. Each time we read, we should keep track of 'dirty' bit on the archive, and write the data if 'dirty' is true.
1.0
If a class is upgraded to a new schema, it does not save its data - Do this: 1. Save ${CACHE}/prices/poloniex.bin 2. remove ${CACHE}/prices/poloniex.bin 3. Run test for ethprice There will be a discrepancy in the schema from 0.2.0 to 0.3.0. This happens because even though we do upgrade the data on read, we don't re-write that data to hard drive. Each time we read, we should keep track of 'dirty' bit on the archive, and write the data if 'dirty' is true.
process
if a class is upgraded to a new schema it does not save its data do this save cache prices poloniex bin remove cache prices poloniex bin run test for ethprice there will be a discrepancy in the schema from to this happens because even though we do upgrade the data on read we don t re write that data to hard drive each time we read we should keep track of dirty bit on the archive and write the data if dirty is true
1
155,653
24,496,222,620
IssuesEvent
2022-10-10 08:54:56
anubhab1710/FLOSS-Graphics-Templates
https://api.github.com/repos/anubhab1710/FLOSS-Graphics-Templates
opened
Create Social Media Templates using GIMP
good first issue hacktoberfest design beginner-friendly
Use GIMP to design visually attractive Social Media Templates!
1.0
Create Social Media Templates using GIMP - Use GIMP to design visually attractive Social Media Templates!
non_process
create social media templates using gimp use gimp to design visually attractive social media templates
0
446,815
12,879,129,569
IssuesEvent
2020-07-11 20:19:30
clinwiki-org/clinwiki
https://api.github.com/repos/clinwiki-org/clinwiki
closed
restore navigation back to search results from study page
Priority 3
used to be able to get back to search results from within study page and maintain filter/sort - need this from anywhere on study pages
1.0
restore navigation back to search results from study page - used to be able to get back to search results from within study page and maintain filter/sort - need this from anywhere on study pages
non_process
restore navigation back to search results from study page used to be able to get back to search results from within study page and maintain filter sort need this from anywhere on study pages
0
318,293
27,297,075,801
IssuesEvent
2023-02-23 21:21:23
nucleus-security/Test-repo
https://api.github.com/repos/nucleus-security/Test-repo
closed
Nucleus - [High] - 440057
Test
Source: QUALYS Finding Description: CentOS has released security update for kernel to fix the vulnerabilities. Affected Products: centos 6 Impact: Successful exploitation allows attacker to compromise the system. Target(s): Asset name: 192.168.56.103 Solution: To resolve this issue, upgrade to the latest packages which contain a patch. Refer to CentOS advisory centos 6 (https://lists.centos.org/pipermail/centos-announce/2016-July/021977.html) for updates and patch information. Patch: Following are links for downloading patches to fix the vulnerabilities: CESA-2016:1406: centos 6 (https://lists.centos.org/pipermail/centos-announce/2016-July/021977.html) References: QID:440057 CVE:CVE-2016-4565 Category:CentOS PCI Flagged:yes Vendor References:CESA-2016:1406 centos 6 Bugtraq IDs:90301 Severity: High Date Discovered: 2022-11-12 08:04:44 Nucleus Notification Rules Triggered: Rule GitHub Project Name: 6716 Please see Nucleus for more information on these vulnerabilities:https://192.168.56.101/nucleus/public/app/index.html#vuln/201000007/NDQwMDU3/UVVBTFlT/VnVsbg--/false/MjAxMDAwMDA3/c3VtbWFyeQ--/false
1.0
Nucleus - [High] - 440057 - Source: QUALYS Finding Description: CentOS has released security update for kernel to fix the vulnerabilities. Affected Products: centos 6 Impact: Successful exploitation allows attacker to compromise the system. Target(s): Asset name: 192.168.56.103 Solution: To resolve this issue, upgrade to the latest packages which contain a patch. Refer to CentOS advisory centos 6 (https://lists.centos.org/pipermail/centos-announce/2016-July/021977.html) for updates and patch information. Patch: Following are links for downloading patches to fix the vulnerabilities: CESA-2016:1406: centos 6 (https://lists.centos.org/pipermail/centos-announce/2016-July/021977.html) References: QID:440057 CVE:CVE-2016-4565 Category:CentOS PCI Flagged:yes Vendor References:CESA-2016:1406 centos 6 Bugtraq IDs:90301 Severity: High Date Discovered: 2022-11-12 08:04:44 Nucleus Notification Rules Triggered: Rule GitHub Project Name: 6716 Please see Nucleus for more information on these vulnerabilities:https://192.168.56.101/nucleus/public/app/index.html#vuln/201000007/NDQwMDU3/UVVBTFlT/VnVsbg--/false/MjAxMDAwMDA3/c3VtbWFyeQ--/false
non_process
nucleus source qualys finding description centos has released security update for kernel to fix the vulnerabilities affected products centos impact successful exploitation allows attacker to compromise the system target s asset name solution to resolve this issue upgrade to the latest packages which contain a patch refer to centos advisory centos for updates and patch information patch following are links for downloading patches to fix the vulnerabilities cesa centos references qid cve cve category centos pci flagged yes vendor references cesa centos bugtraq ids severity high date discovered nucleus notification rules triggered rule github project name please see nucleus for more information on these vulnerabilities
0
19,575
25,895,131,180
IssuesEvent
2022-12-14 21:39:30
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
DISABLED test_success_first_then_exception (__main__.SpawnTest)
high priority triage review module: multiprocessing triaged module: flaky-tests skipped
Platforms: linux This test was disabled because it is failing on trunk. See [recent examples](http://torch-ci.com/failure/test_success_first_then_exception%2C%20SpawnTest) and the most recent [workflow logs](https://github.com/pytorch/pytorch/actions/runs/1820236472). Over the past 6 hours, it has been determined flaky in 1 workflow(s) with 1 red and 3 green. cc @ezyang @gchanan @zou3519 @VitalyFedyunin
1.0
DISABLED test_success_first_then_exception (__main__.SpawnTest) - Platforms: linux This test was disabled because it is failing on trunk. See [recent examples](http://torch-ci.com/failure/test_success_first_then_exception%2C%20SpawnTest) and the most recent [workflow logs](https://github.com/pytorch/pytorch/actions/runs/1820236472). Over the past 6 hours, it has been determined flaky in 1 workflow(s) with 1 red and 3 green. cc @ezyang @gchanan @zou3519 @VitalyFedyunin
process
disabled test success first then exception main spawntest platforms linux this test was disabled because it is failing on trunk see and the most recent over the past hours it has been determined flaky in workflow s with red and green cc ezyang gchanan vitalyfedyunin
1
298,271
22,473,595,190
IssuesEvent
2022-06-22 10:13:29
ices-tools-dev/RDBES
https://api.github.com/repos/ices-tools-dev/RDBES
reopened
out of frame ports in a strata
Documentation selectionMethod Estimation issue Need plenary
Hello, I have read in the documentation that out of frame strata should not be declared. And I have doubts about how to report my data. I will try to explain the situation. In my sampling design, I have considered different fleets as strata (each of them defined with non-overlapping vessel lists). Some strata (fleets) are out of my sampling frame. They are never sampled. I understand that these are not reported in the RDBES. No problems with this. But in every sampled strata, I have some ports which are sampled, and some other ports that are not included in my sampling frame (due to practical constrains) and therefore will never be sampled. I thought that these should be reported with sampling = 0, because they will be need in order to know the total PSU in the strata. But somehow I also think that it would be needed to indicate that they are out of the sampling frame and therefore their sampling probability is zero. Shall include information of these ports out of my sampling frame? Thank you!
1.0
out of frame ports in a strata - Hello, I have read in the documentation that out of frame strata should not be declared. And I have doubts about how to report my data. I will try to explain the situation. In my sampling design, I have considered different fleets as strata (each of them defined with non-overlapping vessel lists). Some strata (fleets) are out of my sampling frame. They are never sampled. I understand that these are not reported in the RDBES. No problems with this. But in every sampled strata, I have some ports which are sampled, and some other ports that are not included in my sampling frame (due to practical constrains) and therefore will never be sampled. I thought that these should be reported with sampling = 0, because they will be need in order to know the total PSU in the strata. But somehow I also think that it would be needed to indicate that they are out of the sampling frame and therefore their sampling probability is zero. Shall include information of these ports out of my sampling frame? Thank you!
non_process
out of frame ports in a strata hello i have read in the documentation that out of frame strata should not be declared and i have doubts about how to report my data i will try to explain the situation in my sampling design i have considered different fleets as strata each of them defined with non overlapping vessel lists some strata fleets are out of my sampling frame they are never sampled i understand that these are not reported in the rdbes no problems with this but in every sampled strata i have some ports which are sampled and some other ports that are not included in my sampling frame due to practical constrains and therefore will never be sampled i thought that these should be reported with sampling because they will be need in order to know the total psu in the strata but somehow i also think that it would be needed to indicate that they are out of the sampling frame and therefore their sampling probability is zero shall include information of these ports out of my sampling frame thank you
0
14,472
9,321,790,968
IssuesEvent
2019-03-27 05:45:17
danielwii/react-exercises
https://api.github.com/repos/danielwii/react-exercises
opened
CVE-2018-1000620 High Severity Vulnerability detected by WhiteSource
security vulnerability
## CVE-2018-1000620 - High Severity Vulnerability <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cryptiles-2.0.5.tgz</b></p></summary> <p>General purpose crypto utilities</p> <p>path: null</p> <p> <p>Library home page: <a href=http://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz>http://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz</a></p> Dependency Hierarchy: - react-scripts-0.8.4.tgz (Root Library) - fsevents-1.0.14.tgz - node-pre-gyp-0.6.29.tgz - request-2.73.0.tgz - hawk-3.1.3.tgz - :x: **cryptiles-2.0.5.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2. <p>Publish Date: 2018-07-09 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000620>CVE-2018-1000620</a></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-1000620 High Severity Vulnerability detected by WhiteSource - ## CVE-2018-1000620 - High Severity Vulnerability <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cryptiles-2.0.5.tgz</b></p></summary> <p>General purpose crypto utilities</p> <p>path: null</p> <p> <p>Library home page: <a href=http://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz>http://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz</a></p> Dependency Hierarchy: - react-scripts-0.8.4.tgz (Root Library) - fsevents-1.0.14.tgz - node-pre-gyp-0.6.29.tgz - request-2.73.0.tgz - hawk-3.1.3.tgz - :x: **cryptiles-2.0.5.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2. <p>Publish Date: 2018-07-09 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000620>CVE-2018-1000620</a></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high severity vulnerability detected by whitesource cve high severity vulnerability vulnerable library cryptiles tgz general purpose crypto utilities path null library home page a href dependency hierarchy react scripts tgz root library fsevents tgz node pre gyp tgz request tgz hawk tgz x cryptiles tgz vulnerable library vulnerability details eran hammer cryptiles version earlier contains a cwe insufficient entropy vulnerability in randomdigits method that can result in an attacker is more likely to be able to brute force something that was supposed to be random this attack appear to be exploitable via depends upon the calling application this vulnerability appears to have been fixed in publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href step up your open source security game with whitesource
0
9,685
12,685,673,065
IssuesEvent
2020-06-20 06:05:33
ClickHouse/ClickHouse
https://api.github.com/repos/ClickHouse/ClickHouse
closed
Master: pipeline stuck
bug comp-processors prio-major v20.5-affected
``` CREATE TABLE pipeline_stuck ( `id` String, `val` SimpleAggregateFunction(anyLast, String), INDEX ind val TYPE bloom_filter() GRANULARITY 1 ) ENGINE = AggregatingMergeTree() ORDER BY id ; insert into pipeline_stuck select toString(number%100), toString(number) from numbers(10000000); select * from pipeline_stuck final limit 10; select * from pipeline_stuck final where val = '9999900'; alter table pipeline_stuck drop index ind; select * from pipeline_stuck final where val = '9999900'; ```
1.0
Master: pipeline stuck - ``` CREATE TABLE pipeline_stuck ( `id` String, `val` SimpleAggregateFunction(anyLast, String), INDEX ind val TYPE bloom_filter() GRANULARITY 1 ) ENGINE = AggregatingMergeTree() ORDER BY id ; insert into pipeline_stuck select toString(number%100), toString(number) from numbers(10000000); select * from pipeline_stuck final limit 10; select * from pipeline_stuck final where val = '9999900'; alter table pipeline_stuck drop index ind; select * from pipeline_stuck final where val = '9999900'; ```
process
master pipeline stuck create table pipeline stuck id string val simpleaggregatefunction anylast string index ind val type bloom filter granularity engine aggregatingmergetree order by id insert into pipeline stuck select tostring number tostring number from numbers select from pipeline stuck final limit select from pipeline stuck final where val alter table pipeline stuck drop index ind select from pipeline stuck final where val
1
20,130
26,666,079,141
IssuesEvent
2023-01-26 04:02:31
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Flaky test TestProcessorConsumeMetricsErrors
bug processor/spanmetrics flaky test
### Component(s) processor/spanmetrics ### What happened? See https://github.com/open-telemetry/opentelemetry-collector-contrib/actions/runs/3993460821/jobs/6850258904 ``` === RUN TestProcessorConsumeMetricsErrors processor_test.go:297: Error Trace: /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/spanmetricsprocessor/processor_test.go:297 Error: Should NOT be empty, but was [] Test: TestProcessorConsumeMetricsErrors --- FAIL: TestProcessorConsumeMetricsErrors (0.01s) ``` ### Collector version 0.70.0 ### Environment information _No response_ ### OpenTelemetry Collector configuration _No response_ ### Log output _No response_ ### Additional context _No response_
1.0
Flaky test TestProcessorConsumeMetricsErrors - ### Component(s) processor/spanmetrics ### What happened? See https://github.com/open-telemetry/opentelemetry-collector-contrib/actions/runs/3993460821/jobs/6850258904 ``` === RUN TestProcessorConsumeMetricsErrors processor_test.go:297: Error Trace: /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/spanmetricsprocessor/processor_test.go:297 Error: Should NOT be empty, but was [] Test: TestProcessorConsumeMetricsErrors --- FAIL: TestProcessorConsumeMetricsErrors (0.01s) ``` ### Collector version 0.70.0 ### Environment information _No response_ ### OpenTelemetry Collector configuration _No response_ ### Log output _No response_ ### Additional context _No response_
process
flaky test testprocessorconsumemetricserrors component s processor spanmetrics what happened see run testprocessorconsumemetricserrors processor test go error trace home runner work opentelemetry collector contrib opentelemetry collector contrib processor spanmetricsprocessor processor test go error should not be empty but was test testprocessorconsumemetricserrors fail testprocessorconsumemetricserrors collector version environment information no response opentelemetry collector configuration no response log output no response additional context no response
1
23,347
6,419,936,737
IssuesEvent
2017-08-08 22:29:49
xhqiao89/HydroDesktop_test
https://api.github.com/repos/xhqiao89/HydroDesktop_test
closed
Add option 'save WaterML files' to search download settings
CodePlex
<b>jirikadlec2[CodePlex]</b> <br />It seems that HydroDesktop is saving all the GetValues WaterML files to the C:\users\user name\appdata\local\temp\HydroDesktop folder. This uses large hard drive space on the user's computer. For example I found that the folder size with the WaterML files was 2 Gigabytes after I downloaded all data from the 'Little Bear River experimental watershed' web service. Improvement suggestion: Add a 'save WaterML files' checkbox to the search download settings. If the checkbox is unchecked then don't save the downloaded WaterML files to the temporary folder. Show the name of the folder where the WaterML files are saved (usually it's C:\users\user name\appdata\local\temp\HydroDesktop)
1.0
Add option 'save WaterML files' to search download settings - <b>jirikadlec2[CodePlex]</b> <br />It seems that HydroDesktop is saving all the GetValues WaterML files to the C:\users\user name\appdata\local\temp\HydroDesktop folder. This uses large hard drive space on the user's computer. For example I found that the folder size with the WaterML files was 2 Gigabytes after I downloaded all data from the 'Little Bear River experimental watershed' web service. Improvement suggestion: Add a 'save WaterML files' checkbox to the search download settings. If the checkbox is unchecked then don't save the downloaded WaterML files to the temporary folder. Show the name of the folder where the WaterML files are saved (usually it's C:\users\user name\appdata\local\temp\HydroDesktop)
non_process
add option save waterml files to search download settings it seems that hydrodesktop is saving all the getvalues waterml files to the c users user name appdata local temp hydrodesktop folder this uses large hard drive space on the user s computer for example i found that the folder size with the waterml files was gigabytes after i downloaded all data from the little bear river experimental watershed web service improvement suggestion add a save waterml files checkbox to the search download settings if the checkbox is unchecked then don t save the downloaded waterml files to the temporary folder show the name of the folder where the waterml files are saved usually it s c users user name appdata local temp hydrodesktop
0
554,111
16,389,382,480
IssuesEvent
2021-05-17 14:23:27
Bibliome/alvisnlp
https://api.github.com/repos/Bibliome/alvisnlp
closed
Function after bug
Priority: High Status: Done Topic: Factory Type: Bug
Function `after:LAYER` should return annotations that start at the same offset than the context annotation. Currently `after` skips the adjacent annotation after the context.
1.0
Function after bug - Function `after:LAYER` should return annotations that start at the same offset than the context annotation. Currently `after` skips the adjacent annotation after the context.
non_process
function after bug function after layer should return annotations that start at the same offset than the context annotation currently after skips the adjacent annotation after the context
0
747,202
26,077,171,352
IssuesEvent
2022-12-24 18:21:32
pystardust/ani-cli
https://api.github.com/repos/pystardust/ani-cli
opened
Episodes not released yet!
type: bug priority 2: medium
Version: 3.4.7 OS: Windows 10 Shell: git bash Anime: all i searched for every time i search for an anime and select the title it gives the error "Episodes not released yet!" **Steps To Reproduce** 1. Run `ani-cli ` 2. search for an anime 3. Choose any one **Expected behavior** it should ask for the episode number **Screenshots (if applicable; you can just drag the image onto github)** ![image](https://user-images.githubusercontent.com/80813371/209447377-2c8d283a-3227-4468-b984-b96999b3afb2.png)
1.0
Episodes not released yet! - Version: 3.4.7 OS: Windows 10 Shell: git bash Anime: all i searched for every time i search for an anime and select the title it gives the error "Episodes not released yet!" **Steps To Reproduce** 1. Run `ani-cli ` 2. search for an anime 3. Choose any one **Expected behavior** it should ask for the episode number **Screenshots (if applicable; you can just drag the image onto github)** ![image](https://user-images.githubusercontent.com/80813371/209447377-2c8d283a-3227-4468-b984-b96999b3afb2.png)
non_process
episodes not released yet version os windows shell git bash anime all i searched for every time i search for an anime and select the title it gives the error episodes not released yet steps to reproduce run ani cli search for an anime choose any one expected behavior it should ask for the episode number screenshots if applicable you can just drag the image onto github
0
8,454
11,628,002,550
IssuesEvent
2020-02-27 17:29:32
qri-io/qri
https://api.github.com/repos/qri-io/qri
closed
executable bit not set on binary releases, date is wrong as well
bug release process
https://github.com/qri-io/qri/releases/download/v0.9.4/qri_linux_amd64.zip Executable bit is not set. Modified date is 12/31/1979. I did not check any other architectures.
1.0
executable bit not set on binary releases, date is wrong as well - https://github.com/qri-io/qri/releases/download/v0.9.4/qri_linux_amd64.zip Executable bit is not set. Modified date is 12/31/1979. I did not check any other architectures.
process
executable bit not set on binary releases date is wrong as well executable bit is not set modified date is i did not check any other architectures
1
8,635
6,599,860,081
IssuesEvent
2017-09-17 02:33:23
thrust/thrust
https://api.github.com/repos/thrust/thrust
closed
slow compilation with CUDA 6.5
performance bug wontfix
I just switched from CUDA 6 to CUDA 6.5 and noticed that compilation of my thrust based code is slowed down **a lot**. Because my own, more complex code did not yet finish compiling I created a simple demo program: https://gist.github.com/dachziegel/c7bc0226fe2f16196f3e When I use the thrust version which is delivered with CUDA 6.5 this takes around 6 seconds: ``` $ time nvcc -arch sm_30 compile_time.cu real 0m6.469s user 0m6.010s sys 0m0.452s ``` When I use the thrust development version from github it takes around 30 (!) seconds: ``` $time nvcc -arch sm_30 -I/tmp/thrust-github/ compile_time.cu real 0m31.051s user 0m30.237s sys 0m0.782s ``` The difference is even higher if the program is a bit more complex, e.g. this one from another bug report: https://gist.github.com/dachziegel/11288269 when using CUDA 6.5: thrust from CUDA 6.5 takes around **7 seconds**, thrust development version from github took **7.44 minutes**(!). when using CUDA 6: thrust development version from github takes **51 seconds**. when using CUDA 7: thrust development version from github takes **41 seconds**. My system is a 64bit Ubuntu 14.04.
True
slow compilation with CUDA 6.5 - I just switched from CUDA 6 to CUDA 6.5 and noticed that compilation of my thrust based code is slowed down **a lot**. Because my own, more complex code did not yet finish compiling I created a simple demo program: https://gist.github.com/dachziegel/c7bc0226fe2f16196f3e When I use the thrust version which is delivered with CUDA 6.5 this takes around 6 seconds: ``` $ time nvcc -arch sm_30 compile_time.cu real 0m6.469s user 0m6.010s sys 0m0.452s ``` When I use the thrust development version from github it takes around 30 (!) seconds: ``` $time nvcc -arch sm_30 -I/tmp/thrust-github/ compile_time.cu real 0m31.051s user 0m30.237s sys 0m0.782s ``` The difference is even higher if the program is a bit more complex, e.g. this one from another bug report: https://gist.github.com/dachziegel/11288269 when using CUDA 6.5: thrust from CUDA 6.5 takes around **7 seconds**, thrust development version from github took **7.44 minutes**(!). when using CUDA 6: thrust development version from github takes **51 seconds**. when using CUDA 7: thrust development version from github takes **41 seconds**. My system is a 64bit Ubuntu 14.04.
non_process
slow compilation with cuda i just switched from cuda to cuda and noticed that compilation of my thrust based code is slowed down a lot because my own more complex code did not yet finish compiling i created a simple demo program when i use the thrust version which is delivered with cuda this takes around seconds time nvcc arch sm compile time cu real user sys when i use the thrust development version from github it takes around seconds time nvcc arch sm i tmp thrust github compile time cu real user sys the difference is even higher if the program is a bit more complex e g this one from another bug report when using cuda thrust from cuda takes around seconds thrust development version from github took minutes when using cuda thrust development version from github takes seconds when using cuda thrust development version from github takes seconds my system is a ubuntu
0
14,058
16,870,223,295
IssuesEvent
2021-06-22 02:49:17
Leviatan-Analytics/LA-data-processing
https://api.github.com/repos/Leviatan-Analytics/LA-data-processing
closed
Test YoloV3 model [2]
Data Processing Sprint 2 Week 4
Estimated time: 1 hs per assignee Test the accuracy of the trained model with different inputs. Output: Document with the result of the different inputs test.
1.0
Test YoloV3 model [2] - Estimated time: 1 hs per assignee Test the accuracy of the trained model with different inputs. Output: Document with the result of the different inputs test.
process
test model estimated time hs per assignee test the accuracy of the trained model with different inputs output document with the result of the different inputs test
1
19,046
25,048,270,715
IssuesEvent
2022-11-05 14:51:54
fluent/fluent-bit
https://api.github.com/repos/fluent/fluent-bit
closed
Changes in v2.0.0+ break static compilation on armhf
work-in-process
## Bug Report **Describe the bug** Changes in v2.0.0+ break static compilation on `armhf`. I have been building statically-compiled Fluent-Bit on the 4 supported platforms (i386, amd64, armhf, arm64) for a while now in [this project](https://github.com/brian-maloney/fluent-bit-static). I use Docker with a Debian bullseye base to build my binaries. With the release of v2.0.0 all my builds broke. After some research, I tracked down that changes in `fb_hash` to directly use OpenSSL instead of Monkey required me to add `libssl-dev` to my build image, which caused further downstream issues requiring me to then disable WASM and LUAJIT. All of these are acceptable for my use case, but while I was able to solve the issues in most general cases, I am stuck on a linking issue with `armhf`. **To Reproduce** - Link error: ``` /usr/bin/ld: /usr/lib/gcc/arm-linux-gnueabihf/10/../../../arm-linux-gnueabihf/libc.a(getcontext.o): in function `__getcontext': (.text+0x74): undefined reference to `_rtld_global_ro' /usr/bin/ld: /usr/lib/gcc/arm-linux-gnueabihf/10/../../../arm-linux-gnueabihf/libc.a(setcontext.o): in function `__startcontext': (.text+0x80): undefined reference to `_rtld_global_ro' ``` - Steps to reproduce the problem: ``` git clone https://github.com/brian-maloney/fluent-bit-static.git cd fluent-bit-static docker build --progress plain --build-arg "TAG=v2.0.0" --build-arg "BUILDARCH=armhf" --platform 'linux/arm/v7' -t fluent-bit-static . ``` **Expected behavior** Successful build and link **Your Environment** <!--- Include as many relevant details about the environment you experienced the bug in --> * Version used: 2.0.0 * Configuration: Build in Docker **Additional context** <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> I understand this isn't a directly supported use case, and I'm looking for some unusual assistance, but I'd love to avoid having to drop `armhf` support on these builds that I supply to the community.
1.0
Changes in v2.0.0+ break static compilation on armhf - ## Bug Report **Describe the bug** Changes in v2.0.0+ break static compilation on `armhf`. I have been building statically-compiled Fluent-Bit on the 4 supported platforms (i386, amd64, armhf, arm64) for a while now in [this project](https://github.com/brian-maloney/fluent-bit-static). I use Docker with a Debian bullseye base to build my binaries. With the release of v2.0.0 all my builds broke. After some research, I tracked down that changes in `fb_hash` to directly use OpenSSL instead of Monkey required me to add `libssl-dev` to my build image, which caused further downstream issues requiring me to then disable WASM and LUAJIT. All of these are acceptable for my use case, but while I was able to solve the issues in most general cases, I am stuck on a linking issue with `armhf`. **To Reproduce** - Link error: ``` /usr/bin/ld: /usr/lib/gcc/arm-linux-gnueabihf/10/../../../arm-linux-gnueabihf/libc.a(getcontext.o): in function `__getcontext': (.text+0x74): undefined reference to `_rtld_global_ro' /usr/bin/ld: /usr/lib/gcc/arm-linux-gnueabihf/10/../../../arm-linux-gnueabihf/libc.a(setcontext.o): in function `__startcontext': (.text+0x80): undefined reference to `_rtld_global_ro' ``` - Steps to reproduce the problem: ``` git clone https://github.com/brian-maloney/fluent-bit-static.git cd fluent-bit-static docker build --progress plain --build-arg "TAG=v2.0.0" --build-arg "BUILDARCH=armhf" --platform 'linux/arm/v7' -t fluent-bit-static . ``` **Expected behavior** Successful build and link **Your Environment** <!--- Include as many relevant details about the environment you experienced the bug in --> * Version used: 2.0.0 * Configuration: Build in Docker **Additional context** <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> I understand this isn't a directly supported use case, and I'm looking for some unusual assistance, but I'd love to avoid having to drop `armhf` support on these builds that I supply to the community.
process
changes in break static compilation on armhf bug report describe the bug changes in break static compilation on armhf i have been building statically compiled fluent bit on the supported platforms armhf for a while now in i use docker with a debian bullseye base to build my binaries with the release of all my builds broke after some research i tracked down that changes in fb hash to directly use openssl instead of monkey required me to add libssl dev to my build image which caused further downstream issues requiring me to then disable wasm and luajit all of these are acceptable for my use case but while i was able to solve the issues in most general cases i am stuck on a linking issue with armhf to reproduce link error usr bin ld usr lib gcc arm linux gnueabihf arm linux gnueabihf libc a getcontext o in function getcontext text undefined reference to rtld global ro usr bin ld usr lib gcc arm linux gnueabihf arm linux gnueabihf libc a setcontext o in function startcontext text undefined reference to rtld global ro steps to reproduce the problem git clone cd fluent bit static docker build progress plain build arg tag build arg buildarch armhf platform linux arm t fluent bit static expected behavior successful build and link your environment version used configuration build in docker additional context i understand this isn t a directly supported use case and i m looking for some unusual assistance but i d love to avoid having to drop armhf support on these builds that i supply to the community
1
6,729
9,842,157,502
IssuesEvent
2019-06-18 08:40:35
syndesisio/syndesis
https://api.github.com/repos/syndesisio/syndesis
opened
Continuous Integration improvement: Improve Elephant carpaccio
cat/process cat/research
We should find ways to cut features in smaller end-to-end deliverables. It would allow us to integrate our software faster and recover feedback sooner.
1.0
Continuous Integration improvement: Improve Elephant carpaccio - We should find ways to cut features in smaller end-to-end deliverables. It would allow us to integrate our software faster and recover feedback sooner.
process
continuous integration improvement improve elephant carpaccio we should find ways to cut features in smaller end to end deliverables it would allow us to integrate our software faster and recover feedback sooner
1
18,251
24,334,011,045
IssuesEvent
2022-09-30 23:08:18
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Log flooded with `string to non-string with op EQ`
bug priority:p2 processor/transform
### What happened? ## Description Upgraded to 0.60.0. Using a transform processor with a clause such as: ``` transform: metrics: queries: - set(attributes["foo"], "bar") where attributes["foo"] == "foobar" ``` ## Steps to Reproduce Set up a pipeline using a transform processor that uses an eq operator. ## Expected Result The pipeline should perform correctly. ## Actual Result Logs are flooded with: ``` string to non-string with op EQ string to non-string with op EQ string to non-string with op EQ string to non-string with op EQ string to non-string with op EQ string to non-string with op EQ string to non-string with op EQ string to non-string with op EQ ``` ### Collector version 0.60.0 ### Environment information ## Environment OS: (e.g., "Ubuntu 20.04") Compiler(if manually compiled): (e.g., "go 14.2") ### OpenTelemetry Collector configuration ```yaml receivers: splunk_hec: exporters: splunk_hec/metrics: # Splunk HTTP Event Collector token. token: "00000000-0000-0000-0000-0000000000000" # URL to a Splunk instance to send data to. endpoint: "https://splunk:8088/services/collector" # Optional Splunk source: https://docs.splunk.com/Splexicon:Source source: "app" # Optional Splunk source type: https://docs.splunk.com/Splexicon:Sourcetype sourcetype: "telegraf" # Splunk index, optional name of the Splunk index targeted. index: "metrics" # Maximum HTTP connections to use simultaneously when sending data. Defaults to 100. max_connections: 20 # Whether to disable gzip compression over HTTP. Defaults to false. disable_compression: false # HTTP timeout when sending data. Defaults to 10s. timeout: 10s # Whether to skip checking the certificate of the HEC endpoint when sending data over HTTPS. Defaults to false. # For this demo, we use a self-signed certificate on the Splunk docker instance, so this flag is set to true. insecure_skip_verify: true processors: batch: transform: metrics: queries: - set(attributes["foo"], "bar") where attributes["foo"] == "foobar" extensions: health_check: endpoint: 0.0.0.0:13133 pprof: endpoint: :1888 zpages: endpoint: :55679 service: telemetry: logs: level: "debug" extensions: [pprof, zpages, health_check] pipelines: metrics: receivers: [splunk_hec] processors: [batch, transform] exporters: [splunk_hec/metrics] ``` ### Log output _No response_ ### Additional context _No response_
1.0
Log flooded with `string to non-string with op EQ` - ### What happened? ## Description Upgraded to 0.60.0. Using a transform processor with a clause such as: ``` transform: metrics: queries: - set(attributes["foo"], "bar") where attributes["foo"] == "foobar" ``` ## Steps to Reproduce Set up a pipeline using a transform processor that uses an eq operator. ## Expected Result The pipeline should perform correctly. ## Actual Result Logs are flooded with: ``` string to non-string with op EQ string to non-string with op EQ string to non-string with op EQ string to non-string with op EQ string to non-string with op EQ string to non-string with op EQ string to non-string with op EQ string to non-string with op EQ ``` ### Collector version 0.60.0 ### Environment information ## Environment OS: (e.g., "Ubuntu 20.04") Compiler(if manually compiled): (e.g., "go 14.2") ### OpenTelemetry Collector configuration ```yaml receivers: splunk_hec: exporters: splunk_hec/metrics: # Splunk HTTP Event Collector token. token: "00000000-0000-0000-0000-0000000000000" # URL to a Splunk instance to send data to. endpoint: "https://splunk:8088/services/collector" # Optional Splunk source: https://docs.splunk.com/Splexicon:Source source: "app" # Optional Splunk source type: https://docs.splunk.com/Splexicon:Sourcetype sourcetype: "telegraf" # Splunk index, optional name of the Splunk index targeted. index: "metrics" # Maximum HTTP connections to use simultaneously when sending data. Defaults to 100. max_connections: 20 # Whether to disable gzip compression over HTTP. Defaults to false. disable_compression: false # HTTP timeout when sending data. Defaults to 10s. timeout: 10s # Whether to skip checking the certificate of the HEC endpoint when sending data over HTTPS. Defaults to false. # For this demo, we use a self-signed certificate on the Splunk docker instance, so this flag is set to true. insecure_skip_verify: true processors: batch: transform: metrics: queries: - set(attributes["foo"], "bar") where attributes["foo"] == "foobar" extensions: health_check: endpoint: 0.0.0.0:13133 pprof: endpoint: :1888 zpages: endpoint: :55679 service: telemetry: logs: level: "debug" extensions: [pprof, zpages, health_check] pipelines: metrics: receivers: [splunk_hec] processors: [batch, transform] exporters: [splunk_hec/metrics] ``` ### Log output _No response_ ### Additional context _No response_
process
log flooded with string to non string with op eq what happened description upgraded to using a transform processor with a clause such as transform metrics queries set attributes bar where attributes foobar steps to reproduce set up a pipeline using a transform processor that uses an eq operator expected result the pipeline should perform correctly actual result logs are flooded with string to non string with op eq string to non string with op eq string to non string with op eq string to non string with op eq string to non string with op eq string to non string with op eq string to non string with op eq string to non string with op eq collector version environment information environment os e g ubuntu compiler if manually compiled e g go opentelemetry collector configuration yaml receivers splunk hec exporters splunk hec metrics splunk http event collector token token url to a splunk instance to send data to endpoint optional splunk source source app optional splunk source type sourcetype telegraf splunk index optional name of the splunk index targeted index metrics maximum http connections to use simultaneously when sending data defaults to max connections whether to disable gzip compression over http defaults to false disable compression false http timeout when sending data defaults to timeout whether to skip checking the certificate of the hec endpoint when sending data over https defaults to false for this demo we use a self signed certificate on the splunk docker instance so this flag is set to true insecure skip verify true processors batch transform metrics queries set attributes bar where attributes foobar extensions health check endpoint pprof endpoint zpages endpoint service telemetry logs level debug extensions pipelines metrics receivers processors exporters log output no response additional context no response
1
1,230
3,769,103,470
IssuesEvent
2016-03-16 09:16:19
e-government-ua/i
https://api.github.com/repos/e-government-ua/i
closed
Поправить футер
active In process of testing test _central-js
1. под последней услугой больше интервал. 2. левый край услуг должен быть по одной линии с левым краем конктента в футере.
1.0
Поправить футер - 1. под последней услугой больше интервал. 2. левый край услуг должен быть по одной линии с левым краем конктента в футере.
process
поправить футер под последней услугой больше интервал левый край услуг должен быть по одной линии с левым краем конктента в футере
1
16,800
5,290,798,002
IssuesEvent
2017-02-08 20:49:58
dotnet/coreclr
https://api.github.com/repos/dotnet/coreclr
opened
Optimize default(T) == null at compile time
area-CodeGen optimization
`default(T) == null` in generic code is a popular* pattern to determine if `T` is a nullable or a reference type. Example: ```csharp using System; using System.Runtime.CompilerServices; class Program { [MethodImpl(MethodImplOptions.NoInlining)] static bool IsNotNullableValueType<T>() { return default(T) != null; } private static void Main(string[] args) { Console.WriteLine(IsNotNullableValueType<string>()); Console.WriteLine(IsNotNullableValueType<int>()); Console.WriteLine(IsNotNullableValueType<int?>()); } } ``` Will print `False True False`. While the codegen for the "it's a reference type" case is pretty efficient: ```nasm xor eax,eax test rax,rax setb al movzx eax,al ret ``` For the "it's a valuetype" and "it's a nullable" case we get a useless allocation: ```nasm sub rsp,28h call __NewHelper_System_Private_CoreLib_System_Int32 xor edx,edx mov dword ptr [rax+8],edx test rax,rax seta al movzx eax,al add rsp,28h ret ``` The places that use this pattern would likely benefit from being able to statically evaluate this because it could lead to elimination of entire branches and smaller code (aside from removing a useless allocation). \* There are 17 hits for the `default\(.*\) == null` regular expression in the CoreFX repo and a couple more for the "not equal to null" case.
1.0
Optimize default(T) == null at compile time - `default(T) == null` in generic code is a popular* pattern to determine if `T` is a nullable or a reference type. Example: ```csharp using System; using System.Runtime.CompilerServices; class Program { [MethodImpl(MethodImplOptions.NoInlining)] static bool IsNotNullableValueType<T>() { return default(T) != null; } private static void Main(string[] args) { Console.WriteLine(IsNotNullableValueType<string>()); Console.WriteLine(IsNotNullableValueType<int>()); Console.WriteLine(IsNotNullableValueType<int?>()); } } ``` Will print `False True False`. While the codegen for the "it's a reference type" case is pretty efficient: ```nasm xor eax,eax test rax,rax setb al movzx eax,al ret ``` For the "it's a valuetype" and "it's a nullable" case we get a useless allocation: ```nasm sub rsp,28h call __NewHelper_System_Private_CoreLib_System_Int32 xor edx,edx mov dword ptr [rax+8],edx test rax,rax seta al movzx eax,al add rsp,28h ret ``` The places that use this pattern would likely benefit from being able to statically evaluate this because it could lead to elimination of entire branches and smaller code (aside from removing a useless allocation). \* There are 17 hits for the `default\(.*\) == null` regular expression in the CoreFX repo and a couple more for the "not equal to null" case.
non_process
optimize default t null at compile time default t null in generic code is a popular pattern to determine if t is a nullable or a reference type example csharp using system using system runtime compilerservices class program static bool isnotnullablevaluetype return default t null private static void main string args console writeline isnotnullablevaluetype console writeline isnotnullablevaluetype console writeline isnotnullablevaluetype will print false true false while the codegen for the it s a reference type case is pretty efficient nasm xor eax eax test rax rax setb al movzx eax al ret for the it s a valuetype and it s a nullable case we get a useless allocation nasm sub rsp call newhelper system private corelib system xor edx edx mov dword ptr edx test rax rax seta al movzx eax al add rsp ret the places that use this pattern would likely benefit from being able to statically evaluate this because it could lead to elimination of entire branches and smaller code aside from removing a useless allocation there are hits for the default null regular expression in the corefx repo and a couple more for the not equal to null case
0
21,229
28,320,587,790
IssuesEvent
2023-04-11 00:31:38
nephio-project/sig-release
https://api.github.com/repos/nephio-project/sig-release
closed
Establish Roles and Responsibilities in the community for helping with Nephio Releases.
area/process-mgmt sig/release
We need to establish clear roles and responsibilities in the community for helping with the Nephio release process. We can start by having a Release Lead (RL) role which will be a role that will be rotated across SIG release members for every minor release. Release lead will be responsible for the following, Coordinate with other SIG(s) on finalizing the issues that are being targeted for the release. Lead all the phases of the release process defined earlier. Lead the release postmortem for the release. Identifying the RL for the next release by working with SIG release. We can define additional roles as required for the release process as we learn from our experience with doing a few releases.
1.0
Establish Roles and Responsibilities in the community for helping with Nephio Releases. - We need to establish clear roles and responsibilities in the community for helping with the Nephio release process. We can start by having a Release Lead (RL) role which will be a role that will be rotated across SIG release members for every minor release. Release lead will be responsible for the following, Coordinate with other SIG(s) on finalizing the issues that are being targeted for the release. Lead all the phases of the release process defined earlier. Lead the release postmortem for the release. Identifying the RL for the next release by working with SIG release. We can define additional roles as required for the release process as we learn from our experience with doing a few releases.
process
establish roles and responsibilities in the community for helping with nephio releases we need to establish clear roles and responsibilities in the community for helping with the nephio release process we can start by having a release lead rl role which will be a role that will be rotated across sig release members for every minor release release lead will be responsible for the following coordinate with other sig s on finalizing the issues that are being targeted for the release lead all the phases of the release process defined earlier lead the release postmortem for the release identifying the rl for the next release by working with sig release we can define additional roles as required for the release process as we learn from our experience with doing a few releases
1
21,191
28,209,461,989
IssuesEvent
2023-04-05 02:00:10
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Wed, 5 Apr 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### Fully Convolutional Networks for Dense Water Flow Intensity Prediction in Swedish Catchment Areas - **Authors:** Aleksis Pirinen, Olof Mogren, Mårten Västerdal - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.01658 - **Pdf link:** https://arxiv.org/pdf/2304.01658 - **Abstract** Intensifying climate change will lead to more extreme weather events, including heavy rainfall and drought. Accurate stream flow prediction models which are adaptable and robust to new circumstances in a changing climate will be an important source of information for decisions on climate adaptation efforts, especially regarding mitigation of the risks of and damages associated with flooding. In this work we propose a machine learning-based approach for predicting water flow intensities in inland watercourses based on the physical characteristics of the catchment areas, obtained from geospatial data (including elevation and soil maps, as well as satellite imagery), in addition to temporal information about past rainfall quantities and temperature variations. We target the one-day-ahead regime, where a fully convolutional neural network model receives spatio-temporal inputs and predicts the water flow intensity in every coordinate of the spatial input for the subsequent day. To the best of our knowledge, we are the first to tackle the task of dense water flow intensity prediction; earlier works have considered predicting flow intensities at a sparse set of locations at a time. An extensive set of model evaluations and ablations are performed, which empirically justify our various design choices. Code and preprocessed data have been made publicly available at https://github.com/aleksispi/fcn-water-flow. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### IterativePFN: True Iterative Point Cloud Filtering - **Authors:** Dasith de Silva Edirimuni, Xuequan Lu, Zhiwen Shao, Gang Li, Antonio Robles-Kelly, Ying He - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.01529 - **Pdf link:** https://arxiv.org/pdf/2304.01529 - **Abstract** The quality of point clouds is often limited by noise introduced during their capture process. Consequently, a fundamental 3D vision task is the removal of noise, known as point cloud filtering or denoising. State-of-the-art learning based methods focus on training neural networks to infer filtered displacements and directly shift noisy points onto the underlying clean surfaces. In high noise conditions, they iterate the filtering process. However, this iterative filtering is only done at test time and is less effective at ensuring points converge quickly onto the clean surfaces. We propose IterativePFN (iterative point cloud filtering network), which consists of multiple IterationModules that model the true iterative filtering process internally, within a single network. We train our IterativePFN network using a novel loss function that utilizes an adaptive ground truth target at each iteration to capture the relationship between intermediate filtering results during training. This ensures that the filtered results converge faster to the clean surfaces. Our method is able to obtain better performance compared to state-of-the-art methods. The source code can be found at: https://github.com/ddsediri/IterativePFN. ### Fully Convolutional Networks for Dense Water Flow Intensity Prediction in Swedish Catchment Areas - **Authors:** Aleksis Pirinen, Olof Mogren, Mårten Västerdal - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.01658 - **Pdf link:** https://arxiv.org/pdf/2304.01658 - **Abstract** Intensifying climate change will lead to more extreme weather events, including heavy rainfall and drought. Accurate stream flow prediction models which are adaptable and robust to new circumstances in a changing climate will be an important source of information for decisions on climate adaptation efforts, especially regarding mitigation of the risks of and damages associated with flooding. In this work we propose a machine learning-based approach for predicting water flow intensities in inland watercourses based on the physical characteristics of the catchment areas, obtained from geospatial data (including elevation and soil maps, as well as satellite imagery), in addition to temporal information about past rainfall quantities and temperature variations. We target the one-day-ahead regime, where a fully convolutional neural network model receives spatio-temporal inputs and predicts the water flow intensity in every coordinate of the spatial input for the subsequent day. To the best of our knowledge, we are the first to tackle the task of dense water flow intensity prediction; earlier works have considered predicting flow intensities at a sparse set of locations at a time. An extensive set of model evaluations and ablations are performed, which empirically justify our various design choices. Code and preprocessed data have been made publicly available at https://github.com/aleksispi/fcn-water-flow. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### A Scale-Invariant Trajectory Simplification Method for Efficient Data Collection in Videos - **Authors:** Yang Liu, Luiz Gustavo Hafemann - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.01340 - **Pdf link:** https://arxiv.org/pdf/2304.01340 - **Abstract** Training data is a critical requirement for machine learning tasks, and labeled training data can be expensive to acquire, often requiring manual or semi-automated data collection pipelines. For tracking applications, the data collection involves drawing bounding boxes around the classes of interest on each frame, and associate detections of the same "instance" over frames. In a semi-automated data collection pipeline, this can be achieved by running a baseline detection and tracking algorithm, and relying on manual correction to add/remove/change bounding boxes on each frame, as well as resolving errors in the associations over frames (track switches). In this paper, we propose a data correction pipeline to generate ground-truth data more efficiently in this semi-automated scenario. Our method simplifies the trajectories from the tracking systems and let the annotator verify and correct the objects in the sampled keyframes. Once the objects in the keyframes are corrected, the bounding boxes in the other frames are obtained by interpolation. Our method achieves substantial reduction in the number of frames requiring manual correction. In the MOT dataset, it reduces the number of frames by 30x while maintaining a HOTA score of 89.61% . Moreover, it reduces the number of frames by a factor of 10x while achieving a HOTA score of 79.24% in the SoccerNet dataset, and 85.79% in the DanceTrack dataset. The project code and data are publicly released at https://github.com/foreverYoungGitHub/trajectory-simplify-benchmark. ### Self-Supervised Image Denoising for Real-World Images with Context-aware Transformer - **Authors:** Dan Zhang, Fangfang Zhou - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2304.01627 - **Pdf link:** https://arxiv.org/pdf/2304.01627 - **Abstract** In recent years, the development of deep learning has been pushing image denoising to a new level. Among them, self-supervised denoising is increasingly popular because it does not require any prior knowledge. Most of the existing self-supervised methods are based on convolutional neural networks (CNN), which are restricted by the locality of the receptive field and would cause color shifts or textures loss. In this paper, we propose a novel Denoise Transformer for real-world image denoising, which is mainly constructed with Context-aware Denoise Transformer (CADT) units and Secondary Noise Extractor (SNE) block. CADT is designed as a dual-branch structure, where the global branch uses a window-based Transformer encoder to extract the global information, while the local branch focuses on the extraction of local features with small receptive field. By incorporating CADT as basic components, we build a hierarchical network to directly learn the noise distribution information through residual learning and obtain the first stage denoised output. Then, we design SNE in low computation for secondary global noise extraction. Finally the blind spots are collected from the Denoise Transformer output and reconstructed, forming the final denoised image. Extensive experiments on the real-world SIDD benchmark achieve 50.62/0.990 for PSNR/SSIM, which is competitive with the current state-of-the-art method and only 0.17/0.001 lower. Visual comparisons on public sRGB, Raw-RGB and greyscale datasets prove that our proposed Denoise Transformer has a competitive performance, especially on blurred textures and low-light images, without using additional knowledge, e.g., noise level or noise type, regarding the underlying unknown noise. ### Trace and Pace: Controllable Pedestrian Animation via Guided Trajectory Diffusion - **Authors:** Davis Rempe, Zhengyi Luo, Xue Bin Peng, Ye Yuan, Kris Kitani, Karsten Kreis, Sanja Fidler, Or Litany - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2304.01893 - **Pdf link:** https://arxiv.org/pdf/2304.01893 - **Abstract** We introduce a method for generating realistic pedestrian trajectories and full-body animations that can be controlled to meet user-defined goals. We draw on recent advances in guided diffusion modeling to achieve test-time controllability of trajectories, which is normally only associated with rule-based systems. Our guided diffusion model allows users to constrain trajectories through target waypoints, speed, and specified social groups while accounting for the surrounding environment context. This trajectory diffusion model is integrated with a novel physics-based humanoid controller to form a closed-loop, full-body pedestrian animation system capable of placing large crowds in a simulated environment with varying terrains. We further propose utilizing the value function learned during RL training of the animation controller to guide diffusion to produce trajectories better suited for particular scenarios such as collision avoidance and traversing uneven terrain. Video results are available on the project page at https://nv-tlabs.github.io/trace-pace . ## Keyword: raw image There is no result
2.0
New submissions for Wed, 5 Apr 23 - ## Keyword: events ### Fully Convolutional Networks for Dense Water Flow Intensity Prediction in Swedish Catchment Areas - **Authors:** Aleksis Pirinen, Olof Mogren, Mårten Västerdal - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.01658 - **Pdf link:** https://arxiv.org/pdf/2304.01658 - **Abstract** Intensifying climate change will lead to more extreme weather events, including heavy rainfall and drought. Accurate stream flow prediction models which are adaptable and robust to new circumstances in a changing climate will be an important source of information for decisions on climate adaptation efforts, especially regarding mitigation of the risks of and damages associated with flooding. In this work we propose a machine learning-based approach for predicting water flow intensities in inland watercourses based on the physical characteristics of the catchment areas, obtained from geospatial data (including elevation and soil maps, as well as satellite imagery), in addition to temporal information about past rainfall quantities and temperature variations. We target the one-day-ahead regime, where a fully convolutional neural network model receives spatio-temporal inputs and predicts the water flow intensity in every coordinate of the spatial input for the subsequent day. To the best of our knowledge, we are the first to tackle the task of dense water flow intensity prediction; earlier works have considered predicting flow intensities at a sparse set of locations at a time. An extensive set of model evaluations and ablations are performed, which empirically justify our various design choices. Code and preprocessed data have been made publicly available at https://github.com/aleksispi/fcn-water-flow. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### IterativePFN: True Iterative Point Cloud Filtering - **Authors:** Dasith de Silva Edirimuni, Xuequan Lu, Zhiwen Shao, Gang Li, Antonio Robles-Kelly, Ying He - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.01529 - **Pdf link:** https://arxiv.org/pdf/2304.01529 - **Abstract** The quality of point clouds is often limited by noise introduced during their capture process. Consequently, a fundamental 3D vision task is the removal of noise, known as point cloud filtering or denoising. State-of-the-art learning based methods focus on training neural networks to infer filtered displacements and directly shift noisy points onto the underlying clean surfaces. In high noise conditions, they iterate the filtering process. However, this iterative filtering is only done at test time and is less effective at ensuring points converge quickly onto the clean surfaces. We propose IterativePFN (iterative point cloud filtering network), which consists of multiple IterationModules that model the true iterative filtering process internally, within a single network. We train our IterativePFN network using a novel loss function that utilizes an adaptive ground truth target at each iteration to capture the relationship between intermediate filtering results during training. This ensures that the filtered results converge faster to the clean surfaces. Our method is able to obtain better performance compared to state-of-the-art methods. The source code can be found at: https://github.com/ddsediri/IterativePFN. ### Fully Convolutional Networks for Dense Water Flow Intensity Prediction in Swedish Catchment Areas - **Authors:** Aleksis Pirinen, Olof Mogren, Mårten Västerdal - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.01658 - **Pdf link:** https://arxiv.org/pdf/2304.01658 - **Abstract** Intensifying climate change will lead to more extreme weather events, including heavy rainfall and drought. Accurate stream flow prediction models which are adaptable and robust to new circumstances in a changing climate will be an important source of information for decisions on climate adaptation efforts, especially regarding mitigation of the risks of and damages associated with flooding. In this work we propose a machine learning-based approach for predicting water flow intensities in inland watercourses based on the physical characteristics of the catchment areas, obtained from geospatial data (including elevation and soil maps, as well as satellite imagery), in addition to temporal information about past rainfall quantities and temperature variations. We target the one-day-ahead regime, where a fully convolutional neural network model receives spatio-temporal inputs and predicts the water flow intensity in every coordinate of the spatial input for the subsequent day. To the best of our knowledge, we are the first to tackle the task of dense water flow intensity prediction; earlier works have considered predicting flow intensities at a sparse set of locations at a time. An extensive set of model evaluations and ablations are performed, which empirically justify our various design choices. Code and preprocessed data have been made publicly available at https://github.com/aleksispi/fcn-water-flow. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### A Scale-Invariant Trajectory Simplification Method for Efficient Data Collection in Videos - **Authors:** Yang Liu, Luiz Gustavo Hafemann - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.01340 - **Pdf link:** https://arxiv.org/pdf/2304.01340 - **Abstract** Training data is a critical requirement for machine learning tasks, and labeled training data can be expensive to acquire, often requiring manual or semi-automated data collection pipelines. For tracking applications, the data collection involves drawing bounding boxes around the classes of interest on each frame, and associate detections of the same "instance" over frames. In a semi-automated data collection pipeline, this can be achieved by running a baseline detection and tracking algorithm, and relying on manual correction to add/remove/change bounding boxes on each frame, as well as resolving errors in the associations over frames (track switches). In this paper, we propose a data correction pipeline to generate ground-truth data more efficiently in this semi-automated scenario. Our method simplifies the trajectories from the tracking systems and let the annotator verify and correct the objects in the sampled keyframes. Once the objects in the keyframes are corrected, the bounding boxes in the other frames are obtained by interpolation. Our method achieves substantial reduction in the number of frames requiring manual correction. In the MOT dataset, it reduces the number of frames by 30x while maintaining a HOTA score of 89.61% . Moreover, it reduces the number of frames by a factor of 10x while achieving a HOTA score of 79.24% in the SoccerNet dataset, and 85.79% in the DanceTrack dataset. The project code and data are publicly released at https://github.com/foreverYoungGitHub/trajectory-simplify-benchmark. ### Self-Supervised Image Denoising for Real-World Images with Context-aware Transformer - **Authors:** Dan Zhang, Fangfang Zhou - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2304.01627 - **Pdf link:** https://arxiv.org/pdf/2304.01627 - **Abstract** In recent years, the development of deep learning has been pushing image denoising to a new level. Among them, self-supervised denoising is increasingly popular because it does not require any prior knowledge. Most of the existing self-supervised methods are based on convolutional neural networks (CNN), which are restricted by the locality of the receptive field and would cause color shifts or textures loss. In this paper, we propose a novel Denoise Transformer for real-world image denoising, which is mainly constructed with Context-aware Denoise Transformer (CADT) units and Secondary Noise Extractor (SNE) block. CADT is designed as a dual-branch structure, where the global branch uses a window-based Transformer encoder to extract the global information, while the local branch focuses on the extraction of local features with small receptive field. By incorporating CADT as basic components, we build a hierarchical network to directly learn the noise distribution information through residual learning and obtain the first stage denoised output. Then, we design SNE in low computation for secondary global noise extraction. Finally the blind spots are collected from the Denoise Transformer output and reconstructed, forming the final denoised image. Extensive experiments on the real-world SIDD benchmark achieve 50.62/0.990 for PSNR/SSIM, which is competitive with the current state-of-the-art method and only 0.17/0.001 lower. Visual comparisons on public sRGB, Raw-RGB and greyscale datasets prove that our proposed Denoise Transformer has a competitive performance, especially on blurred textures and low-light images, without using additional knowledge, e.g., noise level or noise type, regarding the underlying unknown noise. ### Trace and Pace: Controllable Pedestrian Animation via Guided Trajectory Diffusion - **Authors:** Davis Rempe, Zhengyi Luo, Xue Bin Peng, Ye Yuan, Kris Kitani, Karsten Kreis, Sanja Fidler, Or Litany - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2304.01893 - **Pdf link:** https://arxiv.org/pdf/2304.01893 - **Abstract** We introduce a method for generating realistic pedestrian trajectories and full-body animations that can be controlled to meet user-defined goals. We draw on recent advances in guided diffusion modeling to achieve test-time controllability of trajectories, which is normally only associated with rule-based systems. Our guided diffusion model allows users to constrain trajectories through target waypoints, speed, and specified social groups while accounting for the surrounding environment context. This trajectory diffusion model is integrated with a novel physics-based humanoid controller to form a closed-loop, full-body pedestrian animation system capable of placing large crowds in a simulated environment with varying terrains. We further propose utilizing the value function learned during RL training of the animation controller to guide diffusion to produce trajectories better suited for particular scenarios such as collision avoidance and traversing uneven terrain. Video results are available on the project page at https://nv-tlabs.github.io/trace-pace . ## Keyword: raw image There is no result
process
new submissions for wed apr keyword events fully convolutional networks for dense water flow intensity prediction in swedish catchment areas authors aleksis pirinen olof mogren mårten västerdal subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract intensifying climate change will lead to more extreme weather events including heavy rainfall and drought accurate stream flow prediction models which are adaptable and robust to new circumstances in a changing climate will be an important source of information for decisions on climate adaptation efforts especially regarding mitigation of the risks of and damages associated with flooding in this work we propose a machine learning based approach for predicting water flow intensities in inland watercourses based on the physical characteristics of the catchment areas obtained from geospatial data including elevation and soil maps as well as satellite imagery in addition to temporal information about past rainfall quantities and temperature variations we target the one day ahead regime where a fully convolutional neural network model receives spatio temporal inputs and predicts the water flow intensity in every coordinate of the spatial input for the subsequent day to the best of our knowledge we are the first to tackle the task of dense water flow intensity prediction earlier works have considered predicting flow intensities at a sparse set of locations at a time an extensive set of model evaluations and ablations are performed which empirically justify our various design choices code and preprocessed data have been made publicly available at keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp iterativepfn true iterative point cloud filtering authors dasith de silva edirimuni xuequan lu zhiwen shao gang li antonio robles kelly ying he subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract the quality of point clouds is often limited by noise introduced during their capture process consequently a fundamental vision task is the removal of noise known as point cloud filtering or denoising state of the art learning based methods focus on training neural networks to infer filtered displacements and directly shift noisy points onto the underlying clean surfaces in high noise conditions they iterate the filtering process however this iterative filtering is only done at test time and is less effective at ensuring points converge quickly onto the clean surfaces we propose iterativepfn iterative point cloud filtering network which consists of multiple iterationmodules that model the true iterative filtering process internally within a single network we train our iterativepfn network using a novel loss function that utilizes an adaptive ground truth target at each iteration to capture the relationship between intermediate filtering results during training this ensures that the filtered results converge faster to the clean surfaces our method is able to obtain better performance compared to state of the art methods the source code can be found at fully convolutional networks for dense water flow intensity prediction in swedish catchment areas authors aleksis pirinen olof mogren mårten västerdal subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract intensifying climate change will lead to more extreme weather events including heavy rainfall and drought accurate stream flow prediction models which are adaptable and robust to new circumstances in a changing climate will be an important source of information for decisions on climate adaptation efforts especially regarding mitigation of the risks of and damages associated with flooding in this work we propose a machine learning based approach for predicting water flow intensities in inland watercourses based on the physical characteristics of the catchment areas obtained from geospatial data including elevation and soil maps as well as satellite imagery in addition to temporal information about past rainfall quantities and temperature variations we target the one day ahead regime where a fully convolutional neural network model receives spatio temporal inputs and predicts the water flow intensity in every coordinate of the spatial input for the subsequent day to the best of our knowledge we are the first to tackle the task of dense water flow intensity prediction earlier works have considered predicting flow intensities at a sparse set of locations at a time an extensive set of model evaluations and ablations are performed which empirically justify our various design choices code and preprocessed data have been made publicly available at keyword image signal processing there is no result keyword image signal process there is no result keyword compression there is no result keyword raw a scale invariant trajectory simplification method for efficient data collection in videos authors yang liu luiz gustavo hafemann subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract training data is a critical requirement for machine learning tasks and labeled training data can be expensive to acquire often requiring manual or semi automated data collection pipelines for tracking applications the data collection involves drawing bounding boxes around the classes of interest on each frame and associate detections of the same instance over frames in a semi automated data collection pipeline this can be achieved by running a baseline detection and tracking algorithm and relying on manual correction to add remove change bounding boxes on each frame as well as resolving errors in the associations over frames track switches in this paper we propose a data correction pipeline to generate ground truth data more efficiently in this semi automated scenario our method simplifies the trajectories from the tracking systems and let the annotator verify and correct the objects in the sampled keyframes once the objects in the keyframes are corrected the bounding boxes in the other frames are obtained by interpolation our method achieves substantial reduction in the number of frames requiring manual correction in the mot dataset it reduces the number of frames by while maintaining a hota score of moreover it reduces the number of frames by a factor of while achieving a hota score of in the soccernet dataset and in the dancetrack dataset the project code and data are publicly released at self supervised image denoising for real world images with context aware transformer authors dan zhang fangfang zhou subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract in recent years the development of deep learning has been pushing image denoising to a new level among them self supervised denoising is increasingly popular because it does not require any prior knowledge most of the existing self supervised methods are based on convolutional neural networks cnn which are restricted by the locality of the receptive field and would cause color shifts or textures loss in this paper we propose a novel denoise transformer for real world image denoising which is mainly constructed with context aware denoise transformer cadt units and secondary noise extractor sne block cadt is designed as a dual branch structure where the global branch uses a window based transformer encoder to extract the global information while the local branch focuses on the extraction of local features with small receptive field by incorporating cadt as basic components we build a hierarchical network to directly learn the noise distribution information through residual learning and obtain the first stage denoised output then we design sne in low computation for secondary global noise extraction finally the blind spots are collected from the denoise transformer output and reconstructed forming the final denoised image extensive experiments on the real world sidd benchmark achieve for psnr ssim which is competitive with the current state of the art method and only lower visual comparisons on public srgb raw rgb and greyscale datasets prove that our proposed denoise transformer has a competitive performance especially on blurred textures and low light images without using additional knowledge e g noise level or noise type regarding the underlying unknown noise trace and pace controllable pedestrian animation via guided trajectory diffusion authors davis rempe zhengyi luo xue bin peng ye yuan kris kitani karsten kreis sanja fidler or litany subjects computer vision and pattern recognition cs cv graphics cs gr machine learning cs lg arxiv link pdf link abstract we introduce a method for generating realistic pedestrian trajectories and full body animations that can be controlled to meet user defined goals we draw on recent advances in guided diffusion modeling to achieve test time controllability of trajectories which is normally only associated with rule based systems our guided diffusion model allows users to constrain trajectories through target waypoints speed and specified social groups while accounting for the surrounding environment context this trajectory diffusion model is integrated with a novel physics based humanoid controller to form a closed loop full body pedestrian animation system capable of placing large crowds in a simulated environment with varying terrains we further propose utilizing the value function learned during rl training of the animation controller to guide diffusion to produce trajectories better suited for particular scenarios such as collision avoidance and traversing uneven terrain video results are available on the project page at keyword raw image there is no result
1
4,053
6,988,245,602
IssuesEvent
2017-12-14 12:12:04
nlbdev/pipeline
https://api.github.com/repos/nlbdev/pipeline
closed
move colophon, hide titlepage and print toc
enhancement pre-processing Priority:3 - High
Maybe delete most of it and keep only dedications and forewords?
1.0
move colophon, hide titlepage and print toc - Maybe delete most of it and keep only dedications and forewords?
process
move colophon hide titlepage and print toc maybe delete most of it and keep only dedications and forewords
1
22,528
31,627,016,475
IssuesEvent
2023-09-06 06:25:37
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Backwards Compatibility: Combine always_sample with a not
question processor/tailsampling
### Component(s) processor/tailsampling ### Is your feature request related to a problem? Please describe. We want to start implementing tail based sampling in our organization in a backwards compatible way. Some services will use tail sampling, for the others we want to start by using the always_sample policy. Is there a straightforward way to implement this policy: ``` if service.name not in ['list', 'of', 'services', 'already', 'using', 'tail', 'sampling'] then always_sample ``` Here are a few things I'm trying ### Option 1 ``` { name: backwards-compatibility-policy-v1, type: and, and: { and_sub_policy: [ { name: services-using-tail_sampling-policy, type: string_attribute, string_attribute: { key: service.name, values: [ list, of, services, already, using, tail_sampling ], invert_match: true, }, }, { name: sample-all-policy, type: always_sample }, ], }, } ``` However, this might not work, because this will lead to an inverted not sample decision - meaning the actual policies for these services will have no impact. Did I understand this correctly ? ----------------------------------------- ### Option 2 ``` { name: backwards-compatibility-policy-v2, type: and, and: { and_sub_policy: [ { name: services-not-using-tail_sampling-policy, type: string_attribute, string_attribute: { key: service.name, values: [ list, of, services, not, using, tail_sampling ], }, }, { name: sample-all-policy, type: always_sample }, ], }, } ``` This might work - however, this means that the entire list of legacy services would need to be added here, and any new service that is not using tail sampling would also need to be added here. This unfortunately means we need to maintain a redundant list of services here, thus shifting the burden on teams that are not ready to use tail sampling, and onto those that are just trying out Otel in the first place. --------------------- ### Option 3 ``` { name: backwards-compatibility-policy-v3, type: and, and: { and_sub_policy: [ { name: services-using-tail_sampling-policy, type: ottl_condition, ottl_condition: { span: [ "attributes[\"service.name\"] != \"list\" ", "attributes[\"service.name\"] != \"of\" ", "attributes[\"service.name\"] != \"services\" ", "attributes[\"service.name\"] != \"already\" ", "attributes[\"service.name\"] != \"using\" ", "attributes[\"service.name\"] != \"tail_sampling\" ", ], spanevent: [ "attributes[\"service.name\"] != \"list\" ", "attributes[\"service.name\"] != \"of\" ", "attributes[\"service.name\"] != \"services\" ", "attributes[\"service.name\"] != \"already\" ", "attributes[\"service.name\"] != \"using\" ", "attributes[\"service.name\"] != \"tail_sampling\" ", ], } }, { name: sample-all-policy, type: always_sample }, ], }, } ``` Does this approach work ? Where can I find reference to the syntax of OTTL conditionals. I'm hoping this will create an OR between the list of services - Is that correct ? ---------------------------- Could someone help out with this problem. Is there a more straightforward way of solving this ? Thanks! ### Describe the solution you'd like Either some help would the above 3 approaches would be appreciated. Or it would also be useful, if we could use a simple `not` statement with `string_attribute` policy - that leads to a sample decision for string values that do not match - rather than leading to an explicit inverted not sampled decision. Something like ``` not: { string_attribute: { key: http.route, values: [/v1/providers/.+], }, }, ``` ### Describe alternatives you've considered Described above. ### Additional context _No response_
1.0
Backwards Compatibility: Combine always_sample with a not - ### Component(s) processor/tailsampling ### Is your feature request related to a problem? Please describe. We want to start implementing tail based sampling in our organization in a backwards compatible way. Some services will use tail sampling, for the others we want to start by using the always_sample policy. Is there a straightforward way to implement this policy: ``` if service.name not in ['list', 'of', 'services', 'already', 'using', 'tail', 'sampling'] then always_sample ``` Here are a few things I'm trying ### Option 1 ``` { name: backwards-compatibility-policy-v1, type: and, and: { and_sub_policy: [ { name: services-using-tail_sampling-policy, type: string_attribute, string_attribute: { key: service.name, values: [ list, of, services, already, using, tail_sampling ], invert_match: true, }, }, { name: sample-all-policy, type: always_sample }, ], }, } ``` However, this might not work, because this will lead to an inverted not sample decision - meaning the actual policies for these services will have no impact. Did I understand this correctly ? ----------------------------------------- ### Option 2 ``` { name: backwards-compatibility-policy-v2, type: and, and: { and_sub_policy: [ { name: services-not-using-tail_sampling-policy, type: string_attribute, string_attribute: { key: service.name, values: [ list, of, services, not, using, tail_sampling ], }, }, { name: sample-all-policy, type: always_sample }, ], }, } ``` This might work - however, this means that the entire list of legacy services would need to be added here, and any new service that is not using tail sampling would also need to be added here. This unfortunately means we need to maintain a redundant list of services here, thus shifting the burden on teams that are not ready to use tail sampling, and onto those that are just trying out Otel in the first place. --------------------- ### Option 3 ``` { name: backwards-compatibility-policy-v3, type: and, and: { and_sub_policy: [ { name: services-using-tail_sampling-policy, type: ottl_condition, ottl_condition: { span: [ "attributes[\"service.name\"] != \"list\" ", "attributes[\"service.name\"] != \"of\" ", "attributes[\"service.name\"] != \"services\" ", "attributes[\"service.name\"] != \"already\" ", "attributes[\"service.name\"] != \"using\" ", "attributes[\"service.name\"] != \"tail_sampling\" ", ], spanevent: [ "attributes[\"service.name\"] != \"list\" ", "attributes[\"service.name\"] != \"of\" ", "attributes[\"service.name\"] != \"services\" ", "attributes[\"service.name\"] != \"already\" ", "attributes[\"service.name\"] != \"using\" ", "attributes[\"service.name\"] != \"tail_sampling\" ", ], } }, { name: sample-all-policy, type: always_sample }, ], }, } ``` Does this approach work ? Where can I find reference to the syntax of OTTL conditionals. I'm hoping this will create an OR between the list of services - Is that correct ? ---------------------------- Could someone help out with this problem. Is there a more straightforward way of solving this ? Thanks! ### Describe the solution you'd like Either some help would the above 3 approaches would be appreciated. Or it would also be useful, if we could use a simple `not` statement with `string_attribute` policy - that leads to a sample decision for string values that do not match - rather than leading to an explicit inverted not sampled decision. Something like ``` not: { string_attribute: { key: http.route, values: [/v1/providers/.+], }, }, ``` ### Describe alternatives you've considered Described above. ### Additional context _No response_
process
backwards compatibility combine always sample with a not component s processor tailsampling is your feature request related to a problem please describe we want to start implementing tail based sampling in our organization in a backwards compatible way some services will use tail sampling for the others we want to start by using the always sample policy is there a straightforward way to implement this policy if service name not in then always sample here are a few things i m trying option name backwards compatibility policy type and and and sub policy name services using tail sampling policy type string attribute string attribute key service name values list of services already using tail sampling invert match true name sample all policy type always sample however this might not work because this will lead to an inverted not sample decision meaning the actual policies for these services will have no impact did i understand this correctly option name backwards compatibility policy type and and and sub policy name services not using tail sampling policy type string attribute string attribute key service name values list of services not using tail sampling name sample all policy type always sample this might work however this means that the entire list of legacy services would need to be added here and any new service that is not using tail sampling would also need to be added here this unfortunately means we need to maintain a redundant list of services here thus shifting the burden on teams that are not ready to use tail sampling and onto those that are just trying out otel in the first place option name backwards compatibility policy type and and and sub policy name services using tail sampling policy type ottl condition ottl condition span attributes list attributes of attributes services attributes already attributes using attributes tail sampling spanevent attributes list attributes of attributes services attributes already attributes using attributes tail sampling name sample all policy type always sample does this approach work where can i find reference to the syntax of ottl conditionals i m hoping this will create an or between the list of services is that correct could someone help out with this problem is there a more straightforward way of solving this thanks describe the solution you d like either some help would the above approaches would be appreciated or it would also be useful if we could use a simple not statement with string attribute policy that leads to a sample decision for string values that do not match rather than leading to an explicit inverted not sampled decision something like not string attribute key http route values describe alternatives you ve considered described above additional context no response
1
399,517
11,756,416,501
IssuesEvent
2020-03-13 11:31:22
TrinityCore/TrinityCore
https://api.github.com/repos/TrinityCore/TrinityCore
closed
DB/Creature: Kirin Tor Mage in Ulduar
Comp-Database Priority-Cosmetic Sub-Spawns
Description: Due to the error of NPC spawn point, the characters are overlapped. .go xyz -814.109985 -201.125 429.924988 603 NPC ID: 33672 Kirin Tor Mage in Ulduar Raid must be summoned via an event https://www.youtube.com/watch?v=UuPPW5jJaP4&feature=youtu.be Branch(es): 3.3.5 TC rev. hash/commit: 961adea3727e32c068245735e76db2445c2fcd2d Operating system: CentOS
1.0
DB/Creature: Kirin Tor Mage in Ulduar - Description: Due to the error of NPC spawn point, the characters are overlapped. .go xyz -814.109985 -201.125 429.924988 603 NPC ID: 33672 Kirin Tor Mage in Ulduar Raid must be summoned via an event https://www.youtube.com/watch?v=UuPPW5jJaP4&feature=youtu.be Branch(es): 3.3.5 TC rev. hash/commit: 961adea3727e32c068245735e76db2445c2fcd2d Operating system: CentOS
non_process
db creature kirin tor mage in ulduar description due to the error of npc spawn point the characters are overlapped go xyz npc id kirin tor mage in ulduar raid must be summoned via an event branch es tc rev hash commit operating system centos
0
7,235
10,384,414,931
IssuesEvent
2019-09-10 11:57:45
RIOT-OS/RIOT
https://api.github.com/repos/RIOT-OS/RIOT
closed
periph/timer: API inconstency
Area: drivers Process: API change State: stale Type: enhancement
Reviewing #4089 I realized that `periph/timer` API seems to be inconsistent. `timer_init()` returns 0 on success, while all other functions return 1 on success. I think all periph functions which just return a boolean value should return the same one. Ideally this would be rather a macro like `PERIPH_SUCCESS` than a literal.
1.0
periph/timer: API inconstency - Reviewing #4089 I realized that `periph/timer` API seems to be inconsistent. `timer_init()` returns 0 on success, while all other functions return 1 on success. I think all periph functions which just return a boolean value should return the same one. Ideally this would be rather a macro like `PERIPH_SUCCESS` than a literal.
process
periph timer api inconstency reviewing i realized that periph timer api seems to be inconsistent timer init returns on success while all other functions return on success i think all periph functions which just return a boolean value should return the same one ideally this would be rather a macro like periph success than a literal
1
79,143
22,623,271,115
IssuesEvent
2022-06-30 08:28:22
adoptium/temurin-build
https://api.github.com/repos/adoptium/temurin-build
closed
JDK11 / Win 32 Build Broken
windows buildbreak
The JDK11 build has been failing since 21st June. https://ci.adoptopenjdk.net/view/Failing%20Temurin%20jobs/job/build-scripts/job/jobs/job/jdk11u/job/jdk11u-windows-x86-32-temurin/ Believe this may be due to the change to VS2019, currently investigating New build : https://ci.adoptopenjdk.net/view/Failing%20Temurin%20jobs/job/build-scripts/job/jobs/job/jdk11u/job/jdk11u-windows-x86-32-temurin/77/
1.0
JDK11 / Win 32 Build Broken - The JDK11 build has been failing since 21st June. https://ci.adoptopenjdk.net/view/Failing%20Temurin%20jobs/job/build-scripts/job/jobs/job/jdk11u/job/jdk11u-windows-x86-32-temurin/ Believe this may be due to the change to VS2019, currently investigating New build : https://ci.adoptopenjdk.net/view/Failing%20Temurin%20jobs/job/build-scripts/job/jobs/job/jdk11u/job/jdk11u-windows-x86-32-temurin/77/
non_process
win build broken the build has been failing since june believe this may be due to the change to currently investigating new build
0
5,174
7,959,437,504
IssuesEvent
2018-07-13 00:47:51
GoogleCloudPlatform/google-cloud-cpp
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-cpp
closed
Document the CurlRequest API as a concept.
status: will not fix storage type: process
This is two parts: one is to write a comment or brief document describing the requirements for `CurlRequest`. The second is to write a series of compile-time assertions that check this concept, so we can enforce the requirements in `storage::DefaultClient<>` and generate good error messages when we make a mistake.
1.0
Document the CurlRequest API as a concept. - This is two parts: one is to write a comment or brief document describing the requirements for `CurlRequest`. The second is to write a series of compile-time assertions that check this concept, so we can enforce the requirements in `storage::DefaultClient<>` and generate good error messages when we make a mistake.
process
document the curlrequest api as a concept this is two parts one is to write a comment or brief document describing the requirements for curlrequest the second is to write a series of compile time assertions that check this concept so we can enforce the requirements in storage defaultclient and generate good error messages when we make a mistake
1
13,852
16,614,007,117
IssuesEvent
2021-06-02 14:40:47
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
In raster calculator, the AND and OR buttons insert the translated words
Bug Processing
When using Spanish language and raster calculator the AND and OR buttons insert the translated words Y (=AND) and O (=OR) into the expression instead of the correct, untranslated ones, resulting in a syntax error. 1. Set language to Spanish 2. Open raster calculator 3. Hit AND or OR buttons (labelled Y or O) 4. The spanish words Y / O are inserted, instead of AND / OR, the expression is incorrect **QGIS and OS versions** Versión de QGIS | 3.16.2-Hannover | Revisión del código de QGIS | f1660f9da5 -- | -- | -- | -- Compilado contra Qt | 5.11.2 | Ejecutándose contra Qt | 5.11.2 Compilado contra GDAL/OGR | 3.1.4 | Ejecutándose contra GDAL/OGR | 3.1.4 Compilado contra GEOS | 3.8.1-CAPI-1.13.3 | Ejecutándose contra GEOS | 3.8.1-CAPI-1.13.3 Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0 Versión del cliente PostgreSQL | 11.5 | Versión de SpatiaLite | 4.3.0 Versión de QWT | 6.1.3 | Versión de QScintilla2 | 2.10.8 Compilado contra PROJ | 6.3.2 | Ejecutándose contra PROJ | Rel. 6.3.2, May 1st, 2020 Versión del SO | Windows 10 (10.0) Active python plugins | gdb2gpkg; mapswipetool_plugin; openlayers_plugin; db_manager; MetaSearch; processing
1.0
In raster calculator, the AND and OR buttons insert the translated words - When using Spanish language and raster calculator the AND and OR buttons insert the translated words Y (=AND) and O (=OR) into the expression instead of the correct, untranslated ones, resulting in a syntax error. 1. Set language to Spanish 2. Open raster calculator 3. Hit AND or OR buttons (labelled Y or O) 4. The spanish words Y / O are inserted, instead of AND / OR, the expression is incorrect **QGIS and OS versions** Versión de QGIS | 3.16.2-Hannover | Revisión del código de QGIS | f1660f9da5 -- | -- | -- | -- Compilado contra Qt | 5.11.2 | Ejecutándose contra Qt | 5.11.2 Compilado contra GDAL/OGR | 3.1.4 | Ejecutándose contra GDAL/OGR | 3.1.4 Compilado contra GEOS | 3.8.1-CAPI-1.13.3 | Ejecutándose contra GEOS | 3.8.1-CAPI-1.13.3 Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0 Versión del cliente PostgreSQL | 11.5 | Versión de SpatiaLite | 4.3.0 Versión de QWT | 6.1.3 | Versión de QScintilla2 | 2.10.8 Compilado contra PROJ | 6.3.2 | Ejecutándose contra PROJ | Rel. 6.3.2, May 1st, 2020 Versión del SO | Windows 10 (10.0) Active python plugins | gdb2gpkg; mapswipetool_plugin; openlayers_plugin; db_manager; MetaSearch; processing
process
in raster calculator the and and or buttons insert the translated words when using spanish language and raster calculator the and and or buttons insert the translated words y and and o or into the expression instead of the correct untranslated ones resulting in a syntax error set language to spanish open raster calculator hit and or or buttons labelled y or o the spanish words y o are inserted instead of and or the expression is incorrect qgis and os versions versión de qgis hannover revisión del código de qgis compilado contra qt ejecutándose contra qt compilado contra gdal ogr ejecutándose contra gdal ogr compilado contra geos capi ejecutándose contra geos capi compiled against sqlite running against sqlite versión del cliente postgresql versión de spatialite versión de qwt versión de compilado contra proj ejecutándose contra proj rel may versión del so windows active python plugins mapswipetool plugin openlayers plugin db manager metasearch processing
1
25,504
6,668,088,604
IssuesEvent
2017-10-03 14:44:26
rust-lang/rust
https://api.github.com/repos/rust-lang/rust
closed
enable mir inlining across trait methods
A-codegen A-mir C-enhancement T-compiler
As the fix to https://github.com/rust-lang/rust/issues/40473, @qmx disabled MIR inlining of trait calls altogether. This issue tracks the task of re-enabling the inlining, where possible! This is a slightly more involved task. Here is the example text: ```rust pub trait Foo { fn bar(&self) {} } impl Foo for () { fn bar(&self) { println!("Hello, World!"); } } pub fn main() { ().bar(); } ``` The goal would be to have the call to `bar()` inlined. But we should draw up some more interesting tasks showing edge-cases: for example, generic functions and so forth (we should be able to use `RevealMode::All` to "see through" specialization, since inlining executes quite late). I think roughly speaking the steps to solve this issue "properly" are to: - Refactor `Instance` and `InstanceDef` out of `librustc_trans` and (probably) into `librustc` - Rework inlining to use those mechanisms to resolve method calls (including trait ones)
1.0
enable mir inlining across trait methods - As the fix to https://github.com/rust-lang/rust/issues/40473, @qmx disabled MIR inlining of trait calls altogether. This issue tracks the task of re-enabling the inlining, where possible! This is a slightly more involved task. Here is the example text: ```rust pub trait Foo { fn bar(&self) {} } impl Foo for () { fn bar(&self) { println!("Hello, World!"); } } pub fn main() { ().bar(); } ``` The goal would be to have the call to `bar()` inlined. But we should draw up some more interesting tasks showing edge-cases: for example, generic functions and so forth (we should be able to use `RevealMode::All` to "see through" specialization, since inlining executes quite late). I think roughly speaking the steps to solve this issue "properly" are to: - Refactor `Instance` and `InstanceDef` out of `librustc_trans` and (probably) into `librustc` - Rework inlining to use those mechanisms to resolve method calls (including trait ones)
non_process
enable mir inlining across trait methods as the fix to qmx disabled mir inlining of trait calls altogether this issue tracks the task of re enabling the inlining where possible this is a slightly more involved task here is the example text rust pub trait foo fn bar self impl foo for fn bar self println hello world pub fn main bar the goal would be to have the call to bar inlined but we should draw up some more interesting tasks showing edge cases for example generic functions and so forth we should be able to use revealmode all to see through specialization since inlining executes quite late i think roughly speaking the steps to solve this issue properly are to refactor instance and instancedef out of librustc trans and probably into librustc rework inlining to use those mechanisms to resolve method calls including trait ones
0
71,388
3,356,379,343
IssuesEvent
2015-11-18 20:14:48
TechReborn/TechReborn
https://api.github.com/repos/TechReborn/TechReborn
closed
Missing texture with standard machine casing.
bug Medium priority
Techreborn: 0.5.6.1004 reborncore:1.0.0.9 forge:10.13.4.1558 ic2/3:2.2.2.791 # Enable Connected textures B:"Enable Connected textures"=false ![2015-11-12_22 39 56](https://cloud.githubusercontent.com/assets/8199121/11138944/ac85ef34-898e-11e5-8830-dce26ac8b0ae.png)
1.0
Missing texture with standard machine casing. - Techreborn: 0.5.6.1004 reborncore:1.0.0.9 forge:10.13.4.1558 ic2/3:2.2.2.791 # Enable Connected textures B:"Enable Connected textures"=false ![2015-11-12_22 39 56](https://cloud.githubusercontent.com/assets/8199121/11138944/ac85ef34-898e-11e5-8830-dce26ac8b0ae.png)
non_process
missing texture with standard machine casing techreborn reborncore forge enable connected textures b enable connected textures false
0
16,278
20,884,553,965
IssuesEvent
2022-03-23 02:34:50
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add Da Derp Dee Derp Da Teetley Derpee Derpee Dumb
suggested title in process
Please add as much of the following info as you can: Title: Da Derp Dee Derp Da Teetley Derpee Derpee Dumb Type (film/tv show): Film Film or show in which it appears: South Park (https://www.imdb.com/title/tt0705968/ Season 06 Episode 15) Is the parent film/show streaming anywhere? Amazon Prime in the UK. About when in the parent film/show does it appear? 15 minute Actual footage of the film/show can be seen (yes/no)? https://www.youtube.com/watch?v=ZxYA6duF-9E
1.0
Add Da Derp Dee Derp Da Teetley Derpee Derpee Dumb - Please add as much of the following info as you can: Title: Da Derp Dee Derp Da Teetley Derpee Derpee Dumb Type (film/tv show): Film Film or show in which it appears: South Park (https://www.imdb.com/title/tt0705968/ Season 06 Episode 15) Is the parent film/show streaming anywhere? Amazon Prime in the UK. About when in the parent film/show does it appear? 15 minute Actual footage of the film/show can be seen (yes/no)? https://www.youtube.com/watch?v=ZxYA6duF-9E
process
add da derp dee derp da teetley derpee derpee dumb please add as much of the following info as you can title da derp dee derp da teetley derpee derpee dumb type film tv show film film or show in which it appears south park season episode is the parent film show streaming anywhere amazon prime in the uk about when in the parent film show does it appear minute actual footage of the film show can be seen yes no
1
303,338
22,971,499,485
IssuesEvent
2022-07-20 03:53:59
a-happy-dev-day/yanawa-server
https://api.github.com/repos/a-happy-dev-day/yanawa-server
closed
[Feature] Search Service
documentation feature
### 용어 사전 용어 | 영어 | 설명 -- | -- | -- 지역 | location | 참여 가능한 위치 시작 일자 | startDate | 참여 가능한 매칭 시작 일자 종료 일자 | endDate | 참여 가능한 매칭 마지막 일자 시작 시간 | startTime | 참여 가능한 매칭 시작 시간 종료 시간 | endTime | 참여 가능한 매칭 마지막 시간 구력 | annual | 원하는 매칭의 구력 ### 행위 - 회원은 지역 정보와 예정 날짜, 예정 시간, 구력을 이용해 매칭을 검색할 수 있다. - 지역 정보는 코트장의 지역 정보이다. - 예정 날짜는 찾으려는 날짜의 기간이 입력된다. - 예정 시간은 시작 시간과 종료 시간이 입력된다. - 구력 정보는 각 연차별 이상 및 미만의 정보가 입력된다. - **값이 들어가지 않으면 해당 조건은 검색하지 않는다.** ### 다이어그램 ```mermaid sequenceDiagram Server -->> Search : request search Search -->> Matching : search matching Matching ->> Matching : search Matching ->> Search : return matching json value Search ->> Server : response matching ``` ### 이미지 <img width="904" alt="스크린샷 2022-07-06 오후 12 40 51" src="https://user-images.githubusercontent.com/92219795/177463244-abcc7625-b93c-4999-affc-f2280998ebb7.png">
1.0
[Feature] Search Service - ### 용어 사전 용어 | 영어 | 설명 -- | -- | -- 지역 | location | 참여 가능한 위치 시작 일자 | startDate | 참여 가능한 매칭 시작 일자 종료 일자 | endDate | 참여 가능한 매칭 마지막 일자 시작 시간 | startTime | 참여 가능한 매칭 시작 시간 종료 시간 | endTime | 참여 가능한 매칭 마지막 시간 구력 | annual | 원하는 매칭의 구력 ### 행위 - 회원은 지역 정보와 예정 날짜, 예정 시간, 구력을 이용해 매칭을 검색할 수 있다. - 지역 정보는 코트장의 지역 정보이다. - 예정 날짜는 찾으려는 날짜의 기간이 입력된다. - 예정 시간은 시작 시간과 종료 시간이 입력된다. - 구력 정보는 각 연차별 이상 및 미만의 정보가 입력된다. - **값이 들어가지 않으면 해당 조건은 검색하지 않는다.** ### 다이어그램 ```mermaid sequenceDiagram Server -->> Search : request search Search -->> Matching : search matching Matching ->> Matching : search Matching ->> Search : return matching json value Search ->> Server : response matching ``` ### 이미지 <img width="904" alt="스크린샷 2022-07-06 오후 12 40 51" src="https://user-images.githubusercontent.com/92219795/177463244-abcc7625-b93c-4999-affc-f2280998ebb7.png">
non_process
search service 용어 사전 용어 영어 설명 지역 location 참여 가능한 위치 시작 일자 startdate 참여 가능한 매칭 시작 일자 종료 일자 enddate 참여 가능한 매칭 마지막 일자 시작 시간 starttime 참여 가능한 매칭 시작 시간 종료 시간 endtime 참여 가능한 매칭 마지막 시간 구력 annual 원하는 매칭의 구력 행위 회원은 지역 정보와 예정 날짜 예정 시간 구력을 이용해 매칭을 검색할 수 있다 지역 정보는 코트장의 지역 정보이다 예정 날짜는 찾으려는 날짜의 기간이 입력된다 예정 시간은 시작 시간과 종료 시간이 입력된다 구력 정보는 각 연차별 이상 및 미만의 정보가 입력된다 값이 들어가지 않으면 해당 조건은 검색하지 않는다 다이어그램 mermaid sequencediagram server search request search search matching search matching matching matching search matching search return matching json value search server response matching 이미지 img width alt 스크린샷 오후 src
0
13,921
16,677,496,367
IssuesEvent
2021-06-07 18:09:57
ORNL-AMO/AMO-Tools-Suite
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Suite
closed
Heat Cascade Error
Needs Verification Process Heating bug
Issue overview -------------- Fix error from the Algo doc... currently: line 21 ` const double energySavings = eqEnergySupply * effOpHours;` Should be: line 21 ` const double energySavings = std::min(eqEnergySupply, secFiringRate) * effOpHours;` Looks like secFiringRate may need bindings Also pls make a binding for "hourlySavings" which = std::min(eqEnergySupply, secFiringRate) If you want to make that tidier (add a line for hourlySavings and use that in energySavings), that is fine
1.0
Heat Cascade Error - Issue overview -------------- Fix error from the Algo doc... currently: line 21 ` const double energySavings = eqEnergySupply * effOpHours;` Should be: line 21 ` const double energySavings = std::min(eqEnergySupply, secFiringRate) * effOpHours;` Looks like secFiringRate may need bindings Also pls make a binding for "hourlySavings" which = std::min(eqEnergySupply, secFiringRate) If you want to make that tidier (add a line for hourlySavings and use that in energySavings), that is fine
process
heat cascade error issue overview fix error from the algo doc currently line const double energysavings eqenergysupply effophours should be line const double energysavings std min eqenergysupply secfiringrate effophours looks like secfiringrate may need bindings also pls make a binding for hourlysavings which std min eqenergysupply secfiringrate if you want to make that tidier add a line for hourlysavings and use that in energysavings that is fine
1
203,984
15,396,544,827
IssuesEvent
2021-03-03 20:47:49
projectcontour/contour
https://api.github.com/repos/projectcontour/contour
closed
SNI integration test is flaky
area/testing
See https://travis-ci.com/github/projectcontour/contour/jobs/365814535 Some relevant output: ``` # reading document from ./_integration/testsuite/httpproxy/002-header-condition-match.yaml ... ok 46 - running object update check # Fatal - internal/check/grnqoUyHmVtJ:5: eval_conflict_error: complete rules must not produce multiple outputs not ok 47 - running Rego check lines 160-172 --- - Message: 'internal/check/grnqoUyHmVtJ:5: eval_conflict_error: complete rules must not produce multiple outputs' Severity: Fatal Timestamp: "2020-07-28T18:28:52.766611607Z" ... # skipping ok 48 - deleting test objects 1..48 ... # reading document from ./_integration/testsuite/httpproxy/003-path-condition-match.yaml ... ok 31 - running Rego check lines 98-106 # Error - raised predicate "error_proxy_is_not_valid" # Error - HTTP proxy for 'conditions.projectcontour.io' is not valid # Error - status: invalid # Error - desc: fqdn "conditions.projectcontour.io" is used in multiple HTTPProxies: default/header-conditions, default/path-conditions not ok 32 - running Rego check lines 108-121 --- - Message: |- raised predicate "error_proxy_is_not_valid" HTTP proxy for 'conditions.projectcontour.io' is not valid status: invalid desc: fqdn "conditions.projectcontour.io" is used in multiple HTTPProxies: default/header-conditions, default/path-conditions Severity: Error Timestamp: "0001-01-01T00:00:00Z" ... # Error - internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45072->127.0.0.1:9080: read: connection reset by peer # Error - internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45076->127.0.0.1:9080: read: connection reset by peer # Error - internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45080->127.0.0.1:9080: read: connection reset by peer not ok 33 - running Rego check lines 123-182 --- - Message: 'internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45072->127.0.0.1:9080: read: connection reset by peer' Severity: Error Timestamp: "0001-01-01T00:00:00Z" - Message: 'internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45076->127.0.0.1:9080: read: connection reset by peer' Severity: Error Timestamp: "0001-01-01T00:00:00Z" - Message: 'internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45080->127.0.0.1:9080: read: connection reset by peer' Severity: Error Timestamp: "0001-01-01T00:00:00Z" ... ok 34 - deleting test objects 1..34 ```
1.0
SNI integration test is flaky - See https://travis-ci.com/github/projectcontour/contour/jobs/365814535 Some relevant output: ``` # reading document from ./_integration/testsuite/httpproxy/002-header-condition-match.yaml ... ok 46 - running object update check # Fatal - internal/check/grnqoUyHmVtJ:5: eval_conflict_error: complete rules must not produce multiple outputs not ok 47 - running Rego check lines 160-172 --- - Message: 'internal/check/grnqoUyHmVtJ:5: eval_conflict_error: complete rules must not produce multiple outputs' Severity: Fatal Timestamp: "2020-07-28T18:28:52.766611607Z" ... # skipping ok 48 - deleting test objects 1..48 ... # reading document from ./_integration/testsuite/httpproxy/003-path-condition-match.yaml ... ok 31 - running Rego check lines 98-106 # Error - raised predicate "error_proxy_is_not_valid" # Error - HTTP proxy for 'conditions.projectcontour.io' is not valid # Error - status: invalid # Error - desc: fqdn "conditions.projectcontour.io" is used in multiple HTTPProxies: default/header-conditions, default/path-conditions not ok 32 - running Rego check lines 108-121 --- - Message: |- raised predicate "error_proxy_is_not_valid" HTTP proxy for 'conditions.projectcontour.io' is not valid status: invalid desc: fqdn "conditions.projectcontour.io" is used in multiple HTTPProxies: default/header-conditions, default/path-conditions Severity: Error Timestamp: "0001-01-01T00:00:00Z" ... # Error - internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45072->127.0.0.1:9080: read: connection reset by peer # Error - internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45076->127.0.0.1:9080: read: connection reset by peer # Error - internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45080->127.0.0.1:9080: read: connection reset by peer not ok 33 - running Rego check lines 123-182 --- - Message: 'internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45072->127.0.0.1:9080: read: connection reset by peer' Severity: Error Timestamp: "0001-01-01T00:00:00Z" - Message: 'internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45076->127.0.0.1:9080: read: connection reset by peer' Severity: Error Timestamp: "0001-01-01T00:00:00Z" - Message: 'internal/check/RaYbFFqqeGji:31: eval_builtin_error: http.send: Get "http://127.0.0.1:9080/": read tcp 127.0.0.1:45080->127.0.0.1:9080: read: connection reset by peer' Severity: Error Timestamp: "0001-01-01T00:00:00Z" ... ok 34 - deleting test objects 1..34 ```
non_process
sni integration test is flaky see some relevant output reading document from integration testsuite httpproxy header condition match yaml ok running object update check fatal internal check grnqouyhmvtj eval conflict error complete rules must not produce multiple outputs not ok running rego check lines message internal check grnqouyhmvtj eval conflict error complete rules must not produce multiple outputs severity fatal timestamp skipping ok deleting test objects reading document from integration testsuite httpproxy path condition match yaml ok running rego check lines error raised predicate error proxy is not valid error http proxy for conditions projectcontour io is not valid error status invalid error desc fqdn conditions projectcontour io is used in multiple httpproxies default header conditions default path conditions not ok running rego check lines message raised predicate error proxy is not valid http proxy for conditions projectcontour io is not valid status invalid desc fqdn conditions projectcontour io is used in multiple httpproxies default header conditions default path conditions severity error timestamp error internal check raybffqqegji eval builtin error http send get read tcp read connection reset by peer error internal check raybffqqegji eval builtin error http send get read tcp read connection reset by peer error internal check raybffqqegji eval builtin error http send get read tcp read connection reset by peer not ok running rego check lines message internal check raybffqqegji eval builtin error http send get read tcp read connection reset by peer severity error timestamp message internal check raybffqqegji eval builtin error http send get read tcp read connection reset by peer severity error timestamp message internal check raybffqqegji eval builtin error http send get read tcp read connection reset by peer severity error timestamp ok deleting test objects
0
22,142
11,694,116,267
IssuesEvent
2020-03-06 02:52:59
letpoc/faq
https://api.github.com/repos/letpoc/faq
closed
Restful Service - User login
P01-S01 PI-01 Restful Service
As a developer, develop user login functionalities with security feature with secret token, authorization code Acceptance Criteria: Only Valid user can login in to the application
1.0
Restful Service - User login - As a developer, develop user login functionalities with security feature with secret token, authorization code Acceptance Criteria: Only Valid user can login in to the application
non_process
restful service user login as a developer develop user login functionalities with security feature with secret token authorization code acceptance criteria only valid user can login in to the application
0
17,197
22,773,833,309
IssuesEvent
2022-07-08 12:41:41
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Missing "Networking" tab
automation/svc triaged assigned-to-author doc-bug process-automation/subsvc Pri2
Hello, the document doesn't include details on the "Networking" tab: <img width="732" alt="Screen Shot 2022-05-24 at 10 49 39" src="https://user-images.githubusercontent.com/1591516/170065395-e791d6c3-33c5-4e9a-94b5-658eea018e96.png"> --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 87f330cc-8989-9721-7d02-965dce0e988d * Version Independent ID: 4242498b-4503-2f16-9922-3c401e155b11 * Content: [Quickstart - Create an Azure Automation account using the portal](https://docs.microsoft.com/en-us/azure/automation/quickstarts/create-account-portal) * Content Source: [articles/automation/quickstarts/create-account-portal.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/quickstarts/create-account-portal.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @SGSneha * Microsoft Alias: **sudhirsneha**
1.0
Missing "Networking" tab - Hello, the document doesn't include details on the "Networking" tab: <img width="732" alt="Screen Shot 2022-05-24 at 10 49 39" src="https://user-images.githubusercontent.com/1591516/170065395-e791d6c3-33c5-4e9a-94b5-658eea018e96.png"> --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 87f330cc-8989-9721-7d02-965dce0e988d * Version Independent ID: 4242498b-4503-2f16-9922-3c401e155b11 * Content: [Quickstart - Create an Azure Automation account using the portal](https://docs.microsoft.com/en-us/azure/automation/quickstarts/create-account-portal) * Content Source: [articles/automation/quickstarts/create-account-portal.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/quickstarts/create-account-portal.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @SGSneha * Microsoft Alias: **sudhirsneha**
process
missing networking tab hello the document doesn t include details on the networking tab img width alt screen shot at src document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login sgsneha microsoft alias sudhirsneha
1
604
2,577,791,304
IssuesEvent
2015-02-12 19:09:33
chrsmith/quake2-gwt-port
https://api.github.com/repos/chrsmith/quake2-gwt-port
opened
can't connect to multiplayer from a computer other than the server
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. start the server 2. hit the url from another machine 3. try to join the multiplayer game What is the expected output? What do you see instead? Being in the game. Connecting <hostname> repeatedly. What version of the product are you using? On what operating system? Current Webkit nightly and Chromium nightly. ``` ----- Original issue reported on code.google.com by b...@getdown.org on 2 Apr 2010 at 2:09
1.0
can't connect to multiplayer from a computer other than the server - ``` What steps will reproduce the problem? 1. start the server 2. hit the url from another machine 3. try to join the multiplayer game What is the expected output? What do you see instead? Being in the game. Connecting <hostname> repeatedly. What version of the product are you using? On what operating system? Current Webkit nightly and Chromium nightly. ``` ----- Original issue reported on code.google.com by b...@getdown.org on 2 Apr 2010 at 2:09
non_process
can t connect to multiplayer from a computer other than the server what steps will reproduce the problem start the server hit the url from another machine try to join the multiplayer game what is the expected output what do you see instead being in the game connecting repeatedly what version of the product are you using on what operating system current webkit nightly and chromium nightly original issue reported on code google com by b getdown org on apr at
0
11,519
14,400,980,262
IssuesEvent
2020-12-03 13:08:07
gain-wyj/gain-wyj.github.io
https://api.github.com/repos/gain-wyj/gain-wyj.github.io
opened
数学分析模型(一):数据的无量纲处理方法及示例(附完整代码) | 梦魇回生的博客
2020/12/01/dimensionless-processing-of-data/ Gitalk
https://gain-wyj.cn/2020/12/01/dimensionless-processing-of-data/ 在对实际问题建模过程中,特别是在建立指标评价体系时,常常会面临不同类型的数据处理及融合。而各个指标之间由于计量单位和数量级的不尽相同,从而使得各指标间不具有可比性。在数据分析之前,通常需要先将数据标准化,利用标准化后的数据进行分析。数据标准
1.0
数学分析模型(一):数据的无量纲处理方法及示例(附完整代码) | 梦魇回生的博客 - https://gain-wyj.cn/2020/12/01/dimensionless-processing-of-data/ 在对实际问题建模过程中,特别是在建立指标评价体系时,常常会面临不同类型的数据处理及融合。而各个指标之间由于计量单位和数量级的不尽相同,从而使得各指标间不具有可比性。在数据分析之前,通常需要先将数据标准化,利用标准化后的数据进行分析。数据标准
process
数学分析模型(一):数据的无量纲处理方法及示例(附完整代码) 梦魇回生的博客 在对实际问题建模过程中,特别是在建立指标评价体系时,常常会面临不同类型的数据处理及融合。而各个指标之间由于计量单位和数量级的不尽相同,从而使得各指标间不具有可比性。在数据分析之前,通常需要先将数据标准化,利用标准化后的数据进行分析。数据标准
1
10,038
13,044,161,595
IssuesEvent
2020-07-29 03:47:24
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `AddDateDatetimeReal` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `AddDateDatetimeReal` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @lonng ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `AddDateDatetimeReal` from TiDB - ## Description Port the scalar function `AddDateDatetimeReal` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @lonng ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function adddatedatetimereal from tidb description port the scalar function adddatedatetimereal from tidb to coprocessor score mentor s lonng recommended skills rust programming learning materials already implemented expressions ported from tidb
1
121,523
25,982,875,242
IssuesEvent
2022-12-19 20:35:02
Clueless-Community/seamless-ui
https://api.github.com/repos/Clueless-Community/seamless-ui
closed
account options dropdown
codepeak 22
One need to make this component using `HTML` and `Tailwind CSS`. I would suggest to use [Tailwind Playgrounds](https://play.tailwindcss.com/) to make things faster and quicker. Here is a reference to the component. <img width="235" alt="Screenshot 2022-12-20 at 12 00 15 AM" src="https://user-images.githubusercontent.com/107763774/208495181-1181c129-dc48-43df-9db4-3040cdf17c5d.png"> After building the component please raise a PR with a screenshot of the component and add the component in `path of the file`. If you need to use any icon please use it from [Hero Icons](https://heroicons.com/) Good luck.
1.0
account options dropdown - One need to make this component using `HTML` and `Tailwind CSS`. I would suggest to use [Tailwind Playgrounds](https://play.tailwindcss.com/) to make things faster and quicker. Here is a reference to the component. <img width="235" alt="Screenshot 2022-12-20 at 12 00 15 AM" src="https://user-images.githubusercontent.com/107763774/208495181-1181c129-dc48-43df-9db4-3040cdf17c5d.png"> After building the component please raise a PR with a screenshot of the component and add the component in `path of the file`. If you need to use any icon please use it from [Hero Icons](https://heroicons.com/) Good luck.
non_process
account options dropdown one need to make this component using html and tailwind css i would suggest to use to make things faster and quicker here is a reference to the component img width alt screenshot at am src after building the component please raise a pr with a screenshot of the component and add the component in path of the file if you need to use any icon please use it from good luck
0
21,146
28,125,460,482
IssuesEvent
2023-03-31 17:18:57
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
NTR: maternal contribution terms
New term request RNA processes development organism-level process Needed for GO-CAM
general reference for all of these https://www.ncbi.nlm.nih.gov/books/NBK53192/ maternal contribution of gene product to offspring part_of oocyte development is_a developmental process involved in reproduction def: The process whose specific outcome is the contribution of maternal gene products to offspring for use until the offspring's genome is activated. maternally deposition of RNA in oocyte is_a maternal contribution of gene product to offspring def: The process whose specific outcome is the deposition of maternal RNA in the developing oocyte. maternally deposition of protein in oocyte is_amaternal contribution of gene product to offspring def: The process whose specific outcome is the deposition of maternal protein in the developing oocyte. regulation of maternally deposited RNA stability is_a regulation of mRNA stability def: Any process that modulates the propensity of maternally deposited RNA molecules to degradation. Includes processes that both stabilize and destabilize RNAs regulation of maternally deposited protein stability is_a regulation of protein stability def: Any process that affects the structure and integrity of a maternally deposited protein, altering the likelihood of its degradation or aggregation. Includes processes that both stabilize and destabilize proteins.
2.0
NTR: maternal contribution terms - general reference for all of these https://www.ncbi.nlm.nih.gov/books/NBK53192/ maternal contribution of gene product to offspring part_of oocyte development is_a developmental process involved in reproduction def: The process whose specific outcome is the contribution of maternal gene products to offspring for use until the offspring's genome is activated. maternally deposition of RNA in oocyte is_a maternal contribution of gene product to offspring def: The process whose specific outcome is the deposition of maternal RNA in the developing oocyte. maternally deposition of protein in oocyte is_amaternal contribution of gene product to offspring def: The process whose specific outcome is the deposition of maternal protein in the developing oocyte. regulation of maternally deposited RNA stability is_a regulation of mRNA stability def: Any process that modulates the propensity of maternally deposited RNA molecules to degradation. Includes processes that both stabilize and destabilize RNAs regulation of maternally deposited protein stability is_a regulation of protein stability def: Any process that affects the structure and integrity of a maternally deposited protein, altering the likelihood of its degradation or aggregation. Includes processes that both stabilize and destabilize proteins.
process
ntr maternal contribution terms general reference for all of these maternal contribution of gene product to offspring part of oocyte development is a developmental process involved in reproduction def the process whose specific outcome is the contribution of maternal gene products to offspring for use until the offspring s genome is activated maternally deposition of rna in oocyte is a maternal contribution of gene product to offspring def the process whose specific outcome is the deposition of maternal rna in the developing oocyte maternally deposition of protein in oocyte is amaternal contribution of gene product to offspring def the process whose specific outcome is the deposition of maternal protein in the developing oocyte regulation of maternally deposited rna stability is a regulation of mrna stability def any process that modulates the propensity of maternally deposited rna molecules to degradation includes processes that both stabilize and destabilize rnas regulation of maternally deposited protein stability is a regulation of protein stability def any process that affects the structure and integrity of a maternally deposited protein altering the likelihood of its degradation or aggregation includes processes that both stabilize and destabilize proteins
1
398,760
11,742,308,471
IssuesEvent
2020-03-12 00:18:56
thaliawww/concrexit
https://api.github.com/repos/thaliawww/concrexit
closed
'become a member' forms not present on website
bug priority: medium
In GitLab by @thomwiggers on Dec 21, 2017, 18:07 ### One-sentence description They're not on https://thalia.nu/admin/members/becomeamemberdocument/ / https://thalia.nu/become-a-member/
1.0
'become a member' forms not present on website - In GitLab by @thomwiggers on Dec 21, 2017, 18:07 ### One-sentence description They're not on https://thalia.nu/admin/members/becomeamemberdocument/ / https://thalia.nu/become-a-member/
non_process
become a member forms not present on website in gitlab by thomwiggers on dec one sentence description they re not on
0
115,737
11,886,450,348
IssuesEvent
2020-03-27 21:58:14
generaliinformatik/mycoradar
https://api.github.com/repos/generaliinformatik/mycoradar
closed
Translate README to english
documentation
Translate the README to english to address english speaking contributors. This is one of the tasks from the team vote of 25th March 2020. With the development of the website mycoradar.eu, which is also to be written in English, English-speaking interested parties may be increasingly addressed
1.0
Translate README to english - Translate the README to english to address english speaking contributors. This is one of the tasks from the team vote of 25th March 2020. With the development of the website mycoradar.eu, which is also to be written in English, English-speaking interested parties may be increasingly addressed
non_process
translate readme to english translate the readme to english to address english speaking contributors this is one of the tasks from the team vote of march with the development of the website mycoradar eu which is also to be written in english english speaking interested parties may be increasingly addressed
0
475,500
13,711,607,676
IssuesEvent
2020-10-02 05:05:00
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
cse.google.com - see bug description
browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-critical
<!-- @browser: Firefox 82.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:82.0) Gecko/20100101 Firefox/82.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/59149 --> **URL**: https://cse.google.com/cse?q=gjkfshsg&sa=Search&ie=UTF-8&cx=partner-pub%2D6638247779433690:3873384991#%9C&gsc.tab=0&gsc.q=youtube **Browser / Version**: Firefox 82.0 **Operating System**: Windows 7 **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: programable scarch engine started how to fix **Steps to Reproduce**: progamable screch engine <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/10/2267ed51-92aa-4513-bf6c-d1169c122ee0.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200926073307</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/10/2d30d28d-2dde-43b9-941c-0e58574e0704) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
cse.google.com - see bug description - <!-- @browser: Firefox 82.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:82.0) Gecko/20100101 Firefox/82.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/59149 --> **URL**: https://cse.google.com/cse?q=gjkfshsg&sa=Search&ie=UTF-8&cx=partner-pub%2D6638247779433690:3873384991#%9C&gsc.tab=0&gsc.q=youtube **Browser / Version**: Firefox 82.0 **Operating System**: Windows 7 **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: programable scarch engine started how to fix **Steps to Reproduce**: progamable screch engine <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/10/2267ed51-92aa-4513-bf6c-d1169c122ee0.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200926073307</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/10/2d30d28d-2dde-43b9-941c-0e58574e0704) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
cse google com see bug description url browser version firefox operating system windows tested another browser yes chrome problem type something else description programable scarch engine started how to fix steps to reproduce progamable screch engine view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
18,681
24,594,565,706
IssuesEvent
2022-10-14 07:08:40
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[FHIR] Questionnaire resource > JSON > Branching logic data is not getting mapped into the JSON file for text choice + Other option
Bug P1 Response datastore Process: Fixed Process: Tested QA Process: Tested dev
**AR:** Questionnaire resource > JSON > Branching logic data is not getting mapped into the JSON file for text choice + Other option **ER:** Branching logic data should get mapped into the JSON file as configured in the SB
3.0
[FHIR] Questionnaire resource > JSON > Branching logic data is not getting mapped into the JSON file for text choice + Other option - **AR:** Questionnaire resource > JSON > Branching logic data is not getting mapped into the JSON file for text choice + Other option **ER:** Branching logic data should get mapped into the JSON file as configured in the SB
process
questionnaire resource json branching logic data is not getting mapped into the json file for text choice other option ar questionnaire resource json branching logic data is not getting mapped into the json file for text choice other option er branching logic data should get mapped into the json file as configured in the sb
1
17,412
23,228,628,463
IssuesEvent
2022-08-03 04:44:38
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
Add more advanced actions to processing dialog, including copy as qgis_process command, copy/paste as json (Request in QGIS)
Processing 3.24
### Request for documentation From pull request QGIS/qgis#46514 Author: @nyalldawson QGIS version: 3.24 **Add more advanced actions to processing dialog, including copy as qgis_process command, copy/paste as json** ### PR Description: Together these actions: - Allow for easy generation of qgis_process commands via the QGIS gui, including any tricky parameter values like gpkg outputs with specific layer (refs https://github.com/qgis/QGIS/issues/38272) - Provide a way for users to copy the settings defined in the dialog in a text format, so they can store these easily and then restore later by pasting the value - Provides an easy way for users to copy the settings into the json format consumed by qgis_process (following #46497), so that it is easy for users to see the expected format even for complex parameters (like tin interpolation parameters) Refs paleolimbot/qgisprocess#56 Refs paleolimbot/qgisprocess#44 Sponsored by the Research Institute for Nature and Forest, Flemish Govt ### Commits tagged with [need-docs] or [FEATURE] "[feature] Add action in advanced menu for processing algorithm to copy\nequivalent qgis_process command\n\nAllows for easy generation of qgis_process commands via the QGIS gui" "[feature] Add action to processing algorithm advanced menu to copy the\ncurrent settings as JSON and paste json settings from clipboard\n\nThe utility of this is two fold:\n\n1. It provides a way for users to copy the settings defined in the\ndialog in a text format, so they can store these easily and then\nrestore later by pasting the value\n\n2. It provides an easy way for users to copy the settings into\nthe json format consumed by qgis_process (following\nhttps://github.com/qgis/QGIS/pull/46497), so that it is easy\nfor users to see the expected format even for complex parameters\n(like tin interpolation parameters)"
1.0
Add more advanced actions to processing dialog, including copy as qgis_process command, copy/paste as json (Request in QGIS) - ### Request for documentation From pull request QGIS/qgis#46514 Author: @nyalldawson QGIS version: 3.24 **Add more advanced actions to processing dialog, including copy as qgis_process command, copy/paste as json** ### PR Description: Together these actions: - Allow for easy generation of qgis_process commands via the QGIS gui, including any tricky parameter values like gpkg outputs with specific layer (refs https://github.com/qgis/QGIS/issues/38272) - Provide a way for users to copy the settings defined in the dialog in a text format, so they can store these easily and then restore later by pasting the value - Provides an easy way for users to copy the settings into the json format consumed by qgis_process (following #46497), so that it is easy for users to see the expected format even for complex parameters (like tin interpolation parameters) Refs paleolimbot/qgisprocess#56 Refs paleolimbot/qgisprocess#44 Sponsored by the Research Institute for Nature and Forest, Flemish Govt ### Commits tagged with [need-docs] or [FEATURE] "[feature] Add action in advanced menu for processing algorithm to copy\nequivalent qgis_process command\n\nAllows for easy generation of qgis_process commands via the QGIS gui" "[feature] Add action to processing algorithm advanced menu to copy the\ncurrent settings as JSON and paste json settings from clipboard\n\nThe utility of this is two fold:\n\n1. It provides a way for users to copy the settings defined in the\ndialog in a text format, so they can store these easily and then\nrestore later by pasting the value\n\n2. It provides an easy way for users to copy the settings into\nthe json format consumed by qgis_process (following\nhttps://github.com/qgis/QGIS/pull/46497), so that it is easy\nfor users to see the expected format even for complex parameters\n(like tin interpolation parameters)"
process
add more advanced actions to processing dialog including copy as qgis process command copy paste as json request in qgis request for documentation from pull request qgis qgis author nyalldawson qgis version add more advanced actions to processing dialog including copy as qgis process command copy paste as json pr description together these actions allow for easy generation of qgis process commands via the qgis gui including any tricky parameter values like gpkg outputs with specific layer refs provide a way for users to copy the settings defined in the dialog in a text format so they can store these easily and then restore later by pasting the value provides an easy way for users to copy the settings into the json format consumed by qgis process following so that it is easy for users to see the expected format even for complex parameters like tin interpolation parameters refs paleolimbot qgisprocess refs paleolimbot qgisprocess sponsored by the research institute for nature and forest flemish govt commits tagged with or add action in advanced menu for processing algorithm to copy nequivalent qgis process command n nallows for easy generation of qgis process commands via the qgis gui add action to processing algorithm advanced menu to copy the ncurrent settings as json and paste json settings from clipboard n nthe utility of this is two fold n it provides a way for users to copy the settings defined in the ndialog in a text format so they can store these easily and then nrestore later by pasting the value n it provides an easy way for users to copy the settings into nthe json format consumed by qgis process following n so that it is easy nfor users to see the expected format even for complex parameters n like tin interpolation parameters
1
8,287
11,453,542,964
IssuesEvent
2020-02-06 15:34:33
ION28/BLUESPAWN
https://api.github.com/repos/ION28/BLUESPAWN
closed
T1055, T1093, T1186 - Implement Process Related Hunt Techniques
difficulty/hard lang/c++ module/processes priority/low type/enhancement
Will implement after @jnmcd finishes up process analytics testing
1.0
T1055, T1093, T1186 - Implement Process Related Hunt Techniques - Will implement after @jnmcd finishes up process analytics testing
process
implement process related hunt techniques will implement after jnmcd finishes up process analytics testing
1
22,604
31,825,171,562
IssuesEvent
2023-09-14 07:00:02
0xPolygonMiden/miden-vm
https://api.github.com/repos/0xPolygonMiden/miden-vm
opened
Expand capabilities of the `debug` instruction
assembly processor
In #1069 we introduced the basic `debug` instruction which prints out the current state of the stack. We should expand the capabilities of this instruction to also enable printing of memory and procedure locals (as described in https://github.com/0xPolygonMiden/miden-vm/issues/988#issuecomment-1711099264).
1.0
Expand capabilities of the `debug` instruction - In #1069 we introduced the basic `debug` instruction which prints out the current state of the stack. We should expand the capabilities of this instruction to also enable printing of memory and procedure locals (as described in https://github.com/0xPolygonMiden/miden-vm/issues/988#issuecomment-1711099264).
process
expand capabilities of the debug instruction in we introduced the basic debug instruction which prints out the current state of the stack we should expand the capabilities of this instruction to also enable printing of memory and procedure locals as described in
1
11,657
14,519,299,281
IssuesEvent
2020-12-14 02:28:20
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Pseudohyphal growth definition is S. cerevisie-specific
cellular processes textual definition
This is a side issue brought up by Karen in #13157 so it should be a separate ticket. The current definition for GO:0007124 'pseudohyphal growth' specifically describes the situation in S. cerevisiae diploid cells. Other fungi do not necessarily require nitrogen limitation and an abundant fermentable carbon source to grow pseudohyphally. These conditions are specific to S. cerevisiae diploid strains. In S. cerevisiae haploids, these same conditions stimulate "(haploid) invasive growth."' I'm not sure if S. pombe requires the same conditions/limitations. Where it states that cells become elongated, this implies a non-elongated (yeast) switches but some fungi (Candida albicans) can switch from true hyphal growth to pdeudohyphal growth. In this case, "becomes elongated" is not correct. Both SGD and CGD have many annotations to this term and it is appropriate but the definition is too narrow and specific to S. cerevisiae. See PMID: 19347739. This definition should be broadened to remove the S. cerevisiae-specific facts. "The ability of cells to grow as a chain of physically attached, elongated cells in response to an environmental stimulus or stimuli." This broader definition is preferable because orthologs in Sc and Ca are often involved in this process and the definition should be correct for more than just S. cerevisiae. This or a similar definition for GO:0007124 and related will not cause any need for reannotation. @mah11 @srengel
1.0
Pseudohyphal growth definition is S. cerevisie-specific - This is a side issue brought up by Karen in #13157 so it should be a separate ticket. The current definition for GO:0007124 'pseudohyphal growth' specifically describes the situation in S. cerevisiae diploid cells. Other fungi do not necessarily require nitrogen limitation and an abundant fermentable carbon source to grow pseudohyphally. These conditions are specific to S. cerevisiae diploid strains. In S. cerevisiae haploids, these same conditions stimulate "(haploid) invasive growth."' I'm not sure if S. pombe requires the same conditions/limitations. Where it states that cells become elongated, this implies a non-elongated (yeast) switches but some fungi (Candida albicans) can switch from true hyphal growth to pdeudohyphal growth. In this case, "becomes elongated" is not correct. Both SGD and CGD have many annotations to this term and it is appropriate but the definition is too narrow and specific to S. cerevisiae. See PMID: 19347739. This definition should be broadened to remove the S. cerevisiae-specific facts. "The ability of cells to grow as a chain of physically attached, elongated cells in response to an environmental stimulus or stimuli." This broader definition is preferable because orthologs in Sc and Ca are often involved in this process and the definition should be correct for more than just S. cerevisiae. This or a similar definition for GO:0007124 and related will not cause any need for reannotation. @mah11 @srengel
process
pseudohyphal growth definition is s cerevisie specific this is a side issue brought up by karen in so it should be a separate ticket the current definition for go pseudohyphal growth specifically describes the situation in s cerevisiae diploid cells other fungi do not necessarily require nitrogen limitation and an abundant fermentable carbon source to grow pseudohyphally these conditions are specific to s cerevisiae diploid strains in s cerevisiae haploids these same conditions stimulate haploid invasive growth i m not sure if s pombe requires the same conditions limitations where it states that cells become elongated this implies a non elongated yeast switches but some fungi candida albicans can switch from true hyphal growth to pdeudohyphal growth in this case becomes elongated is not correct both sgd and cgd have many annotations to this term and it is appropriate but the definition is too narrow and specific to s cerevisiae see pmid this definition should be broadened to remove the s cerevisiae specific facts the ability of cells to grow as a chain of physically attached elongated cells in response to an environmental stimulus or stimuli this broader definition is preferable because orthologs in sc and ca are often involved in this process and the definition should be correct for more than just s cerevisiae this or a similar definition for go and related will not cause any need for reannotation srengel
1
568,183
16,961,668,191
IssuesEvent
2021-06-29 05:21:24
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Speedreader and reader mode icon is too small
OS/Desktop QA/No feature/speedreader priority/P3 release-notes/exclude
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> The Speedreader and reader mode icon is too small. ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Go to any readable page 2. See that the icon is too small ## Actual result: <!--Please add screenshots if needed--> ![image](https://user-images.githubusercontent.com/2388823/123704192-48610800-d81a-11eb-88a7-1a84eaadc3c6.png) ![image](https://user-images.githubusercontent.com/2388823/123704202-4c8d2580-d81a-11eb-8cb6-407757177a9f.png) ## Expected result: <img width="201" alt="Screen Shot 2021-06-28 at 2 08 35 PM" src="https://user-images.githubusercontent.com/2388823/123704236-56168d80-d81a-11eb-83f1-a9a5be2d7c74.png"> <img width="257" alt="Screen Shot 2021-06-28 at 2 08 44 PM" src="https://user-images.githubusercontent.com/2388823/123704264-5adb4180-d81a-11eb-8ebe-4249df61d6ae.png"> <img width="300" alt="Screen Shot 2021-06-28 at 2 08 56 PM" src="https://user-images.githubusercontent.com/2388823/123704312-62024f80-d81a-11eb-961e-306c13b29c4f.png"> <img width="337" alt="Screen Shot 2021-06-28 at 2 09 03 PM" src="https://user-images.githubusercontent.com/2388823/123704334-6595d680-d81a-11eb-8154-f12c0fa3cea7.png"> ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Easily ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? n/a - Can you reproduce this issue with the beta channel? n/a - Can you reproduce this issue with the nightly channel? Yes ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? n/a - Does the issue resolve itself when disabling Brave Rewards? n/a - Is the issue reproducible on the latest version of Chrome? n/a ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
1.0
Speedreader and reader mode icon is too small - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> The Speedreader and reader mode icon is too small. ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Go to any readable page 2. See that the icon is too small ## Actual result: <!--Please add screenshots if needed--> ![image](https://user-images.githubusercontent.com/2388823/123704192-48610800-d81a-11eb-88a7-1a84eaadc3c6.png) ![image](https://user-images.githubusercontent.com/2388823/123704202-4c8d2580-d81a-11eb-8cb6-407757177a9f.png) ## Expected result: <img width="201" alt="Screen Shot 2021-06-28 at 2 08 35 PM" src="https://user-images.githubusercontent.com/2388823/123704236-56168d80-d81a-11eb-83f1-a9a5be2d7c74.png"> <img width="257" alt="Screen Shot 2021-06-28 at 2 08 44 PM" src="https://user-images.githubusercontent.com/2388823/123704264-5adb4180-d81a-11eb-8ebe-4249df61d6ae.png"> <img width="300" alt="Screen Shot 2021-06-28 at 2 08 56 PM" src="https://user-images.githubusercontent.com/2388823/123704312-62024f80-d81a-11eb-961e-306c13b29c4f.png"> <img width="337" alt="Screen Shot 2021-06-28 at 2 09 03 PM" src="https://user-images.githubusercontent.com/2388823/123704334-6595d680-d81a-11eb-8154-f12c0fa3cea7.png"> ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Easily ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? n/a - Can you reproduce this issue with the beta channel? n/a - Can you reproduce this issue with the nightly channel? Yes ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? n/a - Does the issue resolve itself when disabling Brave Rewards? n/a - Is the issue reproducible on the latest version of Chrome? n/a ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
non_process
speedreader and reader mode icon is too small have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description the speedreader and reader mode icon is too small steps to reproduce go to any readable page see that the icon is too small actual result expected result img width alt screen shot at pm src img width alt screen shot at pm src img width alt screen shot at pm src img width alt screen shot at pm src reproduces how often easily brave version brave version info version channel information can you reproduce this issue with the current release n a can you reproduce this issue with the beta channel n a can you reproduce this issue with the nightly channel yes other additional information does the issue resolve itself when disabling brave shields n a does the issue resolve itself when disabling brave rewards n a is the issue reproducible on the latest version of chrome n a miscellaneous information
0
22,544
31,717,828,087
IssuesEvent
2023-09-10 03:49:28
Flow-Glow/Code-Jam-2023-Async-Aggregators
https://api.github.com/repos/Flow-Glow/Code-Jam-2023-Async-Aggregators
closed
Add real time image updating to UI
enhancement image processing
When performing edits to an image in the UI, we want the image to change in real time inside of the Pyqt window.
1.0
Add real time image updating to UI - When performing edits to an image in the UI, we want the image to change in real time inside of the Pyqt window.
process
add real time image updating to ui when performing edits to an image in the ui we want the image to change in real time inside of the pyqt window
1
1,885
4,712,358,396
IssuesEvent
2016-10-14 16:30:07
material-motion/material-motion-family-pop-swift
https://api.github.com/repos/material-motion/material-motion-family-pop-swift
closed
Publish the v1.0.0 release
Process
Must be ran by a member of the @material-motion/core-team `mdm release publish 1.0.0`
1.0
Publish the v1.0.0 release - Must be ran by a member of the @material-motion/core-team `mdm release publish 1.0.0`
process
publish the release must be ran by a member of the material motion core team mdm release publish
1
775,692
27,235,410,346
IssuesEvent
2023-02-21 15:59:13
ascheid/itsg33-pbmm-issue-gen
https://api.github.com/repos/ascheid/itsg33-pbmm-issue-gen
closed
AC-2(-1): Account Management
Priority: P2 Class: Technical ITSG-33 Control: AC-2 Suggested Assignment: IT Operations Group
#Control Definition ACCOUNT MANAGEMENT | AUTOMATED SYSTEM ACCOUNT MANAGEMENT The organization employs automated mechanisms to support the management of information system accounts. #Class Technical #Supplemental Guidance The use of automated mechanisms can include, for example: using email or text messaging to automatically notify account managers when users are terminated or transferred; using the information system to monitor account usage; and using telephonic notification to report atypical system account usage. #General Guide This security control/enhancement can be met using readily available Commercial-Off-The-Shelf (COTS) components. Consequently, inclusion of this security control/enhancement is strongly encouraged in most cases. #Suggested Placeholder Values X #Suggested Assignment IT Operations Group #Support Teams IT Projects
1.0
AC-2(-1): Account Management - #Control Definition ACCOUNT MANAGEMENT | AUTOMATED SYSTEM ACCOUNT MANAGEMENT The organization employs automated mechanisms to support the management of information system accounts. #Class Technical #Supplemental Guidance The use of automated mechanisms can include, for example: using email or text messaging to automatically notify account managers when users are terminated or transferred; using the information system to monitor account usage; and using telephonic notification to report atypical system account usage. #General Guide This security control/enhancement can be met using readily available Commercial-Off-The-Shelf (COTS) components. Consequently, inclusion of this security control/enhancement is strongly encouraged in most cases. #Suggested Placeholder Values X #Suggested Assignment IT Operations Group #Support Teams IT Projects
non_process
ac account management control definition account management automated system account management the organization employs automated mechanisms to support the management of information system accounts class technical supplemental guidance the use of automated mechanisms can include for example using email or text messaging to automatically notify account managers when users are terminated or transferred using the information system to monitor account usage and using telephonic notification to report atypical system account usage general guide this security control enhancement can be met using readily available commercial off the shelf cots components consequently inclusion of this security control enhancement is strongly encouraged in most cases suggested placeholder values x suggested assignment it operations group support teams it projects
0
630,704
20,116,707,105
IssuesEvent
2022-02-07 20:20:57
googleapis/python-api-core
https://api.github.com/repos/googleapis/python-api-core
closed
Field mask helper ignores optional fields explicitly set to default values
type: bug priority: p2 :rotating_light:
#### Environment details - OS type and version: glinux - Python version: `3.7.0` - pip version: `10.0.1` - `google-api-core` version: `1.31.0` #### Steps to reproduce Run the below code example. #### Code example ```python import proto from google.api_core.protobuf_helpers import field_mask class TestMessage(proto.Message): description = proto.Field(proto.STRING, number=1, optional=True) optimization_score = proto.Field(proto.INT64, number=2, optional=True) test = TestMessage() # Possibly related to Issue #230. When an optional field is explicitly set to its default value, # i.e. empty string or zero, the field mask helper does not recognize that it's been set. # Users must manually add the field path in these cases. test.description = "" test.optimization_score = 0 fm = field_mask(None, test._pb) assert fm.paths == ["description", "optimization_score"] ``` #### Stack trace ``` Traceback (most recent call last): File "meow.py", line 15, in <module> assert fm.paths == ["description", "optimization_score"] AssertionError ```
1.0
Field mask helper ignores optional fields explicitly set to default values - #### Environment details - OS type and version: glinux - Python version: `3.7.0` - pip version: `10.0.1` - `google-api-core` version: `1.31.0` #### Steps to reproduce Run the below code example. #### Code example ```python import proto from google.api_core.protobuf_helpers import field_mask class TestMessage(proto.Message): description = proto.Field(proto.STRING, number=1, optional=True) optimization_score = proto.Field(proto.INT64, number=2, optional=True) test = TestMessage() # Possibly related to Issue #230. When an optional field is explicitly set to its default value, # i.e. empty string or zero, the field mask helper does not recognize that it's been set. # Users must manually add the field path in these cases. test.description = "" test.optimization_score = 0 fm = field_mask(None, test._pb) assert fm.paths == ["description", "optimization_score"] ``` #### Stack trace ``` Traceback (most recent call last): File "meow.py", line 15, in <module> assert fm.paths == ["description", "optimization_score"] AssertionError ```
non_process
field mask helper ignores optional fields explicitly set to default values environment details os type and version glinux python version pip version google api core version steps to reproduce run the below code example code example python import proto from google api core protobuf helpers import field mask class testmessage proto message description proto field proto string number optional true optimization score proto field proto number optional true test testmessage possibly related to issue when an optional field is explicitly set to its default value i e empty string or zero the field mask helper does not recognize that it s been set users must manually add the field path in these cases test description test optimization score fm field mask none test pb assert fm paths stack trace traceback most recent call last file meow py line in assert fm paths assertionerror
0
17,401
23,218,983,949
IssuesEvent
2022-08-02 16:19:53
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
opened
testdata/runfiles_test/WORKSPACE.mock pins to a very old rules_cc
type: process
### Description of the bug: src/test/py/bazel/testdata/runfiles_test/WORKSPACE.mock uses an ancient commit from rules_cc. https://github.com/bazelbuild/rules_cc/commit/8bd6cd75d03c01bb82561a96d9c1f9f7157b13d0 We should generate this file from distdir_deps.bzl using gen_workspace_stanza. That will align the tests using this declaration with the other integration tests that use the rules_cc we build with. ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. No need to reproduce. This is a build file change. ### Which operating system are you running Bazel on? n/a ### What is the output of `bazel info release`? n/a ### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel. n/a ### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ? ```text n/a ``` ### Have you found anything relevant by searching the web? _No response_ ### Any other information, logs, or outputs that you want to share? _No response_
1.0
testdata/runfiles_test/WORKSPACE.mock pins to a very old rules_cc - ### Description of the bug: src/test/py/bazel/testdata/runfiles_test/WORKSPACE.mock uses an ancient commit from rules_cc. https://github.com/bazelbuild/rules_cc/commit/8bd6cd75d03c01bb82561a96d9c1f9f7157b13d0 We should generate this file from distdir_deps.bzl using gen_workspace_stanza. That will align the tests using this declaration with the other integration tests that use the rules_cc we build with. ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. No need to reproduce. This is a build file change. ### Which operating system are you running Bazel on? n/a ### What is the output of `bazel info release`? n/a ### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel. n/a ### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ? ```text n/a ``` ### Have you found anything relevant by searching the web? _No response_ ### Any other information, logs, or outputs that you want to share? _No response_
process
testdata runfiles test workspace mock pins to a very old rules cc description of the bug src test py bazel testdata runfiles test workspace mock uses an ancient commit from rules cc we should generate this file from distdir deps bzl using gen workspace stanza that will align the tests using this declaration with the other integration tests that use the rules cc we build with what s the simplest easiest way to reproduce this bug please provide a minimal example if possible no need to reproduce this is a build file change which operating system are you running bazel on n a what is the output of bazel info release n a if bazel info release returns development version or non git tell us how you built bazel n a what s the output of git remote get url origin git rev parse master git rev parse head text n a have you found anything relevant by searching the web no response any other information logs or outputs that you want to share no response
1
1,528
4,118,762,877
IssuesEvent
2016-06-08 12:48:44
World4Fly/Interface-for-Arduino
https://api.github.com/repos/World4Fly/Interface-for-Arduino
closed
Design complex system to process commands
process
Create the concept on how the firmware handles messages from the interface and sends them to it by considering verification and so on...
1.0
Design complex system to process commands - Create the concept on how the firmware handles messages from the interface and sends them to it by considering verification and so on...
process
design complex system to process commands create the concept on how the firmware handles messages from the interface and sends them to it by considering verification and so on
1
4,890
7,763,757,039
IssuesEvent
2018-06-01 17:43:53
StrikeNP/trac_test
https://api.github.com/repos/StrikeNP/trac_test
closed
Add a function to convert.m to changed a pressure profile into altitude (Trac #4)
Migrated from Trac enhancement fasching@uwm.edu post_processing
Add a function to convert.m to changed a pressure profile into altitude. This would be useful for cases that do not specify things in terms of altitude. Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/4 ```json { "status": "closed", "changetime": "2009-05-16T10:07:24", "description": "Add a function to convert.m to changed a pressure profile into altitude. This would be useful for cases that do not specify things in terms of altitude.", "reporter": "fasching@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1242468444000000", "component": "post_processing", "summary": "Add a function to convert.m to changed a pressure profile into altitude", "priority": "minor", "keywords": "conversion, MATLAB", "time": "2009-05-01T21:20:08", "milestone": "", "owner": "fasching@uwm.edu", "type": "enhancement" } ```
1.0
Add a function to convert.m to changed a pressure profile into altitude (Trac #4) - Add a function to convert.m to changed a pressure profile into altitude. This would be useful for cases that do not specify things in terms of altitude. Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/4 ```json { "status": "closed", "changetime": "2009-05-16T10:07:24", "description": "Add a function to convert.m to changed a pressure profile into altitude. This would be useful for cases that do not specify things in terms of altitude.", "reporter": "fasching@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1242468444000000", "component": "post_processing", "summary": "Add a function to convert.m to changed a pressure profile into altitude", "priority": "minor", "keywords": "conversion, MATLAB", "time": "2009-05-01T21:20:08", "milestone": "", "owner": "fasching@uwm.edu", "type": "enhancement" } ```
process
add a function to convert m to changed a pressure profile into altitude trac add a function to convert m to changed a pressure profile into altitude this would be useful for cases that do not specify things in terms of altitude migrated from json status closed changetime description add a function to convert m to changed a pressure profile into altitude this would be useful for cases that do not specify things in terms of altitude reporter fasching uwm edu cc resolution verified by v larson ts component post processing summary add a function to convert m to changed a pressure profile into altitude priority minor keywords conversion matlab time milestone owner fasching uwm edu type enhancement
1
172,199
27,243,694,636
IssuesEvent
2023-02-21 23:05:15
SAP/fundamental
https://api.github.com/repos/SAP/fundamental
closed
Website Improvement: Foundation Landing Page
Improvement Design Website Documentation UX Visual
Create some kind of information and visual treatment for this landing page.
1.0
Website Improvement: Foundation Landing Page - Create some kind of information and visual treatment for this landing page.
non_process
website improvement foundation landing page create some kind of information and visual treatment for this landing page
0
5,782
8,632,439,914
IssuesEvent
2018-11-22 10:47:56
kiwicom/orbit-components
https://api.github.com/repos/kiwicom/orbit-components
closed
<TripSector>'s 100% height breaks layout
Bug Processing
`<TripSector>` breaks layout as it has rule of 100% height. ## Expected Behavior This component shouldn't break layout <img width="622" alt="screen shot 2018-11-13 at 13 44 19" src="https://user-images.githubusercontent.com/3975660/48416306-16c1ce00-e750-11e8-954d-f6f33c2ce4b3.png">
1.0
<TripSector>'s 100% height breaks layout - `<TripSector>` breaks layout as it has rule of 100% height. ## Expected Behavior This component shouldn't break layout <img width="622" alt="screen shot 2018-11-13 at 13 44 19" src="https://user-images.githubusercontent.com/3975660/48416306-16c1ce00-e750-11e8-954d-f6f33c2ce4b3.png">
process
s height breaks layout breaks layout as it has rule of height expected behavior this component shouldn t break layout img width alt screen shot at src
1
15,187
18,956,009,498
IssuesEvent
2021-11-18 20:20:59
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
closed
Add `pkgconfig` aliasing to release notes
2.status: stale 6.topic: release process
Follow up item to #109887 Add in the release notes that it's been properly included into `aliases.nix` and will not be available if `config.allowAliases` is set to `false`.
1.0
Add `pkgconfig` aliasing to release notes - Follow up item to #109887 Add in the release notes that it's been properly included into `aliases.nix` and will not be available if `config.allowAliases` is set to `false`.
process
add pkgconfig aliasing to release notes follow up item to add in the release notes that it s been properly included into aliases nix and will not be available if config allowaliases is set to false
1
147,238
5,635,838,877
IssuesEvent
2017-04-06 02:34:17
Polymer/polymer-analyzer
https://api.github.com/repos/Polymer/polymer-analyzer
closed
Analyze @namespace jsdoc annotations
Priority: Critical Status: Accepted Type: Enhancement
We need to find variable declarations with a `@namespace` annotation and create ScannedNamespace and Namespace features. Examples: ```javascript /** * @namespace */ const Polymer = function() { … } /** * @namespace * @memberof Polymer */ Polymer.dom = function() { … } /** * @memberof Polymer.dom */ Polymer.dom.foo = …; /** * @memberof Polymer.dom */ Polymer.dom.bar = …; /** * @namespace */ Polymer.Async = { run() {...}, }; ``` `@namespace` can take a name to the right of it, otherwise we can get the name from the left side of the assignment. `@memberof` should be handled as a separate issue.
1.0
Analyze @namespace jsdoc annotations - We need to find variable declarations with a `@namespace` annotation and create ScannedNamespace and Namespace features. Examples: ```javascript /** * @namespace */ const Polymer = function() { … } /** * @namespace * @memberof Polymer */ Polymer.dom = function() { … } /** * @memberof Polymer.dom */ Polymer.dom.foo = …; /** * @memberof Polymer.dom */ Polymer.dom.bar = …; /** * @namespace */ Polymer.Async = { run() {...}, }; ``` `@namespace` can take a name to the right of it, otherwise we can get the name from the left side of the assignment. `@memberof` should be handled as a separate issue.
non_process
analyze namespace jsdoc annotations we need to find variable declarations with a namespace annotation and create scannednamespace and namespace features examples javascript namespace const polymer function … namespace memberof polymer polymer dom function … memberof polymer dom polymer dom foo … memberof polymer dom polymer dom bar … namespace polymer async run namespace can take a name to the right of it otherwise we can get the name from the left side of the assignment memberof should be handled as a separate issue
0
6,669
9,784,009,318
IssuesEvent
2019-06-08 15:30:11
initc3/HoneyBadgerMPC
https://api.github.com/repos/initc3/HoneyBadgerMPC
closed
Blockchain integration
applications avss secret sharing feature request hyperledger integration preprocessing
hbMPC should ship with three blockchain integrations: - [x] the built-in `commonsubset` from hbBFT - [ ] the hyperledger fabric integration - [x] web3 / solidity integration A blockchain integration consists of essentially four components: 1. Coordination of client input (through hbAVSS or preprocess-AVSS) 2. Initiation of MPC online phase 3. Managing offline phase 4. Making results available to the blockchain
1.0
Blockchain integration - hbMPC should ship with three blockchain integrations: - [x] the built-in `commonsubset` from hbBFT - [ ] the hyperledger fabric integration - [x] web3 / solidity integration A blockchain integration consists of essentially four components: 1. Coordination of client input (through hbAVSS or preprocess-AVSS) 2. Initiation of MPC online phase 3. Managing offline phase 4. Making results available to the blockchain
process
blockchain integration hbmpc should ship with three blockchain integrations the built in commonsubset from hbbft the hyperledger fabric integration solidity integration a blockchain integration consists of essentially four components coordination of client input through hbavss or preprocess avss initiation of mpc online phase managing offline phase making results available to the blockchain
1
372,788
11,028,168,950
IssuesEvent
2019-12-06 11:02:17
jdi-testing/jdi-light
https://api.github.com/repos/jdi-testing/jdi-light
closed
Attach screenshots to Allure
priority:high
While option screenshot.strategy=on fail in test.properties Attach screenshots to Allure report by default
1.0
Attach screenshots to Allure - While option screenshot.strategy=on fail in test.properties Attach screenshots to Allure report by default
non_process
attach screenshots to allure while option screenshot strategy on fail in test properties attach screenshots to allure report by default
0
19,217
25,352,684,285
IssuesEvent
2022-11-20 00:21:25
ethereum/EIPs
https://api.github.com/repos/ethereum/EIPs
closed
EIP editors should explain that ERC was renamed to EIP
w-stale question r-process
### Pull Request _No response_ ### What happened? The EIP editors have apparently decided that ERCs shall be renamed to EIPs. (I can see this in the CI bots' notices.) That's fine and I respect your control of this naming. Before I go around telling everybody that ERC-721 is now renamed to EIP-721, it will be helpful if the EIP official documentation does not contradict me. Please do these: - [ ] Update https://github.com/ethereum/EIPs/blob/master/README.md to use this new wording - [ ] Make a blog post documenting this significant change and link to it prominently from the README and EIP-1 ### Relevant log output _No response_
1.0
EIP editors should explain that ERC was renamed to EIP - ### Pull Request _No response_ ### What happened? The EIP editors have apparently decided that ERCs shall be renamed to EIPs. (I can see this in the CI bots' notices.) That's fine and I respect your control of this naming. Before I go around telling everybody that ERC-721 is now renamed to EIP-721, it will be helpful if the EIP official documentation does not contradict me. Please do these: - [ ] Update https://github.com/ethereum/EIPs/blob/master/README.md to use this new wording - [ ] Make a blog post documenting this significant change and link to it prominently from the README and EIP-1 ### Relevant log output _No response_
process
eip editors should explain that erc was renamed to eip pull request no response what happened the eip editors have apparently decided that ercs shall be renamed to eips i can see this in the ci bots notices that s fine and i respect your control of this naming before i go around telling everybody that erc is now renamed to eip it will be helpful if the eip official documentation does not contradict me please do these update to use this new wording make a blog post documenting this significant change and link to it prominently from the readme and eip relevant log output no response
1
6,346
9,389,961,017
IssuesEvent
2019-04-06 00:19:47
natario1/CameraView
https://api.github.com/repos/natario1/CameraView
closed
Throw errors when released frames are accessed
about:frame processing is:enhancement
### Problem to be addressed After releasing, `Frame`s are invalid but can still be inspected. ### Describe the solution you'd like `Frame` access after releasing should throw a readable error.
1.0
Throw errors when released frames are accessed - ### Problem to be addressed After releasing, `Frame`s are invalid but can still be inspected. ### Describe the solution you'd like `Frame` access after releasing should throw a readable error.
process
throw errors when released frames are accessed problem to be addressed after releasing frame s are invalid but can still be inspected describe the solution you d like frame access after releasing should throw a readable error
1
16,946
5,311,045,660
IssuesEvent
2017-02-13 01:03:15
WayofTime/BloodMagic
https://api.github.com/repos/WayofTime/BloodMagic
closed
PacketNoSpamChat isn't handled on main thread
1.10 1.11 bug code complete
#### Issue Description: `PacketNoSpamChat`'s handler doesn't seem to schedule itself on the main thread. Instead, it adds to chat on the current thread. #### What happens: Crash with a ConcurrentModificationException when rendering the chat (from tabbychat). I wrote some code to catch it, so now the stacktrace is more useful. https://gist.github.com/killjoy1221/efede90225e4511a03a483a602dddfe6 It's been fixed on my side, but it's gonna spam the log with stacktraces each time it happens outside the minecraft thread. #### Steps to reproduce: 1. Install TabbyChat 2. Hold right click with a sigil on the blood altar ____ #### Affected Versions (Do *not* use "latest"): - BloodMagic: commit: https://github.com/WayofTime/BloodMagic/commit/11fe41c654678cbf6d96427b82b73ec431081d31 (It's been seen in 1.8) - Minecraft: 1.11.2 - Forge: 13.20.0.2226
1.0
PacketNoSpamChat isn't handled on main thread - #### Issue Description: `PacketNoSpamChat`'s handler doesn't seem to schedule itself on the main thread. Instead, it adds to chat on the current thread. #### What happens: Crash with a ConcurrentModificationException when rendering the chat (from tabbychat). I wrote some code to catch it, so now the stacktrace is more useful. https://gist.github.com/killjoy1221/efede90225e4511a03a483a602dddfe6 It's been fixed on my side, but it's gonna spam the log with stacktraces each time it happens outside the minecraft thread. #### Steps to reproduce: 1. Install TabbyChat 2. Hold right click with a sigil on the blood altar ____ #### Affected Versions (Do *not* use "latest"): - BloodMagic: commit: https://github.com/WayofTime/BloodMagic/commit/11fe41c654678cbf6d96427b82b73ec431081d31 (It's been seen in 1.8) - Minecraft: 1.11.2 - Forge: 13.20.0.2226
non_process
packetnospamchat isn t handled on main thread issue description packetnospamchat s handler doesn t seem to schedule itself on the main thread instead it adds to chat on the current thread what happens crash with a concurrentmodificationexception when rendering the chat from tabbychat i wrote some code to catch it so now the stacktrace is more useful it s been fixed on my side but it s gonna spam the log with stacktraces each time it happens outside the minecraft thread steps to reproduce install tabbychat hold right click with a sigil on the blood altar affected versions do not use latest bloodmagic commit it s been seen in minecraft forge
0
90,188
3,812,655,548
IssuesEvent
2016-03-27 18:58:32
HubTurbo/HubTurbo
https://api.github.com/repos/HubTurbo/HubTurbo
closed
Create Issue Picker to choose issue easily
aspect-ui priority.high
## UI Design - text field for searching - suggestion menu that will pop-up when typing
1.0
Create Issue Picker to choose issue easily - ## UI Design - text field for searching - suggestion menu that will pop-up when typing
non_process
create issue picker to choose issue easily ui design text field for searching suggestion menu that will pop up when typing
0
9,494
12,487,035,615
IssuesEvent
2020-05-31 06:35:06
oppia/oppia-android
https://api.github.com/repos/oppia/oppia-android
closed
Add DragDropAndSort case in StateReceiver
Priority: Essential Status: Finished Type: Process Where: Exploration player
Add testing exploration file with handling rules for Drag & Drop in StateReciever Target PR date: 19 May 2020 Target completion date: 20 May 2020
1.0
Add DragDropAndSort case in StateReceiver - Add testing exploration file with handling rules for Drag & Drop in StateReciever Target PR date: 19 May 2020 Target completion date: 20 May 2020
process
add dragdropandsort case in statereceiver add testing exploration file with handling rules for drag drop in statereciever target pr date may target completion date may
1
501,899
14,536,166,260
IssuesEvent
2020-12-15 07:08:23
kubesphere/kubesphere
https://api.github.com/repos/kubesphere/kubesphere
closed
The submission record is not displayed on the pipeline activity details page
area/console kind/bug priority/medium stale
**Describe the Bug** The submission record is not displayed on the pipeline activity details page ![image](https://user-images.githubusercontent.com/36271543/90000814-34c0b780-dcc3-11ea-88e7-85d0001901c3.png) **Versions Used** host-k8s: v1.15.12 member-k8s: v1.18.6 /kind bug /area console /assign @leoendless /milestone 3.1.0 /priority medium
1.0
The submission record is not displayed on the pipeline activity details page - **Describe the Bug** The submission record is not displayed on the pipeline activity details page ![image](https://user-images.githubusercontent.com/36271543/90000814-34c0b780-dcc3-11ea-88e7-85d0001901c3.png) **Versions Used** host-k8s: v1.15.12 member-k8s: v1.18.6 /kind bug /area console /assign @leoendless /milestone 3.1.0 /priority medium
non_process
the submission record is not displayed on the pipeline activity details page describe the bug the submission record is not displayed on the pipeline activity details page versions used host member kind bug area console assign leoendless milestone priority medium
0
49,706
26,268,808,133
IssuesEvent
2023-01-06 15:04:37
NVIDIA/spark-rapids
https://api.github.com/repos/NVIDIA/spark-rapids
closed
[BUG] Parquet Chunked reader causing performance overhead
bug performance
**Describe the bug** I have a customer job where the parquet chunked reader is causing the stage to go from 2.1 minutes to 3 minutes. It looks like with the chunked reader we have about 3 times the number of batches and the shuffle read and/or coalescebatches is taking much longer. Needs more investigation.
True
[BUG] Parquet Chunked reader causing performance overhead - **Describe the bug** I have a customer job where the parquet chunked reader is causing the stage to go from 2.1 minutes to 3 minutes. It looks like with the chunked reader we have about 3 times the number of batches and the shuffle read and/or coalescebatches is taking much longer. Needs more investigation.
non_process
parquet chunked reader causing performance overhead describe the bug i have a customer job where the parquet chunked reader is causing the stage to go from minutes to minutes it looks like with the chunked reader we have about times the number of batches and the shuffle read and or coalescebatches is taking much longer needs more investigation
0
92,506
15,857,087,389
IssuesEvent
2021-04-08 03:56:11
Thezone1975/send
https://api.github.com/repos/Thezone1975/send
opened
CVE-2020-8237 (High) detected in json-bigint-0.3.0.tgz
security vulnerability
## CVE-2020-8237 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-bigint-0.3.0.tgz</b></p></summary> <p>JSON.parse with bigints support</p> <p>Library home page: <a href="https://registry.npmjs.org/json-bigint/-/json-bigint-0.3.0.tgz">https://registry.npmjs.org/json-bigint/-/json-bigint-0.3.0.tgz</a></p> <p>Path to dependency file: /send/package.json</p> <p>Path to vulnerable library: send/node_modules/json-bigint/package.json</p> <p> Dependency Hierarchy: - storage-3.0.1.tgz (Root Library) - common-2.0.1.tgz - google-auth-library-4.2.1.tgz - gcp-metadata-2.0.0.tgz - :x: **json-bigint-0.3.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution in json-bigint npm package < 1.0.0 may lead to a denial-of-service (DoS) attack. <p>Publish Date: 2020-09-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8237>CVE-2020-8237</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://hackerone.com/reports/916430">https://hackerone.com/reports/916430</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: v1.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-8237 (High) detected in json-bigint-0.3.0.tgz - ## CVE-2020-8237 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-bigint-0.3.0.tgz</b></p></summary> <p>JSON.parse with bigints support</p> <p>Library home page: <a href="https://registry.npmjs.org/json-bigint/-/json-bigint-0.3.0.tgz">https://registry.npmjs.org/json-bigint/-/json-bigint-0.3.0.tgz</a></p> <p>Path to dependency file: /send/package.json</p> <p>Path to vulnerable library: send/node_modules/json-bigint/package.json</p> <p> Dependency Hierarchy: - storage-3.0.1.tgz (Root Library) - common-2.0.1.tgz - google-auth-library-4.2.1.tgz - gcp-metadata-2.0.0.tgz - :x: **json-bigint-0.3.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution in json-bigint npm package < 1.0.0 may lead to a denial-of-service (DoS) attack. <p>Publish Date: 2020-09-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8237>CVE-2020-8237</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://hackerone.com/reports/916430">https://hackerone.com/reports/916430</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: v1.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in json bigint tgz cve high severity vulnerability vulnerable library json bigint tgz json parse with bigints support library home page a href path to dependency file send package json path to vulnerable library send node modules json bigint package json dependency hierarchy storage tgz root library common tgz google auth library tgz gcp metadata tgz x json bigint tgz vulnerable library vulnerability details prototype pollution in json bigint npm package may lead to a denial of service dos attack publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
711,352
24,460,116,162
IssuesEvent
2022-10-07 10:20:12
wazuh/wazuh-documentation
https://api.github.com/repos/wazuh/wazuh-documentation
closed
Release notes for 3.13.1/3.13.2 not showing support for Open Distro 1.9
priority: high type: bug
Hi team! The release notes for [3.13.1](https://documentation.wazuh.com/3.13/release-notes/release_3_13_1.html) and [3.13.2](https://documentation.wazuh.com/3.13/release-notes/release_3_13_2.html) are not showing that they need Open Distro 1.9 to work as specify [here](https://github.com/wazuh/wazuh-kibana-app#wazuh---kibana---open-distro-version-compatibility-matrix) Regards.
1.0
Release notes for 3.13.1/3.13.2 not showing support for Open Distro 1.9 - Hi team! The release notes for [3.13.1](https://documentation.wazuh.com/3.13/release-notes/release_3_13_1.html) and [3.13.2](https://documentation.wazuh.com/3.13/release-notes/release_3_13_2.html) are not showing that they need Open Distro 1.9 to work as specify [here](https://github.com/wazuh/wazuh-kibana-app#wazuh---kibana---open-distro-version-compatibility-matrix) Regards.
non_process
release notes for not showing support for open distro hi team the release notes for and are not showing that they need open distro to work as specify regards
0
191,881
15,306,361,749
IssuesEvent
2021-02-24 19:22:19
ryanheise/just_audio
https://api.github.com/repos/ryanheise/just_audio
opened
Tutorial, written/video
1 backlog documentation
<!-- PLEASE READ CAREFULLY! FOR YOUR DOCUMENTATION REQUEST TO BE PROCESSED, YOU WILL NEED TO FILL IN ALL SECTIONS BELOW. DON'T DELETE THE HEADINGS. THANK YOU :-D --> **To which pages does your suggestion apply?** - Direct URL 1 - Direct URL 2 - ... **Quote the sentences(s) from the documentation to be improved (if any)** > Insert here. (Skip if you are proposing an entirely new section.) **Describe your suggestion** I apologize if this is the wrong venue for my question but is there a tutorial, written or video, on how to get started with this audio player? ...
1.0
Tutorial, written/video - <!-- PLEASE READ CAREFULLY! FOR YOUR DOCUMENTATION REQUEST TO BE PROCESSED, YOU WILL NEED TO FILL IN ALL SECTIONS BELOW. DON'T DELETE THE HEADINGS. THANK YOU :-D --> **To which pages does your suggestion apply?** - Direct URL 1 - Direct URL 2 - ... **Quote the sentences(s) from the documentation to be improved (if any)** > Insert here. (Skip if you are proposing an entirely new section.) **Describe your suggestion** I apologize if this is the wrong venue for my question but is there a tutorial, written or video, on how to get started with this audio player? ...
non_process
tutorial written video please read carefully for your documentation request to be processed you will need to fill in all sections below don t delete the headings thank you d to which pages does your suggestion apply direct url direct url quote the sentences s from the documentation to be improved if any insert here skip if you are proposing an entirely new section describe your suggestion i apologize if this is the wrong venue for my question but is there a tutorial written or video on how to get started with this audio player
0
13,381
15,859,241,584
IssuesEvent
2021-04-08 07:47:14
arcus-azure/arcus.messaging
https://api.github.com/repos/arcus-azure/arcus.messaging
closed
Move message handling types to abstractions projects
area:message-processing breaking-change enhancement
**Is your feature request related to a problem? Please describe.** The Azure Service Bus message pump project contains several message handling abstractions that is now tightly-coupled to the message pump project. It would be good to extract these interfaces in a separate project. **Describe the solution you'd like** * Move the `MessageHandling` types from the `Arcus.Messaging.Pumps.Abstractions` to the `Arcus.Messaging.Abstractions` project/namespace * Move the `MessageHandling` types from the `Arcus.Messaging.Pumps.ServiceBus` to the `Arcus.Messaging.ServiceBus.Abstractions` project/namespace.
1.0
Move message handling types to abstractions projects - **Is your feature request related to a problem? Please describe.** The Azure Service Bus message pump project contains several message handling abstractions that is now tightly-coupled to the message pump project. It would be good to extract these interfaces in a separate project. **Describe the solution you'd like** * Move the `MessageHandling` types from the `Arcus.Messaging.Pumps.Abstractions` to the `Arcus.Messaging.Abstractions` project/namespace * Move the `MessageHandling` types from the `Arcus.Messaging.Pumps.ServiceBus` to the `Arcus.Messaging.ServiceBus.Abstractions` project/namespace.
process
move message handling types to abstractions projects is your feature request related to a problem please describe the azure service bus message pump project contains several message handling abstractions that is now tightly coupled to the message pump project it would be good to extract these interfaces in a separate project describe the solution you d like move the messagehandling types from the arcus messaging pumps abstractions to the arcus messaging abstractions project namespace move the messagehandling types from the arcus messaging pumps servicebus to the arcus messaging servicebus abstractions project namespace
1
5,766
8,609,080,822
IssuesEvent
2018-11-18 18:07:12
RosenborgSupporterSoftware/RBKweb-ext
https://api.github.com/repos/RosenborgSupporterSoftware/RBKweb-ext
opened
Håndtering av engelsk/norsk språk i foruminnstillinger
process
Dette vil nok ikke påvirke så mye, men noe kan det være. Vi bør etterstrebe at utvidelsen fungerer helt likt uavhengig av hvilket språk som er valgt i innstillinger på phpbb. Det er enkelte steder i RFS hvor dette ikke var tilfelle (husker ikke hvilke i farten), men dette er noe å være obs på.
1.0
Håndtering av engelsk/norsk språk i foruminnstillinger - Dette vil nok ikke påvirke så mye, men noe kan det være. Vi bør etterstrebe at utvidelsen fungerer helt likt uavhengig av hvilket språk som er valgt i innstillinger på phpbb. Det er enkelte steder i RFS hvor dette ikke var tilfelle (husker ikke hvilke i farten), men dette er noe å være obs på.
process
håndtering av engelsk norsk språk i foruminnstillinger dette vil nok ikke påvirke så mye men noe kan det være vi bør etterstrebe at utvidelsen fungerer helt likt uavhengig av hvilket språk som er valgt i innstillinger på phpbb det er enkelte steder i rfs hvor dette ikke var tilfelle husker ikke hvilke i farten men dette er noe å være obs på
1
11,373
14,215,317,384
IssuesEvent
2020-11-17 07:09:04
microsoft/react-native-windows
https://api.github.com/repos/microsoft/react-native-windows
closed
Update promote-release Script to Mark Most Packages in Stable Branches as Private
Area: Release Process enhancement
There are packages like react-native-windows-init that we publish only from our master branch. Publishing outside of our master branch would break both PR and publish pipelines in our master branch, along with creating confusion. The lack of guardrails for this is an especially sharp edge as of 0.63, as bumps to the react-native-windows version will now bump react-native-windows-init, as the latter takes a devDependency on the former to use its typings. We should prevent the wrong thing from happening by marking most packages as private in stable branches. This can be done pretty easily in the `promote-release` script, to mark any package.json file as private that doesn't add the `promoteRelease: true` flag.
1.0
Update promote-release Script to Mark Most Packages in Stable Branches as Private - There are packages like react-native-windows-init that we publish only from our master branch. Publishing outside of our master branch would break both PR and publish pipelines in our master branch, along with creating confusion. The lack of guardrails for this is an especially sharp edge as of 0.63, as bumps to the react-native-windows version will now bump react-native-windows-init, as the latter takes a devDependency on the former to use its typings. We should prevent the wrong thing from happening by marking most packages as private in stable branches. This can be done pretty easily in the `promote-release` script, to mark any package.json file as private that doesn't add the `promoteRelease: true` flag.
process
update promote release script to mark most packages in stable branches as private there are packages like react native windows init that we publish only from our master branch publishing outside of our master branch would break both pr and publish pipelines in our master branch along with creating confusion the lack of guardrails for this is an especially sharp edge as of as bumps to the react native windows version will now bump react native windows init as the latter takes a devdependency on the former to use its typings we should prevent the wrong thing from happening by marking most packages as private in stable branches this can be done pretty easily in the promote release script to mark any package json file as private that doesn t add the promoterelease true flag
1
140,037
21,002,256,156
IssuesEvent
2022-03-29 18:39:20
zesty-io/nextjs-website
https://api.github.com/repos/zesty-io/nextjs-website
opened
Pricing Page - H1 and H2
CSS Design SEO
https://www.zesty.io/pricing/ Can we please change the title "Headless CMS Pricing" to an H3 and the sentence below it ("Flexible headless CMS pricing for where your business is at and where you plan to go") to an H1? Keep the same styling.
1.0
Pricing Page - H1 and H2 - https://www.zesty.io/pricing/ Can we please change the title "Headless CMS Pricing" to an H3 and the sentence below it ("Flexible headless CMS pricing for where your business is at and where you plan to go") to an H1? Keep the same styling.
non_process
pricing page and can we please change the title headless cms pricing to an and the sentence below it flexible headless cms pricing for where your business is at and where you plan to go to an keep the same styling
0
9,796
12,810,607,752
IssuesEvent
2020-07-03 19:17:06
Maximus5/ConEmu
https://api.github.com/repos/Maximus5/ConEmu
closed
FR: Support variable errorlevel for use in a CMD prompt
processes
Hi, I'm trying to add the errorlevel (saved in the internal variable `%errorlevel%`) to my CMD prompt. The expected result is, that the errorlevel of the last command is displayed if it exits and the prompt appears again. There's already the possibility to add environment variables with an ANSI sequence: `$E]9;8;"USERNAME"$E\` adds the variable `USERNAME` to the prompt, which works fine and is possible for any environment variable displayed by the `SET` command. Anyway `$E]9;8;"ERRORLEVEL"$E\` adds an empty string to the prompt, `ERRORLEVEL` is an internal variable and not displayed by `SET`, but can be accessed in scripts like any environment variable by `%ERRORLEVEL%`
1.0
FR: Support variable errorlevel for use in a CMD prompt - Hi, I'm trying to add the errorlevel (saved in the internal variable `%errorlevel%`) to my CMD prompt. The expected result is, that the errorlevel of the last command is displayed if it exits and the prompt appears again. There's already the possibility to add environment variables with an ANSI sequence: `$E]9;8;"USERNAME"$E\` adds the variable `USERNAME` to the prompt, which works fine and is possible for any environment variable displayed by the `SET` command. Anyway `$E]9;8;"ERRORLEVEL"$E\` adds an empty string to the prompt, `ERRORLEVEL` is an internal variable and not displayed by `SET`, but can be accessed in scripts like any environment variable by `%ERRORLEVEL%`
process
fr support variable errorlevel for use in a cmd prompt hi i m trying to add the errorlevel saved in the internal variable errorlevel to my cmd prompt the expected result is that the errorlevel of the last command is displayed if it exits and the prompt appears again there s already the possibility to add environment variables with an ansi sequence e username e adds the variable username to the prompt which works fine and is possible for any environment variable displayed by the set command anyway e errorlevel e adds an empty string to the prompt errorlevel is an internal variable and not displayed by set but can be accessed in scripts like any environment variable by errorlevel
1
14,599
17,703,578,276
IssuesEvent
2021-08-25 03:19:14
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
closed
Change term - country
Term - change Class - Location non-normative Process - complete
## Change term * Submitter: John Wieczorek (following issue raised by Ian Engelbrecht @ianengelbrecht) * Justification (why is this change necessary?): Clarity * Proponents (who needs this change): Everyone Current Term definition: https://dwc.tdwg.org/list/#dwc_country Proposed new attributes of the term: * Term name (in lowerCamelCase): country * Organized in Class (e.g. Location, Taxon): Location * Definition of the term: The name of the country or major administrative unit in which the Location occurs. * Usage comments (recommendations regarding content, etc.): Recommended best practice is to use a controlled vocabulary such as the Getty Thesaurus of Geographic Names. **Recommended best practice is to leave this field blank if the Location spans multiple entities at this administrative level or if the Location might be in one or another of multiple possible entities at this level. Multiplicity and uncertainty of the geographic entity can be captured either in the term higherGeography or in the term locality, or both.** * Examples: `Denmark`, `Colombia`, `España` * Refines (identifier of the broader term this term refines, if applicable): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): http://rs.tdwg.org/dwc/terms/version/country-2017-10-06 * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG, if applicable): DataSets/DataSet/Units/Unit/Gathering/Country/Name This change proposal arises from discussions in Issue #221 and https://github.com/tdwg/dwc-qa/issues/141. I would like to recommend the same exact amendment to the usage notes for each of the geography terms continent, waterbody, islandGroup, island, stateProvince, country, municipality.
1.0
Change term - country - ## Change term * Submitter: John Wieczorek (following issue raised by Ian Engelbrecht @ianengelbrecht) * Justification (why is this change necessary?): Clarity * Proponents (who needs this change): Everyone Current Term definition: https://dwc.tdwg.org/list/#dwc_country Proposed new attributes of the term: * Term name (in lowerCamelCase): country * Organized in Class (e.g. Location, Taxon): Location * Definition of the term: The name of the country or major administrative unit in which the Location occurs. * Usage comments (recommendations regarding content, etc.): Recommended best practice is to use a controlled vocabulary such as the Getty Thesaurus of Geographic Names. **Recommended best practice is to leave this field blank if the Location spans multiple entities at this administrative level or if the Location might be in one or another of multiple possible entities at this level. Multiplicity and uncertainty of the geographic entity can be captured either in the term higherGeography or in the term locality, or both.** * Examples: `Denmark`, `Colombia`, `España` * Refines (identifier of the broader term this term refines, if applicable): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): http://rs.tdwg.org/dwc/terms/version/country-2017-10-06 * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG, if applicable): DataSets/DataSet/Units/Unit/Gathering/Country/Name This change proposal arises from discussions in Issue #221 and https://github.com/tdwg/dwc-qa/issues/141. I would like to recommend the same exact amendment to the usage notes for each of the geography terms continent, waterbody, islandGroup, island, stateProvince, country, municipality.
process
change term country change term submitter john wieczorek following issue raised by ian engelbrecht ianengelbrecht justification why is this change necessary clarity proponents who needs this change everyone current term definition proposed new attributes of the term term name in lowercamelcase country organized in class e g location taxon location definition of the term the name of the country or major administrative unit in which the location occurs usage comments recommendations regarding content etc recommended best practice is to use a controlled vocabulary such as the getty thesaurus of geographic names recommended best practice is to leave this field blank if the location spans multiple entities at this administrative level or if the location might be in one or another of multiple possible entities at this level multiplicity and uncertainty of the geographic entity can be captured either in the term highergeography or in the term locality or both examples denmark colombia españa refines identifier of the broader term this term refines if applicable none replaces identifier of the existing term that would be deprecated and replaced by this term if applicable abcd xpath of the equivalent term in abcd or efg if applicable datasets dataset units unit gathering country name this change proposal arises from discussions in issue and i would like to recommend the same exact amendment to the usage notes for each of the geography terms continent waterbody islandgroup island stateprovince country municipality
1
175,196
6,547,979,260
IssuesEvent
2017-09-04 17:53:04
commonality/archetypes
https://api.github.com/repos/commonality/archetypes
closed
feat(quantity): specify the Quantity archetype pattern with OpenAPI 2.0
Points: 8 Priority: Medium Status: Completed Type: Docs Type: Feature
## User story As an API consumer, I want to consistently describe the amount of something measured according to some standard of measurement, In order to conduct business unambiguously and clearly. ## Acceptance criteria - [x] 1. `Metric` represents a standard of measurement for quantities. - [x] 2. `Unit` represents a type of `Metrics` that is part of a `SystemOfUnits` - [x] 3. `SystemOfUnits` represents a set of related `Units` defined by a standard such as the International System of Units (SI) - [x] 4. `SiBaseUnit` provides common properties of SI base units - [x] 4.1. `Ampere` represents an SI unit of electric current - [x] 4.2. `Candela` represents an SI unit of luminous intensity - [x] 4.3. `Kelvin` represents an SI unit of temperature - [x] 4.4. `Kilogram` represents an SI unit of mass - [x] 4.5 `Meter` represents an SI unit of length - [x] 4.6. `Mole` represents an SI unit of amount of substance - [x] 4.7. `Second` represents an SI unit of time - [x] 5. `DerivedUnit` represents a combination of one or more base `Units` according to a specific equation - [x] 5.1. `DerivedUnitTerm` represents a term, i.e., a single `Unit` and its power within a `DerivedUnit` - [x] 6. `LaborHour` represents the amount of work corresponding to one person working for one hour. - [x] 7. `Quantity` specifies an amount that is measured in some `Metric` and can be used for - [x] 7.1. Arithmetic operations - [x] 7.2. Comparison operations - [x] 7.3. Rounding operations with a - [x] 7.3.1. `RoundingPolicy` that defines the mathematical semantics of the rounding operation, and a - [x] 7.3.2. `RoundingStrategy` that represents the type of rounding to be applied - [x] 8. Conversion with - [x] 8.1. `StandardConversion` defines a `conversionFactor` that can be used to convert a source `Quantity` to a `Quantity` in a `targetUnit` - [x] 8.2. `UnitConverter` represents a conversion process ## Compliance with standards | Standard | Contents | |----------|-----------| | [SI][si-url] | International System of Units (BIPM) | ## Note All archetypes have been defined, but not all of them have direct RESTful services that can operate on values. [si-url]: http://www.bipm.org/en/measurement-units/
1.0
feat(quantity): specify the Quantity archetype pattern with OpenAPI 2.0 - ## User story As an API consumer, I want to consistently describe the amount of something measured according to some standard of measurement, In order to conduct business unambiguously and clearly. ## Acceptance criteria - [x] 1. `Metric` represents a standard of measurement for quantities. - [x] 2. `Unit` represents a type of `Metrics` that is part of a `SystemOfUnits` - [x] 3. `SystemOfUnits` represents a set of related `Units` defined by a standard such as the International System of Units (SI) - [x] 4. `SiBaseUnit` provides common properties of SI base units - [x] 4.1. `Ampere` represents an SI unit of electric current - [x] 4.2. `Candela` represents an SI unit of luminous intensity - [x] 4.3. `Kelvin` represents an SI unit of temperature - [x] 4.4. `Kilogram` represents an SI unit of mass - [x] 4.5 `Meter` represents an SI unit of length - [x] 4.6. `Mole` represents an SI unit of amount of substance - [x] 4.7. `Second` represents an SI unit of time - [x] 5. `DerivedUnit` represents a combination of one or more base `Units` according to a specific equation - [x] 5.1. `DerivedUnitTerm` represents a term, i.e., a single `Unit` and its power within a `DerivedUnit` - [x] 6. `LaborHour` represents the amount of work corresponding to one person working for one hour. - [x] 7. `Quantity` specifies an amount that is measured in some `Metric` and can be used for - [x] 7.1. Arithmetic operations - [x] 7.2. Comparison operations - [x] 7.3. Rounding operations with a - [x] 7.3.1. `RoundingPolicy` that defines the mathematical semantics of the rounding operation, and a - [x] 7.3.2. `RoundingStrategy` that represents the type of rounding to be applied - [x] 8. Conversion with - [x] 8.1. `StandardConversion` defines a `conversionFactor` that can be used to convert a source `Quantity` to a `Quantity` in a `targetUnit` - [x] 8.2. `UnitConverter` represents a conversion process ## Compliance with standards | Standard | Contents | |----------|-----------| | [SI][si-url] | International System of Units (BIPM) | ## Note All archetypes have been defined, but not all of them have direct RESTful services that can operate on values. [si-url]: http://www.bipm.org/en/measurement-units/
non_process
feat quantity specify the quantity archetype pattern with openapi user story as an api consumer i want to consistently describe the amount of something measured according to some standard of measurement in order to conduct business unambiguously and clearly acceptance criteria metric represents a standard of measurement for quantities unit represents a type of metrics that is part of a systemofunits systemofunits represents a set of related units defined by a standard such as the international system of units si sibaseunit provides common properties of si base units ampere represents an si unit of electric current candela represents an si unit of luminous intensity kelvin represents an si unit of temperature kilogram represents an si unit of mass meter represents an si unit of length mole represents an si unit of amount of substance second represents an si unit of time derivedunit represents a combination of one or more base units according to a specific equation derivedunitterm represents a term i e a single unit and its power within a derivedunit laborhour represents the amount of work corresponding to one person working for one hour quantity specifies an amount that is measured in some metric and can be used for arithmetic operations comparison operations rounding operations with a roundingpolicy that defines the mathematical semantics of the rounding operation and a roundingstrategy that represents the type of rounding to be applied conversion with standardconversion defines a conversionfactor that can be used to convert a source quantity to a quantity in a targetunit unitconverter represents a conversion process compliance with standards standard contents international system of units bipm note all archetypes have been defined but not all of them have direct restful services that can operate on values
0
6,060
8,887,298,730
IssuesEvent
2019-01-15 05:00:45
mick-warehime/sixth_corp
https://api.github.com/repos/mick-warehime/sixth_corp
closed
remove this comment from character_factory.build
development process
# This is an intermediate fix as in future these things will be handled # by the chassis. chassis = data.chassis_type.build() in the future create more issues and link them to these comments or use the TODO comment keyword so we dont forget anything
1.0
remove this comment from character_factory.build - # This is an intermediate fix as in future these things will be handled # by the chassis. chassis = data.chassis_type.build() in the future create more issues and link them to these comments or use the TODO comment keyword so we dont forget anything
process
remove this comment from character factory build this is an intermediate fix as in future these things will be handled by the chassis chassis data chassis type build in the future create more issues and link them to these comments or use the todo comment keyword so we dont forget anything
1