Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
592,475
17,893,711,185
IssuesEvent
2021-09-08 05:13:12
francheska-vicente/cssweng
https://api.github.com/repos/francheska-vicente/cssweng
opened
Discounts do not reflect when booking for Bridal Family Room
bug priority: high issue: back-end severity: high
### Summary: - Only for Bridal Family Room, when inputting discounts, the total cost does not change ### Steps to Reproduce: - Create Booking (Room 999, Sept 8-10, 2 pax, 2 PWD, 2000PHP discount) ### Visual Proof: ![image](https://user-images.githubusercontent.com/75743382/132449619-1a811adb-87dd-4527-bf7b-e8f6cc24dd51.png) ![image](https://user-images.githubusercontent.com/75743382/132449593-61501861-58b6-4f18-97c6-efd9dab4637b.png) ### Expected Results: - Total cost would change ### Actual Results: - Total cost remained the same | Additional Information | | | ----------- | ----------- | | Platform | V8 engine (Google) | | Operating System | Windows 10 |
1.0
Discounts do not reflect when booking for Bridal Family Room - ### Summary: - Only for Bridal Family Room, when inputting discounts, the total cost does not change ### Steps to Reproduce: - Create Booking (Room 999, Sept 8-10, 2 pax, 2 PWD, 2000PHP discount) ### Visual Proof: ![image](https://user-images.githubusercontent.com/75743382/132449619-1a811adb-87dd-4527-bf7b-e8f6cc24dd51.png) ![image](https://user-images.githubusercontent.com/75743382/132449593-61501861-58b6-4f18-97c6-efd9dab4637b.png) ### Expected Results: - Total cost would change ### Actual Results: - Total cost remained the same | Additional Information | | | ----------- | ----------- | | Platform | V8 engine (Google) | | Operating System | Windows 10 |
non_process
discounts do not reflect when booking for bridal family room summary only for bridal family room when inputting discounts the total cost does not change steps to reproduce create booking room sept pax pwd discount visual proof expected results total cost would change actual results total cost remained the same additional information platform engine google operating system windows
0
13,366
15,832,121,377
IssuesEvent
2021-04-06 14:18:40
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Create a utility package
kind/tech process/candidate team/client
Many tools are scattered around in the monorepo. What if we created a package for it? This could improve type safety and centralize updates on those tools.
1.0
Create a utility package - Many tools are scattered around in the monorepo. What if we created a package for it? This could improve type safety and centralize updates on those tools.
process
create a utility package many tools are scattered around in the monorepo what if we created a package for it this could improve type safety and centralize updates on those tools
1
7,464
10,562,919,925
IssuesEvent
2019-10-04 19:34:21
w3c/transitions
https://api.github.com/repos/w3c/transitions
closed
Use the real Github issues templates
Process Issue
It would be a better integration if you could use the real issue templates provided by GitHub See: https://help.github.com/articles/about-issue-and-pull-request-templates/#issue-templates (I discussed this with @deniak)
1.0
Use the real Github issues templates - It would be a better integration if you could use the real issue templates provided by GitHub See: https://help.github.com/articles/about-issue-and-pull-request-templates/#issue-templates (I discussed this with @deniak)
process
use the real github issues templates it would be a better integration if you could use the real issue templates provided by github see i discussed this with deniak
1
15,178
18,951,001,890
IssuesEvent
2021-11-18 15:10:19
googleapis/google-auth-library-python
https://api.github.com/repos/googleapis/google-auth-library-python
closed
Intersphinx URL for 'requests-oauthlib' now 404s
type: process
From [this failed build](https://source.cloud.google.com/results/invocations/56d76e91-941f-45a5-b4e4-6b74a9e86461/targets/cloud-devrel%2Fclient-libraries%2Fpython%2Fgoogleapis%2Fgoogle-auth-library-python%2Fpresubmit%2Fpresubmit/log): ```python nox > Running session docs nox > Creating virtual environment (virtualenv) using python3.8 in .nox/docs nox > python -m pip install -e .[aiohttp] nox > python -m pip install sphinx alabaster recommonmark sphinx-docstring-typing nox > sphinx-build -T -W -N -b html -d docs/_build/doctrees/ docs/ docs/_build/html/ Running Sphinx v4.3.0 making output directory... done loading intersphinx inventory from https://docs.python.org/3.5/objects.inv... loading intersphinx inventory from https://urllib3.readthedocs.io/en/stable/objects.inv... loading intersphinx inventory from https://requests.kennethreitz.org/en/master/objects.inv... loading intersphinx inventory from https://requests-oauthlib.readthedocs.io/en/stable/objects.inv... Traceback (most recent call last): File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/cmd/build.py", line 276, in build_main app = Sphinx(args.sourcedir, args.confdir, args.outputdir, File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/application.py", line 270, in __init__ self._init_builder() File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/application.py", line 329, in _init_builder self.events.emit('builder-inited') File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/events.py", line 101, in emit results.append(listener.handler(self.app, *args)) File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/ext/intersphinx.py", line 240, in load_mappings updated = [f.result() for f in concurrent.futures.as_completed(futures)] File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/ext/intersphinx.py", line 240, in <listcomp> updated = [f.result() for f in concurrent.futures.as_completed(futures)] File "/usr/local/lib/python3.8/concurrent/futures/_base.py", line 437, in result return self.__get_result() File "/usr/local/lib/python3.8/concurrent/futures/_base.py", line 389, in __get_result raise self._exception File "/usr/local/lib/python3.8/concurrent/futures/thread.py", line 57, in run result = self.fn(*self.args, **self.kwargs) File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/ext/intersphinx.py", line 225, in fetch_inventory_group logger.warning(__("failed to reach any of the inventories " File "/usr/local/lib/python3.8/logging/__init__.py", line 1812, in warning self.log(WARNING, msg, *args, **kwargs) File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/util/logging.py", line 123, in log super().log(level, msg, *args, **kwargs) File "/usr/local/lib/python3.8/logging/__init__.py", line 1844, in log self.logger.log(level, msg, *args, **kwargs) File "/usr/local/lib/python3.8/logging/__init__.py", line 1512, in log self._log(level, msg, args, **kwargs) File "/usr/local/lib/python3.8/logging/__init__.py", line 1589, in _log self.handle(record) File "/usr/local/lib/python3.8/logging/__init__.py", line 1599, in handle self.callHandlers(record) File "/usr/local/lib/python3.8/logging/__init__.py", line 1661, in callHandlers hdlr.handle(record) File "/usr/local/lib/python3.8/logging/__init__.py", line 950, in handle rv = self.filter(record) File "/usr/local/lib/python3.8/logging/__init__.py", line 811, in filter result = f.filter(record) File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/util/logging.py", line 430, in filter raise exc sphinx.errors.SphinxWarning: failed to reach any of the inventories with the following issues: intersphinx inventory 'https://requests-oauthlib.readthedocs.io/en/stable/objects.inv' not fetchable due to <class 'requests.exceptions.HTTPError'>: 404 Client Error: Not Found for url: https://requests-oauthlib.readthedocs.io/en/stable/objects.inv ``` I have no idea why they would yank the `stable` docs version. The only versions I can see today for the [`requests-oauthlib` docs](https://requests-oauthlib.readthedocs.io/en/latest/) are `latest` and `v1.3.0-docs`. Given that `1.3.0` is the [latest release on PyPI](https://pypi.org/project/requests-oauthlib/#history), and is two years old, I think we're better off using the corresponding version.
1.0
Intersphinx URL for 'requests-oauthlib' now 404s - From [this failed build](https://source.cloud.google.com/results/invocations/56d76e91-941f-45a5-b4e4-6b74a9e86461/targets/cloud-devrel%2Fclient-libraries%2Fpython%2Fgoogleapis%2Fgoogle-auth-library-python%2Fpresubmit%2Fpresubmit/log): ```python nox > Running session docs nox > Creating virtual environment (virtualenv) using python3.8 in .nox/docs nox > python -m pip install -e .[aiohttp] nox > python -m pip install sphinx alabaster recommonmark sphinx-docstring-typing nox > sphinx-build -T -W -N -b html -d docs/_build/doctrees/ docs/ docs/_build/html/ Running Sphinx v4.3.0 making output directory... done loading intersphinx inventory from https://docs.python.org/3.5/objects.inv... loading intersphinx inventory from https://urllib3.readthedocs.io/en/stable/objects.inv... loading intersphinx inventory from https://requests.kennethreitz.org/en/master/objects.inv... loading intersphinx inventory from https://requests-oauthlib.readthedocs.io/en/stable/objects.inv... Traceback (most recent call last): File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/cmd/build.py", line 276, in build_main app = Sphinx(args.sourcedir, args.confdir, args.outputdir, File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/application.py", line 270, in __init__ self._init_builder() File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/application.py", line 329, in _init_builder self.events.emit('builder-inited') File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/events.py", line 101, in emit results.append(listener.handler(self.app, *args)) File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/ext/intersphinx.py", line 240, in load_mappings updated = [f.result() for f in concurrent.futures.as_completed(futures)] File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/ext/intersphinx.py", line 240, in <listcomp> updated = [f.result() for f in concurrent.futures.as_completed(futures)] File "/usr/local/lib/python3.8/concurrent/futures/_base.py", line 437, in result return self.__get_result() File "/usr/local/lib/python3.8/concurrent/futures/_base.py", line 389, in __get_result raise self._exception File "/usr/local/lib/python3.8/concurrent/futures/thread.py", line 57, in run result = self.fn(*self.args, **self.kwargs) File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/ext/intersphinx.py", line 225, in fetch_inventory_group logger.warning(__("failed to reach any of the inventories " File "/usr/local/lib/python3.8/logging/__init__.py", line 1812, in warning self.log(WARNING, msg, *args, **kwargs) File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/util/logging.py", line 123, in log super().log(level, msg, *args, **kwargs) File "/usr/local/lib/python3.8/logging/__init__.py", line 1844, in log self.logger.log(level, msg, *args, **kwargs) File "/usr/local/lib/python3.8/logging/__init__.py", line 1512, in log self._log(level, msg, args, **kwargs) File "/usr/local/lib/python3.8/logging/__init__.py", line 1589, in _log self.handle(record) File "/usr/local/lib/python3.8/logging/__init__.py", line 1599, in handle self.callHandlers(record) File "/usr/local/lib/python3.8/logging/__init__.py", line 1661, in callHandlers hdlr.handle(record) File "/usr/local/lib/python3.8/logging/__init__.py", line 950, in handle rv = self.filter(record) File "/usr/local/lib/python3.8/logging/__init__.py", line 811, in filter result = f.filter(record) File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/util/logging.py", line 430, in filter raise exc sphinx.errors.SphinxWarning: failed to reach any of the inventories with the following issues: intersphinx inventory 'https://requests-oauthlib.readthedocs.io/en/stable/objects.inv' not fetchable due to <class 'requests.exceptions.HTTPError'>: 404 Client Error: Not Found for url: https://requests-oauthlib.readthedocs.io/en/stable/objects.inv ``` I have no idea why they would yank the `stable` docs version. The only versions I can see today for the [`requests-oauthlib` docs](https://requests-oauthlib.readthedocs.io/en/latest/) are `latest` and `v1.3.0-docs`. Given that `1.3.0` is the [latest release on PyPI](https://pypi.org/project/requests-oauthlib/#history), and is two years old, I think we're better off using the corresponding version.
process
intersphinx url for requests oauthlib now from python nox running session docs nox creating virtual environment virtualenv using in nox docs nox python m pip install e nox python m pip install sphinx alabaster recommonmark sphinx docstring typing nox sphinx build t w n b html d docs build doctrees docs docs build html running sphinx making output directory done loading intersphinx inventory from loading intersphinx inventory from loading intersphinx inventory from loading intersphinx inventory from traceback most recent call last file tmpfs src github google auth library python nox docs lib site packages sphinx cmd build py line in build main app sphinx args sourcedir args confdir args outputdir file tmpfs src github google auth library python nox docs lib site packages sphinx application py line in init self init builder file tmpfs src github google auth library python nox docs lib site packages sphinx application py line in init builder self events emit builder inited file tmpfs src github google auth library python nox docs lib site packages sphinx events py line in emit results append listener handler self app args file tmpfs src github google auth library python nox docs lib site packages sphinx ext intersphinx py line in load mappings updated file tmpfs src github google auth library python nox docs lib site packages sphinx ext intersphinx py line in updated file usr local lib concurrent futures base py line in result return self get result file usr local lib concurrent futures base py line in get result raise self exception file usr local lib concurrent futures thread py line in run result self fn self args self kwargs file tmpfs src github google auth library python nox docs lib site packages sphinx ext intersphinx py line in fetch inventory group logger warning failed to reach any of the inventories file usr local lib logging init py line in warning self log warning msg args kwargs file tmpfs src github google auth library python nox docs lib site packages sphinx util logging py line in log super log level msg args kwargs file usr local lib logging init py line in log self logger log level msg args kwargs file usr local lib logging init py line in log self log level msg args kwargs file usr local lib logging init py line in log self handle record file usr local lib logging init py line in handle self callhandlers record file usr local lib logging init py line in callhandlers hdlr handle record file usr local lib logging init py line in handle rv self filter record file usr local lib logging init py line in filter result f filter record file tmpfs src github google auth library python nox docs lib site packages sphinx util logging py line in filter raise exc sphinx errors sphinxwarning failed to reach any of the inventories with the following issues intersphinx inventory not fetchable due to client error not found for url i have no idea why they would yank the stable docs version the only versions i can see today for the are latest and docs given that is the and is two years old i think we re better off using the corresponding version
1
441,234
12,709,669,513
IssuesEvent
2020-06-23 12:44:54
grpc/grpc
https://api.github.com/repos/grpc/grpc
opened
Invalid TLS handshakes from the grpc cpp client when connecting to a specific IP
kind/bug priority/P2
### What version of gRPC and what language are you using? Latest, C++ ### What operating system (Linux, Windows,...) and version? Linux, Centos 7 ### What runtime / compiler are you using (e.g. python version or version of gcc) n/a ### What did you do? **Background on the TLS Server Name Indication extension** > Server Name Indication (SNI) is an extension to the Transport Layer Security (TLS) computer networking protocol by which a client indicates which hostname it is attempting to connect to at the start of the handshaking process.[1] This allows a server to present multiple certificates on the same IP address and TCP port number and hence allows multiple secure (HTTPS) websites (or any other service over TLS) to be served by the same IP address without requiring all those sites to use the same certificate. https://en.wikipedia.org/wiki/Server_Name_Indication According to the SNI TLS extension spec the SNI in a TLS handshake must be set to a fully qualified domain name, not an IP address: > "HostName" contains the fully qualified DNS hostname of the server, as understood by the client. ... Literal IPv4 and IPv6 addresses are not permitted in "HostName". https://tools.ietf.org/html/rfc6066#section-3 **Bug** In our setup we have a grpc server running with a valid certificate for the IPv6 address its bound to. We're attempting to use the grpc c++ client to connect directly to this server via its IPv6 address and port directly, rather than using a fully qualified domain name. When doing so the grpc client incorrectly sets the Server Name Indicator (SNI) in its TLS handshake to the target IPv6 address. Our grpc server implementation then rejects the request due to the SNI being invalid. ### What did you expect to see? I expect grpc to only use the SNI TLS extension when its connecting to a domain name. This will make grpc clients use of SNI spec compliant. ### What did you see instead? grpc always uses the SNI extension, regardless of whether its connecting to an IP address or a hostname. ### Anything else we should know about your project / environment? If changing grpc so that it doesn't set the SNI when connecting directly to an IP address sounds like a valid solution I'm happy to implement it. Code pointers or advice is appreciated.
1.0
Invalid TLS handshakes from the grpc cpp client when connecting to a specific IP - ### What version of gRPC and what language are you using? Latest, C++ ### What operating system (Linux, Windows,...) and version? Linux, Centos 7 ### What runtime / compiler are you using (e.g. python version or version of gcc) n/a ### What did you do? **Background on the TLS Server Name Indication extension** > Server Name Indication (SNI) is an extension to the Transport Layer Security (TLS) computer networking protocol by which a client indicates which hostname it is attempting to connect to at the start of the handshaking process.[1] This allows a server to present multiple certificates on the same IP address and TCP port number and hence allows multiple secure (HTTPS) websites (or any other service over TLS) to be served by the same IP address without requiring all those sites to use the same certificate. https://en.wikipedia.org/wiki/Server_Name_Indication According to the SNI TLS extension spec the SNI in a TLS handshake must be set to a fully qualified domain name, not an IP address: > "HostName" contains the fully qualified DNS hostname of the server, as understood by the client. ... Literal IPv4 and IPv6 addresses are not permitted in "HostName". https://tools.ietf.org/html/rfc6066#section-3 **Bug** In our setup we have a grpc server running with a valid certificate for the IPv6 address its bound to. We're attempting to use the grpc c++ client to connect directly to this server via its IPv6 address and port directly, rather than using a fully qualified domain name. When doing so the grpc client incorrectly sets the Server Name Indicator (SNI) in its TLS handshake to the target IPv6 address. Our grpc server implementation then rejects the request due to the SNI being invalid. ### What did you expect to see? I expect grpc to only use the SNI TLS extension when its connecting to a domain name. This will make grpc clients use of SNI spec compliant. ### What did you see instead? grpc always uses the SNI extension, regardless of whether its connecting to an IP address or a hostname. ### Anything else we should know about your project / environment? If changing grpc so that it doesn't set the SNI when connecting directly to an IP address sounds like a valid solution I'm happy to implement it. Code pointers or advice is appreciated.
non_process
invalid tls handshakes from the grpc cpp client when connecting to a specific ip what version of grpc and what language are you using latest c what operating system linux windows and version linux centos what runtime compiler are you using e g python version or version of gcc n a what did you do background on the tls server name indication extension server name indication sni is an extension to the transport layer security tls computer networking protocol by which a client indicates which hostname it is attempting to connect to at the start of the handshaking process this allows a server to present multiple certificates on the same ip address and tcp port number and hence allows multiple secure https websites or any other service over tls to be served by the same ip address without requiring all those sites to use the same certificate according to the sni tls extension spec the sni in a tls handshake must be set to a fully qualified domain name not an ip address hostname contains the fully qualified dns hostname of the server as understood by the client literal and addresses are not permitted in hostname bug in our setup we have a grpc server running with a valid certificate for the address its bound to we re attempting to use the grpc c client to connect directly to this server via its address and port directly rather than using a fully qualified domain name when doing so the grpc client incorrectly sets the server name indicator sni in its tls handshake to the target address our grpc server implementation then rejects the request due to the sni being invalid what did you expect to see i expect grpc to only use the sni tls extension when its connecting to a domain name this will make grpc clients use of sni spec compliant what did you see instead grpc always uses the sni extension regardless of whether its connecting to an ip address or a hostname anything else we should know about your project environment if changing grpc so that it doesn t set the sni when connecting directly to an ip address sounds like a valid solution i m happy to implement it code pointers or advice is appreciated
0
485,842
14,000,311,017
IssuesEvent
2020-10-28 12:08:11
EyeSeeTea/training-app
https://api.github.com/repos/EyeSeeTea/training-app
closed
Close behavior on tutorial
complexity - low (1 hr) priority - low priority - medium status - blocked type - cosmetic type - feedback
- Pop-up window on shutting down tutorial telling users what is happening and how they can relaunch the tool (or minimize if they prefer)
2.0
Close behavior on tutorial - - Pop-up window on shutting down tutorial telling users what is happening and how they can relaunch the tool (or minimize if they prefer)
non_process
close behavior on tutorial pop up window on shutting down tutorial telling users what is happening and how they can relaunch the tool or minimize if they prefer
0
6,027
7,468,143,929
IssuesEvent
2018-04-02 17:55:45
acelabini/ibinex-option-one
https://api.github.com/repos/acelabini/ibinex-option-one
closed
[ SERVICES ] Orientation and size of phone icon in nav/header
DESKTOP/WEB NAV/HEADER SERVICES
Tester: Joel Simpao OS: Windows 10 Browser: Chrome Expected: ![expected nav phone icon](https://user-images.githubusercontent.com/37061492/38185890-cabe55f6-3682-11e8-8eb2-84f3a994b150.PNG) Actual: ![actual nav phone icon](https://user-images.githubusercontent.com/37061492/38185899-d8b8b20a-3682-11e8-840e-0088c18f7263.PNG)
1.0
[ SERVICES ] Orientation and size of phone icon in nav/header - Tester: Joel Simpao OS: Windows 10 Browser: Chrome Expected: ![expected nav phone icon](https://user-images.githubusercontent.com/37061492/38185890-cabe55f6-3682-11e8-8eb2-84f3a994b150.PNG) Actual: ![actual nav phone icon](https://user-images.githubusercontent.com/37061492/38185899-d8b8b20a-3682-11e8-840e-0088c18f7263.PNG)
non_process
orientation and size of phone icon in nav header tester joel simpao os windows browser chrome expected actual
0
1,454
4,029,504,009
IssuesEvent
2016-05-18 10:54:38
nodejs/node
https://api.github.com/repos/nodejs/node
closed
child_process ignores cwd
child_process windows
<!-- Thank you for reporting an issue. Please fill in the template below. If unsure about something, just do as best as you're able. Version: usually output of `node -v` Platform: either `uname -a` output, or if Windows, version and 32 or 64-bit Subsystem: if known, please specify affected core module name It will be much easier for us to fix the issue if a test case that reproduces the problem is provided. Ideally this test case should not have any external dependencies. We understand that it is not always possible to reduce your code to a small test case, but we would appreciate to have as much data as possible. --> * **v4.4.4**: * **Windows 7 64bit**: * **child_process**: <!-- Enter your issue details below this comment. --> I run from the command line using `node test.js` the following file: ``` var exec = require('child_process').exec; exec('cd',{cwd: "C:\\Code\\Misc"},function(error,stdout,stderr) { console.log(stdout); }); ``` gives the output `C:\Users\myname` `test.js` is located in `C:\Code\Misc`. Any Idea why?? Shouldn't it give `C:\Code\Misc`? the cwd I specified? This causes problems when trying to use the Atom text editor 'build' add-on. I have recently re-installed node and npm, but the error persists.
1.0
child_process ignores cwd - <!-- Thank you for reporting an issue. Please fill in the template below. If unsure about something, just do as best as you're able. Version: usually output of `node -v` Platform: either `uname -a` output, or if Windows, version and 32 or 64-bit Subsystem: if known, please specify affected core module name It will be much easier for us to fix the issue if a test case that reproduces the problem is provided. Ideally this test case should not have any external dependencies. We understand that it is not always possible to reduce your code to a small test case, but we would appreciate to have as much data as possible. --> * **v4.4.4**: * **Windows 7 64bit**: * **child_process**: <!-- Enter your issue details below this comment. --> I run from the command line using `node test.js` the following file: ``` var exec = require('child_process').exec; exec('cd',{cwd: "C:\\Code\\Misc"},function(error,stdout,stderr) { console.log(stdout); }); ``` gives the output `C:\Users\myname` `test.js` is located in `C:\Code\Misc`. Any Idea why?? Shouldn't it give `C:\Code\Misc`? the cwd I specified? This causes problems when trying to use the Atom text editor 'build' add-on. I have recently re-installed node and npm, but the error persists.
process
child process ignores cwd thank you for reporting an issue please fill in the template below if unsure about something just do as best as you re able version usually output of node v platform either uname a output or if windows version and or bit subsystem if known please specify affected core module name it will be much easier for us to fix the issue if a test case that reproduces the problem is provided ideally this test case should not have any external dependencies we understand that it is not always possible to reduce your code to a small test case but we would appreciate to have as much data as possible windows child process i run from the command line using node test js the following file var exec require child process exec exec cd cwd c code misc function error stdout stderr console log stdout gives the output c users myname test js is located in c code misc any idea why shouldn t it give c code misc the cwd i specified this causes problems when trying to use the atom text editor build add on i have recently re installed node and npm but the error persists
1
9,181
12,227,846,644
IssuesEvent
2020-05-03 16:54:16
emacs-ess/ESS
https://api.github.com/repos/emacs-ess/ESS
closed
`ess-set-working-directory` should not change `default-directory`
process
I've been confused for years with the behaviour of Emacs regarding the active directories of the files I'm editing. I just realised it's `ess-set-working-directory` that changes it to the new working directory of R. It seems that this behaviour is intended: https://github.com/emacs-ess/ESS/commit/40ebf0f71eb0677895506201fac6da4d6119ffa2. But I'm not sure why changing the REPL directory should affect the buffer I'm switching from. Should we change this to preserve the active directory? The active directory influences UI commands like `M-x find-file`. I suspect I'm not the only one who is confused by this behaviour.
1.0
`ess-set-working-directory` should not change `default-directory` - I've been confused for years with the behaviour of Emacs regarding the active directories of the files I'm editing. I just realised it's `ess-set-working-directory` that changes it to the new working directory of R. It seems that this behaviour is intended: https://github.com/emacs-ess/ESS/commit/40ebf0f71eb0677895506201fac6da4d6119ffa2. But I'm not sure why changing the REPL directory should affect the buffer I'm switching from. Should we change this to preserve the active directory? The active directory influences UI commands like `M-x find-file`. I suspect I'm not the only one who is confused by this behaviour.
process
ess set working directory should not change default directory i ve been confused for years with the behaviour of emacs regarding the active directories of the files i m editing i just realised it s ess set working directory that changes it to the new working directory of r it seems that this behaviour is intended but i m not sure why changing the repl directory should affect the buffer i m switching from should we change this to preserve the active directory the active directory influences ui commands like m x find file i suspect i m not the only one who is confused by this behaviour
1
18,021
24,032,779,074
IssuesEvent
2022-09-15 16:18:44
googleapis/java-beyondcorp-appgateways
https://api.github.com/repos/googleapis/java-beyondcorp-appgateways
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'beyondcorp-appgateways' invalid in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'beyondcorp-appgateways' invalid in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname beyondcorp appgateways invalid in repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
1
65,642
8,824,579,129
IssuesEvent
2019-01-02 17:40:29
LycheeOrg/Lychee-Laravel
https://api.github.com/repos/LycheeOrg/Lychee-Laravel
closed
Add `.env` file
Documentation
`.env.exmple` is helpful for learning how to configure Laravel, but should the repo contain a pre-configured `.env` for the user? Even an empty document so they encryption key is added to the correct place will make installation easier.
1.0
Add `.env` file - `.env.exmple` is helpful for learning how to configure Laravel, but should the repo contain a pre-configured `.env` for the user? Even an empty document so they encryption key is added to the correct place will make installation easier.
non_process
add env file env exmple is helpful for learning how to configure laravel but should the repo contain a pre configured env for the user even an empty document so they encryption key is added to the correct place will make installation easier
0
14,578
17,702,955,893
IssuesEvent
2021-08-25 01:58:12
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
closed
Change term - associatedTaxa
Term - change Class - Occurrence Class - ResourceRelationship non-normative Process - complete
## Change term * Submitter: John Wieczorek * Justification (why is this change necessary?): Consistency and clarity * Proponents (who needs this change): Everyone Current Term definition: https://dwc.tdwg.org/terms/#dwc:associatedTaxa Proposed new attributes of the term: * Term name (in lowerCamelCase): associatedTaxa * Organized in Class (e.g. Location, Taxon): Occurrence * Definition of the term: **A list (concatenated and separated) of identifiers or names of taxa and the associations of this Occurrence to each of them.** * Usage comments (recommendations regarding content, etc.): **This term can be used to provide a list of associations to Taxa other than the one defined in the Occurrence. Note that the ResourceRelationship class is an alternative means of representing associations, and with more detail. This term is not apt for establishing relationships between Taxa, only between specific Occurrences of an Organism with other Taxa. Recommended best practice is to separate the values in a list with space vertical bar space ( | ).** * Examples: `"host":"Quercus alba"`, **`"host":"gbif.org/species/2879737"`**,`"parasitoid of":"Cyclocephala signaticollis" | "predator of":"Apis mellifera"` * Refines (identifier of the broader term this term refines, if applicable): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): http://rs.tdwg.org/dwc/terms/version/associatedTaxa-2017-10-06 * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG, if applicable): DataSets/DataSet/Units/Unit/Gathering/Synecology/AssociatedTaxa Discussions around changes to relationshipOfResource (#194), around a new term relationshipOfResourceID (#186, #283), and changes to associatedOccurrences (Issue #324) suggest that a clarification should also be made in the associatedTaxa definition and usage notes. Specifically, the directionality of the relationship should be made clear.
1.0
Change term - associatedTaxa - ## Change term * Submitter: John Wieczorek * Justification (why is this change necessary?): Consistency and clarity * Proponents (who needs this change): Everyone Current Term definition: https://dwc.tdwg.org/terms/#dwc:associatedTaxa Proposed new attributes of the term: * Term name (in lowerCamelCase): associatedTaxa * Organized in Class (e.g. Location, Taxon): Occurrence * Definition of the term: **A list (concatenated and separated) of identifiers or names of taxa and the associations of this Occurrence to each of them.** * Usage comments (recommendations regarding content, etc.): **This term can be used to provide a list of associations to Taxa other than the one defined in the Occurrence. Note that the ResourceRelationship class is an alternative means of representing associations, and with more detail. This term is not apt for establishing relationships between Taxa, only between specific Occurrences of an Organism with other Taxa. Recommended best practice is to separate the values in a list with space vertical bar space ( | ).** * Examples: `"host":"Quercus alba"`, **`"host":"gbif.org/species/2879737"`**,`"parasitoid of":"Cyclocephala signaticollis" | "predator of":"Apis mellifera"` * Refines (identifier of the broader term this term refines, if applicable): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): http://rs.tdwg.org/dwc/terms/version/associatedTaxa-2017-10-06 * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG, if applicable): DataSets/DataSet/Units/Unit/Gathering/Synecology/AssociatedTaxa Discussions around changes to relationshipOfResource (#194), around a new term relationshipOfResourceID (#186, #283), and changes to associatedOccurrences (Issue #324) suggest that a clarification should also be made in the associatedTaxa definition and usage notes. Specifically, the directionality of the relationship should be made clear.
process
change term associatedtaxa change term submitter john wieczorek justification why is this change necessary consistency and clarity proponents who needs this change everyone current term definition proposed new attributes of the term term name in lowercamelcase associatedtaxa organized in class e g location taxon occurrence definition of the term a list concatenated and separated of identifiers or names of taxa and the associations of this occurrence to each of them usage comments recommendations regarding content etc this term can be used to provide a list of associations to taxa other than the one defined in the occurrence note that the resourcerelationship class is an alternative means of representing associations and with more detail this term is not apt for establishing relationships between taxa only between specific occurrences of an organism with other taxa recommended best practice is to separate the values in a list with space vertical bar space examples host quercus alba host gbif org species parasitoid of cyclocephala signaticollis predator of apis mellifera refines identifier of the broader term this term refines if applicable none replaces identifier of the existing term that would be deprecated and replaced by this term if applicable abcd xpath of the equivalent term in abcd or efg if applicable datasets dataset units unit gathering synecology associatedtaxa discussions around changes to relationshipofresource around a new term relationshipofresourceid and changes to associatedoccurrences issue suggest that a clarification should also be made in the associatedtaxa definition and usage notes specifically the directionality of the relationship should be made clear
1
136,796
12,734,791,564
IssuesEvent
2020-06-25 14:24:56
graknlabs/graql
https://api.github.com/repos/graknlabs/graql
opened
Slack badge displayed in README
type: documentation
## Problem to Solve A Slack badge is displayed in the README. ## Proposed Solution Replace the Slack badge with a Discord badge, like we have in the `grakn` repo.
1.0
Slack badge displayed in README - ## Problem to Solve A Slack badge is displayed in the README. ## Proposed Solution Replace the Slack badge with a Discord badge, like we have in the `grakn` repo.
non_process
slack badge displayed in readme problem to solve a slack badge is displayed in the readme proposed solution replace the slack badge with a discord badge like we have in the grakn repo
0
640
3,098,040,417
IssuesEvent
2015-08-28 08:20:42
deb-sandeep/PHPWebApps
https://api.github.com/repos/deb-sandeep/PHPWebApps
opened
Provide support for multi-choice questions
enhancement jove_notes_db jove_notes_grammar jove_notes_processor jove_notes_server jove_notes_ui
This is an enhancement request for supporting multiple-choice type questions in JoveNotes, which is _sorely_ missing now. ### Definition A multiple-choice question is a type of interactive question, which has: 1. A question 2. Multiple options 3. One or more right answers 4. Explanation [Optional] The user interaction model differs while interacting between a single choice and multiple choice question. For example, in a single choice question the user has to choose only one answer, after which the system evaluates the answer. In case of multiple-choices, system should take over the response for evaluation only when the user has selected all the possible answers and given a hint to the system that he is done choosing the answers. _Should the system not take over when the requisite number of answers have been chosen?_ > This might look like a logical choice, but IMHO is not the right approach for the following reasons. Assume there are 'n' options and 'x' correct answers. {x <= n} > > 1. If the user chooses y answers (y<x) and believes that he has completed it, the system will keep waiting for him to choose x-y remaining answers. > 2. If the user believes there are y answers (y>x), the system will not let him make the mistake. Hence, in case of multiple answer questions, the user should let the system know once he is done choosing all the answers. ### Syntax for digitization ``` @multi_choice "Choose the plants which are a planets in our solar system" { @options { "Sun", "Earth" correct, "Moon", "Neptune correct" } @explanation "Sun is a star and Moon is a satellite and hence they are not planets" } ``` Important thing to note is that the system recognizes whether it should treat the question as single choice or multi choice based on the number of correct answers. The digitization syntax remains the same.
1.0
Provide support for multi-choice questions - This is an enhancement request for supporting multiple-choice type questions in JoveNotes, which is _sorely_ missing now. ### Definition A multiple-choice question is a type of interactive question, which has: 1. A question 2. Multiple options 3. One or more right answers 4. Explanation [Optional] The user interaction model differs while interacting between a single choice and multiple choice question. For example, in a single choice question the user has to choose only one answer, after which the system evaluates the answer. In case of multiple-choices, system should take over the response for evaluation only when the user has selected all the possible answers and given a hint to the system that he is done choosing the answers. _Should the system not take over when the requisite number of answers have been chosen?_ > This might look like a logical choice, but IMHO is not the right approach for the following reasons. Assume there are 'n' options and 'x' correct answers. {x <= n} > > 1. If the user chooses y answers (y<x) and believes that he has completed it, the system will keep waiting for him to choose x-y remaining answers. > 2. If the user believes there are y answers (y>x), the system will not let him make the mistake. Hence, in case of multiple answer questions, the user should let the system know once he is done choosing all the answers. ### Syntax for digitization ``` @multi_choice "Choose the plants which are a planets in our solar system" { @options { "Sun", "Earth" correct, "Moon", "Neptune correct" } @explanation "Sun is a star and Moon is a satellite and hence they are not planets" } ``` Important thing to note is that the system recognizes whether it should treat the question as single choice or multi choice based on the number of correct answers. The digitization syntax remains the same.
process
provide support for multi choice questions this is an enhancement request for supporting multiple choice type questions in jovenotes which is sorely missing now definition a multiple choice question is a type of interactive question which has a question multiple options one or more right answers explanation the user interaction model differs while interacting between a single choice and multiple choice question for example in a single choice question the user has to choose only one answer after which the system evaluates the answer in case of multiple choices system should take over the response for evaluation only when the user has selected all the possible answers and given a hint to the system that he is done choosing the answers should the system not take over when the requisite number of answers have been chosen this might look like a logical choice but imho is not the right approach for the following reasons assume there are n options and x correct answers x n if the user chooses y answers y x and believes that he has completed it the system will keep waiting for him to choose x y remaining answers if the user believes there are y answers y x the system will not let him make the mistake hence in case of multiple answer questions the user should let the system know once he is done choosing all the answers syntax for digitization multi choice choose the plants which are a planets in our solar system options sun earth correct moon neptune correct explanation sun is a star and moon is a satellite and hence they are not planets important thing to note is that the system recognizes whether it should treat the question as single choice or multi choice based on the number of correct answers the digitization syntax remains the same
1
2,295
5,115,485,568
IssuesEvent
2017-01-06 21:58:36
jlm2017/jlm-video-subtitles
https://api.github.com/repos/jlm2017/jlm-video-subtitles
opened
[subtitles] [eng] MÉLENCHON - POURQUOI LE CODE DU TRAVAIL EST GROS ?
Language: English Process: [2] Ready for review (1)
# Video title MÉLENCHON - POURQUOI LE CODE DU TRAVAIL EST GROS ? # URL https://www.youtube.com/watch?v=jnqujbuCNsM # Youtube subtitles language Anglais # Duration 1:58 # Subtitles URL https://www.youtube.com/timedtext_editor?tab=captions&bl=vmp&lang=en-US&action_mde_edit_form=1&ref=player&v=jnqujbuCNsM&ui=hd
1.0
[subtitles] [eng] MÉLENCHON - POURQUOI LE CODE DU TRAVAIL EST GROS ? - # Video title MÉLENCHON - POURQUOI LE CODE DU TRAVAIL EST GROS ? # URL https://www.youtube.com/watch?v=jnqujbuCNsM # Youtube subtitles language Anglais # Duration 1:58 # Subtitles URL https://www.youtube.com/timedtext_editor?tab=captions&bl=vmp&lang=en-US&action_mde_edit_form=1&ref=player&v=jnqujbuCNsM&ui=hd
process
mélenchon pourquoi le code du travail est gros video title mélenchon pourquoi le code du travail est gros url youtube subtitles language anglais duration subtitles url
1
5,900
8,717,581,182
IssuesEvent
2018-12-07 17:35:35
GoogleCloudPlatform/google-cloud-cpp
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-cpp
opened
Decide if we want to support Xcode 7.3.
type: process
We got a bug report, the code fails to compile with [Xcode 7.3](https://en.wikipedia.org/wiki/Xcode#7.x_series), which was released around 2015. The problem is actually in the compiler (e.g. does not support `thread_local`), and will submit a workaround shortly. We need to decide if the compiler is something we want to support long term, thus this bug to make us revisit the decision.
1.0
Decide if we want to support Xcode 7.3. - We got a bug report, the code fails to compile with [Xcode 7.3](https://en.wikipedia.org/wiki/Xcode#7.x_series), which was released around 2015. The problem is actually in the compiler (e.g. does not support `thread_local`), and will submit a workaround shortly. We need to decide if the compiler is something we want to support long term, thus this bug to make us revisit the decision.
process
decide if we want to support xcode we got a bug report the code fails to compile with which was released around the problem is actually in the compiler e g does not support thread local and will submit a workaround shortly we need to decide if the compiler is something we want to support long term thus this bug to make us revisit the decision
1
458,404
13,174,672,593
IssuesEvent
2020-08-11 23:09:00
shimming-toolbox/shimming-toolbox-py
https://api.github.com/repos/shimming-toolbox/shimming-toolbox-py
opened
Improve masking capabilities
Priority: LOW enhancement
## Context The initial issue #47 which was partially fixed by PR #60 added some basic capabilities to mask some data (threshold, shape: square, cube). More masking capability could be implemented to segment the brain and spinal cord as well as other shapes. Algorithms that could help - SCT: for spinal cord + simple shapes (license MIT) - FSL BET (non-commercial): for brain - Others? ## Suggestion - Add disk and ball to `shimmingtoolbox.masking.shape - Add capability to handle phase images to `shimmingtoolbox.masking.threshold - `shimmingtoolbox.masking.bet()` --> input is file name - `shimmingtoolbox.masking.sct()` --> input could be file name or an np.array (if we use SCT's API) output: mask as nd.array (of file in the case of bet)
1.0
Improve masking capabilities - ## Context The initial issue #47 which was partially fixed by PR #60 added some basic capabilities to mask some data (threshold, shape: square, cube). More masking capability could be implemented to segment the brain and spinal cord as well as other shapes. Algorithms that could help - SCT: for spinal cord + simple shapes (license MIT) - FSL BET (non-commercial): for brain - Others? ## Suggestion - Add disk and ball to `shimmingtoolbox.masking.shape - Add capability to handle phase images to `shimmingtoolbox.masking.threshold - `shimmingtoolbox.masking.bet()` --> input is file name - `shimmingtoolbox.masking.sct()` --> input could be file name or an np.array (if we use SCT's API) output: mask as nd.array (of file in the case of bet)
non_process
improve masking capabilities context the initial issue which was partially fixed by pr added some basic capabilities to mask some data threshold shape square cube more masking capability could be implemented to segment the brain and spinal cord as well as other shapes algorithms that could help sct for spinal cord simple shapes license mit fsl bet non commercial for brain others suggestion add disk and ball to shimmingtoolbox masking shape add capability to handle phase images to shimmingtoolbox masking threshold shimmingtoolbox masking bet input is file name shimmingtoolbox masking sct input could be file name or an np array if we use sct s api output mask as nd array of file in the case of bet
0
15,945
20,163,727,854
IssuesEvent
2022-02-10 00:47:37
ooi-data/CE09OSSM-MFD37-01-OPTAAC000-telemetered-optaa_dj_dcl_instrument
https://api.github.com/repos/ooi-data/CE09OSSM-MFD37-01-OPTAAC000-telemetered-optaa_dj_dcl_instrument
opened
🛑 Processing failed: GroupNotFoundError
process
## Overview `GroupNotFoundError` found in `processing_task` task during run ended on 2022-02-10T00:47:36.630149. ## Details Flow name: `CE09OSSM-MFD37-01-OPTAAC000-telemetered-optaa_dj_dcl_instrument` Task name: `processing_task` Error type: `GroupNotFoundError` Error message: group not found at path '' <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 64, in finalize_data_stream final_group = zarr.open_group(final_store, mode='r+') File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/hierarchy.py", line 1168, in open_group raise GroupNotFoundError(path) zarr.errors.GroupNotFoundError: group not found at path '' ``` </details>
1.0
🛑 Processing failed: GroupNotFoundError - ## Overview `GroupNotFoundError` found in `processing_task` task during run ended on 2022-02-10T00:47:36.630149. ## Details Flow name: `CE09OSSM-MFD37-01-OPTAAC000-telemetered-optaa_dj_dcl_instrument` Task name: `processing_task` Error type: `GroupNotFoundError` Error message: group not found at path '' <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 64, in finalize_data_stream final_group = zarr.open_group(final_store, mode='r+') File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/hierarchy.py", line 1168, in open_group raise GroupNotFoundError(path) zarr.errors.GroupNotFoundError: group not found at path '' ``` </details>
process
🛑 processing failed groupnotfounderror overview groupnotfounderror found in processing task task during run ended on details flow name telemetered optaa dj dcl instrument task name processing task error type groupnotfounderror error message group not found at path traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream final group zarr open group final store mode r file srv conda envs notebook lib site packages zarr hierarchy py line in open group raise groupnotfounderror path zarr errors groupnotfounderror group not found at path
1
208,120
15,876,454,812
IssuesEvent
2021-04-09 08:23:42
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
closed
No default blob container name shows in the 'Clone' dialog for one local emulator blob container
:beetle: regression :gear: emulator 🧪 testing
**Storage Explorer Version**: 1.19.0-dev **Build Number**: 20210408.5 **Branch**: main **Platform/OS**: Windows 10 **Emulator Version**: Azurite 3.11.0/ Microsoft Azure Storage Emulator 5.10.0.0 **Architecture**: ia32 **Regression From**: Previous release (1.18.1) ## Steps to Reproduce ## 1. Install and run Azurite 3.11.0 -> Launch Storage Explorer. 2. Expand 'Local & Attached -> Storage Accounts -> (Emulator - Default Ports) (Key)'. 3. Right click the 'Blob Containers' node -> Create a new blob container. 4. Right click the blob container -> Click 'Clone...'. 5. Check whether the default blob container name shows in the dialog. ## Expected Experience ## The default blob container name shows. ![image](https://user-images.githubusercontent.com/41351993/114008315-b5f13e00-9894-11eb-9f69-81023f91dcb6.png) ## Actual Experience ## No default blob container name shows. ![image](https://user-images.githubusercontent.com/41351993/114008306-b25db700-9894-11eb-90b0-7f666cdc6798.png) ## Additional Context ## The error message keeps showing after typing one valid blob container name. ![image](https://user-images.githubusercontent.com/41351993/114008377-c1dd0000-9894-11eb-8c6b-f41b4a53fd74.png)
1.0
No default blob container name shows in the 'Clone' dialog for one local emulator blob container - **Storage Explorer Version**: 1.19.0-dev **Build Number**: 20210408.5 **Branch**: main **Platform/OS**: Windows 10 **Emulator Version**: Azurite 3.11.0/ Microsoft Azure Storage Emulator 5.10.0.0 **Architecture**: ia32 **Regression From**: Previous release (1.18.1) ## Steps to Reproduce ## 1. Install and run Azurite 3.11.0 -> Launch Storage Explorer. 2. Expand 'Local & Attached -> Storage Accounts -> (Emulator - Default Ports) (Key)'. 3. Right click the 'Blob Containers' node -> Create a new blob container. 4. Right click the blob container -> Click 'Clone...'. 5. Check whether the default blob container name shows in the dialog. ## Expected Experience ## The default blob container name shows. ![image](https://user-images.githubusercontent.com/41351993/114008315-b5f13e00-9894-11eb-9f69-81023f91dcb6.png) ## Actual Experience ## No default blob container name shows. ![image](https://user-images.githubusercontent.com/41351993/114008306-b25db700-9894-11eb-90b0-7f666cdc6798.png) ## Additional Context ## The error message keeps showing after typing one valid blob container name. ![image](https://user-images.githubusercontent.com/41351993/114008377-c1dd0000-9894-11eb-8c6b-f41b4a53fd74.png)
non_process
no default blob container name shows in the clone dialog for one local emulator blob container storage explorer version dev build number branch main platform os windows emulator version azurite microsoft azure storage emulator architecture regression from previous release steps to reproduce install and run azurite launch storage explorer expand local attached storage accounts emulator default ports key right click the blob containers node create a new blob container right click the blob container click clone check whether the default blob container name shows in the dialog expected experience the default blob container name shows actual experience no default blob container name shows additional context the error message keeps showing after typing one valid blob container name
0
19,721
26,073,828,975
IssuesEvent
2022-12-24 07:06:26
pyanodon/pybugreports
https://api.github.com/repos/pyanodon/pybugreports
closed
Mod "Packing Tape" incompatible
bug mod:pypostprocessing crash compatibility
### Mod source Factorio Mod Portal ### Which mod are you having an issue with? - [ ] pyalienlife - [ ] pyalternativeenergy - [ ] pycoalprocessing - [ ] pyfusionenergy - [ ] pyhightech - [ ] pyindustry - [ ] pypetroleumhandling - [X] pypostprocessing - [ ] pyrawores ### Operating system >=Windows 10 ### What kind of issue is this? - [X] Compatibility - [ ] Locale (names, descriptions, unknown keys) - [ ] Graphical - [ ] Crash - [ ] Progression - [ ] Balance - [ ] Pypostprocessing failure - [ ] Other ### What is the problem? https://mods.factorio.com/mod/packing-tape Game will not load with the above mod enabled, gives this error; ![afbeelding](https://user-images.githubusercontent.com/41924788/195975719-623ab13c-c6b6-490f-81d6-feea921c77a9.png) When mod is removed it loads fine. I did also try loading the game without any Py mods, but with all my other QOL mods enabled, to confirm it is actually the Py mods causing this. ### Steps to reproduce _No response_ ### Additional context _No response_ ### Log file _No response_
1.0
Mod "Packing Tape" incompatible - ### Mod source Factorio Mod Portal ### Which mod are you having an issue with? - [ ] pyalienlife - [ ] pyalternativeenergy - [ ] pycoalprocessing - [ ] pyfusionenergy - [ ] pyhightech - [ ] pyindustry - [ ] pypetroleumhandling - [X] pypostprocessing - [ ] pyrawores ### Operating system >=Windows 10 ### What kind of issue is this? - [X] Compatibility - [ ] Locale (names, descriptions, unknown keys) - [ ] Graphical - [ ] Crash - [ ] Progression - [ ] Balance - [ ] Pypostprocessing failure - [ ] Other ### What is the problem? https://mods.factorio.com/mod/packing-tape Game will not load with the above mod enabled, gives this error; ![afbeelding](https://user-images.githubusercontent.com/41924788/195975719-623ab13c-c6b6-490f-81d6-feea921c77a9.png) When mod is removed it loads fine. I did also try loading the game without any Py mods, but with all my other QOL mods enabled, to confirm it is actually the Py mods causing this. ### Steps to reproduce _No response_ ### Additional context _No response_ ### Log file _No response_
process
mod packing tape incompatible mod source factorio mod portal which mod are you having an issue with pyalienlife pyalternativeenergy pycoalprocessing pyfusionenergy pyhightech pyindustry pypetroleumhandling pypostprocessing pyrawores operating system windows what kind of issue is this compatibility locale names descriptions unknown keys graphical crash progression balance pypostprocessing failure other what is the problem game will not load with the above mod enabled gives this error when mod is removed it loads fine i did also try loading the game without any py mods but with all my other qol mods enabled to confirm it is actually the py mods causing this steps to reproduce no response additional context no response log file no response
1
6,856
9,993,166,156
IssuesEvent
2019-07-11 14:50:34
eaudeweb/ozone
https://api.github.com/repos/eaudeweb/ozone
closed
Error when importing process agents
Component: Backend Feature: Data migration Feature: Process Agents Priority: Highest
KeyError: 'UNK' File "/var/local/ozone/ozone/core/management/commands/import_procagents.py", line 340, in get_or_create_decision meeting = self.meetings[meeting_id] caused by ProcAgentEmitLimits.Decision = UNK
1.0
Error when importing process agents - KeyError: 'UNK' File "/var/local/ozone/ozone/core/management/commands/import_procagents.py", line 340, in get_or_create_decision meeting = self.meetings[meeting_id] caused by ProcAgentEmitLimits.Decision = UNK
process
error when importing process agents keyerror unk file var local ozone ozone core management commands import procagents py line in get or create decision meeting self meetings caused by procagentemitlimits decision unk
1
1,417
3,984,317,443
IssuesEvent
2016-05-07 03:36:26
opattison/olivermakes
https://api.github.com/repos/opattison/olivermakes
closed
Testing: avoid links that only exist on master (html-proofer)
bug content maintenance process
See: https://travis-ci.org/opattison/olivermakes/builds/127293069 The link is only live when the post is published on the master branch, leading to errors in Travis CI that aren’t actually errors. Either: - `--url-ignore www.github.com` - `<a href="http://notareallink" data-proofer-ignore>Not checked.</a>`
1.0
Testing: avoid links that only exist on master (html-proofer) - See: https://travis-ci.org/opattison/olivermakes/builds/127293069 The link is only live when the post is published on the master branch, leading to errors in Travis CI that aren’t actually errors. Either: - `--url-ignore www.github.com` - `<a href="http://notareallink" data-proofer-ignore>Not checked.</a>`
process
testing avoid links that only exist on master html proofer see the link is only live when the post is published on the master branch leading to errors in travis ci that aren’t actually errors either url ignore not checked
1
193,192
22,216,084,211
IssuesEvent
2022-06-08 01:54:15
maddyCode23/linux-4.1.15
https://api.github.com/repos/maddyCode23/linux-4.1.15
reopened
CVE-2019-15216 (Medium) detected in linux-stable-rtv4.1.33
security vulnerability
## CVE-2019-15216 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/yurex.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/yurex.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.0.14. There is a NULL pointer dereference caused by a malicious USB device in the drivers/usb/misc/yurex.c driver. <p>Publish Date: 2019-08-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-15216>CVE-2019-15216</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15216">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15216</a></p> <p>Release Date: 2019-09-03</p> <p>Fix Resolution: v5.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-15216 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2019-15216 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/yurex.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/yurex.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.0.14. There is a NULL pointer dereference caused by a malicious USB device in the drivers/usb/misc/yurex.c driver. <p>Publish Date: 2019-08-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-15216>CVE-2019-15216</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15216">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15216</a></p> <p>Release Date: 2019-09-03</p> <p>Fix Resolution: v5.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers usb misc yurex c drivers usb misc yurex c vulnerability details an issue was discovered in the linux kernel before there is a null pointer dereference caused by a malicious usb device in the drivers usb misc yurex c driver publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
18,615
24,579,367,699
IssuesEvent
2022-10-13 14:34:16
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[Mobile apps] [FHIR] Enrollment flow is not working in the android mobile apps
Bug Blocker P0 iOS Android Process: Fixed Process: Tested QA Process: Tested dev
**AR:** Enrollment flow is not working in the android mobile apps **ER:** Participants should be able to enroll in the android mobile apps ![Untitled](https://user-images.githubusercontent.com/86007179/165324483-a56576b1-4405-4a7d-a0b2-1dbdd74aa1ac.png)
3.0
[Mobile apps] [FHIR] Enrollment flow is not working in the android mobile apps - **AR:** Enrollment flow is not working in the android mobile apps **ER:** Participants should be able to enroll in the android mobile apps ![Untitled](https://user-images.githubusercontent.com/86007179/165324483-a56576b1-4405-4a7d-a0b2-1dbdd74aa1ac.png)
process
enrollment flow is not working in the android mobile apps ar enrollment flow is not working in the android mobile apps er participants should be able to enroll in the android mobile apps
1
10,522
13,305,042,977
IssuesEvent
2020-08-25 17:54:04
googleapis/python-storage
https://api.github.com/repos/googleapis/python-storage
closed
'test_get_signed_policy_v4' flakes with 500
api: storage flaky testing type: process
From [this Kokoro build](https://source.cloud.google.com/results/invocations/0499a0e3-d3c1-444e-a1a4-e74e80e39dcf/targets/cloud-devrel%2Fclient-libraries%2Fpython%2Fgoogleapis%2Fpython-storage%2Fpresubmit%2Fpresubmit/log): ```python _________________ TestV4POSTPolicies.test_get_signed_policy_v4 _________________ self = <test_system.TestV4POSTPolicies testMethod=test_get_signed_policy_v4> def test_get_signed_policy_v4(self): bucket_name = "post_policy" + unique_resource_id("-") self.assertRaises(exceptions.NotFound, Config.CLIENT.get_bucket, bucket_name) retry_429_503(Config.CLIENT.create_bucket)(bucket_name) self.case_buckets_to_delete.append(bucket_name) blob_name = "post_policy_obj.txt" with open(blob_name, "w") as f: f.write("DEADBEEF") policy = Config.CLIENT.generate_signed_post_policy_v4( bucket_name, blob_name, conditions=[ {"bucket": bucket_name}, ["starts-with", "$Content-Type", "text/pla"], ], expiration=datetime.datetime.now() + datetime.timedelta(hours=1), fields={"content-type": "text/plain"}, ) with open(blob_name, "r") as f: files = {"file": (blob_name, f)} response = requests.post(policy["url"], data=policy["fields"], files=files) os.remove(blob_name) > self.assertEqual(response.status_code, 204) E AssertionError: 500 != 204 tests/system/test_system.py:2347: AssertionError ```
1.0
'test_get_signed_policy_v4' flakes with 500 - From [this Kokoro build](https://source.cloud.google.com/results/invocations/0499a0e3-d3c1-444e-a1a4-e74e80e39dcf/targets/cloud-devrel%2Fclient-libraries%2Fpython%2Fgoogleapis%2Fpython-storage%2Fpresubmit%2Fpresubmit/log): ```python _________________ TestV4POSTPolicies.test_get_signed_policy_v4 _________________ self = <test_system.TestV4POSTPolicies testMethod=test_get_signed_policy_v4> def test_get_signed_policy_v4(self): bucket_name = "post_policy" + unique_resource_id("-") self.assertRaises(exceptions.NotFound, Config.CLIENT.get_bucket, bucket_name) retry_429_503(Config.CLIENT.create_bucket)(bucket_name) self.case_buckets_to_delete.append(bucket_name) blob_name = "post_policy_obj.txt" with open(blob_name, "w") as f: f.write("DEADBEEF") policy = Config.CLIENT.generate_signed_post_policy_v4( bucket_name, blob_name, conditions=[ {"bucket": bucket_name}, ["starts-with", "$Content-Type", "text/pla"], ], expiration=datetime.datetime.now() + datetime.timedelta(hours=1), fields={"content-type": "text/plain"}, ) with open(blob_name, "r") as f: files = {"file": (blob_name, f)} response = requests.post(policy["url"], data=policy["fields"], files=files) os.remove(blob_name) > self.assertEqual(response.status_code, 204) E AssertionError: 500 != 204 tests/system/test_system.py:2347: AssertionError ```
process
test get signed policy flakes with from python test get signed policy self def test get signed policy self bucket name post policy unique resource id self assertraises exceptions notfound config client get bucket bucket name retry config client create bucket bucket name self case buckets to delete append bucket name blob name post policy obj txt with open blob name w as f f write deadbeef policy config client generate signed post policy bucket name blob name conditions bucket bucket name expiration datetime datetime now datetime timedelta hours fields content type text plain with open blob name r as f files file blob name f response requests post policy data policy files files os remove blob name self assertequal response status code e assertionerror tests system test system py assertionerror
1
21,242
28,366,913,256
IssuesEvent
2023-04-12 14:25:54
Deltares/Ribasim
https://api.github.com/repos/Deltares/Ribasim
opened
allow salt injection as boundary condition
physical process
include salt injection as boundary condition to represent salt influx from the sea
1.0
allow salt injection as boundary condition - include salt injection as boundary condition to represent salt influx from the sea
process
allow salt injection as boundary condition include salt injection as boundary condition to represent salt influx from the sea
1
156,523
12,313,199,921
IssuesEvent
2020-05-12 14:59:12
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: version/mixed/nodes=3 failed
C-test-failure O-roachtest O-robot branch-release-20.1 release-blocker
[(roachtest).version/mixed/nodes=3 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1928509&tab=buildLog) on [release-20.1@1affcc3ac8819e133d54668a1ce0fc4a4ce99b4f](https://github.com/cockroachdb/cockroach/commits/1affcc3ac8819e133d54668a1ce0fc4a4ce99b4f): ``` | 579.0s 0 2559.0 2587.3 2.6 7.3 12.6 19.9 write | 580.0s 0 2377.9 2587.0 2.5 8.9 15.2 27.3 write | _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms) | 581.0s 0 2268.0 2586.4 3.0 7.6 12.1 25.2 write | 582.0s 0 2456.1 2586.2 2.6 7.6 11.0 46.1 write | 583.0s 0 2069.9 2585.3 2.8 10.5 17.8 33.6 write | 584.0s 0 2459.0 2585.1 2.6 7.9 11.5 24.1 write | 585.0s 0 2622.9 2585.2 2.5 7.1 11.0 35.7 write | 586.0s 0 2463.2 2585.0 2.8 7.3 11.0 19.9 write | 587.0s 0 2467.9 2584.8 2.8 7.3 11.0 21.0 write | 588.0s 0 2389.1 2584.4 2.8 7.6 11.0 14.7 write | 589.0s 0 2412.0 2584.1 2.8 7.6 11.5 18.9 write | 590.0s 0 2345.9 2583.7 2.8 7.6 12.1 218.1 write | 591.0s 0 2271.0 2583.2 2.9 7.6 12.6 33.6 write | 592.0s 0 2577.1 2583.2 2.6 7.3 11.0 24.1 write | 593.0s 0 2585.0 2583.2 2.6 6.8 10.0 24.1 write | 594.0s 0 2338.0 2582.8 2.9 7.9 12.1 23.1 write | 595.0s 0 2429.0 2582.5 2.6 8.1 11.0 30.4 write | 596.0s 0 2122.0 2581.7 3.1 8.1 12.1 25.2 write | 597.0s 0 2526.9 2581.7 2.6 7.6 10.5 21.0 write | 598.0s 0 1976.1 2580.6 2.9 11.5 18.9 25.2 write | 599.0s 0 2153.0 2579.9 3.1 7.9 12.1 18.9 write Wraps: (9) exit status 30 Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *withstack.withStack (4) *errutil.withMessage (5) *withstack.withStack (6) *safedetails.withSafeDetails (7) *errutil.withMessage (8) *main.withCommandDetails (9) *exec.ExitError cluster.go:1481,context.go:135,cluster.go:1470,test_runner.go:825: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-1928509-1588924553-09-n4cpu4 --oneshot --ignore-empty-nodes: exit status 1 4: skipped 3: 7750 2: 7786 1: dead Error: UNCLASSIFIED_PROBLEM: 1: dead (1) UNCLASSIFIED_PROBLEM Wraps: (2) attached stack trace | main.glob..func13 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1129 | main.wrap.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:272 | github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:766 | github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).ExecuteC | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:852 | github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).Execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:800 | main.main | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1799 | runtime.main | /usr/local/go/src/runtime/proc.go:203 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1357 Wraps: (3) 1: dead Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *errors.errorString ``` <details><summary>More</summary><p> Artifacts: [/version/mixed/nodes=3](https://teamcity.cockroachdb.com/viewLog.html?buildId=1928509&tab=artifacts#/version/mixed/nodes=3) Related: - #48574 roachtest: version/mixed/nodes=3 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #47741 roachtest: version/mixed/nodes=3 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Aversion%2Fmixed%2Fnodes%3D3.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
2.0
roachtest: version/mixed/nodes=3 failed - [(roachtest).version/mixed/nodes=3 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1928509&tab=buildLog) on [release-20.1@1affcc3ac8819e133d54668a1ce0fc4a4ce99b4f](https://github.com/cockroachdb/cockroach/commits/1affcc3ac8819e133d54668a1ce0fc4a4ce99b4f): ``` | 579.0s 0 2559.0 2587.3 2.6 7.3 12.6 19.9 write | 580.0s 0 2377.9 2587.0 2.5 8.9 15.2 27.3 write | _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms) | 581.0s 0 2268.0 2586.4 3.0 7.6 12.1 25.2 write | 582.0s 0 2456.1 2586.2 2.6 7.6 11.0 46.1 write | 583.0s 0 2069.9 2585.3 2.8 10.5 17.8 33.6 write | 584.0s 0 2459.0 2585.1 2.6 7.9 11.5 24.1 write | 585.0s 0 2622.9 2585.2 2.5 7.1 11.0 35.7 write | 586.0s 0 2463.2 2585.0 2.8 7.3 11.0 19.9 write | 587.0s 0 2467.9 2584.8 2.8 7.3 11.0 21.0 write | 588.0s 0 2389.1 2584.4 2.8 7.6 11.0 14.7 write | 589.0s 0 2412.0 2584.1 2.8 7.6 11.5 18.9 write | 590.0s 0 2345.9 2583.7 2.8 7.6 12.1 218.1 write | 591.0s 0 2271.0 2583.2 2.9 7.6 12.6 33.6 write | 592.0s 0 2577.1 2583.2 2.6 7.3 11.0 24.1 write | 593.0s 0 2585.0 2583.2 2.6 6.8 10.0 24.1 write | 594.0s 0 2338.0 2582.8 2.9 7.9 12.1 23.1 write | 595.0s 0 2429.0 2582.5 2.6 8.1 11.0 30.4 write | 596.0s 0 2122.0 2581.7 3.1 8.1 12.1 25.2 write | 597.0s 0 2526.9 2581.7 2.6 7.6 10.5 21.0 write | 598.0s 0 1976.1 2580.6 2.9 11.5 18.9 25.2 write | 599.0s 0 2153.0 2579.9 3.1 7.9 12.1 18.9 write Wraps: (9) exit status 30 Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *withstack.withStack (4) *errutil.withMessage (5) *withstack.withStack (6) *safedetails.withSafeDetails (7) *errutil.withMessage (8) *main.withCommandDetails (9) *exec.ExitError cluster.go:1481,context.go:135,cluster.go:1470,test_runner.go:825: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-1928509-1588924553-09-n4cpu4 --oneshot --ignore-empty-nodes: exit status 1 4: skipped 3: 7750 2: 7786 1: dead Error: UNCLASSIFIED_PROBLEM: 1: dead (1) UNCLASSIFIED_PROBLEM Wraps: (2) attached stack trace | main.glob..func13 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1129 | main.wrap.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:272 | github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:766 | github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).ExecuteC | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:852 | github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).Execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:800 | main.main | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1799 | runtime.main | /usr/local/go/src/runtime/proc.go:203 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1357 Wraps: (3) 1: dead Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *errors.errorString ``` <details><summary>More</summary><p> Artifacts: [/version/mixed/nodes=3](https://teamcity.cockroachdb.com/viewLog.html?buildId=1928509&tab=artifacts#/version/mixed/nodes=3) Related: - #48574 roachtest: version/mixed/nodes=3 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #47741 roachtest: version/mixed/nodes=3 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Aversion%2Fmixed%2Fnodes%3D3.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
non_process
roachtest version mixed nodes failed on write write elapsed errors ops sec inst ops sec cum ms ms ms pmax ms write write write write write write write write write write write write write write write write write write write wraps exit status error types withstack withstack errutil withmessage withstack withstack errutil withmessage withstack withstack safedetails withsafedetails errutil withmessage main withcommanddetails exec exiterror cluster go context go cluster go test runner go dead node detection home agent work go src github com cockroachdb cockroach bin roachprod monitor teamcity oneshot ignore empty nodes exit status skipped dead error unclassified problem dead unclassified problem wraps attached stack trace main glob home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go main wrap home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go github com cockroachdb cockroach vendor github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cockroachdb cockroach vendor github com cobra command executec home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cockroachdb cockroach vendor github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go main main home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps dead error types errors unclassified withstack withstack errors errorstring more artifacts related roachtest version mixed nodes failed roachtest version mixed nodes failed powered by
0
33,405
15,896,458,729
IssuesEvent
2021-04-11 17:30:10
sp614x/optifine
https://api.github.com/repos/sp614x/optifine
closed
FPS decreasing over time with shaders/in F5/in inventory
1.16 details needed performance shaders
Without shaders I get 100fps easily, but when I put them on it goes from 70fps to 20fps in ~10 minutes. It used to work well. Specs : - Ryzen 7 2700 - 2x8gb ram 3200mhz - RTX 2070 Temps are very good, I allocated 6-8gb of ram for mc. I use Forge 1.16.1 and the last optifine update
True
FPS decreasing over time with shaders/in F5/in inventory - Without shaders I get 100fps easily, but when I put them on it goes from 70fps to 20fps in ~10 minutes. It used to work well. Specs : - Ryzen 7 2700 - 2x8gb ram 3200mhz - RTX 2070 Temps are very good, I allocated 6-8gb of ram for mc. I use Forge 1.16.1 and the last optifine update
non_process
fps decreasing over time with shaders in in inventory without shaders i get easily but when i put them on it goes from to in minutes it used to work well specs ryzen ram rtx temps are very good i allocated of ram for mc i use forge and the last optifine update
0
105,996
16,663,950,074
IssuesEvent
2021-06-06 20:50:28
AlexRogalskiy/github-action-charts
https://api.github.com/repos/AlexRogalskiy/github-action-charts
opened
CVE-2020-28469 (Medium) detected in glob-parent-5.1.1.tgz
security vulnerability
## CVE-2020-28469 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-5.1.1.tgz</b></p></summary> <p>Extract the non-magic parent path from a glob string.</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz</a></p> <p>Path to dependency file: github-action-charts/package.json</p> <p>Path to vulnerable library: github-action-charts/node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - eslint-7.19.0.tgz (Root Library) - :x: **glob-parent-5.1.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-charts/commit/c839647436e6519774bf3d4a639525381eedfb20">c839647436e6519774bf3d4a639525381eedfb20</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator. <p>Publish Date: 2021-06-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469>CVE-2020-28469</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469</a></p> <p>Release Date: 2021-06-03</p> <p>Fix Resolution: glob-parent - 5.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-28469 (Medium) detected in glob-parent-5.1.1.tgz - ## CVE-2020-28469 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-5.1.1.tgz</b></p></summary> <p>Extract the non-magic parent path from a glob string.</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz</a></p> <p>Path to dependency file: github-action-charts/package.json</p> <p>Path to vulnerable library: github-action-charts/node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - eslint-7.19.0.tgz (Root Library) - :x: **glob-parent-5.1.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-charts/commit/c839647436e6519774bf3d4a639525381eedfb20">c839647436e6519774bf3d4a639525381eedfb20</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator. <p>Publish Date: 2021-06-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469>CVE-2020-28469</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469</a></p> <p>Release Date: 2021-06-03</p> <p>Fix Resolution: glob-parent - 5.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in glob parent tgz cve medium severity vulnerability vulnerable library glob parent tgz extract the non magic parent path from a glob string library home page a href path to dependency file github action charts package json path to vulnerable library github action charts node modules glob parent package json dependency hierarchy eslint tgz root library x glob parent tgz vulnerable library found in head commit a href vulnerability details this affects the package glob parent before the enclosure regex used to check for strings ending in enclosure containing path separator publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent step up your open source security game with whitesource
0
109,112
9,368,522,340
IssuesEvent
2019-04-03 08:54:15
scality/metalk8s
https://api.github.com/repos/scality/metalk8s
closed
Conflict between control plane network IPs when running tests in CI single node env
bug ci tests
In the CI we run the tests, and especially we retry the bootstrap script to see if it's idempotent. In the CI, in the single node env, we have two IPs on the worker: 1. The one from Openstack: 10.100.A.B 1. The one from the pod network we set: 10.233.C.D In the CI, we define the control plane network as 10.0.0.0/8 (hence including both previous IPs). Then when retrying the bootstrap script, we pick one IP on this network, and that could be either the one from Openstack, or the 10.233 one. And if it's the second one, then it means we bind all our services on it, and nothing works anymore because we still are targeting the 10.100 one to address the API server. Let's be stricter and consider 10.100.0.0/16 as our control plane network (and also workload plane network) in the CI single node env.
1.0
Conflict between control plane network IPs when running tests in CI single node env - In the CI we run the tests, and especially we retry the bootstrap script to see if it's idempotent. In the CI, in the single node env, we have two IPs on the worker: 1. The one from Openstack: 10.100.A.B 1. The one from the pod network we set: 10.233.C.D In the CI, we define the control plane network as 10.0.0.0/8 (hence including both previous IPs). Then when retrying the bootstrap script, we pick one IP on this network, and that could be either the one from Openstack, or the 10.233 one. And if it's the second one, then it means we bind all our services on it, and nothing works anymore because we still are targeting the 10.100 one to address the API server. Let's be stricter and consider 10.100.0.0/16 as our control plane network (and also workload plane network) in the CI single node env.
non_process
conflict between control plane network ips when running tests in ci single node env in the ci we run the tests and especially we retry the bootstrap script to see if it s idempotent in the ci in the single node env we have two ips on the worker the one from openstack a b the one from the pod network we set c d in the ci we define the control plane network as hence including both previous ips then when retrying the bootstrap script we pick one ip on this network and that could be either the one from openstack or the one and if it s the second one then it means we bind all our services on it and nothing works anymore because we still are targeting the one to address the api server let s be stricter and consider as our control plane network and also workload plane network in the ci single node env
0
816
3,290,958,174
IssuesEvent
2015-10-30 04:35:29
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Make strict processing mode fail on error messages
enhancement P2 preprocess
Improve processing mode coverage to treat error messages as fatal errors.
1.0
Make strict processing mode fail on error messages - Improve processing mode coverage to treat error messages as fatal errors.
process
make strict processing mode fail on error messages improve processing mode coverage to treat error messages as fatal errors
1
281,655
8,697,938,816
IssuesEvent
2018-12-04 21:45:09
mplusmuseum/mplusmuseum-collections-explorer
https://api.github.com/repos/mplusmuseum/mplusmuseum-collections-explorer
closed
Inline stylized category text
priority-low question task
Add inline stylized text for categories. Please see design. ![screen shot 2018-11-20 at 8 34 16 pm](https://user-images.githubusercontent.com/302309/48813169-b27cac80-ed03-11e8-9994-8d8ef875ef7d.png) I'm marking this as a low-priority for now until we get everything else done. The simple case of this would be simple. But I worry that the edge cases of long text might turn this task into a more difficult one. **Question**: How do we handle long text that wraps?
1.0
Inline stylized category text - Add inline stylized text for categories. Please see design. ![screen shot 2018-11-20 at 8 34 16 pm](https://user-images.githubusercontent.com/302309/48813169-b27cac80-ed03-11e8-9994-8d8ef875ef7d.png) I'm marking this as a low-priority for now until we get everything else done. The simple case of this would be simple. But I worry that the edge cases of long text might turn this task into a more difficult one. **Question**: How do we handle long text that wraps?
non_process
inline stylized category text add inline stylized text for categories please see design i m marking this as a low priority for now until we get everything else done the simple case of this would be simple but i worry that the edge cases of long text might turn this task into a more difficult one question how do we handle long text that wraps
0
17,001
22,364,193,203
IssuesEvent
2022-06-16 00:59:22
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
opened
Add additional eth_getBlockByNumber acceptance tests
enhancement P2 process
### Problem The initial eth_getBlockByNumber coverage was scarce ### Solution Add move coverage - [ ] Enhance the already existing `getBlockByNumber` test by adding assertion of the returned object that it has `transactions` array **containing the full transaction** and not only the string of their hash (testing the "hydrated transactions flag") - [ ] Add a test that queries **existing block** by **number** but sets `hydrated transactions to **false**. Assert that the response contains `transactions` array but the values are strings of the **transaction hash** - [ ] Add negative test for non-existing block number with `hydrated transactions set to **true**`. The response of the call must be asserted to be: ``` { "jsonrpc": "2.0", "id": 0, "result": null } ``` - [ ] Add negative test for non-existing block number with `hydrated transactions set to **false**`. The response of the call must be asserted to be: ``` { "jsonrpc": "2.0", "id": 0, "result": null } ``` ### Alternatives _No response_
1.0
Add additional eth_getBlockByNumber acceptance tests - ### Problem The initial eth_getBlockByNumber coverage was scarce ### Solution Add move coverage - [ ] Enhance the already existing `getBlockByNumber` test by adding assertion of the returned object that it has `transactions` array **containing the full transaction** and not only the string of their hash (testing the "hydrated transactions flag") - [ ] Add a test that queries **existing block** by **number** but sets `hydrated transactions to **false**. Assert that the response contains `transactions` array but the values are strings of the **transaction hash** - [ ] Add negative test for non-existing block number with `hydrated transactions set to **true**`. The response of the call must be asserted to be: ``` { "jsonrpc": "2.0", "id": 0, "result": null } ``` - [ ] Add negative test for non-existing block number with `hydrated transactions set to **false**`. The response of the call must be asserted to be: ``` { "jsonrpc": "2.0", "id": 0, "result": null } ``` ### Alternatives _No response_
process
add additional eth getblockbynumber acceptance tests problem the initial eth getblockbynumber coverage was scarce solution add move coverage enhance the already existing getblockbynumber test by adding assertion of the returned object that it has transactions array containing the full transaction and not only the string of their hash testing the hydrated transactions flag add a test that queries existing block by number but sets hydrated transactions to false assert that the response contains transactions array but the values are strings of the transaction hash add negative test for non existing block number with hydrated transactions set to true the response of the call must be asserted to be jsonrpc id result null add negative test for non existing block number with hydrated transactions set to false the response of the call must be asserted to be jsonrpc id result null alternatives no response
1
16,584
21,630,841,257
IssuesEvent
2022-05-05 09:32:26
googleapis/google-cloud-dotnet
https://api.github.com/repos/googleapis/google-cloud-dotnet
closed
Docs release fails for IAM (alpha05)
type: process
Filing as process as it isn't a bug that will affect customers, beyond our ability to release. The release build for #8423 is failing in the docs build with: > [22-04-29 01:57:02.625]Warning:[ExtractMetadata](T:/src/github/google-cloud-dotnet/releasebuild/apis/Google.Cloud.Iam.V1/Google.Cloud.Iam.V1/Google.Cloud.Iam.V1.csproj)Workspace failed with: [Failure] Msbuild failed when processing the file 'T:\src\github\google-cloud-dotnet\releasebuild\apis\Google.Cloud.Iam.V1\Google.Cloud.Iam.V1\Google.Cloud.Iam.V1.csproj' with message: The SDK 'Microsoft.NET.Sdk' specified could not be found. T:\src\github\google-cloud-dotnet\releasebuild\apis\Google.Cloud.Iam.V1\Google.Cloud.Iam.V1\Google.Cloud.Iam.V1.csproj This is odd given that alpha02 released with no problems.
1.0
Docs release fails for IAM (alpha05) - Filing as process as it isn't a bug that will affect customers, beyond our ability to release. The release build for #8423 is failing in the docs build with: > [22-04-29 01:57:02.625]Warning:[ExtractMetadata](T:/src/github/google-cloud-dotnet/releasebuild/apis/Google.Cloud.Iam.V1/Google.Cloud.Iam.V1/Google.Cloud.Iam.V1.csproj)Workspace failed with: [Failure] Msbuild failed when processing the file 'T:\src\github\google-cloud-dotnet\releasebuild\apis\Google.Cloud.Iam.V1\Google.Cloud.Iam.V1\Google.Cloud.Iam.V1.csproj' with message: The SDK 'Microsoft.NET.Sdk' specified could not be found. T:\src\github\google-cloud-dotnet\releasebuild\apis\Google.Cloud.Iam.V1\Google.Cloud.Iam.V1\Google.Cloud.Iam.V1.csproj This is odd given that alpha02 released with no problems.
process
docs release fails for iam filing as process as it isn t a bug that will affect customers beyond our ability to release the release build for is failing in the docs build with warning t src github google cloud dotnet releasebuild apis google cloud iam google cloud iam google cloud iam csproj workspace failed with msbuild failed when processing the file t src github google cloud dotnet releasebuild apis google cloud iam google cloud iam google cloud iam csproj with message the sdk microsoft net sdk specified could not be found t src github google cloud dotnet releasebuild apis google cloud iam google cloud iam google cloud iam csproj this is odd given that released with no problems
1
254,067
8,069,590,243
IssuesEvent
2018-08-06 06:38:42
containous/traefik
https://api.github.com/repos/containous/traefik
closed
[acme.dnsChallenge] not working with gandiv5: No or incorrect TXT record
area/acme kind/bug/confirmed priority/P1
<!-- DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS. The issue tracker is for reporting bugs and feature requests only. For end-user related support questions, please refer to one of the following: - Stack Overflow (using the "traefik" tag): https://stackoverflow.com/questions/tagged/traefik - the Traefik community Slack channel: https://traefik.herokuapp.com --> ### Do you want to request a *feature* or report a *bug*? Bug ### What did you do? <!-- HOW TO WRITE A GOOD BUG REPORT? - Respect the issue template as much as possible. - If possible, use the command `traefik bug`. See https://www.youtube.com/watch?v=Lyz62L8m93I. - The title should be short and descriptive. - Explain the conditions which led you to report this issue: the context. - The context should lead to something, an idea or a problem that you’re facing. - Remain clear and concise. - Format your messages to help the reader focus on what matters and understand the structure of your message, use Markdown syntax https://help.github.com/articles/github-flavored-markdown --> After a change of config I noticed the client no longer gets an accepted certificate. Switching back to the previously working config didn't solve the problem, neither did downgrading to 1.6.2 or 1.6.1.. ### What did you expect to see? A working letsencrypt certificate ### What did you see instead? No certificate is generated, but an error appears in the log (see below) ### Output of `traefik version`: (_What version of Traefik are you using?_) <!-- For the Traefik Docker image: docker run [IMAGE] version ex: docker run traefik version For the alpine Traefik Docker image: docker run [IMAGE] traefik version ex: docker run traefik traefik version --> ``` Version: v1.6.3 Codename: tetedemoine Go version: go1.10.2 Built: 2018-06-05_03:29:01PM OS/Arch: linux/amd64 ``` ``` Same results with 1.6.1, 1.6.2 and 1.6.3 ``` ### What is your environment & configuration (arguments, toml, provider, platform, ...)? ```toml [acme] ... caServer = "https://acme-v02.api.letsencrypt.org/directory" [acme.dnsChallenge] provider = "gandiv5" delayBeforeCheck = 0 ... [[acme.domains]] main = "*.example.org" sans = ["example.org"] [[acme.domains]] main = "*.example.guru" sans = ["example.guru"] [[acme.domains]] main = "service.example.com" ``` <!-- Add more configuration information here. --> The results are the same using: ```toml caServer = "https://acme-staging-v02.api.letsencrypt.org/directory" ``` ### If applicable, please paste the log output in DEBUG level (`--logLevel=DEBUG` switch) ``` 6/6/2018 11:31:34 AMGandi DNS: Zone record deleted 6/6/2018 11:31:34 AMERRO[2018-06-06T09:31:34Z] Unable to obtain ACME certificate for domains "service.example.com" : cannot obtain certificates: acme: Error -> One or more domains had a problem: 6/6/2018 11:31:34 AM[service.example.com] acme: Error 400 - urn:ietf:params:acme:error:dns - DNS problem: NXDOMAIN looking up TXT for _acme-challenge.service.example.com 6/6/2018 11:31:34 AM 6/6/2018 11:32:01 AMGandi DNS: Zone record deleted 6/6/2018 11:32:01 AMERRO[2018-06-06T09:32:01Z] Unable to obtain ACME certificate for domains "*.example.guru,example.guru" : cannot obtain certificates: acme: Error -> One or more domains had a problem: 6/6/2018 11:32:01 AM[example.guru] acme: Error 403 - urn:ietf:params:acme:error:unauthorized - No TXT record found at _acme-challenge.example.guru 6/6/2018 11:32:01 AM 6/6/2018 11:32:35 AMGandi DNS: Zone record deleted 6/6/2018 11:32:36 AMGandi DNS: DNS Record Created 6/6/2018 11:33:21 AMGandi DNS: Zone record deleted 6/6/2018 11:33:21 AMERRO[2018-06-06T09:33:21Z] Unable to obtain ACME certificate for domains "*.example.org,example.org" : cannot obtain certificates: acme: Error -> One or more domains had a problem: 6/6/2018 11:33:21 AM[example.org] acme: Error 403 - urn:ietf:params:acme:error:unauthorized - Incorrect TXT record "nYhYDZD1ENYNTDy4_eQ5bP7vhhTVY5xWKrv_8pF3r2U" found at _acme-challenge.example.org ```
1.0
[acme.dnsChallenge] not working with gandiv5: No or incorrect TXT record - <!-- DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS. The issue tracker is for reporting bugs and feature requests only. For end-user related support questions, please refer to one of the following: - Stack Overflow (using the "traefik" tag): https://stackoverflow.com/questions/tagged/traefik - the Traefik community Slack channel: https://traefik.herokuapp.com --> ### Do you want to request a *feature* or report a *bug*? Bug ### What did you do? <!-- HOW TO WRITE A GOOD BUG REPORT? - Respect the issue template as much as possible. - If possible, use the command `traefik bug`. See https://www.youtube.com/watch?v=Lyz62L8m93I. - The title should be short and descriptive. - Explain the conditions which led you to report this issue: the context. - The context should lead to something, an idea or a problem that you’re facing. - Remain clear and concise. - Format your messages to help the reader focus on what matters and understand the structure of your message, use Markdown syntax https://help.github.com/articles/github-flavored-markdown --> After a change of config I noticed the client no longer gets an accepted certificate. Switching back to the previously working config didn't solve the problem, neither did downgrading to 1.6.2 or 1.6.1.. ### What did you expect to see? A working letsencrypt certificate ### What did you see instead? No certificate is generated, but an error appears in the log (see below) ### Output of `traefik version`: (_What version of Traefik are you using?_) <!-- For the Traefik Docker image: docker run [IMAGE] version ex: docker run traefik version For the alpine Traefik Docker image: docker run [IMAGE] traefik version ex: docker run traefik traefik version --> ``` Version: v1.6.3 Codename: tetedemoine Go version: go1.10.2 Built: 2018-06-05_03:29:01PM OS/Arch: linux/amd64 ``` ``` Same results with 1.6.1, 1.6.2 and 1.6.3 ``` ### What is your environment & configuration (arguments, toml, provider, platform, ...)? ```toml [acme] ... caServer = "https://acme-v02.api.letsencrypt.org/directory" [acme.dnsChallenge] provider = "gandiv5" delayBeforeCheck = 0 ... [[acme.domains]] main = "*.example.org" sans = ["example.org"] [[acme.domains]] main = "*.example.guru" sans = ["example.guru"] [[acme.domains]] main = "service.example.com" ``` <!-- Add more configuration information here. --> The results are the same using: ```toml caServer = "https://acme-staging-v02.api.letsencrypt.org/directory" ``` ### If applicable, please paste the log output in DEBUG level (`--logLevel=DEBUG` switch) ``` 6/6/2018 11:31:34 AMGandi DNS: Zone record deleted 6/6/2018 11:31:34 AMERRO[2018-06-06T09:31:34Z] Unable to obtain ACME certificate for domains "service.example.com" : cannot obtain certificates: acme: Error -> One or more domains had a problem: 6/6/2018 11:31:34 AM[service.example.com] acme: Error 400 - urn:ietf:params:acme:error:dns - DNS problem: NXDOMAIN looking up TXT for _acme-challenge.service.example.com 6/6/2018 11:31:34 AM 6/6/2018 11:32:01 AMGandi DNS: Zone record deleted 6/6/2018 11:32:01 AMERRO[2018-06-06T09:32:01Z] Unable to obtain ACME certificate for domains "*.example.guru,example.guru" : cannot obtain certificates: acme: Error -> One or more domains had a problem: 6/6/2018 11:32:01 AM[example.guru] acme: Error 403 - urn:ietf:params:acme:error:unauthorized - No TXT record found at _acme-challenge.example.guru 6/6/2018 11:32:01 AM 6/6/2018 11:32:35 AMGandi DNS: Zone record deleted 6/6/2018 11:32:36 AMGandi DNS: DNS Record Created 6/6/2018 11:33:21 AMGandi DNS: Zone record deleted 6/6/2018 11:33:21 AMERRO[2018-06-06T09:33:21Z] Unable to obtain ACME certificate for domains "*.example.org,example.org" : cannot obtain certificates: acme: Error -> One or more domains had a problem: 6/6/2018 11:33:21 AM[example.org] acme: Error 403 - urn:ietf:params:acme:error:unauthorized - Incorrect TXT record "nYhYDZD1ENYNTDy4_eQ5bP7vhhTVY5xWKrv_8pF3r2U" found at _acme-challenge.example.org ```
non_process
not working with no or incorrect txt record do not file issues for general support questions the issue tracker is for reporting bugs and feature requests only for end user related support questions please refer to one of the following stack overflow using the traefik tag the traefik community slack channel do you want to request a feature or report a bug bug what did you do how to write a good bug report respect the issue template as much as possible if possible use the command traefik bug see the title should be short and descriptive explain the conditions which led you to report this issue the context the context should lead to something an idea or a problem that you’re facing remain clear and concise format your messages to help the reader focus on what matters and understand the structure of your message use markdown syntax after a change of config i noticed the client no longer gets an accepted certificate switching back to the previously working config didn t solve the problem neither did downgrading to or what did you expect to see a working letsencrypt certificate what did you see instead no certificate is generated but an error appears in the log see below output of traefik version what version of traefik are you using for the traefik docker image docker run version ex docker run traefik version for the alpine traefik docker image docker run traefik version ex docker run traefik traefik version version codename tetedemoine go version built os arch linux same results with and what is your environment configuration arguments toml provider platform toml caserver provider delaybeforecheck main example org sans main example guru sans main service example com add more configuration information here the results are the same using toml caserver if applicable please paste the log output in debug level loglevel debug switch amgandi dns zone record deleted amerro unable to obtain acme certificate for domains service example com cannot obtain certificates acme error one or more domains had a problem am acme error urn ietf params acme error dns dns problem nxdomain looking up txt for acme challenge service example com am amgandi dns zone record deleted amerro unable to obtain acme certificate for domains example guru example guru cannot obtain certificates acme error one or more domains had a problem am acme error urn ietf params acme error unauthorized no txt record found at acme challenge example guru am amgandi dns zone record deleted amgandi dns dns record created amgandi dns zone record deleted amerro unable to obtain acme certificate for domains example org example org cannot obtain certificates acme error one or more domains had a problem am acme error urn ietf params acme error unauthorized incorrect txt record found at acme challenge example org
0
8,028
20,611,725,284
IssuesEvent
2022-03-07 09:18:48
dfds/backstage
https://api.github.com/repos/dfds/backstage
closed
Crosplane - Retrospectively add the required role for enabling assuming role in AWS provider
Enhancement Architecture
- [x] Investigate how this is currently done - [x] Complete #480 - [x] Find and replace the referenced infrastructure-modules version in all capability sub-folders of aws-account-manifests and update to the new infrastructure-modules version - [x] Submit PR - [ ] Merge PR
1.0
Crosplane - Retrospectively add the required role for enabling assuming role in AWS provider - - [x] Investigate how this is currently done - [x] Complete #480 - [x] Find and replace the referenced infrastructure-modules version in all capability sub-folders of aws-account-manifests and update to the new infrastructure-modules version - [x] Submit PR - [ ] Merge PR
non_process
crosplane retrospectively add the required role for enabling assuming role in aws provider investigate how this is currently done complete find and replace the referenced infrastructure modules version in all capability sub folders of aws account manifests and update to the new infrastructure modules version submit pr merge pr
0
94,072
15,962,327,718
IssuesEvent
2021-04-16 01:03:49
RG4421/terra-dev-site
https://api.github.com/repos/RG4421/terra-dev-site
opened
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz
security vulnerability
## CVE-2021-23362 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary> <p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p> <p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p> <p>Path to dependency file: terra-dev-site/package.json</p> <p>Path to vulnerable library: terra-dev-site/node_modules/hosted-git-info/package.json</p> <p> Dependency Hierarchy: - check-installed-dependencies-1.0.1.tgz (Root Library) - meow-3.7.0.tgz - normalize-package-data-2.5.0.tgz - :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/hosted-git-info/releases/tag/v3.0.8">https://github.com/npm/hosted-git-info/releases/tag/v3.0.8</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution: hosted-git-info - 3.0.8</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"hosted-git-info","packageVersion":"2.8.8","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"check-installed-dependencies:1.0.1;meow:3.7.0;normalize-package-data:2.5.0;hosted-git-info:2.8.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"hosted-git-info - 3.0.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23362","vulnerabilityDetails":"The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz - ## CVE-2021-23362 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary> <p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p> <p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p> <p>Path to dependency file: terra-dev-site/package.json</p> <p>Path to vulnerable library: terra-dev-site/node_modules/hosted-git-info/package.json</p> <p> Dependency Hierarchy: - check-installed-dependencies-1.0.1.tgz (Root Library) - meow-3.7.0.tgz - normalize-package-data-2.5.0.tgz - :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/hosted-git-info/releases/tag/v3.0.8">https://github.com/npm/hosted-git-info/releases/tag/v3.0.8</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution: hosted-git-info - 3.0.8</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"hosted-git-info","packageVersion":"2.8.8","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"check-installed-dependencies:1.0.1;meow:3.7.0;normalize-package-data:2.5.0;hosted-git-info:2.8.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"hosted-git-info - 3.0.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23362","vulnerabilityDetails":"The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in hosted git info tgz cve medium severity vulnerability vulnerable library hosted git info tgz provides metadata and conversions from repository urls for github bitbucket and gitlab library home page a href path to dependency file terra dev site package json path to vulnerable library terra dev site node modules hosted git info package json dependency hierarchy check installed dependencies tgz root library meow tgz normalize package data tgz x hosted git info tgz vulnerable library found in base branch master vulnerability details the package hosted git info before are vulnerable to regular expression denial of service redos via regular expression shortcutmatch in the fromurl function in index js the affected regular expression exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hosted git info isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree check installed dependencies meow normalize package data hosted git info isminimumfixversionavailable true minimumfixversion hosted git info basebranches vulnerabilityidentifier cve vulnerabilitydetails the package hosted git info before are vulnerable to regular expression denial of service redos via regular expression shortcutmatch in the fromurl function in index js the affected regular expression exhibits polynomial worst case time complexity vulnerabilityurl
0
2,236
5,088,589,846
IssuesEvent
2016-12-31 22:51:59
sw4j-org/tool-jpa-processor
https://api.github.com/repos/sw4j-org/tool-jpa-processor
opened
Handle @ForeignKey Annotation
annotation processor task
Handle the `@ForeignKey` annotation. See [JSR 338: Java Persistence API, Version 2.1](http://download.oracle.com/otn-pub/jcp/persistence-2_1-fr-eval-spec/JavaPersistence.pdf) - 11.1.19 Foreign Key Annotation
1.0
Handle @ForeignKey Annotation - Handle the `@ForeignKey` annotation. See [JSR 338: Java Persistence API, Version 2.1](http://download.oracle.com/otn-pub/jcp/persistence-2_1-fr-eval-spec/JavaPersistence.pdf) - 11.1.19 Foreign Key Annotation
process
handle foreignkey annotation handle the foreignkey annotation see foreign key annotation
1
54,990
13,943,655,069
IssuesEvent
2020-10-22 23:42:37
mlmcd/WebGoat
https://api.github.com/repos/mlmcd/WebGoat
opened
CVE-2019-10747 (High) detected in set-value-2.0.0.tgz, set-value-0.4.3.tgz
security vulnerability
## CVE-2019-10747 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>set-value-2.0.0.tgz</b>, <b>set-value-0.4.3.tgz</b></p></summary> <p> <details><summary><b>set-value-2.0.0.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz</a></p> <p>Path to dependency file: WebGoat/docs/package.json</p> <p>Path to vulnerable library: WebGoat/docs/node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.3.tgz (Root Library) - chokidar-2.0.4.tgz - braces-2.3.2.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - :x: **set-value-2.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>set-value-0.4.3.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz">https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz</a></p> <p>Path to dependency file: WebGoat/docs/package.json</p> <p>Path to vulnerable library: WebGoat/docs/node_modules/union-value/node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.3.tgz (Root Library) - chokidar-2.0.4.tgz - braces-2.3.2.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - union-value-1.0.0.tgz - :x: **set-value-0.4.3.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/mlmcd/WebGoat/commit/19e9dd9297c4eb1dbe74247f3c0eba17b31acc02">19e9dd9297c4eb1dbe74247f3c0eba17b31acc02</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> set-value is vulnerable to Prototype Pollution in versions lower than 3.0.1. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using any of the constructor, prototype and _proto_ payloads. <p>Publish Date: 2019-08-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10747>CVE-2019-10747</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f">https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f</a></p> <p>Release Date: 2019-07-24</p> <p>Fix Resolution: 2.0.1,3.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"set-value","packageVersion":"2.0.0","isTransitiveDependency":true,"dependencyTree":"browser-sync:2.26.3;chokidar:2.0.4;braces:2.3.2;snapdragon:0.8.2;base:0.11.2;cache-base:1.0.1;set-value:2.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.0.1,3.0.1"},{"packageType":"javascript/Node.js","packageName":"set-value","packageVersion":"0.4.3","isTransitiveDependency":true,"dependencyTree":"browser-sync:2.26.3;chokidar:2.0.4;braces:2.3.2;snapdragon:0.8.2;base:0.11.2;cache-base:1.0.1;union-value:1.0.0;set-value:0.4.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.0.1,3.0.1"}],"vulnerabilityIdentifier":"CVE-2019-10747","vulnerabilityDetails":"set-value is vulnerable to Prototype Pollution in versions lower than 3.0.1. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using any of the constructor, prototype and _proto_ payloads.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10747","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-10747 (High) detected in set-value-2.0.0.tgz, set-value-0.4.3.tgz - ## CVE-2019-10747 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>set-value-2.0.0.tgz</b>, <b>set-value-0.4.3.tgz</b></p></summary> <p> <details><summary><b>set-value-2.0.0.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz</a></p> <p>Path to dependency file: WebGoat/docs/package.json</p> <p>Path to vulnerable library: WebGoat/docs/node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.3.tgz (Root Library) - chokidar-2.0.4.tgz - braces-2.3.2.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - :x: **set-value-2.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>set-value-0.4.3.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz">https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz</a></p> <p>Path to dependency file: WebGoat/docs/package.json</p> <p>Path to vulnerable library: WebGoat/docs/node_modules/union-value/node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.3.tgz (Root Library) - chokidar-2.0.4.tgz - braces-2.3.2.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - union-value-1.0.0.tgz - :x: **set-value-0.4.3.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/mlmcd/WebGoat/commit/19e9dd9297c4eb1dbe74247f3c0eba17b31acc02">19e9dd9297c4eb1dbe74247f3c0eba17b31acc02</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> set-value is vulnerable to Prototype Pollution in versions lower than 3.0.1. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using any of the constructor, prototype and _proto_ payloads. <p>Publish Date: 2019-08-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10747>CVE-2019-10747</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f">https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f</a></p> <p>Release Date: 2019-07-24</p> <p>Fix Resolution: 2.0.1,3.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"set-value","packageVersion":"2.0.0","isTransitiveDependency":true,"dependencyTree":"browser-sync:2.26.3;chokidar:2.0.4;braces:2.3.2;snapdragon:0.8.2;base:0.11.2;cache-base:1.0.1;set-value:2.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.0.1,3.0.1"},{"packageType":"javascript/Node.js","packageName":"set-value","packageVersion":"0.4.3","isTransitiveDependency":true,"dependencyTree":"browser-sync:2.26.3;chokidar:2.0.4;braces:2.3.2;snapdragon:0.8.2;base:0.11.2;cache-base:1.0.1;union-value:1.0.0;set-value:0.4.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.0.1,3.0.1"}],"vulnerabilityIdentifier":"CVE-2019-10747","vulnerabilityDetails":"set-value is vulnerable to Prototype Pollution in versions lower than 3.0.1. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using any of the constructor, prototype and _proto_ payloads.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10747","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in set value tgz set value tgz cve high severity vulnerability vulnerable libraries set value tgz set value tgz set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href path to dependency file webgoat docs package json path to vulnerable library webgoat docs node modules set value package json dependency hierarchy browser sync tgz root library chokidar tgz braces tgz snapdragon tgz base tgz cache base tgz x set value tgz vulnerable library set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href path to dependency file webgoat docs package json path to vulnerable library webgoat docs node modules union value node modules set value package json dependency hierarchy browser sync tgz root library chokidar tgz braces tgz snapdragon tgz base tgz cache base tgz union value tgz x set value tgz vulnerable library found in head commit a href found in base branch develop vulnerability details set value is vulnerable to prototype pollution in versions lower than the function mixin deep could be tricked into adding or modifying properties of object prototype using any of the constructor prototype and proto payloads publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails set value is vulnerable to prototype pollution in versions lower than the function mixin deep could be tricked into adding or modifying properties of object prototype using any of the constructor prototype and proto payloads vulnerabilityurl
0
15,080
18,785,546,567
IssuesEvent
2021-11-08 11:44:14
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Drop experimental_multi_threaded_digest
type: process untriaged team-Local-Exec
This options defaults to true since a while and should just be removed, likely with the whole `SsdModule`.
1.0
Drop experimental_multi_threaded_digest - This options defaults to true since a while and should just be removed, likely with the whole `SsdModule`.
process
drop experimental multi threaded digest this options defaults to true since a while and should just be removed likely with the whole ssdmodule
1
22,171
30,721,409,378
IssuesEvent
2023-07-27 16:11:08
SpikeInterface/spikeinterface
https://api.github.com/repos/SpikeInterface/spikeinterface
closed
Questions about the modes "median"/"average" in `remove_artifacts`
question preprocessing
Hi! I am trying to explore the functionality of [`remove_artifacts`](https://github.com/SpikeInterface/spikeinterface/blob/main/src/spikeinterface/preprocessing/remove_artifacts.py) with mode "median" or "average" and have a few questions: (1) Just to confirm if my understanding is correct or not: The input argument `artifacts` gives a set of templates that suggest what each artifact looks like; if `None`, then the artifacts are calculated from the raw traces by indexing the triggers and extracting "waveforms". Additionally the "jitter" parameter is used to find a best match between the recording traces and the templates and the corresponding amplitude scaling factor. Is it the right way to interpret it? The [documentation](https://spikeinterface.readthedocs.io/en/latest/api.html#spikeinterface.preprocessing.remove_artifacts) only says that they are the "values" of the artifacts. I wonder in which cases the artifacts need to be explicitly specified; when specified, it is assumed that all the channels have the same artifact waveform? (2) The documentation says that providing argument `artifacts` can bypass `ms_before` and `ms_after`. But in the implementation of [`get_traces`](https://github.com/SpikeInterface/spikeinterface/blob/main/src/spikeinterface/preprocessing/remove_artifacts.py#L417) in `RemoveArtifactsRecordingSegment`, in "median"/"average" mode, the slicing of traces still depend on the `pad` property, which is derived from `ms_before` and `ms_after`. Then it seems to me that we need to manually make sure arguments `ms_before` and `ms_after` need to be aligned with `artifacts` so that they indicate the same length. Otherwise there may be mismatch when subtracting the artifacts from the recording traces. I am not sure if I understood the logic in source code correctly, nor have I tested. (3) For my own use case, I wanted to mask my stimulation artifacts to zero. But different stimulation artifacts have different durations. It seems like merely using `zeros` mode wouldn't accommodate different artifacts. If I use `median`/`average` mode while providing labels, it will calculate the artifact waveform template by the label and the subtract from raw data, not quite what I wanted to do. Could you please suggest some efficient ways to zero out the artifacts with varying durations? Thank you so much for your help. I really appreciate your active maintenance and development of this framework. It has been making my life a lot easier!
1.0
Questions about the modes "median"/"average" in `remove_artifacts` - Hi! I am trying to explore the functionality of [`remove_artifacts`](https://github.com/SpikeInterface/spikeinterface/blob/main/src/spikeinterface/preprocessing/remove_artifacts.py) with mode "median" or "average" and have a few questions: (1) Just to confirm if my understanding is correct or not: The input argument `artifacts` gives a set of templates that suggest what each artifact looks like; if `None`, then the artifacts are calculated from the raw traces by indexing the triggers and extracting "waveforms". Additionally the "jitter" parameter is used to find a best match between the recording traces and the templates and the corresponding amplitude scaling factor. Is it the right way to interpret it? The [documentation](https://spikeinterface.readthedocs.io/en/latest/api.html#spikeinterface.preprocessing.remove_artifacts) only says that they are the "values" of the artifacts. I wonder in which cases the artifacts need to be explicitly specified; when specified, it is assumed that all the channels have the same artifact waveform? (2) The documentation says that providing argument `artifacts` can bypass `ms_before` and `ms_after`. But in the implementation of [`get_traces`](https://github.com/SpikeInterface/spikeinterface/blob/main/src/spikeinterface/preprocessing/remove_artifacts.py#L417) in `RemoveArtifactsRecordingSegment`, in "median"/"average" mode, the slicing of traces still depend on the `pad` property, which is derived from `ms_before` and `ms_after`. Then it seems to me that we need to manually make sure arguments `ms_before` and `ms_after` need to be aligned with `artifacts` so that they indicate the same length. Otherwise there may be mismatch when subtracting the artifacts from the recording traces. I am not sure if I understood the logic in source code correctly, nor have I tested. (3) For my own use case, I wanted to mask my stimulation artifacts to zero. But different stimulation artifacts have different durations. It seems like merely using `zeros` mode wouldn't accommodate different artifacts. If I use `median`/`average` mode while providing labels, it will calculate the artifact waveform template by the label and the subtract from raw data, not quite what I wanted to do. Could you please suggest some efficient ways to zero out the artifacts with varying durations? Thank you so much for your help. I really appreciate your active maintenance and development of this framework. It has been making my life a lot easier!
process
questions about the modes median average in remove artifacts hi i am trying to explore the functionality of with mode median or average and have a few questions just to confirm if my understanding is correct or not the input argument artifacts gives a set of templates that suggest what each artifact looks like if none then the artifacts are calculated from the raw traces by indexing the triggers and extracting waveforms additionally the jitter parameter is used to find a best match between the recording traces and the templates and the corresponding amplitude scaling factor is it the right way to interpret it the only says that they are the values of the artifacts i wonder in which cases the artifacts need to be explicitly specified when specified it is assumed that all the channels have the same artifact waveform the documentation says that providing argument artifacts can bypass ms before and ms after but in the implementation of in removeartifactsrecordingsegment in median average mode the slicing of traces still depend on the pad property which is derived from ms before and ms after then it seems to me that we need to manually make sure arguments ms before and ms after need to be aligned with artifacts so that they indicate the same length otherwise there may be mismatch when subtracting the artifacts from the recording traces i am not sure if i understood the logic in source code correctly nor have i tested for my own use case i wanted to mask my stimulation artifacts to zero but different stimulation artifacts have different durations it seems like merely using zeros mode wouldn t accommodate different artifacts if i use median average mode while providing labels it will calculate the artifact waveform template by the label and the subtract from raw data not quite what i wanted to do could you please suggest some efficient ways to zero out the artifacts with varying durations thank you so much for your help i really appreciate your active maintenance and development of this framework it has been making my life a lot easier
1
155,414
24,463,073,836
IssuesEvent
2022-10-07 12:54:50
dccabs/housecallmd
https://api.github.com/repos/dccabs/housecallmd
closed
Contact Us - Redesign with tailwind ui
design upgrade
Please replace the current contact us page with tailwind ui https://www.housecallmd.org/contact - Please replace with this component - https://tailwindui.com/components/marketing/sections/contact-sections#component-dec976a631662b509173ba8b848c49cd - Make sure form still works - Show form is submitting indicator - Please show modal that confirms a message was sent - Please include some type of error handling - if the request fails we should know there is an error. - Please replace indigo colors with Rose in tailwind
1.0
Contact Us - Redesign with tailwind ui - Please replace the current contact us page with tailwind ui https://www.housecallmd.org/contact - Please replace with this component - https://tailwindui.com/components/marketing/sections/contact-sections#component-dec976a631662b509173ba8b848c49cd - Make sure form still works - Show form is submitting indicator - Please show modal that confirms a message was sent - Please include some type of error handling - if the request fails we should know there is an error. - Please replace indigo colors with Rose in tailwind
non_process
contact us redesign with tailwind ui please replace the current contact us page with tailwind ui please replace with this component make sure form still works show form is submitting indicator please show modal that confirms a message was sent please include some type of error handling if the request fails we should know there is an error please replace indigo colors with rose in tailwind
0
68,890
7,113,054,759
IssuesEvent
2018-01-17 19:07:03
GalaxyTrail/GFR_bugs
https://api.github.com/repos/GalaxyTrail/GFR_bugs
closed
1-2 eggplant statue house - enemy once got stuck in floor
bug maybe fixed - needs more testing
Submitter: Patrick Murphy Email: italiangamer97@gmail.com In the house of eggplant statues in Course 1-2, I noticed that an enemy managed to get stuck in the floor. Unsure of how I triggered it, but I was able to get the enemy out of the floor and defeat it. Perhaps I did something with one of the statues there?
1.0
1-2 eggplant statue house - enemy once got stuck in floor - Submitter: Patrick Murphy Email: italiangamer97@gmail.com In the house of eggplant statues in Course 1-2, I noticed that an enemy managed to get stuck in the floor. Unsure of how I triggered it, but I was able to get the enemy out of the floor and defeat it. Perhaps I did something with one of the statues there?
non_process
eggplant statue house enemy once got stuck in floor submitter patrick murphy email gmail com in the house of eggplant statues in course i noticed that an enemy managed to get stuck in the floor unsure of how i triggered it but i was able to get the enemy out of the floor and defeat it perhaps i did something with one of the statues there
0
559,706
16,569,422,307
IssuesEvent
2021-05-30 04:40:41
gitextensions/gitextensions
https://api.github.com/repos/gitextensions/gitextensions
closed
GE hangs when viewing submodules in diff viewer
:beetle: type: bug :construction: status: in progress :grey_exclamation: priority: high
## Current behaviour Regression in master in the last days. If a submodule is selected in worktree (at least - seem OK for HEAD though), GE hangs and stops responding. Status is never changed from the default. Mouse just spinning. ## Expected behaviour Submodule is displayed. ## Steps to reproduce Select modified submodule in worktree in the diff viewer. ## Screenshots ![image](https://user-images.githubusercontent.com/6248932/109844938-e79b3780-7c4c-11eb-9443-1ddee6a2fa6d.png) ## Did this work in previous version of GitExtensions - Git Extensions 3.6.0.11114 - Build f54a727a95393dcacfbbd8603dfd6daf931618c4 (Dirty) ## Environment - Git Extensions 3.6.0.11170 - Build 273a0f6fd3e07858f837cdc19d50827871e32319 (Dirty) - Git 2.30.0.windows.1 - Microsoft Windows NT 10.0.17763.0 - .NET Framework 4.8.4300.0 - DPI 96dpi (no scaling)
1.0
GE hangs when viewing submodules in diff viewer - ## Current behaviour Regression in master in the last days. If a submodule is selected in worktree (at least - seem OK for HEAD though), GE hangs and stops responding. Status is never changed from the default. Mouse just spinning. ## Expected behaviour Submodule is displayed. ## Steps to reproduce Select modified submodule in worktree in the diff viewer. ## Screenshots ![image](https://user-images.githubusercontent.com/6248932/109844938-e79b3780-7c4c-11eb-9443-1ddee6a2fa6d.png) ## Did this work in previous version of GitExtensions - Git Extensions 3.6.0.11114 - Build f54a727a95393dcacfbbd8603dfd6daf931618c4 (Dirty) ## Environment - Git Extensions 3.6.0.11170 - Build 273a0f6fd3e07858f837cdc19d50827871e32319 (Dirty) - Git 2.30.0.windows.1 - Microsoft Windows NT 10.0.17763.0 - .NET Framework 4.8.4300.0 - DPI 96dpi (no scaling)
non_process
ge hangs when viewing submodules in diff viewer current behaviour regression in master in the last days if a submodule is selected in worktree at least seem ok for head though ge hangs and stops responding status is never changed from the default mouse just spinning expected behaviour submodule is displayed steps to reproduce select modified submodule in worktree in the diff viewer screenshots did this work in previous version of gitextensions git extensions build dirty environment git extensions build dirty git windows microsoft windows nt net framework dpi no scaling
0
657,658
21,799,590,185
IssuesEvent
2022-05-16 02:30:09
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Add setting to toggle visibility of Bookmark button in the URL bar
priority/P4 feature/settings OS/Desktop
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> It's easy to accidentally click the bookmark button in the URL bar as it's positioned between the refresh button and the URL input - resulting in unwanted bookmarks. Also, some people don't use the native browser bookmarks at all so this button is not always of use. ![image](https://user-images.githubusercontent.com/4908432/104725284-c5476c00-5729-11eb-9cff-d41677c90623.png) Ideally, there should be an Appearance setting for "Show Bookmark button" which works similarly to the "Show Home button" setting. ![image](https://user-images.githubusercontent.com/4908432/104726001-cfb63580-572a-11eb-8a4c-1ea878a46463.png) ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Have poor hand-eye-coordination 2. Accidentally click bookmark button instead of URL input or refresh button 3. Unwanted bookmark saved ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> I do this multiple times per week. ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.18.77 Chromium: 87.0.4280.101 (Official Build) unknown (64-bit) -- | -- Revision | 9407c80213cda69c2b7abcb4fa8e3f74488f4956-refs/branch-heads/4280@{#1807} OS | Linux JavaScript | V8 8.7.220.29 ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? Yes - Can you reproduce this issue with the beta channel? Yes - Can you reproduce this issue with the nightly channel? Yes ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? N/A - Does the issue resolve itself when disabling Brave Rewards? N/A - Is the issue reproducible on the latest version of Chrome? N/A ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> - Related ticket: https://github.com/brave/brave-browser/issues/774 - Related comment: https://github.com/brave/brave-browser/issues/774#issuecomment-760664719 - [Community thread](https://community.brave.com/t/is-there-a-way-to-move-the-bookmarks-button-to-where-it-is-in-chrome/48785/12) P.S. Thanks for all the amazing work on Brave!
1.0
Add setting to toggle visibility of Bookmark button in the URL bar - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> It's easy to accidentally click the bookmark button in the URL bar as it's positioned between the refresh button and the URL input - resulting in unwanted bookmarks. Also, some people don't use the native browser bookmarks at all so this button is not always of use. ![image](https://user-images.githubusercontent.com/4908432/104725284-c5476c00-5729-11eb-9cff-d41677c90623.png) Ideally, there should be an Appearance setting for "Show Bookmark button" which works similarly to the "Show Home button" setting. ![image](https://user-images.githubusercontent.com/4908432/104726001-cfb63580-572a-11eb-8a4c-1ea878a46463.png) ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Have poor hand-eye-coordination 2. Accidentally click bookmark button instead of URL input or refresh button 3. Unwanted bookmark saved ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> I do this multiple times per week. ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.18.77 Chromium: 87.0.4280.101 (Official Build) unknown (64-bit) -- | -- Revision | 9407c80213cda69c2b7abcb4fa8e3f74488f4956-refs/branch-heads/4280@{#1807} OS | Linux JavaScript | V8 8.7.220.29 ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? Yes - Can you reproduce this issue with the beta channel? Yes - Can you reproduce this issue with the nightly channel? Yes ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? N/A - Does the issue resolve itself when disabling Brave Rewards? N/A - Is the issue reproducible on the latest version of Chrome? N/A ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> - Related ticket: https://github.com/brave/brave-browser/issues/774 - Related comment: https://github.com/brave/brave-browser/issues/774#issuecomment-760664719 - [Community thread](https://community.brave.com/t/is-there-a-way-to-move-the-bookmarks-button-to-where-it-is-in-chrome/48785/12) P.S. Thanks for all the amazing work on Brave!
non_process
add setting to toggle visibility of bookmark button in the url bar have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description it s easy to accidentally click the bookmark button in the url bar as it s positioned between the refresh button and the url input resulting in unwanted bookmarks also some people don t use the native browser bookmarks at all so this button is not always of use ideally there should be an appearance setting for show bookmark button which works similarly to the show home button setting steps to reproduce have poor hand eye coordination accidentally click bookmark button instead of url input or refresh button unwanted bookmark saved reproduces how often i do this multiple times per week brave version brave version info brave chromium   official build  unknown  bit revision refs branch heads os linux javascript   version channel information can you reproduce this issue with the current release yes can you reproduce this issue with the beta channel yes can you reproduce this issue with the nightly channel yes other additional information does the issue resolve itself when disabling brave shields n a does the issue resolve itself when disabling brave rewards n a is the issue reproducible on the latest version of chrome n a miscellaneous information related ticket related comment p s thanks for all the amazing work on brave
0
731
3,214,309,873
IssuesEvent
2015-10-07 00:42:40
broadinstitute/hellbender
https://api.github.com/repos/broadinstitute/hellbender
closed
Add mechanism to generate unique ids for data types (reads, variants, etc.)
Dataflow DataflowPreprocessingPipeline
Needed for GroupByKey, since Java serialization is not deterministic. Initial idea is to create IDs based on the source of each record (eg., URI + file offset or record number).
1.0
Add mechanism to generate unique ids for data types (reads, variants, etc.) - Needed for GroupByKey, since Java serialization is not deterministic. Initial idea is to create IDs based on the source of each record (eg., URI + file offset or record number).
process
add mechanism to generate unique ids for data types reads variants etc needed for groupbykey since java serialization is not deterministic initial idea is to create ids based on the source of each record eg uri file offset or record number
1
28,127
11,590,028,152
IssuesEvent
2020-02-24 05:03:37
aksiksi/vaulty
https://api.github.com/repos/aksiksi/vaulty
closed
Add support for address whitelisting
security vaulty-mail
Allow users to specify a list of whitelisted sender addresses a Vaulty address under their control. In addition, it might make sense to allow users to define a "secret word" that **must** be present in the email body or subject to accept the email. The word could be stripped from the email before getting pushed to storage.
True
Add support for address whitelisting - Allow users to specify a list of whitelisted sender addresses a Vaulty address under their control. In addition, it might make sense to allow users to define a "secret word" that **must** be present in the email body or subject to accept the email. The word could be stripped from the email before getting pushed to storage.
non_process
add support for address whitelisting allow users to specify a list of whitelisted sender addresses a vaulty address under their control in addition it might make sense to allow users to define a secret word that must be present in the email body or subject to accept the email the word could be stripped from the email before getting pushed to storage
0
6,410
9,488,866,277
IssuesEvent
2019-04-22 20:45:24
material-components/material-components-ios
https://api.github.com/repos/material-components/material-components-ios
closed
[AppBar] Mark AppBar theming extension ready
[AppBar] type:Process
This was filed as an internal issue. If you are a Googler, please visit [b/130716217](http://b/130716217) for more details. <!-- Auto-generated content below, do not modify --> --- #### Internal data - Associated internal bug: [b/130716217](http://b/130716217)
1.0
[AppBar] Mark AppBar theming extension ready - This was filed as an internal issue. If you are a Googler, please visit [b/130716217](http://b/130716217) for more details. <!-- Auto-generated content below, do not modify --> --- #### Internal data - Associated internal bug: [b/130716217](http://b/130716217)
process
mark appbar theming extension ready this was filed as an internal issue if you are a googler please visit for more details internal data associated internal bug
1
8,987
12,100,407,087
IssuesEvent
2020-04-20 13:45:31
ComposableWeb/poolbase
https://api.github.com/repos/ComposableWeb/poolbase
opened
[💥FEAT] derive color palette from page screenshot
enhancement epic: processing
**Feature request? Please describe.** For frontend display of pages: Derive colors from screenshot using AI **Acceptance Criteria - Describe the solution you'd like** A clear and concise description of what you want to happen in bullet points: * ... **Related issues** #1 And any other context or screenshots about the feature request here.
1.0
[💥FEAT] derive color palette from page screenshot - **Feature request? Please describe.** For frontend display of pages: Derive colors from screenshot using AI **Acceptance Criteria - Describe the solution you'd like** A clear and concise description of what you want to happen in bullet points: * ... **Related issues** #1 And any other context or screenshots about the feature request here.
process
derive color palette from page screenshot feature request please describe for frontend display of pages derive colors from screenshot using ai acceptance criteria describe the solution you d like a clear and concise description of what you want to happen in bullet points related issues and any other context or screenshots about the feature request here
1
67,843
7,065,242,262
IssuesEvent
2018-01-06 17:38:28
pandas-dev/pandas
https://api.github.com/repos/pandas-dev/pandas
closed
`(x + 1) % 1` inside test_multi.py
MultiIndex Testing
#### Code Sample, a copy-pastable example if possible https://github.com/pandas-dev/pandas/blob/93033151a8d8aaa650a81df9f41347758bf6c393/pandas/tests/indexes/test_multi.py#L170-L171 #### Problem description I assume this is supposed to shift the labels cyclically by one. Then `% 4` and `% 2` should be used. Otherwise its at least weird to write `(x + 1) % 1` instead of `0`.
1.0
`(x + 1) % 1` inside test_multi.py - #### Code Sample, a copy-pastable example if possible https://github.com/pandas-dev/pandas/blob/93033151a8d8aaa650a81df9f41347758bf6c393/pandas/tests/indexes/test_multi.py#L170-L171 #### Problem description I assume this is supposed to shift the labels cyclically by one. Then `% 4` and `% 2` should be used. Otherwise its at least weird to write `(x + 1) % 1` instead of `0`.
non_process
x inside test multi py code sample a copy pastable example if possible problem description i assume this is supposed to shift the labels cyclically by one then and should be used otherwise its at least weird to write x instead of
0
3,921
6,843,043,013
IssuesEvent
2017-11-12 10:48:24
pwittchen/ReactiveNetwork
https://api.github.com/repos/pwittchen/ReactiveNetwork
opened
RxJava1.x: release 0.12.2
release process
**Initial realese notes**: - updated project dependencies - updated Gradle to 3.0.0 **Things to do**: - [ ] RxJava1.x branch: - [ ] bump library version - [ ] upload archives to Maven Central - [ ] close and release artifact on Maven Central - [ ] update `CHANGELOG.md` after Maven Sync - [ ] bump library version in `README.md` - [ ] create new GitHub release
1.0
RxJava1.x: release 0.12.2 - **Initial realese notes**: - updated project dependencies - updated Gradle to 3.0.0 **Things to do**: - [ ] RxJava1.x branch: - [ ] bump library version - [ ] upload archives to Maven Central - [ ] close and release artifact on Maven Central - [ ] update `CHANGELOG.md` after Maven Sync - [ ] bump library version in `README.md` - [ ] create new GitHub release
process
x release initial realese notes updated project dependencies updated gradle to things to do x branch bump library version upload archives to maven central close and release artifact on maven central update changelog md after maven sync bump library version in readme md create new github release
1
103,843
11,383,025,622
IssuesEvent
2020-01-29 04:20:55
car12lin12/InstaBaddies
https://api.github.com/repos/car12lin12/InstaBaddies
opened
Update readme to clarify goals of the project
Documentation :memo:
Update the readme to show the description / objective / core features of the project. Core features : - upload a picture - follow a user - receive notifications from that user if they upload a new picture - comment on a picture
1.0
Update readme to clarify goals of the project - Update the readme to show the description / objective / core features of the project. Core features : - upload a picture - follow a user - receive notifications from that user if they upload a new picture - comment on a picture
non_process
update readme to clarify goals of the project update the readme to show the description objective core features of the project core features upload a picture follow a user receive notifications from that user if they upload a new picture comment on a picture
0
238,054
19,694,490,725
IssuesEvent
2022-01-12 10:40:01
dotnet/sdk
https://api.github.com/repos/dotnet/sdk
closed
With .NET 7 x86 SDK, Testhost process exited with error: It was not possible to find any compatible framework version
Area-DotNet Test untriaged
**--Repro Steps---** 1. Install .NET 7 x86 SDK(https://github.com/dotnet/installer) on Win x64 OS 2. Create a UT project and run it **--Expected Result --** 1. dotnet test works fine **--Actual Result--** 1. Testhost process exited with error: It was not possible to find any compatible framework version The framework 'Microsoft.NETCore.App', version '7.0.0-alpha.1.22059.2' (x64) was not found. ![image](https://user-images.githubusercontent.com/1578239/148919893-8af04dbe-0d56-47af-8b46-0a42adb088d1.png) **--Note--** It didnot repro with .NET 6 x86 SDK **Dotnet info** c:\6t>dotnet --info .NET SDK (reflecting any global.json): Version: 7.0.100-alpha.1.22060.7 Commit: 4a2d57a652 Runtime Environment: OS Name: Windows OS Version: 10.0.22000 OS Platform: Windows RID: win10-x86 Base Path: C:\Program Files (x86)\dotnet\sdk\7.0.100-alpha.1.22060.7\ Host (useful for support): Version: 7.0.0-alpha.1.22059.2 Commit: a2af629476 .NET SDKs installed: 6.0.200-preview.22055.18 [C:\Program Files (x86)\dotnet\sdk] 7.0.100-alpha.1.22060.7 [C:\Program Files (x86)\dotnet\sdk] .NET runtimes installed: Microsoft.AspNetCore.App 6.0.1 [C:\Program Files (x86)\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 7.0.0-alpha.1.22057.17 [C:\Program Files (x86)\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.NETCore.App 6.0.1 [C:\Program Files (x86)\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 7.0.0-alpha.1.22059.2 [C:\Program Files (x86)\dotnet\shared\Microsoft.NETCore.App] Microsoft.WindowsDesktop.App 6.0.1 [C:\Program Files (x86)\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 7.0.0-alpha.1.22055.1 [C:\Program Files (x86)\dotnet\shared\Microsoft.WindowsDesktop.App] To install additional .NET runtimes or SDKs: https://aka.ms/dotnet-download
1.0
With .NET 7 x86 SDK, Testhost process exited with error: It was not possible to find any compatible framework version - **--Repro Steps---** 1. Install .NET 7 x86 SDK(https://github.com/dotnet/installer) on Win x64 OS 2. Create a UT project and run it **--Expected Result --** 1. dotnet test works fine **--Actual Result--** 1. Testhost process exited with error: It was not possible to find any compatible framework version The framework 'Microsoft.NETCore.App', version '7.0.0-alpha.1.22059.2' (x64) was not found. ![image](https://user-images.githubusercontent.com/1578239/148919893-8af04dbe-0d56-47af-8b46-0a42adb088d1.png) **--Note--** It didnot repro with .NET 6 x86 SDK **Dotnet info** c:\6t>dotnet --info .NET SDK (reflecting any global.json): Version: 7.0.100-alpha.1.22060.7 Commit: 4a2d57a652 Runtime Environment: OS Name: Windows OS Version: 10.0.22000 OS Platform: Windows RID: win10-x86 Base Path: C:\Program Files (x86)\dotnet\sdk\7.0.100-alpha.1.22060.7\ Host (useful for support): Version: 7.0.0-alpha.1.22059.2 Commit: a2af629476 .NET SDKs installed: 6.0.200-preview.22055.18 [C:\Program Files (x86)\dotnet\sdk] 7.0.100-alpha.1.22060.7 [C:\Program Files (x86)\dotnet\sdk] .NET runtimes installed: Microsoft.AspNetCore.App 6.0.1 [C:\Program Files (x86)\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 7.0.0-alpha.1.22057.17 [C:\Program Files (x86)\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.NETCore.App 6.0.1 [C:\Program Files (x86)\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 7.0.0-alpha.1.22059.2 [C:\Program Files (x86)\dotnet\shared\Microsoft.NETCore.App] Microsoft.WindowsDesktop.App 6.0.1 [C:\Program Files (x86)\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 7.0.0-alpha.1.22055.1 [C:\Program Files (x86)\dotnet\shared\Microsoft.WindowsDesktop.App] To install additional .NET runtimes or SDKs: https://aka.ms/dotnet-download
non_process
with net sdk testhost process exited with error it was not possible to find any compatible framework version repro steps install net sdk on win os create a ut project and run it expected result dotnet test works fine actual result testhost process exited with error it was not possible to find any compatible framework version the framework microsoft netcore app version alpha was not found note it didnot repro with net sdk dotnet info c dotnet info net sdk reflecting any global json version alpha commit runtime environment os name windows os version os platform windows rid base path c program files dotnet sdk alpha host useful for support version alpha commit net sdks installed preview alpha net runtimes installed microsoft aspnetcore app microsoft aspnetcore app alpha microsoft netcore app microsoft netcore app alpha microsoft windowsdesktop app microsoft windowsdesktop app alpha to install additional net runtimes or sdks
0
16,951
22,306,072,081
IssuesEvent
2022-06-13 13:09:39
deepset-ai/haystack
https://api.github.com/repos/deepset-ai/haystack
closed
Crawler: can it handle website with contents dynamically loaded by JavaScript?
type:feature good first issue Contributions wanted! good second issue topic:preprocessing journey:intermediate
**Question** Per captioned, can crawler handle website with contents dynamically loaded by pressing an "expand" button? **Additional context** In today's website designs a lot of them display content dynamically through Javascript. On inspecting Page Source they will be shown as loading in on client side. Some contents are to be loaded when an "expand" button (or eqivalent) is pressed. Tested with the current version of Crawler node and only the statis title/heading can be scraped, but not any of the dynamic text. Is that expected/limitation of Selenium/bug? Any way to scrape from these kind of pages? E.g. [https://www.sfc.hk/en/Forms/Intermediaries/Licensing-forms](url) **FAQ Check** - [X] Have you had a look at [our new FAQ page](https://haystack.deepset.ai/overview/faq)?
1.0
Crawler: can it handle website with contents dynamically loaded by JavaScript? - **Question** Per captioned, can crawler handle website with contents dynamically loaded by pressing an "expand" button? **Additional context** In today's website designs a lot of them display content dynamically through Javascript. On inspecting Page Source they will be shown as loading in on client side. Some contents are to be loaded when an "expand" button (or eqivalent) is pressed. Tested with the current version of Crawler node and only the statis title/heading can be scraped, but not any of the dynamic text. Is that expected/limitation of Selenium/bug? Any way to scrape from these kind of pages? E.g. [https://www.sfc.hk/en/Forms/Intermediaries/Licensing-forms](url) **FAQ Check** - [X] Have you had a look at [our new FAQ page](https://haystack.deepset.ai/overview/faq)?
process
crawler can it handle website with contents dynamically loaded by javascript question per captioned can crawler handle website with contents dynamically loaded by pressing an expand button additional context in today s website designs a lot of them display content dynamically through javascript on inspecting page source they will be shown as loading in on client side some contents are to be loaded when an expand button or eqivalent is pressed tested with the current version of crawler node and only the statis title heading can be scraped but not any of the dynamic text is that expected limitation of selenium bug any way to scrape from these kind of pages e g url faq check have you had a look at
1
11,179
13,957,695,350
IssuesEvent
2020-10-24 08:11:31
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
PT: Harvesting
Geoportal Harvesting process PT - Portugal
Geoportal team, Can you please start a harvesting do the Portuguese catalogue? Thank you!
1.0
PT: Harvesting - Geoportal team, Can you please start a harvesting do the Portuguese catalogue? Thank you!
process
pt harvesting geoportal team can you please start a harvesting do the portuguese catalogue thank you
1
8,761
11,880,979,031
IssuesEvent
2020-03-27 11:45:07
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Test Failed: Process terminated. Assertion failed. "!_fileHandle.IsClosed"
area-System.Diagnostics.Process os-windows test bug test-corefx
**Detail:** https://helix.dot.net/api/2019-06-17/jobs/0c15185a-2331-4dcf-978c-cd7c4a4b3189/workitems/System.Diagnostics.Process.Tests/console **Log:** ``` C:\dotnetbuild\work\0c15185a-2331-4dcf-978c-cd7c4a4b3189\Work\2eeb62f3-774b-43f1-94ac-c844f873d662\Exec>"C:\dotnetbuild\work\0c15185a-2331-4dcf-978c-cd7c4a4b3189\Payload\dotnet.exe" exec --runtimeconfig System.Diagnostics.Process.Tests.runtimeconfig.json xunit.console.dll System.Diagnostics.Process.Tests.dll -xml testResults.xml -nologo -nocolor -notrait category=nonnetcoreapptests -notrait category=nonwindowstests -notrait category=IgnoreForCI -notrait category=OuterLoop -notrait category=failing @C:\dotnetbuild\work\0c15185a-2331-4dcf-978c-cd7c4a4b3189\Payload\CoreFX.issues.rsp Discovering: System.Diagnostics.Process.Tests (method display = ClassAndMethod, method display options = None) Discovered: System.Diagnostics.Process.Tests (found 221 of 241 test cases) Starting: System.Diagnostics.Process.Tests (parallel test collections = on, max threads = 2) System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_WebPage [SKIP] Manual test System.Diagnostics.Tests.ProcessStartInfoTests.ShellExecute_Nano_Fails_Start [SKIP] Condition(s) not met: "IsWindowsNanoServer" Process terminated. Assertion failed. !_fileHandle.IsClosed at System.IO.FileStream.AssertCanRead() in /_/src/System.Private.CoreLib/shared/System/IO/FileStream.Windows.cs:line 482 at System.IO.FileStream.ReadNative(Span`1 buffer) in /_/src/System.Private.CoreLib/shared/System/IO/FileStream.Windows.cs:line 497 at System.IO.FileStream.ReadSpan(Span`1 destination) in /_/src/System.Private.CoreLib/shared/System/IO/FileStream.Windows.cs:line 429 at System.IO.FileStream.Read(Byte[] array, Int32 offset, Int32 count) in /_/src/System.Private.CoreLib/shared/System/IO/FileStream.cs:line 304 at System.IO.Stream.<>c.<BeginReadInternal>b__43_0(Object <p0>) in /_/src/System.Private.CoreLib/shared/System/IO/Stream.cs:line 303 at System.Threading.Tasks.Task`1.InnerInvoke() in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Future.cs:line 512 at System.Threading.Tasks.Task.<>c.<.cctor>b__274_0(Object obj) in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Task.cs:line 2372 at System.Threading.ExecutionContext.RunFromThreadPoolDispatchLoop(Thread threadPoolThread, ExecutionContext executionContext, ContextCallback callback, Object state) in /_/src/System.Private.CoreLib/shared/System/Threading/ExecutionContext.cs:line 289 at System.Threading.Tasks.Task.ExecuteWithThreadLocal(Task& currentTaskSlot, Thread threadPoolThread) in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Task.cs:line 2333 at System.Threading.Tasks.Task.ExecuteEntryUnsafe(Thread threadPoolThread) in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Task.cs:line 2271 at System.Threading.Tasks.Task.ExecuteFromThreadPool(Thread threadPoolThread) in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Task.cs:line 2256 at System.Threading.ThreadPoolWorkQueue.Dispatch() in /_/src/System.Private.CoreLib/shared/System/Threading/ThreadPool.cs:line 664 at System.Threading._ThreadPoolWaitCallback.PerformWaitCallback() in /_/src/System.Private.CoreLib/src/System/Threading/ThreadPool.CoreCLR.cs:line 29 exit code -2146232797 ```
1.0
Test Failed: Process terminated. Assertion failed. "!_fileHandle.IsClosed" - **Detail:** https://helix.dot.net/api/2019-06-17/jobs/0c15185a-2331-4dcf-978c-cd7c4a4b3189/workitems/System.Diagnostics.Process.Tests/console **Log:** ``` C:\dotnetbuild\work\0c15185a-2331-4dcf-978c-cd7c4a4b3189\Work\2eeb62f3-774b-43f1-94ac-c844f873d662\Exec>"C:\dotnetbuild\work\0c15185a-2331-4dcf-978c-cd7c4a4b3189\Payload\dotnet.exe" exec --runtimeconfig System.Diagnostics.Process.Tests.runtimeconfig.json xunit.console.dll System.Diagnostics.Process.Tests.dll -xml testResults.xml -nologo -nocolor -notrait category=nonnetcoreapptests -notrait category=nonwindowstests -notrait category=IgnoreForCI -notrait category=OuterLoop -notrait category=failing @C:\dotnetbuild\work\0c15185a-2331-4dcf-978c-cd7c4a4b3189\Payload\CoreFX.issues.rsp Discovering: System.Diagnostics.Process.Tests (method display = ClassAndMethod, method display options = None) Discovered: System.Diagnostics.Process.Tests (found 221 of 241 test cases) Starting: System.Diagnostics.Process.Tests (parallel test collections = on, max threads = 2) System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_WebPage [SKIP] Manual test System.Diagnostics.Tests.ProcessStartInfoTests.ShellExecute_Nano_Fails_Start [SKIP] Condition(s) not met: "IsWindowsNanoServer" Process terminated. Assertion failed. !_fileHandle.IsClosed at System.IO.FileStream.AssertCanRead() in /_/src/System.Private.CoreLib/shared/System/IO/FileStream.Windows.cs:line 482 at System.IO.FileStream.ReadNative(Span`1 buffer) in /_/src/System.Private.CoreLib/shared/System/IO/FileStream.Windows.cs:line 497 at System.IO.FileStream.ReadSpan(Span`1 destination) in /_/src/System.Private.CoreLib/shared/System/IO/FileStream.Windows.cs:line 429 at System.IO.FileStream.Read(Byte[] array, Int32 offset, Int32 count) in /_/src/System.Private.CoreLib/shared/System/IO/FileStream.cs:line 304 at System.IO.Stream.<>c.<BeginReadInternal>b__43_0(Object <p0>) in /_/src/System.Private.CoreLib/shared/System/IO/Stream.cs:line 303 at System.Threading.Tasks.Task`1.InnerInvoke() in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Future.cs:line 512 at System.Threading.Tasks.Task.<>c.<.cctor>b__274_0(Object obj) in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Task.cs:line 2372 at System.Threading.ExecutionContext.RunFromThreadPoolDispatchLoop(Thread threadPoolThread, ExecutionContext executionContext, ContextCallback callback, Object state) in /_/src/System.Private.CoreLib/shared/System/Threading/ExecutionContext.cs:line 289 at System.Threading.Tasks.Task.ExecuteWithThreadLocal(Task& currentTaskSlot, Thread threadPoolThread) in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Task.cs:line 2333 at System.Threading.Tasks.Task.ExecuteEntryUnsafe(Thread threadPoolThread) in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Task.cs:line 2271 at System.Threading.Tasks.Task.ExecuteFromThreadPool(Thread threadPoolThread) in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Task.cs:line 2256 at System.Threading.ThreadPoolWorkQueue.Dispatch() in /_/src/System.Private.CoreLib/shared/System/Threading/ThreadPool.cs:line 664 at System.Threading._ThreadPoolWaitCallback.PerformWaitCallback() in /_/src/System.Private.CoreLib/src/System/Threading/ThreadPool.CoreCLR.cs:line 29 exit code -2146232797 ```
process
test failed process terminated assertion failed filehandle isclosed detail log c dotnetbuild work work exec c dotnetbuild work payload dotnet exe exec runtimeconfig system diagnostics process tests runtimeconfig json xunit console dll system diagnostics process tests dll xml testresults xml nologo nocolor notrait category nonnetcoreapptests notrait category nonwindowstests notrait category ignoreforci notrait category outerloop notrait category failing c dotnetbuild work payload corefx issues rsp discovering system diagnostics process tests method display classandmethod method display options none discovered system diagnostics process tests found of test cases starting system diagnostics process tests parallel test collections on max threads system diagnostics tests processstartinfotests startinfo webpage manual test system diagnostics tests processstartinfotests shellexecute nano fails start condition s not met iswindowsnanoserver process terminated assertion failed filehandle isclosed at system io filestream assertcanread in src system private corelib shared system io filestream windows cs line at system io filestream readnative span buffer in src system private corelib shared system io filestream windows cs line at system io filestream readspan span destination in src system private corelib shared system io filestream windows cs line at system io filestream read byte array offset count in src system private corelib shared system io filestream cs line at system io stream c b object in src system private corelib shared system io stream cs line at system threading tasks task innerinvoke in src system private corelib shared system threading tasks future cs line at system threading tasks task c b object obj in src system private corelib shared system threading tasks task cs line at system threading executioncontext runfromthreadpooldispatchloop thread threadpoolthread executioncontext executioncontext contextcallback callback object state in src system private corelib shared system threading executioncontext cs line at system threading tasks task executewiththreadlocal task currenttaskslot thread threadpoolthread in src system private corelib shared system threading tasks task cs line at system threading tasks task executeentryunsafe thread threadpoolthread in src system private corelib shared system threading tasks task cs line at system threading tasks task executefromthreadpool thread threadpoolthread in src system private corelib shared system threading tasks task cs line at system threading threadpoolworkqueue dispatch in src system private corelib shared system threading threadpool cs line at system threading threadpoolwaitcallback performwaitcallback in src system private corelib src system threading threadpool coreclr cs line exit code
1
10,234
13,096,025,883
IssuesEvent
2020-08-03 15:02:04
ZbayApp/zbay
https://api.github.com/repos/ZbayApp/zbay
closed
Automate Windows deploy process
dev process
Automate Windows build, signing with cert if possible. Look into #148 first.
1.0
Automate Windows deploy process - Automate Windows build, signing with cert if possible. Look into #148 first.
process
automate windows deploy process automate windows build signing with cert if possible look into first
1
241,155
26,256,663,675
IssuesEvent
2023-01-06 01:46:12
EmpoHQ/empo.im
https://api.github.com/repos/EmpoHQ/empo.im
opened
CVE-2022-24999 (High) detected in qs-6.10.1.tgz
security vulnerability
## CVE-2022-24999 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-6.10.1.tgz</b></p></summary> <p>A querystring parser that supports nesting and arrays, with a depth limit</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.10.1.tgz">https://registry.npmjs.org/qs/-/qs-6.10.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/qs/package.json</p> <p> Dependency Hierarchy: - nuxt-2.15.6.tgz (Root Library) - telemetry-1.3.3.tgz - git-url-parse-11.4.4.tgz - git-up-4.0.2.tgz - parse-url-5.0.2.tgz - parse-path-4.0.3.tgz - :x: **qs-6.10.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/EmpoHQ/empo.im/commit/488d7a9b0ad7df016dd006a95c3b06f4883d8561">488d7a9b0ad7df016dd006a95c3b06f4883d8561</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> qs before 6.10.3, as used in Express before 4.17.3 and other products, allows attackers to cause a Node process hang for an Express application because an __ proto__ key can be used. In many typical Express use cases, an unauthenticated remote attacker can place the attack payload in the query string of the URL that is used to visit the application, such as a[__proto__]=b&a[__proto__]&a[length]=100000000. The fix was backported to qs 6.9.7, 6.8.3, 6.7.3, 6.6.1, 6.5.3, 6.4.1, 6.3.3, and 6.2.4 (and therefore Express 4.17.3, which has "deps: qs@6.9.7" in its release description, is not vulnerable). <p>Publish Date: 2022-11-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24999>CVE-2022-24999</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-24999">https://www.cve.org/CVERecord?id=CVE-2022-24999</a></p> <p>Release Date: 2022-11-26</p> <p>Fix Resolution (qs): 6.10.3</p> <p>Direct dependency fix Resolution (nuxt): 2.15.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-24999 (High) detected in qs-6.10.1.tgz - ## CVE-2022-24999 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-6.10.1.tgz</b></p></summary> <p>A querystring parser that supports nesting and arrays, with a depth limit</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.10.1.tgz">https://registry.npmjs.org/qs/-/qs-6.10.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/qs/package.json</p> <p> Dependency Hierarchy: - nuxt-2.15.6.tgz (Root Library) - telemetry-1.3.3.tgz - git-url-parse-11.4.4.tgz - git-up-4.0.2.tgz - parse-url-5.0.2.tgz - parse-path-4.0.3.tgz - :x: **qs-6.10.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/EmpoHQ/empo.im/commit/488d7a9b0ad7df016dd006a95c3b06f4883d8561">488d7a9b0ad7df016dd006a95c3b06f4883d8561</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> qs before 6.10.3, as used in Express before 4.17.3 and other products, allows attackers to cause a Node process hang for an Express application because an __ proto__ key can be used. In many typical Express use cases, an unauthenticated remote attacker can place the attack payload in the query string of the URL that is used to visit the application, such as a[__proto__]=b&a[__proto__]&a[length]=100000000. The fix was backported to qs 6.9.7, 6.8.3, 6.7.3, 6.6.1, 6.5.3, 6.4.1, 6.3.3, and 6.2.4 (and therefore Express 4.17.3, which has "deps: qs@6.9.7" in its release description, is not vulnerable). <p>Publish Date: 2022-11-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24999>CVE-2022-24999</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-24999">https://www.cve.org/CVERecord?id=CVE-2022-24999</a></p> <p>Release Date: 2022-11-26</p> <p>Fix Resolution (qs): 6.10.3</p> <p>Direct dependency fix Resolution (nuxt): 2.15.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in qs tgz cve high severity vulnerability vulnerable library qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href path to dependency file package json path to vulnerable library node modules qs package json dependency hierarchy nuxt tgz root library telemetry tgz git url parse tgz git up tgz parse url tgz parse path tgz x qs tgz vulnerable library found in head commit a href found in base branch main vulnerability details qs before as used in express before and other products allows attackers to cause a node process hang for an express application because an proto key can be used in many typical express use cases an unauthenticated remote attacker can place the attack payload in the query string of the url that is used to visit the application such as a b a a the fix was backported to qs and and therefore express which has deps qs in its release description is not vulnerable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution qs direct dependency fix resolution nuxt step up your open source security game with mend
0
315,663
23,590,886,821
IssuesEvent
2022-08-23 15:05:59
arturo-lang/arturo
https://api.github.com/repos/arturo-lang/arturo
closed
[Reflection\complex?] add documentation example
documentation library todo easy
[Reflection\complex?] add documentation example https://github.com/arturo-lang/arturo/blob/3a89302d1b1f33da097ed6e7bd6dfb05fc338482/src/library/Reflection.nim#L292 ```text builtin "color?", alias = unaliased, rule = PrefixPrecedence, description = "checks if given value is of type :color", args = { "value" : {Any} }, attrs = NoAttrs, returns = {Logical}, # TODO(Reflection\color?) add documentation example # labels: library, documentation, easy example = """ """: ########################################################## push(newLogical(x.kind==Color)) builtin "complex?", alias = unaliased, rule = PrefixPrecedence, description = "checks if given value is of type :complex", args = { "value" : {Any} }, attrs = NoAttrs, returns = {Logical}, # TODO(Reflection\complex?) add documentation example # labels: library, documentation, easy example = """ """: ########################################################## push(newLogical(x.kind==Complex)) builtin "database?", alias = unaliased, rule = PrefixPrecedence, ``` 618f5b7315934d100b1650e62313cf8d69700ce6
1.0
[Reflection\complex?] add documentation example - [Reflection\complex?] add documentation example https://github.com/arturo-lang/arturo/blob/3a89302d1b1f33da097ed6e7bd6dfb05fc338482/src/library/Reflection.nim#L292 ```text builtin "color?", alias = unaliased, rule = PrefixPrecedence, description = "checks if given value is of type :color", args = { "value" : {Any} }, attrs = NoAttrs, returns = {Logical}, # TODO(Reflection\color?) add documentation example # labels: library, documentation, easy example = """ """: ########################################################## push(newLogical(x.kind==Color)) builtin "complex?", alias = unaliased, rule = PrefixPrecedence, description = "checks if given value is of type :complex", args = { "value" : {Any} }, attrs = NoAttrs, returns = {Logical}, # TODO(Reflection\complex?) add documentation example # labels: library, documentation, easy example = """ """: ########################################################## push(newLogical(x.kind==Complex)) builtin "database?", alias = unaliased, rule = PrefixPrecedence, ``` 618f5b7315934d100b1650e62313cf8d69700ce6
non_process
add documentation example add documentation example text builtin color alias unaliased rule prefixprecedence description checks if given value is of type color args value any attrs noattrs returns logical todo reflection color add documentation example labels library documentation easy example push newlogical x kind color builtin complex alias unaliased rule prefixprecedence description checks if given value is of type complex args value any attrs noattrs returns logical todo reflection complex add documentation example labels library documentation easy example push newlogical x kind complex builtin database alias unaliased rule prefixprecedence
0
10,186
13,044,162,864
IssuesEvent
2020-07-29 03:47:37
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `RoundWithFracReal` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `RoundWithFracReal` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @mapleFU ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `RoundWithFracReal` from TiDB - ## Description Port the scalar function `RoundWithFracReal` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @mapleFU ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function roundwithfracreal from tidb description port the scalar function roundwithfracreal from tidb to coprocessor score mentor s maplefu recommended skills rust programming learning materials already implemented expressions ported from tidb
1
207,244
7,126,399,097
IssuesEvent
2018-01-20 09:45:55
wordpress-mobile/AztecEditor-Android
https://api.github.com/repos/wordpress-mobile/AztecEditor-Android
reopened
Repeating characters in title
bug high priority
### Expected If I type "ABCD" in the title box, the title should say "ABCD" ### Observed I type A, it says A. I type B, the title becomes "AAB". I type C, I get "AABABC", etc. You can see a video and screenshot of the issue in action here: https://cloudup.com/cE18S9bfbWY Also, note if I add a space, it resets, as you can see in the screenshot. Pressing backspace also does wonky things, but I can eventually clear out the title. I can type normally in the content area and paste to the title area. ### Reproduced 1. Make a new post using the beta editor 2. Type a title Note: We were not able to duplicate this on all phones. #### Tested Tested on LG G4, Android 6.0 with alpha 89
1.0
Repeating characters in title - ### Expected If I type "ABCD" in the title box, the title should say "ABCD" ### Observed I type A, it says A. I type B, the title becomes "AAB". I type C, I get "AABABC", etc. You can see a video and screenshot of the issue in action here: https://cloudup.com/cE18S9bfbWY Also, note if I add a space, it resets, as you can see in the screenshot. Pressing backspace also does wonky things, but I can eventually clear out the title. I can type normally in the content area and paste to the title area. ### Reproduced 1. Make a new post using the beta editor 2. Type a title Note: We were not able to duplicate this on all phones. #### Tested Tested on LG G4, Android 6.0 with alpha 89
non_process
repeating characters in title expected if i type abcd in the title box the title should say abcd observed i type a it says a i type b the title becomes aab i type c i get aababc etc you can see a video and screenshot of the issue in action here also note if i add a space it resets as you can see in the screenshot pressing backspace also does wonky things but i can eventually clear out the title i can type normally in the content area and paste to the title area reproduced make a new post using the beta editor type a title note we were not able to duplicate this on all phones tested tested on lg android with alpha
0
16,821
22,060,943,988
IssuesEvent
2022-05-30 17:44:09
bitPogo/kmock
https://api.github.com/repos/bitPogo/kmock
closed
Relaxation fails for Generics
bug kmock-processor
## Description <!--- Provide a detailed introduction to the issue itself, and why you consider it to be a bug --> Currently Relaxation for Interfaces with generics causes an Compiler Error. This only applies for inline functions. (The need additional arguments and cannot be inline in this case)
1.0
Relaxation fails for Generics - ## Description <!--- Provide a detailed introduction to the issue itself, and why you consider it to be a bug --> Currently Relaxation for Interfaces with generics causes an Compiler Error. This only applies for inline functions. (The need additional arguments and cannot be inline in this case)
process
relaxation fails for generics description currently relaxation for interfaces with generics causes an compiler error this only applies for inline functions the need additional arguments and cannot be inline in this case
1
345,270
30,794,983,040
IssuesEvent
2023-07-31 19:07:10
boltlabs-inc/tss-ecdsa
https://api.github.com/repos/boltlabs-inc/tss-ecdsa
closed
Write tests for PiSchProof
tests
A `PiSchProof` has two elements and must satisfy one constraint. Some ideas for tests: - Choose bad secret input `x` (not the dlog of `X`) - Swap the commitment `A` for a random element - Swap the response `z` for a random element - Try bad common input, e.g. where `g` is not a generator of `G` - Try verifying with different common input that was used to generate the proof Related issue: #52
1.0
Write tests for PiSchProof - A `PiSchProof` has two elements and must satisfy one constraint. Some ideas for tests: - Choose bad secret input `x` (not the dlog of `X`) - Swap the commitment `A` for a random element - Swap the response `z` for a random element - Try bad common input, e.g. where `g` is not a generator of `G` - Try verifying with different common input that was used to generate the proof Related issue: #52
non_process
write tests for pischproof a pischproof has two elements and must satisfy one constraint some ideas for tests choose bad secret input x not the dlog of x swap the commitment a for a random element swap the response z for a random element try bad common input e g where g is not a generator of g try verifying with different common input that was used to generate the proof related issue
0
57,650
7,087,363,575
IssuesEvent
2018-01-11 17:33:05
Nersent/wexond
https://api.github.com/repos/Nersent/wexond
closed
What should be in a new tab page?
design question
What should be in a new tab page? Please give me your suggestions below. I'm wondering if add bookmarks cards or recently visited websites or something different.
1.0
What should be in a new tab page? - What should be in a new tab page? Please give me your suggestions below. I'm wondering if add bookmarks cards or recently visited websites or something different.
non_process
what should be in a new tab page what should be in a new tab page please give me your suggestions below i m wondering if add bookmarks cards or recently visited websites or something different
0
317,094
23,663,707,592
IssuesEvent
2022-08-26 18:18:30
dockstore/dockstore
https://api.github.com/repos/dockstore/dockstore
closed
Remove 1.13 stuff in documentation stable branch
bug documentation
**Describe the bug** On our documentation repo, develop got merged/rebased/cherrypicked into the hotfix/1.12.2 branch. There's 99% chance I'm the guy who did this, so I'm going to be the guy who fixes it. The rebase/merge/whatever doesn't break anything, but it does include a few small details about the next release (such as which version of Cromwell 1.13 will be using). **To Reproduce** You can see 1.13 referenced in https://docs.dockstore.org/en/stable/advanced-topics/dockstore-cli/advanced-features.html?highlight=cromwell#wdl-launcher-configuration **Expected behavior** ~Ash is banned from using `git rebase`~ 1.13 should not be referenced on master/stable **Additional context** I found this issue this afternoon. Charles and I did some digging, using the addition of images to index.rst (which should be on only develop due to [not being on 1.12's release branch](https://github.com/dockstore/dockstore-documentation/pull/170#issuecomment-1075628831)) as our clue. Charles found the comment "merged commit 9748946" in https://github.com/dockstore/dockstore-documentation/pull/206, so that's probably where the rebase happened. Knowing me, I did it to bring in a linkcheck fix. ┆Issue is synchronized with this [Jira Story](https://ucsc-cgl.atlassian.net/browse/DOCK-2206) ┆fixVersions: Dockstore 1.12.x ┆friendlyId: DOCK-2206 ┆sprint: 94- Vampire Pleco ┆taskType: Story
1.0
Remove 1.13 stuff in documentation stable branch - **Describe the bug** On our documentation repo, develop got merged/rebased/cherrypicked into the hotfix/1.12.2 branch. There's 99% chance I'm the guy who did this, so I'm going to be the guy who fixes it. The rebase/merge/whatever doesn't break anything, but it does include a few small details about the next release (such as which version of Cromwell 1.13 will be using). **To Reproduce** You can see 1.13 referenced in https://docs.dockstore.org/en/stable/advanced-topics/dockstore-cli/advanced-features.html?highlight=cromwell#wdl-launcher-configuration **Expected behavior** ~Ash is banned from using `git rebase`~ 1.13 should not be referenced on master/stable **Additional context** I found this issue this afternoon. Charles and I did some digging, using the addition of images to index.rst (which should be on only develop due to [not being on 1.12's release branch](https://github.com/dockstore/dockstore-documentation/pull/170#issuecomment-1075628831)) as our clue. Charles found the comment "merged commit 9748946" in https://github.com/dockstore/dockstore-documentation/pull/206, so that's probably where the rebase happened. Knowing me, I did it to bring in a linkcheck fix. ┆Issue is synchronized with this [Jira Story](https://ucsc-cgl.atlassian.net/browse/DOCK-2206) ┆fixVersions: Dockstore 1.12.x ┆friendlyId: DOCK-2206 ┆sprint: 94- Vampire Pleco ┆taskType: Story
non_process
remove stuff in documentation stable branch describe the bug on our documentation repo develop got merged rebased cherrypicked into the hotfix branch there s chance i m the guy who did this so i m going to be the guy who fixes it the rebase merge whatever doesn t break anything but it does include a few small details about the next release such as which version of cromwell will be using to reproduce you can see referenced in expected behavior ash is banned from using git rebase should not be referenced on master stable additional context i found this issue this afternoon charles and i did some digging using the addition of images to index rst which should be on only develop due to as our clue charles found the comment merged commit in so that s probably where the rebase happened knowing me i did it to bring in a linkcheck fix ┆issue is synchronized with this ┆fixversions dockstore x ┆friendlyid dock ┆sprint vampire pleco ┆tasktype story
0
20,148
26,695,474,296
IssuesEvent
2023-01-27 09:59:22
ppy/osu-web
https://api.github.com/repos/ppy/osu-web
closed
Error downloading old beatmap
area:beatmap-processing
This map is unable to download: https://osu.ppy.sh/beatmapsets/13758#osu/50686 If you try download from beatmap page, it returns 500 error. But if you try doing it form beatmap search page it returns this: ``` Warning: sprintf(): Too few arguments in /var/www/html/S3.php on line 742 Warning: Cannot modify header information - headers already sent by (output started at /var/www/html/S3.php:742) in /var/www/html/d.php on line 109 ```
1.0
Error downloading old beatmap - This map is unable to download: https://osu.ppy.sh/beatmapsets/13758#osu/50686 If you try download from beatmap page, it returns 500 error. But if you try doing it form beatmap search page it returns this: ``` Warning: sprintf(): Too few arguments in /var/www/html/S3.php on line 742 Warning: Cannot modify header information - headers already sent by (output started at /var/www/html/S3.php:742) in /var/www/html/d.php on line 109 ```
process
error downloading old beatmap this map is unable to download if you try download from beatmap page it returns error but if you try doing it form beatmap search page it returns this warning sprintf too few arguments in var www html php on line warning cannot modify header information headers already sent by output started at var www html php in var www html d php on line
1
79,228
15,166,238,917
IssuesEvent
2021-02-12 16:06:09
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
closed
Fix pings for lsif WAUs
team/code-intelligence
We currently aren't getting code intelligence WAUs. <img width="1482" alt="Screenshot 2021-02-02 at 16 38 10" src="https://user-images.githubusercontent.com/1657213/106653045-99840d00-6596-11eb-97a7-8562c1feb4ab.png">
1.0
Fix pings for lsif WAUs - We currently aren't getting code intelligence WAUs. <img width="1482" alt="Screenshot 2021-02-02 at 16 38 10" src="https://user-images.githubusercontent.com/1657213/106653045-99840d00-6596-11eb-97a7-8562c1feb4ab.png">
non_process
fix pings for lsif waus we currently aren t getting code intelligence waus img width alt screenshot at src
0
416,243
28,074,149,565
IssuesEvent
2023-03-29 21:35:03
sujuka99/dotfiles
https://api.github.com/repos/sujuka99/dotfiles
reopened
Document how to use repo
documentation dotfiles
Document: - [x] Clone repo - [x] Configure git for bare repository - [ ] List requirements
1.0
Document how to use repo - Document: - [x] Clone repo - [x] Configure git for bare repository - [ ] List requirements
non_process
document how to use repo document clone repo configure git for bare repository list requirements
0
682,863
23,360,272,544
IssuesEvent
2022-08-10 11:03:33
microsoft/fluentui
https://api.github.com/repos/microsoft/fluentui
opened
[Feature]: Fluent React v9 Preview Banner 🎉
Type: Feature Priority 1: High Area: Website Partner Ask
### Library React Components / v9 (@fluentui/react-components) ### Describe the feature that you would like added In mid to late August the Microsite is planned to go live to the public. To let our customers know of the new site we would like to have a top banner added to the existing [Fluent Design System](https://www.microsoft.com/design/fluent/) to direct customers to the Microsite. ### Solution ![image](https://user-images.githubusercontent.com/78806055/183884822-f14b1f8a-e5ce-4443-a75c-c6a4026a9200.png) ### Asset Resources - For more information into the styling and design, please, view the `website banner` pages in the [Figma file](https://www.figma.com/file/Jm7YiqB9W4lyMA3NxPlK9R/V9-Build-website?node-id=958%3A1908) - The `Go To Fluent React v9` button in the banner should direct to https://react.fluentui.dev/ ### Have you discussed this feature with our team yes (Design and Website Team) ### Additional context _No response_ ### Validations - [X] Check that there isn't already an issue that request the same feature to avoid creating a duplicate.
1.0
[Feature]: Fluent React v9 Preview Banner 🎉 - ### Library React Components / v9 (@fluentui/react-components) ### Describe the feature that you would like added In mid to late August the Microsite is planned to go live to the public. To let our customers know of the new site we would like to have a top banner added to the existing [Fluent Design System](https://www.microsoft.com/design/fluent/) to direct customers to the Microsite. ### Solution ![image](https://user-images.githubusercontent.com/78806055/183884822-f14b1f8a-e5ce-4443-a75c-c6a4026a9200.png) ### Asset Resources - For more information into the styling and design, please, view the `website banner` pages in the [Figma file](https://www.figma.com/file/Jm7YiqB9W4lyMA3NxPlK9R/V9-Build-website?node-id=958%3A1908) - The `Go To Fluent React v9` button in the banner should direct to https://react.fluentui.dev/ ### Have you discussed this feature with our team yes (Design and Website Team) ### Additional context _No response_ ### Validations - [X] Check that there isn't already an issue that request the same feature to avoid creating a duplicate.
non_process
fluent react preview banner 🎉 library react components fluentui react components describe the feature that you would like added in mid to late august the microsite is planned to go live to the public to let our customers know of the new site we would like to have a top banner added to the existing to direct customers to the microsite solution asset resources for more information into the styling and design please view the website banner pages in the the go to fluent react button in the banner should direct to have you discussed this feature with our team yes design and website team additional context no response validations check that there isn t already an issue that request the same feature to avoid creating a duplicate
0
654,114
21,638,101,550
IssuesEvent
2022-05-05 15:55:14
internetarchive/openlibrary
https://api.github.com/repos/internetarchive/openlibrary
opened
Enriching Reading Log data exports (more fields)
Theme: Reading Log export Module: Data dumps Priority: 2 Affects: Data Lead: @mekarpeles
* Currently Reading Log dumps are limited to IDs because large Reading Logs (500+ items) can cause performance issues aggregating results in memory and and serving as large files. Currently, Lists impose an export limit if there are more than ~1k items * The manual work around is to look over each Reading Log entry (e.g. OL16801714W) and to fetch the json for the corresponding item from Open Library API or the data dumps: https://openlibrary.org/works/OL16801714W.json * 2 possible solutions to this problem are: * Give an option to download all work data but limit to e.g. 500 results (and show a notice, similar to lists page) * Offer a js option which allows one (from the OL website) to download pages of ~500 results from Reading Log (calculated based on total number of patron's books on Reading Log) Questions for community: * What format of data would you prefer/expect? CSV? JSON?
1.0
Enriching Reading Log data exports (more fields) - * Currently Reading Log dumps are limited to IDs because large Reading Logs (500+ items) can cause performance issues aggregating results in memory and and serving as large files. Currently, Lists impose an export limit if there are more than ~1k items * The manual work around is to look over each Reading Log entry (e.g. OL16801714W) and to fetch the json for the corresponding item from Open Library API or the data dumps: https://openlibrary.org/works/OL16801714W.json * 2 possible solutions to this problem are: * Give an option to download all work data but limit to e.g. 500 results (and show a notice, similar to lists page) * Offer a js option which allows one (from the OL website) to download pages of ~500 results from Reading Log (calculated based on total number of patron's books on Reading Log) Questions for community: * What format of data would you prefer/expect? CSV? JSON?
non_process
enriching reading log data exports more fields currently reading log dumps are limited to ids because large reading logs items can cause performance issues aggregating results in memory and and serving as large files currently lists impose an export limit if there are more than items the manual work around is to look over each reading log entry e g and to fetch the json for the corresponding item from open library api or the data dumps possible solutions to this problem are give an option to download all work data but limit to e g results and show a notice similar to lists page offer a js option which allows one from the ol website to download pages of results from reading log calculated based on total number of patron s books on reading log questions for community what format of data would you prefer expect csv json
0
10,482
13,252,913,227
IssuesEvent
2020-08-20 06:33:10
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
Deadline check in table / index scan executor
sig/coprocessor status/discussion
## Feature Request Currently we check deadline in DAG next(). However there are many scenarios that this doesn't work, i.e. there is a selection executor that filtered a lot of lines, or an aggregation executor that emits a row only after collecting all rows. By checking deadline in the innermost executor, i.e. table scan or index scan executor, we can avoid this issue and make deadline check work correctly in all scenarios.
1.0
Deadline check in table / index scan executor - ## Feature Request Currently we check deadline in DAG next(). However there are many scenarios that this doesn't work, i.e. there is a selection executor that filtered a lot of lines, or an aggregation executor that emits a row only after collecting all rows. By checking deadline in the innermost executor, i.e. table scan or index scan executor, we can avoid this issue and make deadline check work correctly in all scenarios.
process
deadline check in table index scan executor feature request currently we check deadline in dag next however there are many scenarios that this doesn t work i e there is a selection executor that filtered a lot of lines or an aggregation executor that emits a row only after collecting all rows by checking deadline in the innermost executor i e table scan or index scan executor we can avoid this issue and make deadline check work correctly in all scenarios
1
6,835
9,978,102,343
IssuesEvent
2019-07-09 18:59:30
TileDB-Inc/TileDB
https://api.github.com/repos/TileDB-Inc/TileDB
closed
Lock release with multiple contexts
bug process safety vfs
The following program demonstrates a failure to release the lockfile when a local URI is opened from multiple contexts. <details> ``` import tiledb import numpy as np import shutil, os if os.path.isdir("C:/tmp/foo"): shutil.rmtree("C:/tmp/foo") ctx = tiledb.Ctx() ctx2= tiledb.Ctx() dom = tiledb.Domain(tiledb.Dim(name="rows", domain=(1, 4), tile=4, dtype=np.int32, ctx=ctx), tiledb.Dim(name="cols", domain=(1, 4), tile=4, dtype=np.int32, ctx=ctx), ctx=ctx) schema = tiledb.ArraySchema(domain=dom, sparse=False, attrs=[tiledb.Attr(name="a", dtype=np.int32)], ctx=ctx) uri = "C:/tmp/foo" tiledb.Array.create("C:/tmp/foo", schema) a = tiledb.DenseArray(uri, 'w', ctx=ctx) a[:] = np.random.rand(4,4) b = tiledb.DenseArray(uri, 'r', ctx=ctx2) a.close() b.close() # the call below causes a PermissionDenied error because the handle for # C:/tmp/foo/__lock.tdb is still open. shutil.rmtree(uri) ``` </details> The issue arises due to the following sequence: - array 1, ctx 1, the storage manager calls `OpenArray::file_lock`, which calls `vfs->filelock_lock` w/ a reference to the `filelock_` member: https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/storage_manager/open_array.cc#L93-L99 - ctx 1 increments the global counter (URI count ==1) and calls the platform lock, which sets the `OpenArray::filelock_` member passed above to the FD: https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/filesystem/vfs.cc#L365-L372 - array 2, ctx 2, increments the counter in `VFS::filelock_lock` (now URI count == 2) but *does not* call the platform file lock, so the `OpenArray::filelock_` member for this instance continues to have the value `INVALID_FILELOCK`. https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/filesystem/vfs.cc#L367 - when `tiledb_array_close` is called the first time, the storage manager calls `OpenArray::file_unlock` https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/storage_manager/open_array.cc#L101-L108 - which calls `VFS::file_unlock` and decrements the counter -- but doesn't unlock because the URI global count == 1. https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/filesystem/vfs.cc#L408-L412 - when `tiledb_array_close` is called the 2nd time, we reach `OpenArray::file_unlock` again, but the `filelock_` member for this `OpenArray` instance was not initialized previously, so the call to `vfs->file_unlock` is skipped, leaving the handle open: https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/storage_manager/open_array.cc#L101-L108
1.0
Lock release with multiple contexts - The following program demonstrates a failure to release the lockfile when a local URI is opened from multiple contexts. <details> ``` import tiledb import numpy as np import shutil, os if os.path.isdir("C:/tmp/foo"): shutil.rmtree("C:/tmp/foo") ctx = tiledb.Ctx() ctx2= tiledb.Ctx() dom = tiledb.Domain(tiledb.Dim(name="rows", domain=(1, 4), tile=4, dtype=np.int32, ctx=ctx), tiledb.Dim(name="cols", domain=(1, 4), tile=4, dtype=np.int32, ctx=ctx), ctx=ctx) schema = tiledb.ArraySchema(domain=dom, sparse=False, attrs=[tiledb.Attr(name="a", dtype=np.int32)], ctx=ctx) uri = "C:/tmp/foo" tiledb.Array.create("C:/tmp/foo", schema) a = tiledb.DenseArray(uri, 'w', ctx=ctx) a[:] = np.random.rand(4,4) b = tiledb.DenseArray(uri, 'r', ctx=ctx2) a.close() b.close() # the call below causes a PermissionDenied error because the handle for # C:/tmp/foo/__lock.tdb is still open. shutil.rmtree(uri) ``` </details> The issue arises due to the following sequence: - array 1, ctx 1, the storage manager calls `OpenArray::file_lock`, which calls `vfs->filelock_lock` w/ a reference to the `filelock_` member: https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/storage_manager/open_array.cc#L93-L99 - ctx 1 increments the global counter (URI count ==1) and calls the platform lock, which sets the `OpenArray::filelock_` member passed above to the FD: https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/filesystem/vfs.cc#L365-L372 - array 2, ctx 2, increments the counter in `VFS::filelock_lock` (now URI count == 2) but *does not* call the platform file lock, so the `OpenArray::filelock_` member for this instance continues to have the value `INVALID_FILELOCK`. https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/filesystem/vfs.cc#L367 - when `tiledb_array_close` is called the first time, the storage manager calls `OpenArray::file_unlock` https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/storage_manager/open_array.cc#L101-L108 - which calls `VFS::file_unlock` and decrements the counter -- but doesn't unlock because the URI global count == 1. https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/filesystem/vfs.cc#L408-L412 - when `tiledb_array_close` is called the 2nd time, we reach `OpenArray::file_unlock` again, but the `filelock_` member for this `OpenArray` instance was not initialized previously, so the call to `vfs->file_unlock` is skipped, leaving the handle open: https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/storage_manager/open_array.cc#L101-L108
process
lock release with multiple contexts the following program demonstrates a failure to release the lockfile when a local uri is opened from multiple contexts import tiledb import numpy as np import shutil os if os path isdir c tmp foo shutil rmtree c tmp foo ctx tiledb ctx tiledb ctx dom tiledb domain tiledb dim name rows domain tile dtype np ctx ctx tiledb dim name cols domain tile dtype np ctx ctx ctx ctx schema tiledb arrayschema domain dom sparse false attrs ctx ctx uri c tmp foo tiledb array create c tmp foo schema a tiledb densearray uri w ctx ctx a np random rand b tiledb densearray uri r ctx a close b close the call below causes a permissiondenied error because the handle for c tmp foo lock tdb is still open shutil rmtree uri the issue arises due to the following sequence array ctx the storage manager calls openarray file lock which calls vfs filelock lock w a reference to the filelock member ctx increments the global counter uri count and calls the platform lock which sets the openarray filelock member passed above to the fd array ctx increments the counter in vfs filelock lock now uri count but does not call the platform file lock so the openarray filelock member for this instance continues to have the value invalid filelock when tiledb array close is called the first time the storage manager calls openarray file unlock which calls vfs file unlock and decrements the counter but doesn t unlock because the uri global count when tiledb array close is called the time we reach openarray file unlock again but the filelock member for this openarray instance was not initialized previously so the call to vfs file unlock is skipped leaving the handle open
1
99,060
30,268,069,549
IssuesEvent
2023-07-07 13:23:20
cms-sw/cmssw
https://api.github.com/repos/cms-sw/cmssw
closed
Build CMSSW_13_0_10
release-notes-requested release-announced release-build-request slc7_amd64_gcc11-finished el8_amd64_gcc11-finished el8_aarch64_gcc11-finished el8_ppc64le_gcc11-finished el9_amd64_gcc11-finished
To start the MC production campaign for Run3 2023 The build will go in parallel with the IB tests in CMSSW_13_0_X_2023-07-05-1100, to speed up the procedure: the release will get uploaded only if those tests show no issues.
1.0
Build CMSSW_13_0_10 - To start the MC production campaign for Run3 2023 The build will go in parallel with the IB tests in CMSSW_13_0_X_2023-07-05-1100, to speed up the procedure: the release will get uploaded only if those tests show no issues.
non_process
build cmssw to start the mc production campaign for the build will go in parallel with the ib tests in cmssw x to speed up the procedure the release will get uploaded only if those tests show no issues
0
16,624
21,678,126,955
IssuesEvent
2022-05-09 01:23:15
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add Super in America
suggested title in process
Please add as much of the following info as you can: Title: Super in America Type (film/tv show): reality TV show Film or show in which it appears: The Boys Is the parent film/show streaming anywhere? Yes - Amazon Prime About when in the parent film/show does it appear? Throughout ep. 1x06; "The Innocents." Timestamps: 0:15 - 0:52, 4:24- 5:53 (peppered with behind the scenes footage), 12:30 - 14:30 (peppered with behind the scenes footage), 34:57 - 35:29, 37:54 - 39:33 (peppered with behind the scenes footage). Actual footage of the film/show can be seen (yes/no)? Yes Production Company: Vought Studios
1.0
Add Super in America - Please add as much of the following info as you can: Title: Super in America Type (film/tv show): reality TV show Film or show in which it appears: The Boys Is the parent film/show streaming anywhere? Yes - Amazon Prime About when in the parent film/show does it appear? Throughout ep. 1x06; "The Innocents." Timestamps: 0:15 - 0:52, 4:24- 5:53 (peppered with behind the scenes footage), 12:30 - 14:30 (peppered with behind the scenes footage), 34:57 - 35:29, 37:54 - 39:33 (peppered with behind the scenes footage). Actual footage of the film/show can be seen (yes/no)? Yes Production Company: Vought Studios
process
add super in america please add as much of the following info as you can title super in america type film tv show reality tv show film or show in which it appears the boys is the parent film show streaming anywhere yes amazon prime about when in the parent film show does it appear throughout ep the innocents timestamps peppered with behind the scenes footage peppered with behind the scenes footage peppered with behind the scenes footage actual footage of the film show can be seen yes no yes production company vought studios
1
64,212
18,279,840,558
IssuesEvent
2021-10-05 00:47:17
microsoft/STL
https://api.github.com/repos/microsoft/STL
opened
P2418R2 Add Support For `std::generator`-like Types To `std::format`
cxx20 format defect report
[P2418R2](https://wg21.link/P2418R2) Add Support For `std::generator`-like Types To `std::format` This increases the value of `__cpp_lib_format`, also increased by [P2372R3](https://wg21.link/P2372R3) (see #2237).
1.0
P2418R2 Add Support For `std::generator`-like Types To `std::format` - [P2418R2](https://wg21.link/P2418R2) Add Support For `std::generator`-like Types To `std::format` This increases the value of `__cpp_lib_format`, also increased by [P2372R3](https://wg21.link/P2372R3) (see #2237).
non_process
add support for std generator like types to std format add support for std generator like types to std format this increases the value of cpp lib format also increased by see
0
348,012
31,392,433,350
IssuesEvent
2023-08-26 14:06:54
void-linux/void-packages
https://api.github.com/repos/void-linux/void-packages
opened
`docker build` fails with "ERROR: http: invalid Host header" with docker-buildx-0.10.3_3
bug needs-testing
### Is this a new report? Yes ### System Info Void 6.3.13_1 x86_64 AuthenticAMD notuptodate DFFF ### Package(s) Affected docker-buildx-0.10.3_3 ### Does a report exist for this bug with the project's home (upstream) and/or another distro? https://github.com/moby/moby/issues/45935 ### Expected behaviour work with ``` docker-24.0.5_2 docker-cli-24.0.5_2 moby-24.0.5_2 ``` and lower versions (tested 24.0.5_1) ### Actual behaviour "ERROR: http: invalid Host header" when building downgrade to docker-buildx-0.10.3_2 solves the issue. (docker-buildx-0.10.3_2 works with docker-24.0.5_2 and moby-24.0.5_2, but docker-buildx-0.10.3_3 doesn't) ### Steps to reproduce 1. `xbps-install -S docker-buildx-0.10.3_3` 2. `docker build .` fails 3. `xdowngrade /var/cache/xbps/docker-buildx-0.10.3_2.x86_64.xbps` 4. `docker build .` is fine "FROM node:18-alpine" is the only line in Dockerfile
1.0
`docker build` fails with "ERROR: http: invalid Host header" with docker-buildx-0.10.3_3 - ### Is this a new report? Yes ### System Info Void 6.3.13_1 x86_64 AuthenticAMD notuptodate DFFF ### Package(s) Affected docker-buildx-0.10.3_3 ### Does a report exist for this bug with the project's home (upstream) and/or another distro? https://github.com/moby/moby/issues/45935 ### Expected behaviour work with ``` docker-24.0.5_2 docker-cli-24.0.5_2 moby-24.0.5_2 ``` and lower versions (tested 24.0.5_1) ### Actual behaviour "ERROR: http: invalid Host header" when building downgrade to docker-buildx-0.10.3_2 solves the issue. (docker-buildx-0.10.3_2 works with docker-24.0.5_2 and moby-24.0.5_2, but docker-buildx-0.10.3_3 doesn't) ### Steps to reproduce 1. `xbps-install -S docker-buildx-0.10.3_3` 2. `docker build .` fails 3. `xdowngrade /var/cache/xbps/docker-buildx-0.10.3_2.x86_64.xbps` 4. `docker build .` is fine "FROM node:18-alpine" is the only line in Dockerfile
non_process
docker build fails with error http invalid host header with docker buildx is this a new report yes system info void authenticamd notuptodate dfff package s affected docker buildx does a report exist for this bug with the project s home upstream and or another distro expected behaviour work with docker docker cli moby and lower versions tested actual behaviour error http invalid host header when building downgrade to docker buildx solves the issue docker buildx works with docker and moby but docker buildx doesn t steps to reproduce xbps install s docker buildx docker build fails xdowngrade var cache xbps docker buildx xbps docker build is fine from node alpine is the only line in dockerfile
0
1,561
2,645,217,029
IssuesEvent
2015-03-12 21:20:23
editorconfig/editorconfig
https://api.github.com/repos/editorconfig/editorconfig
closed
Add support for existing `**` expansion and bash-like `**` expansion
core code core code (C) core code (python) feature request
Tests have been added for this in the [v0.11.0-development](https://github.com/editorconfig/editorconfig-core-test/tree/v0.11.0-development) branch. In reference to editorconfig/editorconfig-core-js#1 <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/381233-add-support-for-existing-expansion-and-bash-like-expansion?utm_campaign=plugin&utm_content=tracker%2F76099&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F76099&utm_medium=issues&utm_source=github). </bountysource-plugin>
3.0
Add support for existing `**` expansion and bash-like `**` expansion - Tests have been added for this in the [v0.11.0-development](https://github.com/editorconfig/editorconfig-core-test/tree/v0.11.0-development) branch. In reference to editorconfig/editorconfig-core-js#1 <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/381233-add-support-for-existing-expansion-and-bash-like-expansion?utm_campaign=plugin&utm_content=tracker%2F76099&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F76099&utm_medium=issues&utm_source=github). </bountysource-plugin>
non_process
add support for existing expansion and bash like expansion tests have been added for this in the branch in reference to editorconfig editorconfig core js want to back this issue we accept bounties via
0
21,420
17,038,387,673
IssuesEvent
2021-07-05 10:07:18
indico/indico
https://api.github.com/repos/indico/indico
closed
Timetable: add warning in case of draft mode
enhancement usability
We have a "Draft" switch in the list of contributions, which decides whether the Timetable is published or not. ![image](https://user-images.githubusercontent.com/2699/123974449-e55ba680-d9bc-11eb-900a-e326130b53a8.png) This is at times confusing for users. One nice improvement would be to add a warning at the top of the "Timetable management" page, something like "The timetable is in draft move". (followed by a "publish it" button)
True
Timetable: add warning in case of draft mode - We have a "Draft" switch in the list of contributions, which decides whether the Timetable is published or not. ![image](https://user-images.githubusercontent.com/2699/123974449-e55ba680-d9bc-11eb-900a-e326130b53a8.png) This is at times confusing for users. One nice improvement would be to add a warning at the top of the "Timetable management" page, something like "The timetable is in draft move". (followed by a "publish it" button)
non_process
timetable add warning in case of draft mode we have a draft switch in the list of contributions which decides whether the timetable is published or not this is at times confusing for users one nice improvement would be to add a warning at the top of the timetable management page something like the timetable is in draft move followed by a publish it button
0
76,667
15,496,164,254
IssuesEvent
2021-03-11 02:10:43
jinuem/Shopping-Cart-POC
https://api.github.com/repos/jinuem/Shopping-Cart-POC
opened
CVE-2021-23337 (High) detected in lodash-4.17.11.tgz
security vulnerability
## CVE-2021-23337 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.11.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: /Shopping-Cart-POC/rejsx/package.json</p> <p>Path to vulnerable library: Shopping-Cart-POC/rejsx/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - react-scripts-2.1.5.tgz (Root Library) - core-7.2.2.tgz - :x: **lodash-4.17.11.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of package lodash; all versions of package org.fujion.webjars:lodash are vulnerable to Command Injection via template. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: lodash - 4.17.21</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-23337 (High) detected in lodash-4.17.11.tgz - ## CVE-2021-23337 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.11.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: /Shopping-Cart-POC/rejsx/package.json</p> <p>Path to vulnerable library: Shopping-Cart-POC/rejsx/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - react-scripts-2.1.5.tgz (Root Library) - core-7.2.2.tgz - :x: **lodash-4.17.11.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of package lodash; all versions of package org.fujion.webjars:lodash are vulnerable to Command Injection via template. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: lodash - 4.17.21</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in lodash tgz cve high severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file shopping cart poc rejsx package json path to vulnerable library shopping cart poc rejsx node modules lodash package json dependency hierarchy react scripts tgz root library core tgz x lodash tgz vulnerable library vulnerability details all versions of package lodash all versions of package org fujion webjars lodash are vulnerable to command injection via template publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash step up your open source security game with whitesource
0
5,129
7,896,221,382
IssuesEvent
2018-06-29 07:47:46
decidim/decidim
https://api.github.com/repos/decidim/decidim
closed
Special content can be added to the process or superprocess home (2nd iteration)
Fridge lot: mods space: processes status: Needs-definition
- [ ] Active surveys (we are missing) - [ ] Active sortitions * Related issues: #2273
1.0
Special content can be added to the process or superprocess home (2nd iteration) - - [ ] Active surveys (we are missing) - [ ] Active sortitions * Related issues: #2273
process
special content can be added to the process or superprocess home iteration active surveys we are missing active sortitions related issues
1
249,894
18,858,247,870
IssuesEvent
2021-11-12 09:33:02
Lemonsr/pe
https://api.github.com/repos/Lemonsr/pe
opened
"Lifeline should end at the destroy marker (X)" is mentioned for only one specific diagram
severity.Low type.DocumentationBug
![image.png](https://raw.githubusercontent.com/Lemonsr/pe/main/files/be60d89a-0c00-47dc-9632-e0e91f9d56ee.png) I believe it is unclear to readers if you only mention that for that specific sequence diagram, the lifeline should be destroyed. For every other sequence diagram, it should be mentioned as well. Otherwise, put this warning at the very top to encompass all sequence diagrams in one go. ![image.png](https://raw.githubusercontent.com/Lemonsr/pe/main/files/22321731-8560-47b3-9f35-b193ff3b2731.png) <!--session: 1636703785006-31579939-b7a4-4e4a-ad80-94aa4ad34348--> <!--Version: Web v3.4.1-->
1.0
"Lifeline should end at the destroy marker (X)" is mentioned for only one specific diagram - ![image.png](https://raw.githubusercontent.com/Lemonsr/pe/main/files/be60d89a-0c00-47dc-9632-e0e91f9d56ee.png) I believe it is unclear to readers if you only mention that for that specific sequence diagram, the lifeline should be destroyed. For every other sequence diagram, it should be mentioned as well. Otherwise, put this warning at the very top to encompass all sequence diagrams in one go. ![image.png](https://raw.githubusercontent.com/Lemonsr/pe/main/files/22321731-8560-47b3-9f35-b193ff3b2731.png) <!--session: 1636703785006-31579939-b7a4-4e4a-ad80-94aa4ad34348--> <!--Version: Web v3.4.1-->
non_process
lifeline should end at the destroy marker x is mentioned for only one specific diagram i believe it is unclear to readers if you only mention that for that specific sequence diagram the lifeline should be destroyed for every other sequence diagram it should be mentioned as well otherwise put this warning at the very top to encompass all sequence diagrams in one go
0
21,545
29,865,284,736
IssuesEvent
2023-06-20 02:53:33
cncf/tag-security
https://api.github.com/repos/cncf/tag-security
closed
[Suggestion] assesment outline: does "operation" belong in "intended use"?
assessment-process suggestion inactive
Description: I was reading the https://github.com/cncf/sig-security/blob/master/assessments/guide/outline.md doc, and I saw that under "Intended Use", which describes the use cases and personas in high level, we ask for "Operation" which describes operational considerations of the application. This is a bit confusing because at this place in the document the system design hadn't yet introduced so diving into operational aspects, which will include internal terminology and familiarity with the design, might be confusing for the reader. (I'm also not sure semantic the relation to the use cases section). I think that the "Operation" part could belongs under "Configuration and Set-Up" which comes right after the "Design" section, or possibly in the design section. Impact: Describe your hopes for how this would reduce risk for the cloud native ecosystem. Who will this help? How will it help them? Make the security assessment doc more readable and understandable Scope: How much effort will this take? ok to provide a range of options if or "not yet determined" easy fix
1.0
[Suggestion] assesment outline: does "operation" belong in "intended use"? - Description: I was reading the https://github.com/cncf/sig-security/blob/master/assessments/guide/outline.md doc, and I saw that under "Intended Use", which describes the use cases and personas in high level, we ask for "Operation" which describes operational considerations of the application. This is a bit confusing because at this place in the document the system design hadn't yet introduced so diving into operational aspects, which will include internal terminology and familiarity with the design, might be confusing for the reader. (I'm also not sure semantic the relation to the use cases section). I think that the "Operation" part could belongs under "Configuration and Set-Up" which comes right after the "Design" section, or possibly in the design section. Impact: Describe your hopes for how this would reduce risk for the cloud native ecosystem. Who will this help? How will it help them? Make the security assessment doc more readable and understandable Scope: How much effort will this take? ok to provide a range of options if or "not yet determined" easy fix
process
assesment outline does operation belong in intended use description i was reading the doc and i saw that under intended use which describes the use cases and personas in high level we ask for operation which describes operational considerations of the application this is a bit confusing because at this place in the document the system design hadn t yet introduced so diving into operational aspects which will include internal terminology and familiarity with the design might be confusing for the reader i m also not sure semantic the relation to the use cases section i think that the operation part could belongs under configuration and set up which comes right after the design section or possibly in the design section impact describe your hopes for how this would reduce risk for the cloud native ecosystem who will this help how will it help them make the security assessment doc more readable and understandable scope how much effort will this take ok to provide a range of options if or not yet determined easy fix
1
2,258
5,089,817,186
IssuesEvent
2017-01-01 22:04:46
NJDaeger/EssentialCommands
https://api.github.com/repos/NJDaeger/EssentialCommands
opened
Add messages.yml
feature in process
I want messages to be able to be customized. Also create a parser that will find keywords and replace them. This may take some time.
1.0
Add messages.yml - I want messages to be able to be customized. Also create a parser that will find keywords and replace them. This may take some time.
process
add messages yml i want messages to be able to be customized also create a parser that will find keywords and replace them this may take some time
1
394,998
27,049,809,527
IssuesEvent
2023-02-13 12:27:39
grace287/uxui-issue-practice
https://api.github.com/repos/grace287/uxui-issue-practice
opened
유저 모델링 작업
documentation enhancement
## 설명 우리 서비스에 가입할 유저의 속성을 정의한다. ## 할 일 - [ ] 유저 모델 초기화 - [ ] 유저 모델 정의 - [ ] 유저 모델 검증 ## 참조 [Go to google](https://www.google.com)
1.0
유저 모델링 작업 - ## 설명 우리 서비스에 가입할 유저의 속성을 정의한다. ## 할 일 - [ ] 유저 모델 초기화 - [ ] 유저 모델 정의 - [ ] 유저 모델 검증 ## 참조 [Go to google](https://www.google.com)
non_process
유저 모델링 작업 설명 우리 서비스에 가입할 유저의 속성을 정의한다 할 일 유저 모델 초기화 유저 모델 정의 유저 모델 검증 참조
0
397,997
27,183,412,602
IssuesEvent
2023-02-18 23:00:20
Vurv78/WebAudio
https://api.github.com/repos/Vurv78/WebAudio
closed
Document `wa_sc_compat` or enable by default
documentation enhancement Complexity: Low
**Is your feature request related to a problem? Please describe.** Need to document our convars **Describe the solution you'd like** Document it **Describe alternatives you've considered** Enabling it by default (not good) **Additional context** A bunch of people don't like using webaudio and prefer streamcore's api, `wa_sc_compat` adds backwards compatibility, exposing streamcore's api but using webaudio internally.
1.0
Document `wa_sc_compat` or enable by default - **Is your feature request related to a problem? Please describe.** Need to document our convars **Describe the solution you'd like** Document it **Describe alternatives you've considered** Enabling it by default (not good) **Additional context** A bunch of people don't like using webaudio and prefer streamcore's api, `wa_sc_compat` adds backwards compatibility, exposing streamcore's api but using webaudio internally.
non_process
document wa sc compat or enable by default is your feature request related to a problem please describe need to document our convars describe the solution you d like document it describe alternatives you ve considered enabling it by default not good additional context a bunch of people don t like using webaudio and prefer streamcore s api wa sc compat adds backwards compatibility exposing streamcore s api but using webaudio internally
0
4,967
7,806,309,630
IssuesEvent
2018-06-11 13:44:09
sysown/proxysql
https://api.github.com/repos/sysown/proxysql
closed
Query_Processor_time_nsec per rules
QUERY PROCESSOR STATISTICS enhancement
Hi, as we already discussed measuring the process time per rule would handy when you want to find a "bad" rules or you just want to see which one is the slowest. Of course this could have some overhead so this should be optional. Thans.
1.0
Query_Processor_time_nsec per rules - Hi, as we already discussed measuring the process time per rule would handy when you want to find a "bad" rules or you just want to see which one is the slowest. Of course this could have some overhead so this should be optional. Thans.
process
query processor time nsec per rules hi as we already discussed measuring the process time per rule would handy when you want to find a bad rules or you just want to see which one is the slowest of course this could have some overhead so this should be optional thans
1
20,445
27,102,418,666
IssuesEvent
2023-02-15 09:39:06
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Add integration tests for java annotation plugins.
P2 type: process team-Rules-Java stale
Currently we don't test the `plugins` attribute on the java rules. Testing this feature would have caught https://github.com/bazelbuild/intellij/issues/845 at presubmit time and would have avoided a bazel patch release.
1.0
Add integration tests for java annotation plugins. - Currently we don't test the `plugins` attribute on the java rules. Testing this feature would have caught https://github.com/bazelbuild/intellij/issues/845 at presubmit time and would have avoided a bazel patch release.
process
add integration tests for java annotation plugins currently we don t test the plugins attribute on the java rules testing this feature would have caught at presubmit time and would have avoided a bazel patch release
1
18,646
24,580,930,703
IssuesEvent
2022-10-13 15:33:26
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[FHIR] JSON > Text response type > Some other values are getting displayed
Bug P1 Response datastore Process: Fixed Process: Tested dev
JSON > Text response type > Some other values are getting displayed 1. Configured values in SB ![image](https://user-images.githubusercontent.com/71445210/182408094-8a6a3b84-ccf4-414c-8bec-9a874b9c7eb2.png) 2. JSON values ![image](https://user-images.githubusercontent.com/71445210/182408563-4fadebda-f076-429f-9e4c-81cea65853a9.png)
2.0
[FHIR] JSON > Text response type > Some other values are getting displayed - JSON > Text response type > Some other values are getting displayed 1. Configured values in SB ![image](https://user-images.githubusercontent.com/71445210/182408094-8a6a3b84-ccf4-414c-8bec-9a874b9c7eb2.png) 2. JSON values ![image](https://user-images.githubusercontent.com/71445210/182408563-4fadebda-f076-429f-9e4c-81cea65853a9.png)
process
json text response type some other values are getting displayed json text response type some other values are getting displayed configured values in sb json values
1
417,096
12,155,777,366
IssuesEvent
2020-04-25 14:36:57
FMS-Cat/automaton-with-gui
https://api.github.com/repos/FMS-Cat/automaton-with-gui
closed
Create Channel -> Create Item -> Undo -> Undo -> Redo -> Redo causes an error
category: 🐛 Bug priority: 0-LIT
same goes to Curves
1.0
Create Channel -> Create Item -> Undo -> Undo -> Redo -> Redo causes an error - same goes to Curves
non_process
create channel create item undo undo redo redo causes an error same goes to curves
0
19,252
25,447,718,654
IssuesEvent
2022-11-24 07:57:26
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Having "Split with lines" as in-place editing.
Editing Processing Feature Request
### Feature description It would be good to have "Split by lines" algorithm as in-place editing too. It will allow users carry out complex feature splitting without needing to create other layers. ### Additional context _No response_
1.0
Having "Split with lines" as in-place editing. - ### Feature description It would be good to have "Split by lines" algorithm as in-place editing too. It will allow users carry out complex feature splitting without needing to create other layers. ### Additional context _No response_
process
having split with lines as in place editing feature description it would be good to have split by lines algorithm as in place editing too it will allow users carry out complex feature splitting without needing to create other layers additional context no response
1
85,690
16,719,034,165
IssuesEvent
2021-06-10 03:44:24
ArctosDB/arctos
https://api.github.com/repos/ArctosDB/arctos
closed
Code Table Request - sort code table categories alphabetically
Curatorial Search Display/Interface Enhancement Function-CodeTables Priority-Normal
We just organized the code tables by category and added descriptions, which is great, but the categories are not alphabetical as I would have expected. Can we list these alphabetically so that all of the code tables for the same category (e.g., locality) are together? The current (random?) sorting makes it harder to find the table you're looking for.
1.0
Code Table Request - sort code table categories alphabetically - We just organized the code tables by category and added descriptions, which is great, but the categories are not alphabetical as I would have expected. Can we list these alphabetically so that all of the code tables for the same category (e.g., locality) are together? The current (random?) sorting makes it harder to find the table you're looking for.
non_process
code table request sort code table categories alphabetically we just organized the code tables by category and added descriptions which is great but the categories are not alphabetical as i would have expected can we list these alphabetically so that all of the code tables for the same category e g locality are together the current random sorting makes it harder to find the table you re looking for
0
5,313
8,128,775,538
IssuesEvent
2018-08-17 13:04:59
w3c/transitions
https://api.github.com/repos/w3c/transitions
closed
Automatically Superseding previous versions
Process Issue
Add for PR transition: Does the Working Group believes this specification should supersede a previous Recommendation? Add a note for WBS to highlight this.
1.0
Automatically Superseding previous versions - Add for PR transition: Does the Working Group believes this specification should supersede a previous Recommendation? Add a note for WBS to highlight this.
process
automatically superseding previous versions add for pr transition does the working group believes this specification should supersede a previous recommendation add a note for wbs to highlight this
1
86,232
24,796,315,002
IssuesEvent
2022-10-24 17:35:17
LLNL/serac
https://api.github.com/repos/LLNL/serac
closed
Fix device initialization call on Lassen
CS build
The lassen builds are currently broken due to a device initialization call. We need to figure out the root cause and re-enable this. _Originally posted by @jamiebramwell in https://github.com/LLNL/serac/pull/565#r721618590_
1.0
Fix device initialization call on Lassen - The lassen builds are currently broken due to a device initialization call. We need to figure out the root cause and re-enable this. _Originally posted by @jamiebramwell in https://github.com/LLNL/serac/pull/565#r721618590_
non_process
fix device initialization call on lassen the lassen builds are currently broken due to a device initialization call we need to figure out the root cause and re enable this originally posted by jamiebramwell in
0
15,227
19,099,316,082
IssuesEvent
2021-11-29 20:23:47
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Reprojecting -180 to 180 dg raster to polar projection bilinear/cubic blurry
Feedback Processing Bug
### What is the bug or the crash? Reprojecting a raster with a -180 to 180 degrees longitudal extent, to a polar projection, using the Warp tool, and with e.g. bilinear or cubic rather than nearest neighbour resampling method, creates blurry/stretched segments around the 180dg meridian. In the attached screenshots, the upper result is 3995 nearest neighbour, which is all good, followed by buggy 3995 bilinear, both viewed in a 3995 project CRS. The third result is 3413 bilinear viewed in a 3413 project CRS. The bug seems to affect only the warp tool, while when setting bilinear resampling on zoomed in/out rendering of the original 4326 raster dataset in the Symbology properties, it looks all good. ### Steps to reproduce the issue 1. Extract a -180 to 180 dg northern section (65 to 70 dg N) of e.g. ETOPO1 GeoTIFF. 2. Select Raster / Projections / Warp 3. Set output CRS to e.g. 3995, resampling method Bilinear, and resolution 1000 m, and run. ### Versions <!--StartFragment--><!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd"> <html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /><style type="text/css"> p, li { white-space: pre-wrap; } </style></head><body> QGIS version | 3.22.0-Białowieża | QGIS code revision | d9022691f1 -- | -- | -- | -- Qt version | 5.15.2 Python version | 3.9.5 GDAL/OGR version | 3.4.0 PROJ version | 8.2.0 EPSG Registry database version | v10.038 (2021-10-21) GEOS version | 3.10.0-CAPI-1.16.0 SQLite version | 3.35.2 PDAL version | 2.3.0 PostgreSQL client version | 13.0 SpatiaLite version | 5.0.1 QWT version | 6.1.3 QScintilla2 version | 2.11.5 OS version | Windows 10 Version 2009   |   |   |   Active Python plugins latlontools | 3.5.9 grassprovider | 2.12.99 processing | 2.12.99 sagaprovider | 2.12.99 ![polar](https://user-images.githubusercontent.com/76013985/143786623-a67cb8dc-a820-4133-a512-3bb334bac2ef.png) ### Supported QGIS version - [ ] I'm running a supported QGIS version according to the roadmap. ### New profile - [ ] I tried with a new QGIS profile ### Additional context _No response_
1.0
Reprojecting -180 to 180 dg raster to polar projection bilinear/cubic blurry - ### What is the bug or the crash? Reprojecting a raster with a -180 to 180 degrees longitudal extent, to a polar projection, using the Warp tool, and with e.g. bilinear or cubic rather than nearest neighbour resampling method, creates blurry/stretched segments around the 180dg meridian. In the attached screenshots, the upper result is 3995 nearest neighbour, which is all good, followed by buggy 3995 bilinear, both viewed in a 3995 project CRS. The third result is 3413 bilinear viewed in a 3413 project CRS. The bug seems to affect only the warp tool, while when setting bilinear resampling on zoomed in/out rendering of the original 4326 raster dataset in the Symbology properties, it looks all good. ### Steps to reproduce the issue 1. Extract a -180 to 180 dg northern section (65 to 70 dg N) of e.g. ETOPO1 GeoTIFF. 2. Select Raster / Projections / Warp 3. Set output CRS to e.g. 3995, resampling method Bilinear, and resolution 1000 m, and run. ### Versions <!--StartFragment--><!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd"> <html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /><style type="text/css"> p, li { white-space: pre-wrap; } </style></head><body> QGIS version | 3.22.0-Białowieża | QGIS code revision | d9022691f1 -- | -- | -- | -- Qt version | 5.15.2 Python version | 3.9.5 GDAL/OGR version | 3.4.0 PROJ version | 8.2.0 EPSG Registry database version | v10.038 (2021-10-21) GEOS version | 3.10.0-CAPI-1.16.0 SQLite version | 3.35.2 PDAL version | 2.3.0 PostgreSQL client version | 13.0 SpatiaLite version | 5.0.1 QWT version | 6.1.3 QScintilla2 version | 2.11.5 OS version | Windows 10 Version 2009   |   |   |   Active Python plugins latlontools | 3.5.9 grassprovider | 2.12.99 processing | 2.12.99 sagaprovider | 2.12.99 ![polar](https://user-images.githubusercontent.com/76013985/143786623-a67cb8dc-a820-4133-a512-3bb334bac2ef.png) ### Supported QGIS version - [ ] I'm running a supported QGIS version according to the roadmap. ### New profile - [ ] I tried with a new QGIS profile ### Additional context _No response_
process
reprojecting to dg raster to polar projection bilinear cubic blurry what is the bug or the crash reprojecting a raster with a to degrees longitudal extent to a polar projection using the warp tool and with e g bilinear or cubic rather than nearest neighbour resampling method creates blurry stretched segments around the meridian in the attached screenshots the upper result is nearest neighbour which is all good followed by buggy bilinear both viewed in a project crs the third result is bilinear viewed in a project crs the bug seems to affect only the warp tool while when setting bilinear resampling on zoomed in out rendering of the original raster dataset in the symbology properties it looks all good steps to reproduce the issue extract a to dg northern section to dg n of e g geotiff select raster projections warp set output crs to e g resampling method bilinear and resolution m and run versions doctype html public dtd html en p li white space pre wrap qgis version białowieża qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version spatialite version qwt version version os version windows version         active python plugins latlontools grassprovider processing sagaprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
1
17,112
22,634,729,217
IssuesEvent
2022-06-30 17:45:54
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
closed
Deploy Integration relay instance using Helm
enhancement P1 process
### Problem We currently lack a flow to easily deploy to integration using the new Helm Chart flow. ### Solution 2 phase approach Phase 1 : From a local repo, deploy to an integration cluster in a fashion that updates the existing instances. Phase 2: Automate so CI flow automatically does this on every merge to main, potentially using [release integration Github workflow](https://github.com/hashgraph/hedera-json-rpc-relay/blob/main/.github/workflows/release-integration.yml) ### Alternatives _No response_
1.0
Deploy Integration relay instance using Helm - ### Problem We currently lack a flow to easily deploy to integration using the new Helm Chart flow. ### Solution 2 phase approach Phase 1 : From a local repo, deploy to an integration cluster in a fashion that updates the existing instances. Phase 2: Automate so CI flow automatically does this on every merge to main, potentially using [release integration Github workflow](https://github.com/hashgraph/hedera-json-rpc-relay/blob/main/.github/workflows/release-integration.yml) ### Alternatives _No response_
process
deploy integration relay instance using helm problem we currently lack a flow to easily deploy to integration using the new helm chart flow solution phase approach phase from a local repo deploy to an integration cluster in a fashion that updates the existing instances phase automate so ci flow automatically does this on every merge to main potentially using alternatives no response
1
9,709
12,705,222,941
IssuesEvent
2020-06-23 03:55:20
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
Deadlock with shared CUDA tensors and multiprocessing (spawn)
high priority module: cuda module: multiprocessing triaged
## 🐛 Bug Reported by iffiX in https://discuss.pytorch.org/t/freezing-problem-while-using-cuda-tensor-in-multiprocessing-environment/80000. (I have moved the code into a `main()` function to rule out some other multiprocessing bugs) ```python3 import itertools as it import torch as t import torch.multiprocessing as mp def infer(id, tensor): print(id) print(tensor) # del tensor immediately doesn't solve the problem del tensor def main(): # some global tensor g_tensor = t.full([1000, 1000], 2, device="cuda:0") g_tensor.share_memory_() ctx = mp.get_context("spawn") pool = ctx.Pool(2) for i in range(10000000): print("start") pool.starmap(infer, zip(range(5), it.repeat(g_tensor))) # cpu tensors work just fine # for cuda tensors: # if I delete the global tensor, reassign it with a new cuda tensor # or if I use a tensor created dynamically in each iteration # the program freezes after 2 iterations. # Comment out the following lines and everything will work fine. del g_tensor g_tensor = t.full([1000, 1000], 2, device="cuda:0") g_tensor.share_memory_() if __name__ == "__main__": main() ``` The backtrace for the deadlocked thread is: ``` #0 __lll_lock_wait () at ../sysdeps/unix/sysv/linux/x86_64/lowlevellock.S:135 #1 0x00007ffff7bc0023 in __GI___pthread_mutex_lock (mutex=0x7fffd105aaa8 <torch::(anonymous namespace)::cuda_ipc_global_entities+136>) at ../nptl/pthread_mutex_lock.c:78 #2 0x00007fffd04135b7 in torch::(anonymous namespace)::CudaIPCSentDataLimbo::collect() () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so #3 0x00007fffd04138ec in torch::(anonymous namespace)::CudaIPCSentDataDelete(void*) () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so #4 0x00007fffd0410aae in torch::CudaIPCSentData::~CudaIPCSentData() () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so #5 0x00007fffd0413615 in torch::(anonymous namespace)::CudaIPCSentDataLimbo::collect() () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so #6 0x00007fffd04138c9 in torch::(anonymous namespace)::CudaIPCSentDataDelete(void*) () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so #7 0x00007fffb67b40bd in c10::TensorImpl::release_resources() [clone .localalias.204] () from /scratch/sgross/pytorch/torch/lib/libc10.so #8 0x00007fffd0215649 in c10::intrusive_ptr<c10::TensorImpl, c10::UndefinedTensorImpl>::reset_() () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so #9 0x00007fffd0519437 in THPVariable_dealloc(THPVariable*) () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so #10 0x00005555556f584f in subtype_dealloc () at /tmp/build/80754af9/python_1565725737370/work/Objects/typeobject.c:1278 #11 0x0000555555725cbe in _PyEval_EvalFrameDefault () at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:2156 #12 0x00005555556b7ccb in function_code_fastcall (globals=<optimized out>, nargs=0, args=<optimized out>, co=<optimized out>) at /tmp/build/80754af9/python_1565725737370/work/Objects/call.c:283 #13 _PyFunction_FastCallKeywords () at /tmp/build/80754af9/python_1565725737370/work/Objects/call.c:408 #14 0x000055555571f806 in call_function (kwnames=0x0, oparg=<optimized out>, pp_stack=<synthetic pointer>) at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:4616 #15 _PyEval_EvalFrameDefault () at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:3124 #16 0x0000555555668539 in _PyEval_EvalCodeWithName () at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:3930 #17 0x0000555555669424 in PyEval_EvalCodeEx () at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:3959 #18 0x000055555566944c in PyEval_EvalCode (co=<optimized out>, globals=<optimized out>, locals=<optimized out>) at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:524 #19 0x000055555577eb74 in run_mod () at /tmp/build/80754af9/python_1565725737370/work/Python/pythonrun.c:1035 #20 0x0000555555788eb1 in PyRun_FileExFlags () at /tmp/build/80754af9/python_1565725737370/work/Python/pythonrun.c:988 #21 0x00005555557890a3 in PyRun_SimpleFileExFlags () at /tmp/build/80754af9/python_1565725737370/work/Python/pythonrun.c:429 #22 0x000055555578a195 in pymain_run_file (p_cf=0x7fffffffd310, filename=0x5555558c06a0 L"prog.py", fp=0x555555907310) at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:433 #23 pymain_run_filename (cf=0x7fffffffd310, pymain=0x7fffffffd420) at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:1612 #24 pymain_run_python (pymain=0x7fffffffd420) at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:2873 #25 pymain_main () at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:3413 #26 0x000055555578a2bc in _Py_UnixMain () at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:3448 #27 0x00007ffff77e6b97 in __libc_start_main (main=0x5555556496c0 <main>, argc=2, argv=0x7fffffffd578, init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized out>, stack_end=0x7fffffffd568) at ../csu/libc-start.c:310 #28 0x000055555572f062 in _start () at ../sysdeps/x86_64/elf/start.S:103 ``` The original report was for PyTorch 1.1.0, reproduced in 1.5.0 as well. I see it in 1.6.0a0+324dc16. cc @ezyang @gchanan @zou3519 @ngimel @VitalyFedyunin since this is in `CudaIPCSentData`
1.0
Deadlock with shared CUDA tensors and multiprocessing (spawn) - ## 🐛 Bug Reported by iffiX in https://discuss.pytorch.org/t/freezing-problem-while-using-cuda-tensor-in-multiprocessing-environment/80000. (I have moved the code into a `main()` function to rule out some other multiprocessing bugs) ```python3 import itertools as it import torch as t import torch.multiprocessing as mp def infer(id, tensor): print(id) print(tensor) # del tensor immediately doesn't solve the problem del tensor def main(): # some global tensor g_tensor = t.full([1000, 1000], 2, device="cuda:0") g_tensor.share_memory_() ctx = mp.get_context("spawn") pool = ctx.Pool(2) for i in range(10000000): print("start") pool.starmap(infer, zip(range(5), it.repeat(g_tensor))) # cpu tensors work just fine # for cuda tensors: # if I delete the global tensor, reassign it with a new cuda tensor # or if I use a tensor created dynamically in each iteration # the program freezes after 2 iterations. # Comment out the following lines and everything will work fine. del g_tensor g_tensor = t.full([1000, 1000], 2, device="cuda:0") g_tensor.share_memory_() if __name__ == "__main__": main() ``` The backtrace for the deadlocked thread is: ``` #0 __lll_lock_wait () at ../sysdeps/unix/sysv/linux/x86_64/lowlevellock.S:135 #1 0x00007ffff7bc0023 in __GI___pthread_mutex_lock (mutex=0x7fffd105aaa8 <torch::(anonymous namespace)::cuda_ipc_global_entities+136>) at ../nptl/pthread_mutex_lock.c:78 #2 0x00007fffd04135b7 in torch::(anonymous namespace)::CudaIPCSentDataLimbo::collect() () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so #3 0x00007fffd04138ec in torch::(anonymous namespace)::CudaIPCSentDataDelete(void*) () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so #4 0x00007fffd0410aae in torch::CudaIPCSentData::~CudaIPCSentData() () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so #5 0x00007fffd0413615 in torch::(anonymous namespace)::CudaIPCSentDataLimbo::collect() () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so #6 0x00007fffd04138c9 in torch::(anonymous namespace)::CudaIPCSentDataDelete(void*) () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so #7 0x00007fffb67b40bd in c10::TensorImpl::release_resources() [clone .localalias.204] () from /scratch/sgross/pytorch/torch/lib/libc10.so #8 0x00007fffd0215649 in c10::intrusive_ptr<c10::TensorImpl, c10::UndefinedTensorImpl>::reset_() () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so #9 0x00007fffd0519437 in THPVariable_dealloc(THPVariable*) () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so #10 0x00005555556f584f in subtype_dealloc () at /tmp/build/80754af9/python_1565725737370/work/Objects/typeobject.c:1278 #11 0x0000555555725cbe in _PyEval_EvalFrameDefault () at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:2156 #12 0x00005555556b7ccb in function_code_fastcall (globals=<optimized out>, nargs=0, args=<optimized out>, co=<optimized out>) at /tmp/build/80754af9/python_1565725737370/work/Objects/call.c:283 #13 _PyFunction_FastCallKeywords () at /tmp/build/80754af9/python_1565725737370/work/Objects/call.c:408 #14 0x000055555571f806 in call_function (kwnames=0x0, oparg=<optimized out>, pp_stack=<synthetic pointer>) at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:4616 #15 _PyEval_EvalFrameDefault () at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:3124 #16 0x0000555555668539 in _PyEval_EvalCodeWithName () at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:3930 #17 0x0000555555669424 in PyEval_EvalCodeEx () at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:3959 #18 0x000055555566944c in PyEval_EvalCode (co=<optimized out>, globals=<optimized out>, locals=<optimized out>) at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:524 #19 0x000055555577eb74 in run_mod () at /tmp/build/80754af9/python_1565725737370/work/Python/pythonrun.c:1035 #20 0x0000555555788eb1 in PyRun_FileExFlags () at /tmp/build/80754af9/python_1565725737370/work/Python/pythonrun.c:988 #21 0x00005555557890a3 in PyRun_SimpleFileExFlags () at /tmp/build/80754af9/python_1565725737370/work/Python/pythonrun.c:429 #22 0x000055555578a195 in pymain_run_file (p_cf=0x7fffffffd310, filename=0x5555558c06a0 L"prog.py", fp=0x555555907310) at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:433 #23 pymain_run_filename (cf=0x7fffffffd310, pymain=0x7fffffffd420) at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:1612 #24 pymain_run_python (pymain=0x7fffffffd420) at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:2873 #25 pymain_main () at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:3413 #26 0x000055555578a2bc in _Py_UnixMain () at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:3448 #27 0x00007ffff77e6b97 in __libc_start_main (main=0x5555556496c0 <main>, argc=2, argv=0x7fffffffd578, init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized out>, stack_end=0x7fffffffd568) at ../csu/libc-start.c:310 #28 0x000055555572f062 in _start () at ../sysdeps/x86_64/elf/start.S:103 ``` The original report was for PyTorch 1.1.0, reproduced in 1.5.0 as well. I see it in 1.6.0a0+324dc16. cc @ezyang @gchanan @zou3519 @ngimel @VitalyFedyunin since this is in `CudaIPCSentData`
process
deadlock with shared cuda tensors and multiprocessing spawn 🐛 bug reported by iffix in i have moved the code into a main function to rule out some other multiprocessing bugs import itertools as it import torch as t import torch multiprocessing as mp def infer id tensor print id print tensor del tensor immediately doesn t solve the problem del tensor def main some global tensor g tensor t full device cuda g tensor share memory ctx mp get context spawn pool ctx pool for i in range print start pool starmap infer zip range it repeat g tensor cpu tensors work just fine for cuda tensors if i delete the global tensor reassign it with a new cuda tensor or if i use a tensor created dynamically in each iteration the program freezes after iterations comment out the following lines and everything will work fine del g tensor g tensor t full device cuda g tensor share memory if name main main the backtrace for the deadlocked thread is lll lock wait at sysdeps unix sysv linux lowlevellock s in gi pthread mutex lock mutex at nptl pthread mutex lock c in torch anonymous namespace cudaipcsentdatalimbo collect from scratch sgross pytorch torch lib libtorch python so in torch anonymous namespace cudaipcsentdatadelete void from scratch sgross pytorch torch lib libtorch python so in torch cudaipcsentdata cudaipcsentdata from scratch sgross pytorch torch lib libtorch python so in torch anonymous namespace cudaipcsentdatalimbo collect from scratch sgross pytorch torch lib libtorch python so in torch anonymous namespace cudaipcsentdatadelete void from scratch sgross pytorch torch lib libtorch python so in tensorimpl release resources from scratch sgross pytorch torch lib so in intrusive ptr reset from scratch sgross pytorch torch lib libtorch python so in thpvariable dealloc thpvariable from scratch sgross pytorch torch lib libtorch python so in subtype dealloc at tmp build python work objects typeobject c in pyeval evalframedefault at tmp build python work python ceval c in function code fastcall globals nargs args co at tmp build python work objects call c pyfunction fastcallkeywords at tmp build python work objects call c in call function kwnames oparg pp stack at tmp build python work python ceval c pyeval evalframedefault at tmp build python work python ceval c in pyeval evalcodewithname at tmp build python work python ceval c in pyeval evalcodeex at tmp build python work python ceval c in pyeval evalcode co globals locals at tmp build python work python ceval c in run mod at tmp build python work python pythonrun c in pyrun fileexflags at tmp build python work python pythonrun c in pyrun simplefileexflags at tmp build python work python pythonrun c in pymain run file p cf filename l prog py fp at tmp build python work modules main c pymain run filename cf pymain at tmp build python work modules main c pymain run python pymain at tmp build python work modules main c pymain main at tmp build python work modules main c in py unixmain at tmp build python work modules main c in libc start main main argc argv init fini rtld fini stack end at csu libc start c in start at sysdeps elf start s the original report was for pytorch reproduced in as well i see it in cc ezyang gchanan ngimel vitalyfedyunin since this is in cudaipcsentdata
1
264,104
28,106,214,316
IssuesEvent
2023-03-31 01:03:56
Nivaskumark/CVE-2020-0097-frameworks_base_after
https://api.github.com/repos/Nivaskumark/CVE-2020-0097-frameworks_base_after
reopened
CVE-2019-9377 (Low) detected in baseandroid-10.0.0_r14
Mend: dependency security vulnerability
## CVE-2019-9377 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-10.0.0_r14</b></p></summary> <p> <p>Android framework classes and services</p> <p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p> <p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/CVE-2020-0097-frameworks_base/commit/ab450da4821e38b32262bb7c728210e89468ce0e">ab450da4821e38b32262bb7c728210e89468ce0e</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/core/java/com/android/server/biometrics/fingerprint/FingerprintService.java</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In FingerprintService, there is a possible bypass for operating system protections that isolate user profiles from each other due to a missing permission check. This could lead to a local information disclosure of metadata about the biometrics of another user on the device with no additional execution privileges needed. User interaction is not needed for exploitation. Product: AndroidVersions: Android-10Android ID: A-128599663 <p>Publish Date: 2019-09-27 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-9377>CVE-2019-9377</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-9377">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-9377</a></p> <p>Release Date: 2019-09-27</p> <p>Fix Resolution: android-10.0.0_r30</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-9377 (Low) detected in baseandroid-10.0.0_r14 - ## CVE-2019-9377 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-10.0.0_r14</b></p></summary> <p> <p>Android framework classes and services</p> <p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p> <p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/CVE-2020-0097-frameworks_base/commit/ab450da4821e38b32262bb7c728210e89468ce0e">ab450da4821e38b32262bb7c728210e89468ce0e</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/core/java/com/android/server/biometrics/fingerprint/FingerprintService.java</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In FingerprintService, there is a possible bypass for operating system protections that isolate user profiles from each other due to a missing permission check. This could lead to a local information disclosure of metadata about the biometrics of another user on the device with no additional execution privileges needed. User interaction is not needed for exploitation. Product: AndroidVersions: Android-10Android ID: A-128599663 <p>Publish Date: 2019-09-27 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-9377>CVE-2019-9377</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-9377">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-9377</a></p> <p>Release Date: 2019-09-27</p> <p>Fix Resolution: android-10.0.0_r30</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve low detected in baseandroid cve low severity vulnerability vulnerable library baseandroid android framework classes and services library home page a href found in head commit a href found in base branch master vulnerable source files core java com android server biometrics fingerprint fingerprintservice java vulnerability details in fingerprintservice there is a possible bypass for operating system protections that isolate user profiles from each other due to a missing permission check this could lead to a local information disclosure of metadata about the biometrics of another user on the device with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android step up your open source security game with mend
0
15,163
18,913,017,122
IssuesEvent
2021-11-16 15:49:26
googleapis/sphinx-docfx-yaml
https://api.github.com/repos/googleapis/sphinx-docfx-yaml
closed
Suppress blib2to3.pgen2.driver output...
type: process priority: p3
After adding `format_code` support in #144, there's a lot of debugging output like `blib2to3.pgen2.driver > Stop.` that is littered over the logs, cluttering that doesn't help with debugging through the logs. Will investigate to suppress this 😢
1.0
Suppress blib2to3.pgen2.driver output... - After adding `format_code` support in #144, there's a lot of debugging output like `blib2to3.pgen2.driver > Stop.` that is littered over the logs, cluttering that doesn't help with debugging through the logs. Will investigate to suppress this 😢
process
suppress driver output after adding format code support in there s a lot of debugging output like driver stop that is littered over the logs cluttering that doesn t help with debugging through the logs will investigate to suppress this 😢
1
231,364
7,631,589,272
IssuesEvent
2018-05-05 03:55:01
jahirfiquitiva/Blueprint
https://api.github.com/repos/jahirfiquitiva/Blueprint
closed
Please update FAQs to be include bottom navigation
accepted enhancement low priority
<!-- Any HTML comment will be stripped when the markdown is rendered, so you don't need to delete them. Put an x inside the [] like this: [x] to mark the checkbox. --> - [x] I have verified there are no duplicate active or recent bugs, questions, or requests - [x] I have verified that I am using the latest version of Blueprint. ### Device/dashboard info: - Blueprint version: `?` - Android version: `?` - Device Manufacturer: `?` - Device Name: `?` <!-- Please wrap code with correct syntax highlighting. You can remove it if you think it isn't necessary. --> ```kotlin println("Hello, world!") ``` <!-- If you are getting an error in the LogCat, paste here the stack trace. Please wrap logs with Gradle syntax highlighting (it makes them look better). --> ```Gradle java.lang.RuntimeException: This is an example Exception log at jahirfiquitiva.libs.blueprint.HelloWorld at jahirfiquitiva.libs.blueprint.HelloWorld$ThisIsNotARealLog at android.app.Instrumentation.callActivityOnResume(Instrumentation.kt) ``` <!-- The required steps to reproduce it. --> ### Reproduction Steps 1. 2. 3. ### Expected Result ### Actual Result <!-- Adding pictures/screenshots/videos of the expected/actual result is always helpful :) -->
1.0
Please update FAQs to be include bottom navigation - <!-- Any HTML comment will be stripped when the markdown is rendered, so you don't need to delete them. Put an x inside the [] like this: [x] to mark the checkbox. --> - [x] I have verified there are no duplicate active or recent bugs, questions, or requests - [x] I have verified that I am using the latest version of Blueprint. ### Device/dashboard info: - Blueprint version: `?` - Android version: `?` - Device Manufacturer: `?` - Device Name: `?` <!-- Please wrap code with correct syntax highlighting. You can remove it if you think it isn't necessary. --> ```kotlin println("Hello, world!") ``` <!-- If you are getting an error in the LogCat, paste here the stack trace. Please wrap logs with Gradle syntax highlighting (it makes them look better). --> ```Gradle java.lang.RuntimeException: This is an example Exception log at jahirfiquitiva.libs.blueprint.HelloWorld at jahirfiquitiva.libs.blueprint.HelloWorld$ThisIsNotARealLog at android.app.Instrumentation.callActivityOnResume(Instrumentation.kt) ``` <!-- The required steps to reproduce it. --> ### Reproduction Steps 1. 2. 3. ### Expected Result ### Actual Result <!-- Adding pictures/screenshots/videos of the expected/actual result is always helpful :) -->
non_process
please update faqs to be include bottom navigation any html comment will be stripped when the markdown is rendered so you don t need to delete them put an x inside the like this to mark the checkbox i have verified there are no duplicate active or recent bugs questions or requests i have verified that i am using the latest version of blueprint device dashboard info blueprint version android version device manufacturer device name please wrap code with correct syntax highlighting you can remove it if you think it isn t necessary kotlin println hello world if you are getting an error in the logcat paste here the stack trace please wrap logs with gradle syntax highlighting it makes them look better gradle java lang runtimeexception this is an example exception log at jahirfiquitiva libs blueprint helloworld at jahirfiquitiva libs blueprint helloworld thisisnotareallog at android app instrumentation callactivityonresume instrumentation kt the required steps to reproduce it reproduction steps expected result actual result adding pictures screenshots videos of the expected actual result is always helpful
0
118,964
15,383,340,552
IssuesEvent
2021-03-03 02:32:19
openrewrite/rewrite
https://api.github.com/repos/openrewrite/rewrite
closed
Unable to resolve dependencies from internal repository for unknown reason
bug design-partner-1
All this says is "`MavenPomDownloader.download()` returned null, which it does if the dependency cannot be downloaded for any reason". First step is to capture more information about why this could have failed. ``` org.openrewrite.maven.internal.MavenParsingException: Unable to download commons-logging:commons-logging:1.2. Including POM is at pom.xml at org.openrewrite.maven.internal.RawMavenResolver.lambda$processDependencies$4 (RawMavenResolver.java:300) at java.util.stream.ReferencePipeline$3$1.accept (ReferencePipeline.java:193) at java.util.stream.ReferencePipeline$2$1.accept (ReferencePipeline.java:175) at java.util.ArrayList$ArrayListSpliterator.forEachRemaining (ArrayList.java:1382) at java.util.stream.AbstractPipeline.copyInto (AbstractPipeline.java:481) at java.util.stream.AbstractPipeline.wrapAndCopyInto (AbstractPipeline.java:471) at java.util.stream.ReduceOps$ReduceOp.evaluateSequential (ReduceOps.java:708) at java.util.stream.AbstractPipeline.evaluate (AbstractPipeline.java:234) at java.util.stream.ReferencePipeline.collect (ReferencePipeline.java:499) at org.openrewrite.maven.internal.RawMavenResolver.processDependencies (RawMavenResolver.java:335) at org.openrewrite.maven.internal.RawMavenResolver.processTask (RawMavenResolver.java:122) at org.openrewrite.maven.internal.RawMavenResolver.resolve (RawMavenResolver.java:103) at org.openrewrite.maven.internal.RawMavenResolver.resolve (RawMavenResolver.java:91) at org.openrewrite.maven.internal.RawMavenResolver.resolve (RawMavenResolver.java:74) at org.openrewrite.maven.MavenParser.parseInputs (MavenParser.java:88) at org.openrewrite.Parser.parse (Parser.java:39) at org.openrewrite.maven.AbstractRewriteMojo.parseMaven (AbstractRewriteMojo.java:148) at org.openrewrite.maven.AbstractRewriteMojo.listResults (AbstractRewriteMojo.java:252) at org.openrewrite.maven.RewriteFixMojo.execute (RewriteFixMojo.java:36) at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo (DefaultBuildPluginManager.java:137) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:208) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:154) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:146) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:81) at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build (SingleThreadedBuilder.java:56) at org.apache.maven.lifecycle.internal.LifecycleStarter.execute (LifecycleStarter.java:128) at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:305) at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:192) at org.apache.maven.DefaultMaven.execute (DefaultMaven.java:105) at org.apache.maven.cli.MavenCli.execute (MavenCli.java:954) at org.apache.maven.cli.MavenCli.doMain (MavenCli.java:288) at org.apache.maven.cli.MavenCli.main (MavenCli.java:192) at sun.reflect.NativeMethodAccessorImpl.invoke0 (Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke (NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke (DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke (Method.java:498) at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced (Launcher.java:289) at org.codehaus.plexus.classworlds.launcher.Launcher.launch (Launcher.java:229) at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode (Launcher.java:415) at org.codehaus.plexus.classworlds.launcher.Launcher.main (Launcher.java:356) ```
1.0
Unable to resolve dependencies from internal repository for unknown reason - All this says is "`MavenPomDownloader.download()` returned null, which it does if the dependency cannot be downloaded for any reason". First step is to capture more information about why this could have failed. ``` org.openrewrite.maven.internal.MavenParsingException: Unable to download commons-logging:commons-logging:1.2. Including POM is at pom.xml at org.openrewrite.maven.internal.RawMavenResolver.lambda$processDependencies$4 (RawMavenResolver.java:300) at java.util.stream.ReferencePipeline$3$1.accept (ReferencePipeline.java:193) at java.util.stream.ReferencePipeline$2$1.accept (ReferencePipeline.java:175) at java.util.ArrayList$ArrayListSpliterator.forEachRemaining (ArrayList.java:1382) at java.util.stream.AbstractPipeline.copyInto (AbstractPipeline.java:481) at java.util.stream.AbstractPipeline.wrapAndCopyInto (AbstractPipeline.java:471) at java.util.stream.ReduceOps$ReduceOp.evaluateSequential (ReduceOps.java:708) at java.util.stream.AbstractPipeline.evaluate (AbstractPipeline.java:234) at java.util.stream.ReferencePipeline.collect (ReferencePipeline.java:499) at org.openrewrite.maven.internal.RawMavenResolver.processDependencies (RawMavenResolver.java:335) at org.openrewrite.maven.internal.RawMavenResolver.processTask (RawMavenResolver.java:122) at org.openrewrite.maven.internal.RawMavenResolver.resolve (RawMavenResolver.java:103) at org.openrewrite.maven.internal.RawMavenResolver.resolve (RawMavenResolver.java:91) at org.openrewrite.maven.internal.RawMavenResolver.resolve (RawMavenResolver.java:74) at org.openrewrite.maven.MavenParser.parseInputs (MavenParser.java:88) at org.openrewrite.Parser.parse (Parser.java:39) at org.openrewrite.maven.AbstractRewriteMojo.parseMaven (AbstractRewriteMojo.java:148) at org.openrewrite.maven.AbstractRewriteMojo.listResults (AbstractRewriteMojo.java:252) at org.openrewrite.maven.RewriteFixMojo.execute (RewriteFixMojo.java:36) at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo (DefaultBuildPluginManager.java:137) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:208) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:154) at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:146) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:81) at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build (SingleThreadedBuilder.java:56) at org.apache.maven.lifecycle.internal.LifecycleStarter.execute (LifecycleStarter.java:128) at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:305) at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:192) at org.apache.maven.DefaultMaven.execute (DefaultMaven.java:105) at org.apache.maven.cli.MavenCli.execute (MavenCli.java:954) at org.apache.maven.cli.MavenCli.doMain (MavenCli.java:288) at org.apache.maven.cli.MavenCli.main (MavenCli.java:192) at sun.reflect.NativeMethodAccessorImpl.invoke0 (Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke (NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke (DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke (Method.java:498) at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced (Launcher.java:289) at org.codehaus.plexus.classworlds.launcher.Launcher.launch (Launcher.java:229) at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode (Launcher.java:415) at org.codehaus.plexus.classworlds.launcher.Launcher.main (Launcher.java:356) ```
non_process
unable to resolve dependencies from internal repository for unknown reason all this says is mavenpomdownloader download returned null which it does if the dependency cannot be downloaded for any reason first step is to capture more information about why this could have failed org openrewrite maven internal mavenparsingexception unable to download commons logging commons logging including pom is at pom xml at org openrewrite maven internal rawmavenresolver lambda processdependencies rawmavenresolver java at java util stream referencepipeline accept referencepipeline java at java util stream referencepipeline accept referencepipeline java at java util arraylist arraylistspliterator foreachremaining arraylist java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream reduceops reduceop evaluatesequential reduceops java at java util stream abstractpipeline evaluate abstractpipeline java at java util stream referencepipeline collect referencepipeline java at org openrewrite maven internal rawmavenresolver processdependencies rawmavenresolver java at org openrewrite maven internal rawmavenresolver processtask rawmavenresolver java at org openrewrite maven internal rawmavenresolver resolve rawmavenresolver java at org openrewrite maven internal rawmavenresolver resolve rawmavenresolver java at org openrewrite maven internal rawmavenresolver resolve rawmavenresolver java at org openrewrite maven mavenparser parseinputs mavenparser java at org openrewrite parser parse parser java at org openrewrite maven abstractrewritemojo parsemaven abstractrewritemojo java at org openrewrite maven abstractrewritemojo listresults abstractrewritemojo java at org openrewrite maven rewritefixmojo execute rewritefixmojo java at org apache maven plugin defaultbuildpluginmanager executemojo defaultbuildpluginmanager java at org apache maven lifecycle internal mojoexecutor execute mojoexecutor java at org apache maven lifecycle internal mojoexecutor execute mojoexecutor java at org apache maven lifecycle internal mojoexecutor execute mojoexecutor java at org apache maven lifecycle internal lifecyclemodulebuilder buildproject lifecyclemodulebuilder java at org apache maven lifecycle internal lifecyclemodulebuilder buildproject lifecyclemodulebuilder java at org apache maven lifecycle internal builder singlethreaded singlethreadedbuilder build singlethreadedbuilder java at org apache maven lifecycle internal lifecyclestarter execute lifecyclestarter java at org apache maven defaultmaven doexecute defaultmaven java at org apache maven defaultmaven doexecute defaultmaven java at org apache maven defaultmaven execute defaultmaven java at org apache maven cli mavencli execute mavencli java at org apache maven cli mavencli domain mavencli java at org apache maven cli mavencli main mavencli java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org codehaus plexus classworlds launcher launcher launchenhanced launcher java at org codehaus plexus classworlds launcher launcher launch launcher java at org codehaus plexus classworlds launcher launcher mainwithexitcode launcher java at org codehaus plexus classworlds launcher launcher main launcher java
0