Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
592,475
| 17,893,711,185
|
IssuesEvent
|
2021-09-08 05:13:12
|
francheska-vicente/cssweng
|
https://api.github.com/repos/francheska-vicente/cssweng
|
opened
|
Discounts do not reflect when booking for Bridal Family Room
|
bug priority: high issue: back-end severity: high
|
### Summary:
- Only for Bridal Family Room, when inputting discounts, the total cost does not change
### Steps to Reproduce:
- Create Booking (Room 999, Sept 8-10, 2 pax, 2 PWD, 2000PHP discount)
### Visual Proof:


### Expected Results:
- Total cost would change
### Actual Results:
- Total cost remained the same
| Additional Information | |
| ----------- | ----------- |
| Platform | V8 engine (Google) |
| Operating System | Windows 10 |
|
1.0
|
Discounts do not reflect when booking for Bridal Family Room - ### Summary:
- Only for Bridal Family Room, when inputting discounts, the total cost does not change
### Steps to Reproduce:
- Create Booking (Room 999, Sept 8-10, 2 pax, 2 PWD, 2000PHP discount)
### Visual Proof:


### Expected Results:
- Total cost would change
### Actual Results:
- Total cost remained the same
| Additional Information | |
| ----------- | ----------- |
| Platform | V8 engine (Google) |
| Operating System | Windows 10 |
|
non_process
|
discounts do not reflect when booking for bridal family room summary only for bridal family room when inputting discounts the total cost does not change steps to reproduce create booking room sept pax pwd discount visual proof expected results total cost would change actual results total cost remained the same additional information platform engine google operating system windows
| 0
|
13,366
| 15,832,121,377
|
IssuesEvent
|
2021-04-06 14:18:40
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Create a utility package
|
kind/tech process/candidate team/client
|
Many tools are scattered around in the monorepo. What if we created a package for it? This could improve type safety and centralize updates on those tools.
|
1.0
|
Create a utility package - Many tools are scattered around in the monorepo. What if we created a package for it? This could improve type safety and centralize updates on those tools.
|
process
|
create a utility package many tools are scattered around in the monorepo what if we created a package for it this could improve type safety and centralize updates on those tools
| 1
|
7,464
| 10,562,919,925
|
IssuesEvent
|
2019-10-04 19:34:21
|
w3c/transitions
|
https://api.github.com/repos/w3c/transitions
|
closed
|
Use the real Github issues templates
|
Process Issue
|
It would be a better integration if you could use the real issue templates provided by GitHub
See: https://help.github.com/articles/about-issue-and-pull-request-templates/#issue-templates
(I discussed this with @deniak)
|
1.0
|
Use the real Github issues templates - It would be a better integration if you could use the real issue templates provided by GitHub
See: https://help.github.com/articles/about-issue-and-pull-request-templates/#issue-templates
(I discussed this with @deniak)
|
process
|
use the real github issues templates it would be a better integration if you could use the real issue templates provided by github see i discussed this with deniak
| 1
|
15,178
| 18,951,001,890
|
IssuesEvent
|
2021-11-18 15:10:19
|
googleapis/google-auth-library-python
|
https://api.github.com/repos/googleapis/google-auth-library-python
|
closed
|
Intersphinx URL for 'requests-oauthlib' now 404s
|
type: process
|
From [this failed build](https://source.cloud.google.com/results/invocations/56d76e91-941f-45a5-b4e4-6b74a9e86461/targets/cloud-devrel%2Fclient-libraries%2Fpython%2Fgoogleapis%2Fgoogle-auth-library-python%2Fpresubmit%2Fpresubmit/log):
```python
nox > Running session docs
nox > Creating virtual environment (virtualenv) using python3.8 in .nox/docs
nox > python -m pip install -e .[aiohttp]
nox > python -m pip install sphinx alabaster recommonmark sphinx-docstring-typing
nox > sphinx-build -T -W -N -b html -d docs/_build/doctrees/ docs/ docs/_build/html/
Running Sphinx v4.3.0
making output directory... done
loading intersphinx inventory from https://docs.python.org/3.5/objects.inv...
loading intersphinx inventory from https://urllib3.readthedocs.io/en/stable/objects.inv...
loading intersphinx inventory from https://requests.kennethreitz.org/en/master/objects.inv...
loading intersphinx inventory from https://requests-oauthlib.readthedocs.io/en/stable/objects.inv...
Traceback (most recent call last):
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/cmd/build.py", line 276, in build_main
app = Sphinx(args.sourcedir, args.confdir, args.outputdir,
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/application.py", line 270, in __init__
self._init_builder()
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/application.py", line 329, in _init_builder
self.events.emit('builder-inited')
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/events.py", line 101, in emit
results.append(listener.handler(self.app, *args))
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/ext/intersphinx.py", line 240, in load_mappings
updated = [f.result() for f in concurrent.futures.as_completed(futures)]
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/ext/intersphinx.py", line 240, in <listcomp>
updated = [f.result() for f in concurrent.futures.as_completed(futures)]
File "/usr/local/lib/python3.8/concurrent/futures/_base.py", line 437, in result
return self.__get_result()
File "/usr/local/lib/python3.8/concurrent/futures/_base.py", line 389, in __get_result
raise self._exception
File "/usr/local/lib/python3.8/concurrent/futures/thread.py", line 57, in run
result = self.fn(*self.args, **self.kwargs)
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/ext/intersphinx.py", line 225, in fetch_inventory_group
logger.warning(__("failed to reach any of the inventories "
File "/usr/local/lib/python3.8/logging/__init__.py", line 1812, in warning
self.log(WARNING, msg, *args, **kwargs)
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/util/logging.py", line 123, in log
super().log(level, msg, *args, **kwargs)
File "/usr/local/lib/python3.8/logging/__init__.py", line 1844, in log
self.logger.log(level, msg, *args, **kwargs)
File "/usr/local/lib/python3.8/logging/__init__.py", line 1512, in log
self._log(level, msg, args, **kwargs)
File "/usr/local/lib/python3.8/logging/__init__.py", line 1589, in _log
self.handle(record)
File "/usr/local/lib/python3.8/logging/__init__.py", line 1599, in handle
self.callHandlers(record)
File "/usr/local/lib/python3.8/logging/__init__.py", line 1661, in callHandlers
hdlr.handle(record)
File "/usr/local/lib/python3.8/logging/__init__.py", line 950, in handle
rv = self.filter(record)
File "/usr/local/lib/python3.8/logging/__init__.py", line 811, in filter
result = f.filter(record)
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/util/logging.py", line 430, in filter
raise exc
sphinx.errors.SphinxWarning: failed to reach any of the inventories with the following issues:
intersphinx inventory 'https://requests-oauthlib.readthedocs.io/en/stable/objects.inv' not fetchable due to <class 'requests.exceptions.HTTPError'>: 404 Client Error: Not Found for url: https://requests-oauthlib.readthedocs.io/en/stable/objects.inv
```
I have no idea why they would yank the `stable` docs version. The only versions I can see today for the [`requests-oauthlib` docs](https://requests-oauthlib.readthedocs.io/en/latest/) are `latest` and `v1.3.0-docs`. Given that `1.3.0` is the [latest release on PyPI](https://pypi.org/project/requests-oauthlib/#history), and is two years old, I think we're better off using the corresponding version.
|
1.0
|
Intersphinx URL for 'requests-oauthlib' now 404s - From [this failed build](https://source.cloud.google.com/results/invocations/56d76e91-941f-45a5-b4e4-6b74a9e86461/targets/cloud-devrel%2Fclient-libraries%2Fpython%2Fgoogleapis%2Fgoogle-auth-library-python%2Fpresubmit%2Fpresubmit/log):
```python
nox > Running session docs
nox > Creating virtual environment (virtualenv) using python3.8 in .nox/docs
nox > python -m pip install -e .[aiohttp]
nox > python -m pip install sphinx alabaster recommonmark sphinx-docstring-typing
nox > sphinx-build -T -W -N -b html -d docs/_build/doctrees/ docs/ docs/_build/html/
Running Sphinx v4.3.0
making output directory... done
loading intersphinx inventory from https://docs.python.org/3.5/objects.inv...
loading intersphinx inventory from https://urllib3.readthedocs.io/en/stable/objects.inv...
loading intersphinx inventory from https://requests.kennethreitz.org/en/master/objects.inv...
loading intersphinx inventory from https://requests-oauthlib.readthedocs.io/en/stable/objects.inv...
Traceback (most recent call last):
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/cmd/build.py", line 276, in build_main
app = Sphinx(args.sourcedir, args.confdir, args.outputdir,
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/application.py", line 270, in __init__
self._init_builder()
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/application.py", line 329, in _init_builder
self.events.emit('builder-inited')
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/events.py", line 101, in emit
results.append(listener.handler(self.app, *args))
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/ext/intersphinx.py", line 240, in load_mappings
updated = [f.result() for f in concurrent.futures.as_completed(futures)]
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/ext/intersphinx.py", line 240, in <listcomp>
updated = [f.result() for f in concurrent.futures.as_completed(futures)]
File "/usr/local/lib/python3.8/concurrent/futures/_base.py", line 437, in result
return self.__get_result()
File "/usr/local/lib/python3.8/concurrent/futures/_base.py", line 389, in __get_result
raise self._exception
File "/usr/local/lib/python3.8/concurrent/futures/thread.py", line 57, in run
result = self.fn(*self.args, **self.kwargs)
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/ext/intersphinx.py", line 225, in fetch_inventory_group
logger.warning(__("failed to reach any of the inventories "
File "/usr/local/lib/python3.8/logging/__init__.py", line 1812, in warning
self.log(WARNING, msg, *args, **kwargs)
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/util/logging.py", line 123, in log
super().log(level, msg, *args, **kwargs)
File "/usr/local/lib/python3.8/logging/__init__.py", line 1844, in log
self.logger.log(level, msg, *args, **kwargs)
File "/usr/local/lib/python3.8/logging/__init__.py", line 1512, in log
self._log(level, msg, args, **kwargs)
File "/usr/local/lib/python3.8/logging/__init__.py", line 1589, in _log
self.handle(record)
File "/usr/local/lib/python3.8/logging/__init__.py", line 1599, in handle
self.callHandlers(record)
File "/usr/local/lib/python3.8/logging/__init__.py", line 1661, in callHandlers
hdlr.handle(record)
File "/usr/local/lib/python3.8/logging/__init__.py", line 950, in handle
rv = self.filter(record)
File "/usr/local/lib/python3.8/logging/__init__.py", line 811, in filter
result = f.filter(record)
File "/tmpfs/src/github/google-auth-library-python/.nox/docs/lib/python3.8/site-packages/sphinx/util/logging.py", line 430, in filter
raise exc
sphinx.errors.SphinxWarning: failed to reach any of the inventories with the following issues:
intersphinx inventory 'https://requests-oauthlib.readthedocs.io/en/stable/objects.inv' not fetchable due to <class 'requests.exceptions.HTTPError'>: 404 Client Error: Not Found for url: https://requests-oauthlib.readthedocs.io/en/stable/objects.inv
```
I have no idea why they would yank the `stable` docs version. The only versions I can see today for the [`requests-oauthlib` docs](https://requests-oauthlib.readthedocs.io/en/latest/) are `latest` and `v1.3.0-docs`. Given that `1.3.0` is the [latest release on PyPI](https://pypi.org/project/requests-oauthlib/#history), and is two years old, I think we're better off using the corresponding version.
|
process
|
intersphinx url for requests oauthlib now from python nox running session docs nox creating virtual environment virtualenv using in nox docs nox python m pip install e nox python m pip install sphinx alabaster recommonmark sphinx docstring typing nox sphinx build t w n b html d docs build doctrees docs docs build html running sphinx making output directory done loading intersphinx inventory from loading intersphinx inventory from loading intersphinx inventory from loading intersphinx inventory from traceback most recent call last file tmpfs src github google auth library python nox docs lib site packages sphinx cmd build py line in build main app sphinx args sourcedir args confdir args outputdir file tmpfs src github google auth library python nox docs lib site packages sphinx application py line in init self init builder file tmpfs src github google auth library python nox docs lib site packages sphinx application py line in init builder self events emit builder inited file tmpfs src github google auth library python nox docs lib site packages sphinx events py line in emit results append listener handler self app args file tmpfs src github google auth library python nox docs lib site packages sphinx ext intersphinx py line in load mappings updated file tmpfs src github google auth library python nox docs lib site packages sphinx ext intersphinx py line in updated file usr local lib concurrent futures base py line in result return self get result file usr local lib concurrent futures base py line in get result raise self exception file usr local lib concurrent futures thread py line in run result self fn self args self kwargs file tmpfs src github google auth library python nox docs lib site packages sphinx ext intersphinx py line in fetch inventory group logger warning failed to reach any of the inventories file usr local lib logging init py line in warning self log warning msg args kwargs file tmpfs src github google auth library python nox docs lib site packages sphinx util logging py line in log super log level msg args kwargs file usr local lib logging init py line in log self logger log level msg args kwargs file usr local lib logging init py line in log self log level msg args kwargs file usr local lib logging init py line in log self handle record file usr local lib logging init py line in handle self callhandlers record file usr local lib logging init py line in callhandlers hdlr handle record file usr local lib logging init py line in handle rv self filter record file usr local lib logging init py line in filter result f filter record file tmpfs src github google auth library python nox docs lib site packages sphinx util logging py line in filter raise exc sphinx errors sphinxwarning failed to reach any of the inventories with the following issues intersphinx inventory not fetchable due to client error not found for url i have no idea why they would yank the stable docs version the only versions i can see today for the are latest and docs given that is the and is two years old i think we re better off using the corresponding version
| 1
|
441,234
| 12,709,669,513
|
IssuesEvent
|
2020-06-23 12:44:54
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
opened
|
Invalid TLS handshakes from the grpc cpp client when connecting to a specific IP
|
kind/bug priority/P2
|
### What version of gRPC and what language are you using?
Latest, C++
### What operating system (Linux, Windows,...) and version?
Linux, Centos 7
### What runtime / compiler are you using (e.g. python version or version of gcc)
n/a
### What did you do?
**Background on the TLS Server Name Indication extension**
> Server Name Indication (SNI) is an extension to the Transport Layer Security (TLS) computer networking protocol by which a client indicates which hostname it is attempting to connect to at the start of the handshaking process.[1] This allows a server to present multiple certificates on the same IP address and TCP port number and hence allows multiple secure (HTTPS) websites (or any other service over TLS) to be served by the same IP address without requiring all those sites to use the same certificate.
https://en.wikipedia.org/wiki/Server_Name_Indication
According to the SNI TLS extension spec the SNI in a TLS handshake must be set to a fully qualified domain name, not an IP address:
> "HostName" contains the fully qualified DNS hostname of the server, as understood by the client. ... Literal IPv4 and IPv6 addresses are not permitted in "HostName".
https://tools.ietf.org/html/rfc6066#section-3
**Bug**
In our setup we have a grpc server running with a valid certificate for the IPv6 address its bound to.
We're attempting to use the grpc c++ client to connect directly to this server via its IPv6 address and port directly, rather than using a fully qualified domain name. When doing so the grpc client incorrectly sets the Server Name Indicator (SNI) in its TLS handshake to the target IPv6 address.
Our grpc server implementation then rejects the request due to the SNI being invalid.
### What did you expect to see?
I expect grpc to only use the SNI TLS extension when its connecting to a domain name. This will make grpc clients use of SNI spec compliant.
### What did you see instead?
grpc always uses the SNI extension, regardless of whether its connecting to an IP address or a hostname.
### Anything else we should know about your project / environment?
If changing grpc so that it doesn't set the SNI when connecting directly to an IP address sounds like a valid solution I'm happy to implement it. Code pointers or advice is appreciated.
|
1.0
|
Invalid TLS handshakes from the grpc cpp client when connecting to a specific IP - ### What version of gRPC and what language are you using?
Latest, C++
### What operating system (Linux, Windows,...) and version?
Linux, Centos 7
### What runtime / compiler are you using (e.g. python version or version of gcc)
n/a
### What did you do?
**Background on the TLS Server Name Indication extension**
> Server Name Indication (SNI) is an extension to the Transport Layer Security (TLS) computer networking protocol by which a client indicates which hostname it is attempting to connect to at the start of the handshaking process.[1] This allows a server to present multiple certificates on the same IP address and TCP port number and hence allows multiple secure (HTTPS) websites (or any other service over TLS) to be served by the same IP address without requiring all those sites to use the same certificate.
https://en.wikipedia.org/wiki/Server_Name_Indication
According to the SNI TLS extension spec the SNI in a TLS handshake must be set to a fully qualified domain name, not an IP address:
> "HostName" contains the fully qualified DNS hostname of the server, as understood by the client. ... Literal IPv4 and IPv6 addresses are not permitted in "HostName".
https://tools.ietf.org/html/rfc6066#section-3
**Bug**
In our setup we have a grpc server running with a valid certificate for the IPv6 address its bound to.
We're attempting to use the grpc c++ client to connect directly to this server via its IPv6 address and port directly, rather than using a fully qualified domain name. When doing so the grpc client incorrectly sets the Server Name Indicator (SNI) in its TLS handshake to the target IPv6 address.
Our grpc server implementation then rejects the request due to the SNI being invalid.
### What did you expect to see?
I expect grpc to only use the SNI TLS extension when its connecting to a domain name. This will make grpc clients use of SNI spec compliant.
### What did you see instead?
grpc always uses the SNI extension, regardless of whether its connecting to an IP address or a hostname.
### Anything else we should know about your project / environment?
If changing grpc so that it doesn't set the SNI when connecting directly to an IP address sounds like a valid solution I'm happy to implement it. Code pointers or advice is appreciated.
|
non_process
|
invalid tls handshakes from the grpc cpp client when connecting to a specific ip what version of grpc and what language are you using latest c what operating system linux windows and version linux centos what runtime compiler are you using e g python version or version of gcc n a what did you do background on the tls server name indication extension server name indication sni is an extension to the transport layer security tls computer networking protocol by which a client indicates which hostname it is attempting to connect to at the start of the handshaking process this allows a server to present multiple certificates on the same ip address and tcp port number and hence allows multiple secure https websites or any other service over tls to be served by the same ip address without requiring all those sites to use the same certificate according to the sni tls extension spec the sni in a tls handshake must be set to a fully qualified domain name not an ip address hostname contains the fully qualified dns hostname of the server as understood by the client literal and addresses are not permitted in hostname bug in our setup we have a grpc server running with a valid certificate for the address its bound to we re attempting to use the grpc c client to connect directly to this server via its address and port directly rather than using a fully qualified domain name when doing so the grpc client incorrectly sets the server name indicator sni in its tls handshake to the target address our grpc server implementation then rejects the request due to the sni being invalid what did you expect to see i expect grpc to only use the sni tls extension when its connecting to a domain name this will make grpc clients use of sni spec compliant what did you see instead grpc always uses the sni extension regardless of whether its connecting to an ip address or a hostname anything else we should know about your project environment if changing grpc so that it doesn t set the sni when connecting directly to an ip address sounds like a valid solution i m happy to implement it code pointers or advice is appreciated
| 0
|
485,842
| 14,000,311,017
|
IssuesEvent
|
2020-10-28 12:08:11
|
EyeSeeTea/training-app
|
https://api.github.com/repos/EyeSeeTea/training-app
|
closed
|
Close behavior on tutorial
|
complexity - low (1 hr) priority - low priority - medium status - blocked type - cosmetic type - feedback
|
- Pop-up window on shutting down tutorial telling users what is happening and how they can relaunch the tool (or minimize if they prefer)
|
2.0
|
Close behavior on tutorial - - Pop-up window on shutting down tutorial telling users what is happening and how they can relaunch the tool (or minimize if they prefer)
|
non_process
|
close behavior on tutorial pop up window on shutting down tutorial telling users what is happening and how they can relaunch the tool or minimize if they prefer
| 0
|
6,027
| 7,468,143,929
|
IssuesEvent
|
2018-04-02 17:55:45
|
acelabini/ibinex-option-one
|
https://api.github.com/repos/acelabini/ibinex-option-one
|
closed
|
[ SERVICES ] Orientation and size of phone icon in nav/header
|
DESKTOP/WEB NAV/HEADER SERVICES
|
Tester: Joel Simpao
OS: Windows 10
Browser: Chrome
Expected:

Actual:

|
1.0
|
[ SERVICES ] Orientation and size of phone icon in nav/header - Tester: Joel Simpao
OS: Windows 10
Browser: Chrome
Expected:

Actual:

|
non_process
|
orientation and size of phone icon in nav header tester joel simpao os windows browser chrome expected actual
| 0
|
1,454
| 4,029,504,009
|
IssuesEvent
|
2016-05-18 10:54:38
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
child_process ignores cwd
|
child_process windows
|
<!--
Thank you for reporting an issue. Please fill in the template below. If unsure
about something, just do as best as you're able.
Version: usually output of `node -v`
Platform: either `uname -a` output, or if Windows, version and 32 or 64-bit
Subsystem: if known, please specify affected core module name
It will be much easier for us to fix the issue if a test case that reproduces
the problem is provided. Ideally this test case should not have any external
dependencies. We understand that it is not always possible to reduce your code
to a small test case, but we would appreciate to have as
much data as possible.
-->
* **v4.4.4**:
* **Windows 7 64bit**:
* **child_process**:
<!-- Enter your issue details below this comment. -->
I run from the command line using `node test.js` the following file:
```
var exec = require('child_process').exec;
exec('cd',{cwd: "C:\\Code\\Misc"},function(error,stdout,stderr) {
console.log(stdout);
});
```
gives the output `C:\Users\myname`
`test.js` is located in `C:\Code\Misc`.
Any Idea why?? Shouldn't it give `C:\Code\Misc`? the cwd I specified?
This causes problems when trying to use the Atom text editor 'build' add-on. I have recently re-installed node and npm, but the error persists.
|
1.0
|
child_process ignores cwd - <!--
Thank you for reporting an issue. Please fill in the template below. If unsure
about something, just do as best as you're able.
Version: usually output of `node -v`
Platform: either `uname -a` output, or if Windows, version and 32 or 64-bit
Subsystem: if known, please specify affected core module name
It will be much easier for us to fix the issue if a test case that reproduces
the problem is provided. Ideally this test case should not have any external
dependencies. We understand that it is not always possible to reduce your code
to a small test case, but we would appreciate to have as
much data as possible.
-->
* **v4.4.4**:
* **Windows 7 64bit**:
* **child_process**:
<!-- Enter your issue details below this comment. -->
I run from the command line using `node test.js` the following file:
```
var exec = require('child_process').exec;
exec('cd',{cwd: "C:\\Code\\Misc"},function(error,stdout,stderr) {
console.log(stdout);
});
```
gives the output `C:\Users\myname`
`test.js` is located in `C:\Code\Misc`.
Any Idea why?? Shouldn't it give `C:\Code\Misc`? the cwd I specified?
This causes problems when trying to use the Atom text editor 'build' add-on. I have recently re-installed node and npm, but the error persists.
|
process
|
child process ignores cwd thank you for reporting an issue please fill in the template below if unsure about something just do as best as you re able version usually output of node v platform either uname a output or if windows version and or bit subsystem if known please specify affected core module name it will be much easier for us to fix the issue if a test case that reproduces the problem is provided ideally this test case should not have any external dependencies we understand that it is not always possible to reduce your code to a small test case but we would appreciate to have as much data as possible windows child process i run from the command line using node test js the following file var exec require child process exec exec cd cwd c code misc function error stdout stderr console log stdout gives the output c users myname test js is located in c code misc any idea why shouldn t it give c code misc the cwd i specified this causes problems when trying to use the atom text editor build add on i have recently re installed node and npm but the error persists
| 1
|
9,181
| 12,227,846,644
|
IssuesEvent
|
2020-05-03 16:54:16
|
emacs-ess/ESS
|
https://api.github.com/repos/emacs-ess/ESS
|
closed
|
`ess-set-working-directory` should not change `default-directory`
|
process
|
I've been confused for years with the behaviour of Emacs regarding the active directories of the files I'm editing. I just realised it's `ess-set-working-directory` that changes it to the new working directory of R.
It seems that this behaviour is intended: https://github.com/emacs-ess/ESS/commit/40ebf0f71eb0677895506201fac6da4d6119ffa2. But I'm not sure why changing the REPL directory should affect the buffer I'm switching from.
Should we change this to preserve the active directory? The active directory influences UI commands like `M-x find-file`. I suspect I'm not the only one who is confused by this behaviour.
|
1.0
|
`ess-set-working-directory` should not change `default-directory` - I've been confused for years with the behaviour of Emacs regarding the active directories of the files I'm editing. I just realised it's `ess-set-working-directory` that changes it to the new working directory of R.
It seems that this behaviour is intended: https://github.com/emacs-ess/ESS/commit/40ebf0f71eb0677895506201fac6da4d6119ffa2. But I'm not sure why changing the REPL directory should affect the buffer I'm switching from.
Should we change this to preserve the active directory? The active directory influences UI commands like `M-x find-file`. I suspect I'm not the only one who is confused by this behaviour.
|
process
|
ess set working directory should not change default directory i ve been confused for years with the behaviour of emacs regarding the active directories of the files i m editing i just realised it s ess set working directory that changes it to the new working directory of r it seems that this behaviour is intended but i m not sure why changing the repl directory should affect the buffer i m switching from should we change this to preserve the active directory the active directory influences ui commands like m x find file i suspect i m not the only one who is confused by this behaviour
| 1
|
18,021
| 24,032,779,074
|
IssuesEvent
|
2022-09-15 16:18:44
|
googleapis/java-beyondcorp-appgateways
|
https://api.github.com/repos/googleapis/java-beyondcorp-appgateways
|
opened
|
Your .repo-metadata.json file has a problem 🤒
|
type: process repo-metadata: lint
|
You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'beyondcorp-appgateways' invalid in .repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'beyondcorp-appgateways' invalid in .repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname beyondcorp appgateways invalid in repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
| 1
|
65,642
| 8,824,579,129
|
IssuesEvent
|
2019-01-02 17:40:29
|
LycheeOrg/Lychee-Laravel
|
https://api.github.com/repos/LycheeOrg/Lychee-Laravel
|
closed
|
Add `.env` file
|
Documentation
|
`.env.exmple` is helpful for learning how to configure Laravel, but should the repo contain a pre-configured `.env` for the user? Even an empty document so they encryption key is added to the correct place will make installation easier.
|
1.0
|
Add `.env` file - `.env.exmple` is helpful for learning how to configure Laravel, but should the repo contain a pre-configured `.env` for the user? Even an empty document so they encryption key is added to the correct place will make installation easier.
|
non_process
|
add env file env exmple is helpful for learning how to configure laravel but should the repo contain a pre configured env for the user even an empty document so they encryption key is added to the correct place will make installation easier
| 0
|
14,578
| 17,702,955,893
|
IssuesEvent
|
2021-08-25 01:58:12
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
Change term - associatedTaxa
|
Term - change Class - Occurrence Class - ResourceRelationship non-normative Process - complete
|
## Change term
* Submitter: John Wieczorek
* Justification (why is this change necessary?): Consistency and clarity
* Proponents (who needs this change): Everyone
Current Term definition: https://dwc.tdwg.org/terms/#dwc:associatedTaxa
Proposed new attributes of the term:
* Term name (in lowerCamelCase): associatedTaxa
* Organized in Class (e.g. Location, Taxon): Occurrence
* Definition of the term: **A list (concatenated and separated) of identifiers or names of taxa and the associations of this Occurrence to each of them.**
* Usage comments (recommendations regarding content, etc.): **This term can be used to provide a list of associations to Taxa other than the one defined in the Occurrence. Note that the ResourceRelationship class is an alternative means of representing associations, and with more detail. This term is not apt for establishing relationships between Taxa, only between specific Occurrences of an Organism with other Taxa. Recommended best practice is to separate the values in a list with space vertical bar space ( | ).**
* Examples: `"host":"Quercus alba"`, **`"host":"gbif.org/species/2879737"`**,`"parasitoid of":"Cyclocephala signaticollis" | "predator of":"Apis mellifera"`
* Refines (identifier of the broader term this term refines, if applicable): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): http://rs.tdwg.org/dwc/terms/version/associatedTaxa-2017-10-06
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG, if applicable): DataSets/DataSet/Units/Unit/Gathering/Synecology/AssociatedTaxa
Discussions around changes to relationshipOfResource (#194), around a new term relationshipOfResourceID (#186, #283), and changes to associatedOccurrences (Issue #324) suggest that a clarification should also be made in the associatedTaxa definition and usage notes. Specifically, the directionality of the relationship should be made clear.
|
1.0
|
Change term - associatedTaxa - ## Change term
* Submitter: John Wieczorek
* Justification (why is this change necessary?): Consistency and clarity
* Proponents (who needs this change): Everyone
Current Term definition: https://dwc.tdwg.org/terms/#dwc:associatedTaxa
Proposed new attributes of the term:
* Term name (in lowerCamelCase): associatedTaxa
* Organized in Class (e.g. Location, Taxon): Occurrence
* Definition of the term: **A list (concatenated and separated) of identifiers or names of taxa and the associations of this Occurrence to each of them.**
* Usage comments (recommendations regarding content, etc.): **This term can be used to provide a list of associations to Taxa other than the one defined in the Occurrence. Note that the ResourceRelationship class is an alternative means of representing associations, and with more detail. This term is not apt for establishing relationships between Taxa, only between specific Occurrences of an Organism with other Taxa. Recommended best practice is to separate the values in a list with space vertical bar space ( | ).**
* Examples: `"host":"Quercus alba"`, **`"host":"gbif.org/species/2879737"`**,`"parasitoid of":"Cyclocephala signaticollis" | "predator of":"Apis mellifera"`
* Refines (identifier of the broader term this term refines, if applicable): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term, if applicable): http://rs.tdwg.org/dwc/terms/version/associatedTaxa-2017-10-06
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG, if applicable): DataSets/DataSet/Units/Unit/Gathering/Synecology/AssociatedTaxa
Discussions around changes to relationshipOfResource (#194), around a new term relationshipOfResourceID (#186, #283), and changes to associatedOccurrences (Issue #324) suggest that a clarification should also be made in the associatedTaxa definition and usage notes. Specifically, the directionality of the relationship should be made clear.
|
process
|
change term associatedtaxa change term submitter john wieczorek justification why is this change necessary consistency and clarity proponents who needs this change everyone current term definition proposed new attributes of the term term name in lowercamelcase associatedtaxa organized in class e g location taxon occurrence definition of the term a list concatenated and separated of identifiers or names of taxa and the associations of this occurrence to each of them usage comments recommendations regarding content etc this term can be used to provide a list of associations to taxa other than the one defined in the occurrence note that the resourcerelationship class is an alternative means of representing associations and with more detail this term is not apt for establishing relationships between taxa only between specific occurrences of an organism with other taxa recommended best practice is to separate the values in a list with space vertical bar space examples host quercus alba host gbif org species parasitoid of cyclocephala signaticollis predator of apis mellifera refines identifier of the broader term this term refines if applicable none replaces identifier of the existing term that would be deprecated and replaced by this term if applicable abcd xpath of the equivalent term in abcd or efg if applicable datasets dataset units unit gathering synecology associatedtaxa discussions around changes to relationshipofresource around a new term relationshipofresourceid and changes to associatedoccurrences issue suggest that a clarification should also be made in the associatedtaxa definition and usage notes specifically the directionality of the relationship should be made clear
| 1
|
136,796
| 12,734,791,564
|
IssuesEvent
|
2020-06-25 14:24:56
|
graknlabs/graql
|
https://api.github.com/repos/graknlabs/graql
|
opened
|
Slack badge displayed in README
|
type: documentation
|
## Problem to Solve
A Slack badge is displayed in the README.
## Proposed Solution
Replace the Slack badge with a Discord badge, like we have in the `grakn` repo.
|
1.0
|
Slack badge displayed in README - ## Problem to Solve
A Slack badge is displayed in the README.
## Proposed Solution
Replace the Slack badge with a Discord badge, like we have in the `grakn` repo.
|
non_process
|
slack badge displayed in readme problem to solve a slack badge is displayed in the readme proposed solution replace the slack badge with a discord badge like we have in the grakn repo
| 0
|
640
| 3,098,040,417
|
IssuesEvent
|
2015-08-28 08:20:42
|
deb-sandeep/PHPWebApps
|
https://api.github.com/repos/deb-sandeep/PHPWebApps
|
opened
|
Provide support for multi-choice questions
|
enhancement jove_notes_db jove_notes_grammar jove_notes_processor jove_notes_server jove_notes_ui
|
This is an enhancement request for supporting multiple-choice type questions in JoveNotes, which is _sorely_ missing now.
### Definition
A multiple-choice question is a type of interactive question, which has:
1. A question
2. Multiple options
3. One or more right answers
4. Explanation [Optional]
The user interaction model differs while interacting between a single choice and multiple choice question. For example, in a single choice question the user has to choose only one answer, after which the system evaluates the answer. In case of multiple-choices, system should take over the response for evaluation only when the user has selected all the possible answers and given a hint to the system that he is done choosing the answers.
_Should the system not take over when the requisite number of answers have been chosen?_
> This might look like a logical choice, but IMHO is not the right approach for the following reasons. Assume there are 'n' options and 'x' correct answers. {x <= n}
>
> 1. If the user chooses y answers (y<x) and believes that he has completed it, the system will keep waiting for him to choose x-y remaining answers.
> 2. If the user believes there are y answers (y>x), the system will not let him make the mistake.
Hence, in case of multiple answer questions, the user should let the system know once he is done choosing all the answers.
### Syntax for digitization
```
@multi_choice "Choose the plants which are a planets in our solar system" {
@options {
"Sun",
"Earth" correct,
"Moon",
"Neptune correct"
}
@explanation "Sun is a star and Moon is a satellite and hence they are not planets"
}
```
Important thing to note is that the system recognizes whether it should treat the question as single choice or multi choice based on the number of correct answers. The digitization syntax remains the same.
|
1.0
|
Provide support for multi-choice questions - This is an enhancement request for supporting multiple-choice type questions in JoveNotes, which is _sorely_ missing now.
### Definition
A multiple-choice question is a type of interactive question, which has:
1. A question
2. Multiple options
3. One or more right answers
4. Explanation [Optional]
The user interaction model differs while interacting between a single choice and multiple choice question. For example, in a single choice question the user has to choose only one answer, after which the system evaluates the answer. In case of multiple-choices, system should take over the response for evaluation only when the user has selected all the possible answers and given a hint to the system that he is done choosing the answers.
_Should the system not take over when the requisite number of answers have been chosen?_
> This might look like a logical choice, but IMHO is not the right approach for the following reasons. Assume there are 'n' options and 'x' correct answers. {x <= n}
>
> 1. If the user chooses y answers (y<x) and believes that he has completed it, the system will keep waiting for him to choose x-y remaining answers.
> 2. If the user believes there are y answers (y>x), the system will not let him make the mistake.
Hence, in case of multiple answer questions, the user should let the system know once he is done choosing all the answers.
### Syntax for digitization
```
@multi_choice "Choose the plants which are a planets in our solar system" {
@options {
"Sun",
"Earth" correct,
"Moon",
"Neptune correct"
}
@explanation "Sun is a star and Moon is a satellite and hence they are not planets"
}
```
Important thing to note is that the system recognizes whether it should treat the question as single choice or multi choice based on the number of correct answers. The digitization syntax remains the same.
|
process
|
provide support for multi choice questions this is an enhancement request for supporting multiple choice type questions in jovenotes which is sorely missing now definition a multiple choice question is a type of interactive question which has a question multiple options one or more right answers explanation the user interaction model differs while interacting between a single choice and multiple choice question for example in a single choice question the user has to choose only one answer after which the system evaluates the answer in case of multiple choices system should take over the response for evaluation only when the user has selected all the possible answers and given a hint to the system that he is done choosing the answers should the system not take over when the requisite number of answers have been chosen this might look like a logical choice but imho is not the right approach for the following reasons assume there are n options and x correct answers x n if the user chooses y answers y x and believes that he has completed it the system will keep waiting for him to choose x y remaining answers if the user believes there are y answers y x the system will not let him make the mistake hence in case of multiple answer questions the user should let the system know once he is done choosing all the answers syntax for digitization multi choice choose the plants which are a planets in our solar system options sun earth correct moon neptune correct explanation sun is a star and moon is a satellite and hence they are not planets important thing to note is that the system recognizes whether it should treat the question as single choice or multi choice based on the number of correct answers the digitization syntax remains the same
| 1
|
2,295
| 5,115,485,568
|
IssuesEvent
|
2017-01-06 21:58:36
|
jlm2017/jlm-video-subtitles
|
https://api.github.com/repos/jlm2017/jlm-video-subtitles
|
opened
|
[subtitles] [eng] MÉLENCHON - POURQUOI LE CODE DU TRAVAIL EST GROS ?
|
Language: English Process: [2] Ready for review (1)
|
# Video title
MÉLENCHON - POURQUOI LE CODE DU TRAVAIL EST GROS ?
# URL
https://www.youtube.com/watch?v=jnqujbuCNsM
# Youtube subtitles language
Anglais
# Duration
1:58
# Subtitles URL
https://www.youtube.com/timedtext_editor?tab=captions&bl=vmp&lang=en-US&action_mde_edit_form=1&ref=player&v=jnqujbuCNsM&ui=hd
|
1.0
|
[subtitles] [eng] MÉLENCHON - POURQUOI LE CODE DU TRAVAIL EST GROS ? - # Video title
MÉLENCHON - POURQUOI LE CODE DU TRAVAIL EST GROS ?
# URL
https://www.youtube.com/watch?v=jnqujbuCNsM
# Youtube subtitles language
Anglais
# Duration
1:58
# Subtitles URL
https://www.youtube.com/timedtext_editor?tab=captions&bl=vmp&lang=en-US&action_mde_edit_form=1&ref=player&v=jnqujbuCNsM&ui=hd
|
process
|
mélenchon pourquoi le code du travail est gros video title mélenchon pourquoi le code du travail est gros url youtube subtitles language anglais duration subtitles url
| 1
|
5,900
| 8,717,581,182
|
IssuesEvent
|
2018-12-07 17:35:35
|
GoogleCloudPlatform/google-cloud-cpp
|
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-cpp
|
opened
|
Decide if we want to support Xcode 7.3.
|
type: process
|
We got a bug report, the code fails to compile with [Xcode 7.3](https://en.wikipedia.org/wiki/Xcode#7.x_series), which was released around 2015. The problem is actually in the compiler (e.g. does not support `thread_local`), and will submit a workaround shortly.
We need to decide if the compiler is something we want to support long term, thus this bug to make us revisit the decision.
|
1.0
|
Decide if we want to support Xcode 7.3. - We got a bug report, the code fails to compile with [Xcode 7.3](https://en.wikipedia.org/wiki/Xcode#7.x_series), which was released around 2015. The problem is actually in the compiler (e.g. does not support `thread_local`), and will submit a workaround shortly.
We need to decide if the compiler is something we want to support long term, thus this bug to make us revisit the decision.
|
process
|
decide if we want to support xcode we got a bug report the code fails to compile with which was released around the problem is actually in the compiler e g does not support thread local and will submit a workaround shortly we need to decide if the compiler is something we want to support long term thus this bug to make us revisit the decision
| 1
|
458,404
| 13,174,672,593
|
IssuesEvent
|
2020-08-11 23:09:00
|
shimming-toolbox/shimming-toolbox-py
|
https://api.github.com/repos/shimming-toolbox/shimming-toolbox-py
|
opened
|
Improve masking capabilities
|
Priority: LOW enhancement
|
## Context
The initial issue #47 which was partially fixed by PR #60 added some basic capabilities to mask some data (threshold, shape: square, cube). More masking capability could be implemented to segment the brain and spinal cord as well as other shapes.
Algorithms that could help
- SCT: for spinal cord + simple shapes (license MIT)
- FSL BET (non-commercial): for brain
- Others?
## Suggestion
- Add disk and ball to `shimmingtoolbox.masking.shape
- Add capability to handle phase images to `shimmingtoolbox.masking.threshold
- `shimmingtoolbox.masking.bet()` --> input is file name
- `shimmingtoolbox.masking.sct()` --> input could be file name or an np.array (if we use SCT's API)
output: mask as nd.array (of file in the case of bet)
|
1.0
|
Improve masking capabilities - ## Context
The initial issue #47 which was partially fixed by PR #60 added some basic capabilities to mask some data (threshold, shape: square, cube). More masking capability could be implemented to segment the brain and spinal cord as well as other shapes.
Algorithms that could help
- SCT: for spinal cord + simple shapes (license MIT)
- FSL BET (non-commercial): for brain
- Others?
## Suggestion
- Add disk and ball to `shimmingtoolbox.masking.shape
- Add capability to handle phase images to `shimmingtoolbox.masking.threshold
- `shimmingtoolbox.masking.bet()` --> input is file name
- `shimmingtoolbox.masking.sct()` --> input could be file name or an np.array (if we use SCT's API)
output: mask as nd.array (of file in the case of bet)
|
non_process
|
improve masking capabilities context the initial issue which was partially fixed by pr added some basic capabilities to mask some data threshold shape square cube more masking capability could be implemented to segment the brain and spinal cord as well as other shapes algorithms that could help sct for spinal cord simple shapes license mit fsl bet non commercial for brain others suggestion add disk and ball to shimmingtoolbox masking shape add capability to handle phase images to shimmingtoolbox masking threshold shimmingtoolbox masking bet input is file name shimmingtoolbox masking sct input could be file name or an np array if we use sct s api output mask as nd array of file in the case of bet
| 0
|
15,945
| 20,163,727,854
|
IssuesEvent
|
2022-02-10 00:47:37
|
ooi-data/CE09OSSM-MFD37-01-OPTAAC000-telemetered-optaa_dj_dcl_instrument
|
https://api.github.com/repos/ooi-data/CE09OSSM-MFD37-01-OPTAAC000-telemetered-optaa_dj_dcl_instrument
|
opened
|
🛑 Processing failed: GroupNotFoundError
|
process
|
## Overview
`GroupNotFoundError` found in `processing_task` task during run ended on 2022-02-10T00:47:36.630149.
## Details
Flow name: `CE09OSSM-MFD37-01-OPTAAC000-telemetered-optaa_dj_dcl_instrument`
Task name: `processing_task`
Error type: `GroupNotFoundError`
Error message: group not found at path ''
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 64, in finalize_data_stream
final_group = zarr.open_group(final_store, mode='r+')
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/hierarchy.py", line 1168, in open_group
raise GroupNotFoundError(path)
zarr.errors.GroupNotFoundError: group not found at path ''
```
</details>
|
1.0
|
🛑 Processing failed: GroupNotFoundError - ## Overview
`GroupNotFoundError` found in `processing_task` task during run ended on 2022-02-10T00:47:36.630149.
## Details
Flow name: `CE09OSSM-MFD37-01-OPTAAC000-telemetered-optaa_dj_dcl_instrument`
Task name: `processing_task`
Error type: `GroupNotFoundError`
Error message: group not found at path ''
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 64, in finalize_data_stream
final_group = zarr.open_group(final_store, mode='r+')
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/hierarchy.py", line 1168, in open_group
raise GroupNotFoundError(path)
zarr.errors.GroupNotFoundError: group not found at path ''
```
</details>
|
process
|
🛑 processing failed groupnotfounderror overview groupnotfounderror found in processing task task during run ended on details flow name telemetered optaa dj dcl instrument task name processing task error type groupnotfounderror error message group not found at path traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream final group zarr open group final store mode r file srv conda envs notebook lib site packages zarr hierarchy py line in open group raise groupnotfounderror path zarr errors groupnotfounderror group not found at path
| 1
|
208,120
| 15,876,454,812
|
IssuesEvent
|
2021-04-09 08:23:42
|
microsoft/AzureStorageExplorer
|
https://api.github.com/repos/microsoft/AzureStorageExplorer
|
closed
|
No default blob container name shows in the 'Clone' dialog for one local emulator blob container
|
:beetle: regression :gear: emulator 🧪 testing
|
**Storage Explorer Version**: 1.19.0-dev
**Build Number**: 20210408.5
**Branch**: main
**Platform/OS**: Windows 10
**Emulator Version**: Azurite 3.11.0/ Microsoft Azure Storage Emulator 5.10.0.0
**Architecture**: ia32
**Regression From**: Previous release (1.18.1)
## Steps to Reproduce ##
1. Install and run Azurite 3.11.0 -> Launch Storage Explorer.
2. Expand 'Local & Attached -> Storage Accounts -> (Emulator - Default Ports) (Key)'.
3. Right click the 'Blob Containers' node -> Create a new blob container.
4. Right click the blob container -> Click 'Clone...'.
5. Check whether the default blob container name shows in the dialog.
## Expected Experience ##
The default blob container name shows.

## Actual Experience ##
No default blob container name shows.

## Additional Context ##
The error message keeps showing after typing one valid blob container name.

|
1.0
|
No default blob container name shows in the 'Clone' dialog for one local emulator blob container - **Storage Explorer Version**: 1.19.0-dev
**Build Number**: 20210408.5
**Branch**: main
**Platform/OS**: Windows 10
**Emulator Version**: Azurite 3.11.0/ Microsoft Azure Storage Emulator 5.10.0.0
**Architecture**: ia32
**Regression From**: Previous release (1.18.1)
## Steps to Reproduce ##
1. Install and run Azurite 3.11.0 -> Launch Storage Explorer.
2. Expand 'Local & Attached -> Storage Accounts -> (Emulator - Default Ports) (Key)'.
3. Right click the 'Blob Containers' node -> Create a new blob container.
4. Right click the blob container -> Click 'Clone...'.
5. Check whether the default blob container name shows in the dialog.
## Expected Experience ##
The default blob container name shows.

## Actual Experience ##
No default blob container name shows.

## Additional Context ##
The error message keeps showing after typing one valid blob container name.

|
non_process
|
no default blob container name shows in the clone dialog for one local emulator blob container storage explorer version dev build number branch main platform os windows emulator version azurite microsoft azure storage emulator architecture regression from previous release steps to reproduce install and run azurite launch storage explorer expand local attached storage accounts emulator default ports key right click the blob containers node create a new blob container right click the blob container click clone check whether the default blob container name shows in the dialog expected experience the default blob container name shows actual experience no default blob container name shows additional context the error message keeps showing after typing one valid blob container name
| 0
|
19,721
| 26,073,828,975
|
IssuesEvent
|
2022-12-24 07:06:26
|
pyanodon/pybugreports
|
https://api.github.com/repos/pyanodon/pybugreports
|
closed
|
Mod "Packing Tape" incompatible
|
bug mod:pypostprocessing crash compatibility
|
### Mod source
Factorio Mod Portal
### Which mod are you having an issue with?
- [ ] pyalienlife
- [ ] pyalternativeenergy
- [ ] pycoalprocessing
- [ ] pyfusionenergy
- [ ] pyhightech
- [ ] pyindustry
- [ ] pypetroleumhandling
- [X] pypostprocessing
- [ ] pyrawores
### Operating system
>=Windows 10
### What kind of issue is this?
- [X] Compatibility
- [ ] Locale (names, descriptions, unknown keys)
- [ ] Graphical
- [ ] Crash
- [ ] Progression
- [ ] Balance
- [ ] Pypostprocessing failure
- [ ] Other
### What is the problem?
https://mods.factorio.com/mod/packing-tape
Game will not load with the above mod enabled, gives this error;

When mod is removed it loads fine.
I did also try loading the game without any Py mods, but with all my other QOL mods enabled, to confirm it is actually the Py mods causing this.
### Steps to reproduce
_No response_
### Additional context
_No response_
### Log file
_No response_
|
1.0
|
Mod "Packing Tape" incompatible - ### Mod source
Factorio Mod Portal
### Which mod are you having an issue with?
- [ ] pyalienlife
- [ ] pyalternativeenergy
- [ ] pycoalprocessing
- [ ] pyfusionenergy
- [ ] pyhightech
- [ ] pyindustry
- [ ] pypetroleumhandling
- [X] pypostprocessing
- [ ] pyrawores
### Operating system
>=Windows 10
### What kind of issue is this?
- [X] Compatibility
- [ ] Locale (names, descriptions, unknown keys)
- [ ] Graphical
- [ ] Crash
- [ ] Progression
- [ ] Balance
- [ ] Pypostprocessing failure
- [ ] Other
### What is the problem?
https://mods.factorio.com/mod/packing-tape
Game will not load with the above mod enabled, gives this error;

When mod is removed it loads fine.
I did also try loading the game without any Py mods, but with all my other QOL mods enabled, to confirm it is actually the Py mods causing this.
### Steps to reproduce
_No response_
### Additional context
_No response_
### Log file
_No response_
|
process
|
mod packing tape incompatible mod source factorio mod portal which mod are you having an issue with pyalienlife pyalternativeenergy pycoalprocessing pyfusionenergy pyhightech pyindustry pypetroleumhandling pypostprocessing pyrawores operating system windows what kind of issue is this compatibility locale names descriptions unknown keys graphical crash progression balance pypostprocessing failure other what is the problem game will not load with the above mod enabled gives this error when mod is removed it loads fine i did also try loading the game without any py mods but with all my other qol mods enabled to confirm it is actually the py mods causing this steps to reproduce no response additional context no response log file no response
| 1
|
6,856
| 9,993,166,156
|
IssuesEvent
|
2019-07-11 14:50:34
|
eaudeweb/ozone
|
https://api.github.com/repos/eaudeweb/ozone
|
closed
|
Error when importing process agents
|
Component: Backend Feature: Data migration Feature: Process Agents Priority: Highest
|
KeyError: 'UNK'
File "/var/local/ozone/ozone/core/management/commands/import_procagents.py", line 340, in get_or_create_decision
meeting = self.meetings[meeting_id]
caused by ProcAgentEmitLimits.Decision = UNK
|
1.0
|
Error when importing process agents - KeyError: 'UNK'
File "/var/local/ozone/ozone/core/management/commands/import_procagents.py", line 340, in get_or_create_decision
meeting = self.meetings[meeting_id]
caused by ProcAgentEmitLimits.Decision = UNK
|
process
|
error when importing process agents keyerror unk file var local ozone ozone core management commands import procagents py line in get or create decision meeting self meetings caused by procagentemitlimits decision unk
| 1
|
1,417
| 3,984,317,443
|
IssuesEvent
|
2016-05-07 03:36:26
|
opattison/olivermakes
|
https://api.github.com/repos/opattison/olivermakes
|
closed
|
Testing: avoid links that only exist on master (html-proofer)
|
bug content maintenance process
|
See: https://travis-ci.org/opattison/olivermakes/builds/127293069
The link is only live when the post is published on the master branch, leading to errors in Travis CI that aren’t actually errors.
Either:
- `--url-ignore www.github.com`
- `<a href="http://notareallink" data-proofer-ignore>Not checked.</a>`
|
1.0
|
Testing: avoid links that only exist on master (html-proofer) - See: https://travis-ci.org/opattison/olivermakes/builds/127293069
The link is only live when the post is published on the master branch, leading to errors in Travis CI that aren’t actually errors.
Either:
- `--url-ignore www.github.com`
- `<a href="http://notareallink" data-proofer-ignore>Not checked.</a>`
|
process
|
testing avoid links that only exist on master html proofer see the link is only live when the post is published on the master branch leading to errors in travis ci that aren’t actually errors either url ignore not checked
| 1
|
193,192
| 22,216,084,211
|
IssuesEvent
|
2022-06-08 01:54:15
|
maddyCode23/linux-4.1.15
|
https://api.github.com/repos/maddyCode23/linux-4.1.15
|
reopened
|
CVE-2019-15216 (Medium) detected in linux-stable-rtv4.1.33
|
security vulnerability
|
## CVE-2019-15216 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/yurex.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/yurex.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.0.14. There is a NULL pointer dereference caused by a malicious USB device in the drivers/usb/misc/yurex.c driver.
<p>Publish Date: 2019-08-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-15216>CVE-2019-15216</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Physical
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15216">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15216</a></p>
<p>Release Date: 2019-09-03</p>
<p>Fix Resolution: v5.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-15216 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2019-15216 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/yurex.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/yurex.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.0.14. There is a NULL pointer dereference caused by a malicious USB device in the drivers/usb/misc/yurex.c driver.
<p>Publish Date: 2019-08-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-15216>CVE-2019-15216</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Physical
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15216">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15216</a></p>
<p>Release Date: 2019-09-03</p>
<p>Fix Resolution: v5.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers usb misc yurex c drivers usb misc yurex c vulnerability details an issue was discovered in the linux kernel before there is a null pointer dereference caused by a malicious usb device in the drivers usb misc yurex c driver publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
18,615
| 24,579,367,699
|
IssuesEvent
|
2022-10-13 14:34:16
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[Mobile apps] [FHIR] Enrollment flow is not working in the android mobile apps
|
Bug Blocker P0 iOS Android Process: Fixed Process: Tested QA Process: Tested dev
|
**AR:** Enrollment flow is not working in the android mobile apps
**ER:** Participants should be able to enroll in the android mobile apps

|
3.0
|
[Mobile apps] [FHIR] Enrollment flow is not working in the android mobile apps - **AR:** Enrollment flow is not working in the android mobile apps
**ER:** Participants should be able to enroll in the android mobile apps

|
process
|
enrollment flow is not working in the android mobile apps ar enrollment flow is not working in the android mobile apps er participants should be able to enroll in the android mobile apps
| 1
|
10,522
| 13,305,042,977
|
IssuesEvent
|
2020-08-25 17:54:04
|
googleapis/python-storage
|
https://api.github.com/repos/googleapis/python-storage
|
closed
|
'test_get_signed_policy_v4' flakes with 500
|
api: storage flaky testing type: process
|
From [this Kokoro build](https://source.cloud.google.com/results/invocations/0499a0e3-d3c1-444e-a1a4-e74e80e39dcf/targets/cloud-devrel%2Fclient-libraries%2Fpython%2Fgoogleapis%2Fpython-storage%2Fpresubmit%2Fpresubmit/log):
```python
_________________ TestV4POSTPolicies.test_get_signed_policy_v4 _________________
self = <test_system.TestV4POSTPolicies testMethod=test_get_signed_policy_v4>
def test_get_signed_policy_v4(self):
bucket_name = "post_policy" + unique_resource_id("-")
self.assertRaises(exceptions.NotFound, Config.CLIENT.get_bucket, bucket_name)
retry_429_503(Config.CLIENT.create_bucket)(bucket_name)
self.case_buckets_to_delete.append(bucket_name)
blob_name = "post_policy_obj.txt"
with open(blob_name, "w") as f:
f.write("DEADBEEF")
policy = Config.CLIENT.generate_signed_post_policy_v4(
bucket_name,
blob_name,
conditions=[
{"bucket": bucket_name},
["starts-with", "$Content-Type", "text/pla"],
],
expiration=datetime.datetime.now() + datetime.timedelta(hours=1),
fields={"content-type": "text/plain"},
)
with open(blob_name, "r") as f:
files = {"file": (blob_name, f)}
response = requests.post(policy["url"], data=policy["fields"], files=files)
os.remove(blob_name)
> self.assertEqual(response.status_code, 204)
E AssertionError: 500 != 204
tests/system/test_system.py:2347: AssertionError
```
|
1.0
|
'test_get_signed_policy_v4' flakes with 500 - From [this Kokoro build](https://source.cloud.google.com/results/invocations/0499a0e3-d3c1-444e-a1a4-e74e80e39dcf/targets/cloud-devrel%2Fclient-libraries%2Fpython%2Fgoogleapis%2Fpython-storage%2Fpresubmit%2Fpresubmit/log):
```python
_________________ TestV4POSTPolicies.test_get_signed_policy_v4 _________________
self = <test_system.TestV4POSTPolicies testMethod=test_get_signed_policy_v4>
def test_get_signed_policy_v4(self):
bucket_name = "post_policy" + unique_resource_id("-")
self.assertRaises(exceptions.NotFound, Config.CLIENT.get_bucket, bucket_name)
retry_429_503(Config.CLIENT.create_bucket)(bucket_name)
self.case_buckets_to_delete.append(bucket_name)
blob_name = "post_policy_obj.txt"
with open(blob_name, "w") as f:
f.write("DEADBEEF")
policy = Config.CLIENT.generate_signed_post_policy_v4(
bucket_name,
blob_name,
conditions=[
{"bucket": bucket_name},
["starts-with", "$Content-Type", "text/pla"],
],
expiration=datetime.datetime.now() + datetime.timedelta(hours=1),
fields={"content-type": "text/plain"},
)
with open(blob_name, "r") as f:
files = {"file": (blob_name, f)}
response = requests.post(policy["url"], data=policy["fields"], files=files)
os.remove(blob_name)
> self.assertEqual(response.status_code, 204)
E AssertionError: 500 != 204
tests/system/test_system.py:2347: AssertionError
```
|
process
|
test get signed policy flakes with from python test get signed policy self def test get signed policy self bucket name post policy unique resource id self assertraises exceptions notfound config client get bucket bucket name retry config client create bucket bucket name self case buckets to delete append bucket name blob name post policy obj txt with open blob name w as f f write deadbeef policy config client generate signed post policy bucket name blob name conditions bucket bucket name expiration datetime datetime now datetime timedelta hours fields content type text plain with open blob name r as f files file blob name f response requests post policy data policy files files os remove blob name self assertequal response status code e assertionerror tests system test system py assertionerror
| 1
|
21,242
| 28,366,913,256
|
IssuesEvent
|
2023-04-12 14:25:54
|
Deltares/Ribasim
|
https://api.github.com/repos/Deltares/Ribasim
|
opened
|
allow salt injection as boundary condition
|
physical process
|
include salt injection as boundary condition to represent salt influx from the sea
|
1.0
|
allow salt injection as boundary condition - include salt injection as boundary condition to represent salt influx from the sea
|
process
|
allow salt injection as boundary condition include salt injection as boundary condition to represent salt influx from the sea
| 1
|
156,523
| 12,313,199,921
|
IssuesEvent
|
2020-05-12 14:59:12
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: version/mixed/nodes=3 failed
|
C-test-failure O-roachtest O-robot branch-release-20.1 release-blocker
|
[(roachtest).version/mixed/nodes=3 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1928509&tab=buildLog) on [release-20.1@1affcc3ac8819e133d54668a1ce0fc4a4ce99b4f](https://github.com/cockroachdb/cockroach/commits/1affcc3ac8819e133d54668a1ce0fc4a4ce99b4f):
```
| 579.0s 0 2559.0 2587.3 2.6 7.3 12.6 19.9 write
| 580.0s 0 2377.9 2587.0 2.5 8.9 15.2 27.3 write
| _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms)
| 581.0s 0 2268.0 2586.4 3.0 7.6 12.1 25.2 write
| 582.0s 0 2456.1 2586.2 2.6 7.6 11.0 46.1 write
| 583.0s 0 2069.9 2585.3 2.8 10.5 17.8 33.6 write
| 584.0s 0 2459.0 2585.1 2.6 7.9 11.5 24.1 write
| 585.0s 0 2622.9 2585.2 2.5 7.1 11.0 35.7 write
| 586.0s 0 2463.2 2585.0 2.8 7.3 11.0 19.9 write
| 587.0s 0 2467.9 2584.8 2.8 7.3 11.0 21.0 write
| 588.0s 0 2389.1 2584.4 2.8 7.6 11.0 14.7 write
| 589.0s 0 2412.0 2584.1 2.8 7.6 11.5 18.9 write
| 590.0s 0 2345.9 2583.7 2.8 7.6 12.1 218.1 write
| 591.0s 0 2271.0 2583.2 2.9 7.6 12.6 33.6 write
| 592.0s 0 2577.1 2583.2 2.6 7.3 11.0 24.1 write
| 593.0s 0 2585.0 2583.2 2.6 6.8 10.0 24.1 write
| 594.0s 0 2338.0 2582.8 2.9 7.9 12.1 23.1 write
| 595.0s 0 2429.0 2582.5 2.6 8.1 11.0 30.4 write
| 596.0s 0 2122.0 2581.7 3.1 8.1 12.1 25.2 write
| 597.0s 0 2526.9 2581.7 2.6 7.6 10.5 21.0 write
| 598.0s 0 1976.1 2580.6 2.9 11.5 18.9 25.2 write
| 599.0s 0 2153.0 2579.9 3.1 7.9 12.1 18.9 write
Wraps: (9) exit status 30
Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *withstack.withStack (4) *errutil.withMessage (5) *withstack.withStack (6) *safedetails.withSafeDetails (7) *errutil.withMessage (8) *main.withCommandDetails (9) *exec.ExitError
cluster.go:1481,context.go:135,cluster.go:1470,test_runner.go:825: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-1928509-1588924553-09-n4cpu4 --oneshot --ignore-empty-nodes: exit status 1 4: skipped
3: 7750
2: 7786
1: dead
Error: UNCLASSIFIED_PROBLEM: 1: dead
(1) UNCLASSIFIED_PROBLEM
Wraps: (2) attached stack trace
| main.glob..func13
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1129
| main.wrap.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:272
| github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).execute
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:766
| github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).ExecuteC
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:852
| github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).Execute
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:800
| main.main
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1799
| runtime.main
| /usr/local/go/src/runtime/proc.go:203
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1357
Wraps: (3) 1: dead
Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *errors.errorString
```
<details><summary>More</summary><p>
Artifacts: [/version/mixed/nodes=3](https://teamcity.cockroachdb.com/viewLog.html?buildId=1928509&tab=artifacts#/version/mixed/nodes=3)
Related:
- #48574 roachtest: version/mixed/nodes=3 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #47741 roachtest: version/mixed/nodes=3 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Aversion%2Fmixed%2Fnodes%3D3.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
2.0
|
roachtest: version/mixed/nodes=3 failed - [(roachtest).version/mixed/nodes=3 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1928509&tab=buildLog) on [release-20.1@1affcc3ac8819e133d54668a1ce0fc4a4ce99b4f](https://github.com/cockroachdb/cockroach/commits/1affcc3ac8819e133d54668a1ce0fc4a4ce99b4f):
```
| 579.0s 0 2559.0 2587.3 2.6 7.3 12.6 19.9 write
| 580.0s 0 2377.9 2587.0 2.5 8.9 15.2 27.3 write
| _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms)
| 581.0s 0 2268.0 2586.4 3.0 7.6 12.1 25.2 write
| 582.0s 0 2456.1 2586.2 2.6 7.6 11.0 46.1 write
| 583.0s 0 2069.9 2585.3 2.8 10.5 17.8 33.6 write
| 584.0s 0 2459.0 2585.1 2.6 7.9 11.5 24.1 write
| 585.0s 0 2622.9 2585.2 2.5 7.1 11.0 35.7 write
| 586.0s 0 2463.2 2585.0 2.8 7.3 11.0 19.9 write
| 587.0s 0 2467.9 2584.8 2.8 7.3 11.0 21.0 write
| 588.0s 0 2389.1 2584.4 2.8 7.6 11.0 14.7 write
| 589.0s 0 2412.0 2584.1 2.8 7.6 11.5 18.9 write
| 590.0s 0 2345.9 2583.7 2.8 7.6 12.1 218.1 write
| 591.0s 0 2271.0 2583.2 2.9 7.6 12.6 33.6 write
| 592.0s 0 2577.1 2583.2 2.6 7.3 11.0 24.1 write
| 593.0s 0 2585.0 2583.2 2.6 6.8 10.0 24.1 write
| 594.0s 0 2338.0 2582.8 2.9 7.9 12.1 23.1 write
| 595.0s 0 2429.0 2582.5 2.6 8.1 11.0 30.4 write
| 596.0s 0 2122.0 2581.7 3.1 8.1 12.1 25.2 write
| 597.0s 0 2526.9 2581.7 2.6 7.6 10.5 21.0 write
| 598.0s 0 1976.1 2580.6 2.9 11.5 18.9 25.2 write
| 599.0s 0 2153.0 2579.9 3.1 7.9 12.1 18.9 write
Wraps: (9) exit status 30
Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *withstack.withStack (4) *errutil.withMessage (5) *withstack.withStack (6) *safedetails.withSafeDetails (7) *errutil.withMessage (8) *main.withCommandDetails (9) *exec.ExitError
cluster.go:1481,context.go:135,cluster.go:1470,test_runner.go:825: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-1928509-1588924553-09-n4cpu4 --oneshot --ignore-empty-nodes: exit status 1 4: skipped
3: 7750
2: 7786
1: dead
Error: UNCLASSIFIED_PROBLEM: 1: dead
(1) UNCLASSIFIED_PROBLEM
Wraps: (2) attached stack trace
| main.glob..func13
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1129
| main.wrap.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:272
| github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).execute
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:766
| github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).ExecuteC
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:852
| github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).Execute
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:800
| main.main
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1799
| runtime.main
| /usr/local/go/src/runtime/proc.go:203
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1357
Wraps: (3) 1: dead
Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *errors.errorString
```
<details><summary>More</summary><p>
Artifacts: [/version/mixed/nodes=3](https://teamcity.cockroachdb.com/viewLog.html?buildId=1928509&tab=artifacts#/version/mixed/nodes=3)
Related:
- #48574 roachtest: version/mixed/nodes=3 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #47741 roachtest: version/mixed/nodes=3 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Aversion%2Fmixed%2Fnodes%3D3.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
non_process
|
roachtest version mixed nodes failed on write write elapsed errors ops sec inst ops sec cum ms ms ms pmax ms write write write write write write write write write write write write write write write write write write write wraps exit status error types withstack withstack errutil withmessage withstack withstack errutil withmessage withstack withstack safedetails withsafedetails errutil withmessage main withcommanddetails exec exiterror cluster go context go cluster go test runner go dead node detection home agent work go src github com cockroachdb cockroach bin roachprod monitor teamcity oneshot ignore empty nodes exit status skipped dead error unclassified problem dead unclassified problem wraps attached stack trace main glob home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go main wrap home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go github com cockroachdb cockroach vendor github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cockroachdb cockroach vendor github com cobra command executec home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cockroachdb cockroach vendor github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go main main home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps dead error types errors unclassified withstack withstack errors errorstring more artifacts related roachtest version mixed nodes failed roachtest version mixed nodes failed powered by
| 0
|
33,405
| 15,896,458,729
|
IssuesEvent
|
2021-04-11 17:30:10
|
sp614x/optifine
|
https://api.github.com/repos/sp614x/optifine
|
closed
|
FPS decreasing over time with shaders/in F5/in inventory
|
1.16 details needed performance shaders
|
Without shaders I get 100fps easily, but when I put them on it goes from 70fps to 20fps in ~10 minutes. It used to work well.
Specs :
- Ryzen 7 2700
- 2x8gb ram 3200mhz
- RTX 2070
Temps are very good, I allocated 6-8gb of ram for mc. I use Forge 1.16.1 and the last optifine update
|
True
|
FPS decreasing over time with shaders/in F5/in inventory - Without shaders I get 100fps easily, but when I put them on it goes from 70fps to 20fps in ~10 minutes. It used to work well.
Specs :
- Ryzen 7 2700
- 2x8gb ram 3200mhz
- RTX 2070
Temps are very good, I allocated 6-8gb of ram for mc. I use Forge 1.16.1 and the last optifine update
|
non_process
|
fps decreasing over time with shaders in in inventory without shaders i get easily but when i put them on it goes from to in minutes it used to work well specs ryzen ram rtx temps are very good i allocated of ram for mc i use forge and the last optifine update
| 0
|
105,996
| 16,663,950,074
|
IssuesEvent
|
2021-06-06 20:50:28
|
AlexRogalskiy/github-action-charts
|
https://api.github.com/repos/AlexRogalskiy/github-action-charts
|
opened
|
CVE-2020-28469 (Medium) detected in glob-parent-5.1.1.tgz
|
security vulnerability
|
## CVE-2020-28469 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-5.1.1.tgz</b></p></summary>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz</a></p>
<p>Path to dependency file: github-action-charts/package.json</p>
<p>Path to vulnerable library: github-action-charts/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- eslint-7.19.0.tgz (Root Library)
- :x: **glob-parent-5.1.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-charts/commit/c839647436e6519774bf3d4a639525381eedfb20">c839647436e6519774bf3d4a639525381eedfb20</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator.
<p>Publish Date: 2021-06-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469>CVE-2020-28469</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469</a></p>
<p>Release Date: 2021-06-03</p>
<p>Fix Resolution: glob-parent - 5.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-28469 (Medium) detected in glob-parent-5.1.1.tgz - ## CVE-2020-28469 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-5.1.1.tgz</b></p></summary>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz</a></p>
<p>Path to dependency file: github-action-charts/package.json</p>
<p>Path to vulnerable library: github-action-charts/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- eslint-7.19.0.tgz (Root Library)
- :x: **glob-parent-5.1.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-charts/commit/c839647436e6519774bf3d4a639525381eedfb20">c839647436e6519774bf3d4a639525381eedfb20</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator.
<p>Publish Date: 2021-06-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469>CVE-2020-28469</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469</a></p>
<p>Release Date: 2021-06-03</p>
<p>Fix Resolution: glob-parent - 5.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in glob parent tgz cve medium severity vulnerability vulnerable library glob parent tgz extract the non magic parent path from a glob string library home page a href path to dependency file github action charts package json path to vulnerable library github action charts node modules glob parent package json dependency hierarchy eslint tgz root library x glob parent tgz vulnerable library found in head commit a href vulnerability details this affects the package glob parent before the enclosure regex used to check for strings ending in enclosure containing path separator publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent step up your open source security game with whitesource
| 0
|
109,112
| 9,368,522,340
|
IssuesEvent
|
2019-04-03 08:54:15
|
scality/metalk8s
|
https://api.github.com/repos/scality/metalk8s
|
closed
|
Conflict between control plane network IPs when running tests in CI single node env
|
bug ci tests
|
In the CI we run the tests, and especially we retry the bootstrap script to see if it's idempotent.
In the CI, in the single node env, we have two IPs on the worker:
1. The one from Openstack: 10.100.A.B
1. The one from the pod network we set: 10.233.C.D
In the CI, we define the control plane network as 10.0.0.0/8 (hence including both previous IPs).
Then when retrying the bootstrap script, we pick one IP on this network, and that could be either the one from Openstack, or the 10.233 one. And if it's the second one, then it means we bind all our services on it, and nothing works anymore because we still are targeting the 10.100 one to address the API server.
Let's be stricter and consider 10.100.0.0/16 as our control plane network (and also workload plane network) in the CI single node env.
|
1.0
|
Conflict between control plane network IPs when running tests in CI single node env - In the CI we run the tests, and especially we retry the bootstrap script to see if it's idempotent.
In the CI, in the single node env, we have two IPs on the worker:
1. The one from Openstack: 10.100.A.B
1. The one from the pod network we set: 10.233.C.D
In the CI, we define the control plane network as 10.0.0.0/8 (hence including both previous IPs).
Then when retrying the bootstrap script, we pick one IP on this network, and that could be either the one from Openstack, or the 10.233 one. And if it's the second one, then it means we bind all our services on it, and nothing works anymore because we still are targeting the 10.100 one to address the API server.
Let's be stricter and consider 10.100.0.0/16 as our control plane network (and also workload plane network) in the CI single node env.
|
non_process
|
conflict between control plane network ips when running tests in ci single node env in the ci we run the tests and especially we retry the bootstrap script to see if it s idempotent in the ci in the single node env we have two ips on the worker the one from openstack a b the one from the pod network we set c d in the ci we define the control plane network as hence including both previous ips then when retrying the bootstrap script we pick one ip on this network and that could be either the one from openstack or the one and if it s the second one then it means we bind all our services on it and nothing works anymore because we still are targeting the one to address the api server let s be stricter and consider as our control plane network and also workload plane network in the ci single node env
| 0
|
816
| 3,290,958,174
|
IssuesEvent
|
2015-10-30 04:35:29
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
Make strict processing mode fail on error messages
|
enhancement P2 preprocess
|
Improve processing mode coverage to treat error messages as fatal errors.
|
1.0
|
Make strict processing mode fail on error messages - Improve processing mode coverage to treat error messages as fatal errors.
|
process
|
make strict processing mode fail on error messages improve processing mode coverage to treat error messages as fatal errors
| 1
|
281,655
| 8,697,938,816
|
IssuesEvent
|
2018-12-04 21:45:09
|
mplusmuseum/mplusmuseum-collections-explorer
|
https://api.github.com/repos/mplusmuseum/mplusmuseum-collections-explorer
|
closed
|
Inline stylized category text
|
priority-low question task
|
Add inline stylized text for categories.
Please see design.

I'm marking this as a low-priority for now until we get everything else done. The simple case of this would be simple. But I worry that the edge cases of long text might turn this task into a more difficult one.
**Question**: How do we handle long text that wraps?
|
1.0
|
Inline stylized category text - Add inline stylized text for categories.
Please see design.

I'm marking this as a low-priority for now until we get everything else done. The simple case of this would be simple. But I worry that the edge cases of long text might turn this task into a more difficult one.
**Question**: How do we handle long text that wraps?
|
non_process
|
inline stylized category text add inline stylized text for categories please see design i m marking this as a low priority for now until we get everything else done the simple case of this would be simple but i worry that the edge cases of long text might turn this task into a more difficult one question how do we handle long text that wraps
| 0
|
17,001
| 22,364,193,203
|
IssuesEvent
|
2022-06-16 00:59:22
|
hashgraph/hedera-json-rpc-relay
|
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
|
opened
|
Add additional eth_getBlockByNumber acceptance tests
|
enhancement P2 process
|
### Problem
The initial eth_getBlockByNumber coverage was scarce
### Solution
Add move coverage
- [ ] Enhance the already existing `getBlockByNumber` test by adding assertion of the returned object that it has `transactions` array **containing the full transaction** and not only the string of their hash (testing the "hydrated transactions flag")
- [ ] Add a test that queries **existing block** by **number** but sets `hydrated transactions to **false**. Assert that the response contains `transactions` array but the values are strings of the **transaction hash**
- [ ] Add negative test for non-existing block number with `hydrated transactions set to **true**`. The response of the call must be asserted to be:
```
{
"jsonrpc": "2.0",
"id": 0,
"result": null
}
```
- [ ] Add negative test for non-existing block number with `hydrated transactions set to **false**`. The response of the call must be asserted to be:
```
{
"jsonrpc": "2.0",
"id": 0,
"result": null
}
```
### Alternatives
_No response_
|
1.0
|
Add additional eth_getBlockByNumber acceptance tests - ### Problem
The initial eth_getBlockByNumber coverage was scarce
### Solution
Add move coverage
- [ ] Enhance the already existing `getBlockByNumber` test by adding assertion of the returned object that it has `transactions` array **containing the full transaction** and not only the string of their hash (testing the "hydrated transactions flag")
- [ ] Add a test that queries **existing block** by **number** but sets `hydrated transactions to **false**. Assert that the response contains `transactions` array but the values are strings of the **transaction hash**
- [ ] Add negative test for non-existing block number with `hydrated transactions set to **true**`. The response of the call must be asserted to be:
```
{
"jsonrpc": "2.0",
"id": 0,
"result": null
}
```
- [ ] Add negative test for non-existing block number with `hydrated transactions set to **false**`. The response of the call must be asserted to be:
```
{
"jsonrpc": "2.0",
"id": 0,
"result": null
}
```
### Alternatives
_No response_
|
process
|
add additional eth getblockbynumber acceptance tests problem the initial eth getblockbynumber coverage was scarce solution add move coverage enhance the already existing getblockbynumber test by adding assertion of the returned object that it has transactions array containing the full transaction and not only the string of their hash testing the hydrated transactions flag add a test that queries existing block by number but sets hydrated transactions to false assert that the response contains transactions array but the values are strings of the transaction hash add negative test for non existing block number with hydrated transactions set to true the response of the call must be asserted to be jsonrpc id result null add negative test for non existing block number with hydrated transactions set to false the response of the call must be asserted to be jsonrpc id result null alternatives no response
| 1
|
16,584
| 21,630,841,257
|
IssuesEvent
|
2022-05-05 09:32:26
|
googleapis/google-cloud-dotnet
|
https://api.github.com/repos/googleapis/google-cloud-dotnet
|
closed
|
Docs release fails for IAM (alpha05)
|
type: process
|
Filing as process as it isn't a bug that will affect customers, beyond our ability to release.
The release build for #8423 is failing in the docs build with:
> [22-04-29 01:57:02.625]Warning:[ExtractMetadata](T:/src/github/google-cloud-dotnet/releasebuild/apis/Google.Cloud.Iam.V1/Google.Cloud.Iam.V1/Google.Cloud.Iam.V1.csproj)Workspace failed with: [Failure] Msbuild failed when processing the file 'T:\src\github\google-cloud-dotnet\releasebuild\apis\Google.Cloud.Iam.V1\Google.Cloud.Iam.V1\Google.Cloud.Iam.V1.csproj' with message: The SDK 'Microsoft.NET.Sdk' specified could not be found. T:\src\github\google-cloud-dotnet\releasebuild\apis\Google.Cloud.Iam.V1\Google.Cloud.Iam.V1\Google.Cloud.Iam.V1.csproj
This is odd given that alpha02 released with no problems.
|
1.0
|
Docs release fails for IAM (alpha05) - Filing as process as it isn't a bug that will affect customers, beyond our ability to release.
The release build for #8423 is failing in the docs build with:
> [22-04-29 01:57:02.625]Warning:[ExtractMetadata](T:/src/github/google-cloud-dotnet/releasebuild/apis/Google.Cloud.Iam.V1/Google.Cloud.Iam.V1/Google.Cloud.Iam.V1.csproj)Workspace failed with: [Failure] Msbuild failed when processing the file 'T:\src\github\google-cloud-dotnet\releasebuild\apis\Google.Cloud.Iam.V1\Google.Cloud.Iam.V1\Google.Cloud.Iam.V1.csproj' with message: The SDK 'Microsoft.NET.Sdk' specified could not be found. T:\src\github\google-cloud-dotnet\releasebuild\apis\Google.Cloud.Iam.V1\Google.Cloud.Iam.V1\Google.Cloud.Iam.V1.csproj
This is odd given that alpha02 released with no problems.
|
process
|
docs release fails for iam filing as process as it isn t a bug that will affect customers beyond our ability to release the release build for is failing in the docs build with warning t src github google cloud dotnet releasebuild apis google cloud iam google cloud iam google cloud iam csproj workspace failed with msbuild failed when processing the file t src github google cloud dotnet releasebuild apis google cloud iam google cloud iam google cloud iam csproj with message the sdk microsoft net sdk specified could not be found t src github google cloud dotnet releasebuild apis google cloud iam google cloud iam google cloud iam csproj this is odd given that released with no problems
| 1
|
254,067
| 8,069,590,243
|
IssuesEvent
|
2018-08-06 06:38:42
|
containous/traefik
|
https://api.github.com/repos/containous/traefik
|
closed
|
[acme.dnsChallenge] not working with gandiv5: No or incorrect TXT record
|
area/acme kind/bug/confirmed priority/P1
|
<!--
DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS.
The issue tracker is for reporting bugs and feature requests only.
For end-user related support questions, please refer to one of the following:
- Stack Overflow (using the "traefik" tag): https://stackoverflow.com/questions/tagged/traefik
- the Traefik community Slack channel: https://traefik.herokuapp.com
-->
### Do you want to request a *feature* or report a *bug*?
Bug
### What did you do?
<!--
HOW TO WRITE A GOOD BUG REPORT?
- Respect the issue template as much as possible.
- If possible, use the command `traefik bug`. See https://www.youtube.com/watch?v=Lyz62L8m93I.
- The title should be short and descriptive.
- Explain the conditions which led you to report this issue: the context.
- The context should lead to something, an idea or a problem that you’re facing.
- Remain clear and concise.
- Format your messages to help the reader focus on what matters and understand the structure of your message, use Markdown syntax https://help.github.com/articles/github-flavored-markdown
-->
After a change of config I noticed the client no longer gets an accepted certificate. Switching back to the previously working config didn't solve the problem, neither did downgrading to 1.6.2 or 1.6.1..
### What did you expect to see?
A working letsencrypt certificate
### What did you see instead?
No certificate is generated, but an error appears in the log (see below)
### Output of `traefik version`: (_What version of Traefik are you using?_)
<!--
For the Traefik Docker image:
docker run [IMAGE] version
ex: docker run traefik version
For the alpine Traefik Docker image:
docker run [IMAGE] traefik version
ex: docker run traefik traefik version
-->
```
Version: v1.6.3
Codename: tetedemoine
Go version: go1.10.2
Built: 2018-06-05_03:29:01PM
OS/Arch: linux/amd64
```
```
Same results with 1.6.1, 1.6.2 and 1.6.3
```
### What is your environment & configuration (arguments, toml, provider, platform, ...)?
```toml
[acme]
...
caServer = "https://acme-v02.api.letsencrypt.org/directory"
[acme.dnsChallenge]
provider = "gandiv5"
delayBeforeCheck = 0
...
[[acme.domains]]
main = "*.example.org"
sans = ["example.org"]
[[acme.domains]]
main = "*.example.guru"
sans = ["example.guru"]
[[acme.domains]]
main = "service.example.com"
```
<!--
Add more configuration information here.
-->
The results are the same using:
```toml
caServer = "https://acme-staging-v02.api.letsencrypt.org/directory"
```
### If applicable, please paste the log output in DEBUG level (`--logLevel=DEBUG` switch)
```
6/6/2018 11:31:34 AMGandi DNS: Zone record deleted
6/6/2018 11:31:34 AMERRO[2018-06-06T09:31:34Z] Unable to obtain ACME certificate for domains "service.example.com" : cannot obtain certificates: acme: Error -> One or more domains had a problem:
6/6/2018 11:31:34 AM[service.example.com] acme: Error 400 - urn:ietf:params:acme:error:dns - DNS problem: NXDOMAIN looking up TXT for _acme-challenge.service.example.com
6/6/2018 11:31:34 AM
6/6/2018 11:32:01 AMGandi DNS: Zone record deleted
6/6/2018 11:32:01 AMERRO[2018-06-06T09:32:01Z] Unable to obtain ACME certificate for domains "*.example.guru,example.guru" : cannot obtain certificates: acme: Error -> One or more domains had a problem:
6/6/2018 11:32:01 AM[example.guru] acme: Error 403 - urn:ietf:params:acme:error:unauthorized - No TXT record found at _acme-challenge.example.guru
6/6/2018 11:32:01 AM
6/6/2018 11:32:35 AMGandi DNS: Zone record deleted
6/6/2018 11:32:36 AMGandi DNS: DNS Record Created
6/6/2018 11:33:21 AMGandi DNS: Zone record deleted
6/6/2018 11:33:21 AMERRO[2018-06-06T09:33:21Z] Unable to obtain ACME certificate for domains "*.example.org,example.org" : cannot obtain certificates: acme: Error -> One or more domains had a problem:
6/6/2018 11:33:21 AM[example.org] acme: Error 403 - urn:ietf:params:acme:error:unauthorized - Incorrect TXT record "nYhYDZD1ENYNTDy4_eQ5bP7vhhTVY5xWKrv_8pF3r2U" found at _acme-challenge.example.org
```
|
1.0
|
[acme.dnsChallenge] not working with gandiv5: No or incorrect TXT record - <!--
DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS.
The issue tracker is for reporting bugs and feature requests only.
For end-user related support questions, please refer to one of the following:
- Stack Overflow (using the "traefik" tag): https://stackoverflow.com/questions/tagged/traefik
- the Traefik community Slack channel: https://traefik.herokuapp.com
-->
### Do you want to request a *feature* or report a *bug*?
Bug
### What did you do?
<!--
HOW TO WRITE A GOOD BUG REPORT?
- Respect the issue template as much as possible.
- If possible, use the command `traefik bug`. See https://www.youtube.com/watch?v=Lyz62L8m93I.
- The title should be short and descriptive.
- Explain the conditions which led you to report this issue: the context.
- The context should lead to something, an idea or a problem that you’re facing.
- Remain clear and concise.
- Format your messages to help the reader focus on what matters and understand the structure of your message, use Markdown syntax https://help.github.com/articles/github-flavored-markdown
-->
After a change of config I noticed the client no longer gets an accepted certificate. Switching back to the previously working config didn't solve the problem, neither did downgrading to 1.6.2 or 1.6.1..
### What did you expect to see?
A working letsencrypt certificate
### What did you see instead?
No certificate is generated, but an error appears in the log (see below)
### Output of `traefik version`: (_What version of Traefik are you using?_)
<!--
For the Traefik Docker image:
docker run [IMAGE] version
ex: docker run traefik version
For the alpine Traefik Docker image:
docker run [IMAGE] traefik version
ex: docker run traefik traefik version
-->
```
Version: v1.6.3
Codename: tetedemoine
Go version: go1.10.2
Built: 2018-06-05_03:29:01PM
OS/Arch: linux/amd64
```
```
Same results with 1.6.1, 1.6.2 and 1.6.3
```
### What is your environment & configuration (arguments, toml, provider, platform, ...)?
```toml
[acme]
...
caServer = "https://acme-v02.api.letsencrypt.org/directory"
[acme.dnsChallenge]
provider = "gandiv5"
delayBeforeCheck = 0
...
[[acme.domains]]
main = "*.example.org"
sans = ["example.org"]
[[acme.domains]]
main = "*.example.guru"
sans = ["example.guru"]
[[acme.domains]]
main = "service.example.com"
```
<!--
Add more configuration information here.
-->
The results are the same using:
```toml
caServer = "https://acme-staging-v02.api.letsencrypt.org/directory"
```
### If applicable, please paste the log output in DEBUG level (`--logLevel=DEBUG` switch)
```
6/6/2018 11:31:34 AMGandi DNS: Zone record deleted
6/6/2018 11:31:34 AMERRO[2018-06-06T09:31:34Z] Unable to obtain ACME certificate for domains "service.example.com" : cannot obtain certificates: acme: Error -> One or more domains had a problem:
6/6/2018 11:31:34 AM[service.example.com] acme: Error 400 - urn:ietf:params:acme:error:dns - DNS problem: NXDOMAIN looking up TXT for _acme-challenge.service.example.com
6/6/2018 11:31:34 AM
6/6/2018 11:32:01 AMGandi DNS: Zone record deleted
6/6/2018 11:32:01 AMERRO[2018-06-06T09:32:01Z] Unable to obtain ACME certificate for domains "*.example.guru,example.guru" : cannot obtain certificates: acme: Error -> One or more domains had a problem:
6/6/2018 11:32:01 AM[example.guru] acme: Error 403 - urn:ietf:params:acme:error:unauthorized - No TXT record found at _acme-challenge.example.guru
6/6/2018 11:32:01 AM
6/6/2018 11:32:35 AMGandi DNS: Zone record deleted
6/6/2018 11:32:36 AMGandi DNS: DNS Record Created
6/6/2018 11:33:21 AMGandi DNS: Zone record deleted
6/6/2018 11:33:21 AMERRO[2018-06-06T09:33:21Z] Unable to obtain ACME certificate for domains "*.example.org,example.org" : cannot obtain certificates: acme: Error -> One or more domains had a problem:
6/6/2018 11:33:21 AM[example.org] acme: Error 403 - urn:ietf:params:acme:error:unauthorized - Incorrect TXT record "nYhYDZD1ENYNTDy4_eQ5bP7vhhTVY5xWKrv_8pF3r2U" found at _acme-challenge.example.org
```
|
non_process
|
not working with no or incorrect txt record do not file issues for general support questions the issue tracker is for reporting bugs and feature requests only for end user related support questions please refer to one of the following stack overflow using the traefik tag the traefik community slack channel do you want to request a feature or report a bug bug what did you do how to write a good bug report respect the issue template as much as possible if possible use the command traefik bug see the title should be short and descriptive explain the conditions which led you to report this issue the context the context should lead to something an idea or a problem that you’re facing remain clear and concise format your messages to help the reader focus on what matters and understand the structure of your message use markdown syntax after a change of config i noticed the client no longer gets an accepted certificate switching back to the previously working config didn t solve the problem neither did downgrading to or what did you expect to see a working letsencrypt certificate what did you see instead no certificate is generated but an error appears in the log see below output of traefik version what version of traefik are you using for the traefik docker image docker run version ex docker run traefik version for the alpine traefik docker image docker run traefik version ex docker run traefik traefik version version codename tetedemoine go version built os arch linux same results with and what is your environment configuration arguments toml provider platform toml caserver provider delaybeforecheck main example org sans main example guru sans main service example com add more configuration information here the results are the same using toml caserver if applicable please paste the log output in debug level loglevel debug switch amgandi dns zone record deleted amerro unable to obtain acme certificate for domains service example com cannot obtain certificates acme error one or more domains had a problem am acme error urn ietf params acme error dns dns problem nxdomain looking up txt for acme challenge service example com am amgandi dns zone record deleted amerro unable to obtain acme certificate for domains example guru example guru cannot obtain certificates acme error one or more domains had a problem am acme error urn ietf params acme error unauthorized no txt record found at acme challenge example guru am amgandi dns zone record deleted amgandi dns dns record created amgandi dns zone record deleted amerro unable to obtain acme certificate for domains example org example org cannot obtain certificates acme error one or more domains had a problem am acme error urn ietf params acme error unauthorized incorrect txt record found at acme challenge example org
| 0
|
8,028
| 20,611,725,284
|
IssuesEvent
|
2022-03-07 09:18:48
|
dfds/backstage
|
https://api.github.com/repos/dfds/backstage
|
closed
|
Crosplane - Retrospectively add the required role for enabling assuming role in AWS provider
|
Enhancement Architecture
|
- [x] Investigate how this is currently done
- [x] Complete #480
- [x] Find and replace the referenced infrastructure-modules version in all capability sub-folders of aws-account-manifests and update to the new infrastructure-modules version
- [x] Submit PR
- [ ] Merge PR
|
1.0
|
Crosplane - Retrospectively add the required role for enabling assuming role in AWS provider - - [x] Investigate how this is currently done
- [x] Complete #480
- [x] Find and replace the referenced infrastructure-modules version in all capability sub-folders of aws-account-manifests and update to the new infrastructure-modules version
- [x] Submit PR
- [ ] Merge PR
|
non_process
|
crosplane retrospectively add the required role for enabling assuming role in aws provider investigate how this is currently done complete find and replace the referenced infrastructure modules version in all capability sub folders of aws account manifests and update to the new infrastructure modules version submit pr merge pr
| 0
|
94,072
| 15,962,327,718
|
IssuesEvent
|
2021-04-16 01:03:49
|
RG4421/terra-dev-site
|
https://api.github.com/repos/RG4421/terra-dev-site
|
opened
|
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz
|
security vulnerability
|
## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p>
<p>Path to dependency file: terra-dev-site/package.json</p>
<p>Path to vulnerable library: terra-dev-site/node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- check-installed-dependencies-1.0.1.tgz (Root Library)
- meow-3.7.0.tgz
- normalize-package-data-2.5.0.tgz
- :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/hosted-git-info/releases/tag/v3.0.8">https://github.com/npm/hosted-git-info/releases/tag/v3.0.8</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: hosted-git-info - 3.0.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"hosted-git-info","packageVersion":"2.8.8","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"check-installed-dependencies:1.0.1;meow:3.7.0;normalize-package-data:2.5.0;hosted-git-info:2.8.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"hosted-git-info - 3.0.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23362","vulnerabilityDetails":"The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz - ## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p>
<p>Path to dependency file: terra-dev-site/package.json</p>
<p>Path to vulnerable library: terra-dev-site/node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- check-installed-dependencies-1.0.1.tgz (Root Library)
- meow-3.7.0.tgz
- normalize-package-data-2.5.0.tgz
- :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/hosted-git-info/releases/tag/v3.0.8">https://github.com/npm/hosted-git-info/releases/tag/v3.0.8</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: hosted-git-info - 3.0.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"hosted-git-info","packageVersion":"2.8.8","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"check-installed-dependencies:1.0.1;meow:3.7.0;normalize-package-data:2.5.0;hosted-git-info:2.8.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"hosted-git-info - 3.0.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23362","vulnerabilityDetails":"The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve medium detected in hosted git info tgz cve medium severity vulnerability vulnerable library hosted git info tgz provides metadata and conversions from repository urls for github bitbucket and gitlab library home page a href path to dependency file terra dev site package json path to vulnerable library terra dev site node modules hosted git info package json dependency hierarchy check installed dependencies tgz root library meow tgz normalize package data tgz x hosted git info tgz vulnerable library found in base branch master vulnerability details the package hosted git info before are vulnerable to regular expression denial of service redos via regular expression shortcutmatch in the fromurl function in index js the affected regular expression exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hosted git info isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree check installed dependencies meow normalize package data hosted git info isminimumfixversionavailable true minimumfixversion hosted git info basebranches vulnerabilityidentifier cve vulnerabilitydetails the package hosted git info before are vulnerable to regular expression denial of service redos via regular expression shortcutmatch in the fromurl function in index js the affected regular expression exhibits polynomial worst case time complexity vulnerabilityurl
| 0
|
2,236
| 5,088,589,846
|
IssuesEvent
|
2016-12-31 22:51:59
|
sw4j-org/tool-jpa-processor
|
https://api.github.com/repos/sw4j-org/tool-jpa-processor
|
opened
|
Handle @ForeignKey Annotation
|
annotation processor task
|
Handle the `@ForeignKey` annotation.
See [JSR 338: Java Persistence API, Version 2.1](http://download.oracle.com/otn-pub/jcp/persistence-2_1-fr-eval-spec/JavaPersistence.pdf)
- 11.1.19 Foreign Key Annotation
|
1.0
|
Handle @ForeignKey Annotation - Handle the `@ForeignKey` annotation.
See [JSR 338: Java Persistence API, Version 2.1](http://download.oracle.com/otn-pub/jcp/persistence-2_1-fr-eval-spec/JavaPersistence.pdf)
- 11.1.19 Foreign Key Annotation
|
process
|
handle foreignkey annotation handle the foreignkey annotation see foreign key annotation
| 1
|
54,990
| 13,943,655,069
|
IssuesEvent
|
2020-10-22 23:42:37
|
mlmcd/WebGoat
|
https://api.github.com/repos/mlmcd/WebGoat
|
opened
|
CVE-2019-10747 (High) detected in set-value-2.0.0.tgz, set-value-0.4.3.tgz
|
security vulnerability
|
## CVE-2019-10747 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>set-value-2.0.0.tgz</b>, <b>set-value-0.4.3.tgz</b></p></summary>
<p>
<details><summary><b>set-value-2.0.0.tgz</b></p></summary>
<p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p>
<p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz</a></p>
<p>Path to dependency file: WebGoat/docs/package.json</p>
<p>Path to vulnerable library: WebGoat/docs/node_modules/set-value/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.26.3.tgz (Root Library)
- chokidar-2.0.4.tgz
- braces-2.3.2.tgz
- snapdragon-0.8.2.tgz
- base-0.11.2.tgz
- cache-base-1.0.1.tgz
- :x: **set-value-2.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>set-value-0.4.3.tgz</b></p></summary>
<p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p>
<p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz">https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz</a></p>
<p>Path to dependency file: WebGoat/docs/package.json</p>
<p>Path to vulnerable library: WebGoat/docs/node_modules/union-value/node_modules/set-value/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.26.3.tgz (Root Library)
- chokidar-2.0.4.tgz
- braces-2.3.2.tgz
- snapdragon-0.8.2.tgz
- base-0.11.2.tgz
- cache-base-1.0.1.tgz
- union-value-1.0.0.tgz
- :x: **set-value-0.4.3.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/mlmcd/WebGoat/commit/19e9dd9297c4eb1dbe74247f3c0eba17b31acc02">19e9dd9297c4eb1dbe74247f3c0eba17b31acc02</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
set-value is vulnerable to Prototype Pollution in versions lower than 3.0.1. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using any of the constructor, prototype and _proto_ payloads.
<p>Publish Date: 2019-08-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10747>CVE-2019-10747</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f">https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f</a></p>
<p>Release Date: 2019-07-24</p>
<p>Fix Resolution: 2.0.1,3.0.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"set-value","packageVersion":"2.0.0","isTransitiveDependency":true,"dependencyTree":"browser-sync:2.26.3;chokidar:2.0.4;braces:2.3.2;snapdragon:0.8.2;base:0.11.2;cache-base:1.0.1;set-value:2.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.0.1,3.0.1"},{"packageType":"javascript/Node.js","packageName":"set-value","packageVersion":"0.4.3","isTransitiveDependency":true,"dependencyTree":"browser-sync:2.26.3;chokidar:2.0.4;braces:2.3.2;snapdragon:0.8.2;base:0.11.2;cache-base:1.0.1;union-value:1.0.0;set-value:0.4.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.0.1,3.0.1"}],"vulnerabilityIdentifier":"CVE-2019-10747","vulnerabilityDetails":"set-value is vulnerable to Prototype Pollution in versions lower than 3.0.1. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using any of the constructor, prototype and _proto_ payloads.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10747","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2019-10747 (High) detected in set-value-2.0.0.tgz, set-value-0.4.3.tgz - ## CVE-2019-10747 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>set-value-2.0.0.tgz</b>, <b>set-value-0.4.3.tgz</b></p></summary>
<p>
<details><summary><b>set-value-2.0.0.tgz</b></p></summary>
<p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p>
<p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz</a></p>
<p>Path to dependency file: WebGoat/docs/package.json</p>
<p>Path to vulnerable library: WebGoat/docs/node_modules/set-value/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.26.3.tgz (Root Library)
- chokidar-2.0.4.tgz
- braces-2.3.2.tgz
- snapdragon-0.8.2.tgz
- base-0.11.2.tgz
- cache-base-1.0.1.tgz
- :x: **set-value-2.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>set-value-0.4.3.tgz</b></p></summary>
<p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p>
<p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz">https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz</a></p>
<p>Path to dependency file: WebGoat/docs/package.json</p>
<p>Path to vulnerable library: WebGoat/docs/node_modules/union-value/node_modules/set-value/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.26.3.tgz (Root Library)
- chokidar-2.0.4.tgz
- braces-2.3.2.tgz
- snapdragon-0.8.2.tgz
- base-0.11.2.tgz
- cache-base-1.0.1.tgz
- union-value-1.0.0.tgz
- :x: **set-value-0.4.3.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/mlmcd/WebGoat/commit/19e9dd9297c4eb1dbe74247f3c0eba17b31acc02">19e9dd9297c4eb1dbe74247f3c0eba17b31acc02</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
set-value is vulnerable to Prototype Pollution in versions lower than 3.0.1. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using any of the constructor, prototype and _proto_ payloads.
<p>Publish Date: 2019-08-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10747>CVE-2019-10747</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f">https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f</a></p>
<p>Release Date: 2019-07-24</p>
<p>Fix Resolution: 2.0.1,3.0.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"set-value","packageVersion":"2.0.0","isTransitiveDependency":true,"dependencyTree":"browser-sync:2.26.3;chokidar:2.0.4;braces:2.3.2;snapdragon:0.8.2;base:0.11.2;cache-base:1.0.1;set-value:2.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.0.1,3.0.1"},{"packageType":"javascript/Node.js","packageName":"set-value","packageVersion":"0.4.3","isTransitiveDependency":true,"dependencyTree":"browser-sync:2.26.3;chokidar:2.0.4;braces:2.3.2;snapdragon:0.8.2;base:0.11.2;cache-base:1.0.1;union-value:1.0.0;set-value:0.4.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.0.1,3.0.1"}],"vulnerabilityIdentifier":"CVE-2019-10747","vulnerabilityDetails":"set-value is vulnerable to Prototype Pollution in versions lower than 3.0.1. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using any of the constructor, prototype and _proto_ payloads.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10747","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in set value tgz set value tgz cve high severity vulnerability vulnerable libraries set value tgz set value tgz set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href path to dependency file webgoat docs package json path to vulnerable library webgoat docs node modules set value package json dependency hierarchy browser sync tgz root library chokidar tgz braces tgz snapdragon tgz base tgz cache base tgz x set value tgz vulnerable library set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href path to dependency file webgoat docs package json path to vulnerable library webgoat docs node modules union value node modules set value package json dependency hierarchy browser sync tgz root library chokidar tgz braces tgz snapdragon tgz base tgz cache base tgz union value tgz x set value tgz vulnerable library found in head commit a href found in base branch develop vulnerability details set value is vulnerable to prototype pollution in versions lower than the function mixin deep could be tricked into adding or modifying properties of object prototype using any of the constructor prototype and proto payloads publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails set value is vulnerable to prototype pollution in versions lower than the function mixin deep could be tricked into adding or modifying properties of object prototype using any of the constructor prototype and proto payloads vulnerabilityurl
| 0
|
15,080
| 18,785,546,567
|
IssuesEvent
|
2021-11-08 11:44:14
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
Drop experimental_multi_threaded_digest
|
type: process untriaged team-Local-Exec
|
This options defaults to true since a while and should just be removed, likely with the whole `SsdModule`.
|
1.0
|
Drop experimental_multi_threaded_digest - This options defaults to true since a while and should just be removed, likely with the whole `SsdModule`.
|
process
|
drop experimental multi threaded digest this options defaults to true since a while and should just be removed likely with the whole ssdmodule
| 1
|
22,171
| 30,721,409,378
|
IssuesEvent
|
2023-07-27 16:11:08
|
SpikeInterface/spikeinterface
|
https://api.github.com/repos/SpikeInterface/spikeinterface
|
closed
|
Questions about the modes "median"/"average" in `remove_artifacts`
|
question preprocessing
|
Hi!
I am trying to explore the functionality of [`remove_artifacts`](https://github.com/SpikeInterface/spikeinterface/blob/main/src/spikeinterface/preprocessing/remove_artifacts.py) with mode "median" or "average" and have a few questions:
(1) Just to confirm if my understanding is correct or not: The input argument `artifacts` gives a set of templates that suggest what each artifact looks like; if `None`, then the artifacts are calculated from the raw traces by indexing the triggers and extracting "waveforms". Additionally the "jitter" parameter is used to find a best match between the recording traces and the templates and the corresponding amplitude scaling factor. Is it the right way to interpret it? The [documentation](https://spikeinterface.readthedocs.io/en/latest/api.html#spikeinterface.preprocessing.remove_artifacts) only says that they are the "values" of the artifacts. I wonder in which cases the artifacts need to be explicitly specified; when specified, it is assumed that all the channels have the same artifact waveform?
(2) The documentation says that providing argument `artifacts` can bypass `ms_before` and `ms_after`. But in the implementation of [`get_traces`](https://github.com/SpikeInterface/spikeinterface/blob/main/src/spikeinterface/preprocessing/remove_artifacts.py#L417) in `RemoveArtifactsRecordingSegment`, in "median"/"average" mode, the slicing of traces still depend on the `pad` property, which is derived from `ms_before` and `ms_after`. Then it seems to me that we need to manually make sure arguments `ms_before` and `ms_after` need to be aligned with `artifacts` so that they indicate the same length. Otherwise there may be mismatch when subtracting the artifacts from the recording traces. I am not sure if I understood the logic in source code correctly, nor have I tested.
(3) For my own use case, I wanted to mask my stimulation artifacts to zero. But different stimulation artifacts have different durations. It seems like merely using `zeros` mode wouldn't accommodate different artifacts. If I use `median`/`average` mode while providing labels, it will calculate the artifact waveform template by the label and the subtract from raw data, not quite what I wanted to do. Could you please suggest some efficient ways to zero out the artifacts with varying durations?
Thank you so much for your help. I really appreciate your active maintenance and development of this framework. It has been making my life a lot easier!
|
1.0
|
Questions about the modes "median"/"average" in `remove_artifacts` - Hi!
I am trying to explore the functionality of [`remove_artifacts`](https://github.com/SpikeInterface/spikeinterface/blob/main/src/spikeinterface/preprocessing/remove_artifacts.py) with mode "median" or "average" and have a few questions:
(1) Just to confirm if my understanding is correct or not: The input argument `artifacts` gives a set of templates that suggest what each artifact looks like; if `None`, then the artifacts are calculated from the raw traces by indexing the triggers and extracting "waveforms". Additionally the "jitter" parameter is used to find a best match between the recording traces and the templates and the corresponding amplitude scaling factor. Is it the right way to interpret it? The [documentation](https://spikeinterface.readthedocs.io/en/latest/api.html#spikeinterface.preprocessing.remove_artifacts) only says that they are the "values" of the artifacts. I wonder in which cases the artifacts need to be explicitly specified; when specified, it is assumed that all the channels have the same artifact waveform?
(2) The documentation says that providing argument `artifacts` can bypass `ms_before` and `ms_after`. But in the implementation of [`get_traces`](https://github.com/SpikeInterface/spikeinterface/blob/main/src/spikeinterface/preprocessing/remove_artifacts.py#L417) in `RemoveArtifactsRecordingSegment`, in "median"/"average" mode, the slicing of traces still depend on the `pad` property, which is derived from `ms_before` and `ms_after`. Then it seems to me that we need to manually make sure arguments `ms_before` and `ms_after` need to be aligned with `artifacts` so that they indicate the same length. Otherwise there may be mismatch when subtracting the artifacts from the recording traces. I am not sure if I understood the logic in source code correctly, nor have I tested.
(3) For my own use case, I wanted to mask my stimulation artifacts to zero. But different stimulation artifacts have different durations. It seems like merely using `zeros` mode wouldn't accommodate different artifacts. If I use `median`/`average` mode while providing labels, it will calculate the artifact waveform template by the label and the subtract from raw data, not quite what I wanted to do. Could you please suggest some efficient ways to zero out the artifacts with varying durations?
Thank you so much for your help. I really appreciate your active maintenance and development of this framework. It has been making my life a lot easier!
|
process
|
questions about the modes median average in remove artifacts hi i am trying to explore the functionality of with mode median or average and have a few questions just to confirm if my understanding is correct or not the input argument artifacts gives a set of templates that suggest what each artifact looks like if none then the artifacts are calculated from the raw traces by indexing the triggers and extracting waveforms additionally the jitter parameter is used to find a best match between the recording traces and the templates and the corresponding amplitude scaling factor is it the right way to interpret it the only says that they are the values of the artifacts i wonder in which cases the artifacts need to be explicitly specified when specified it is assumed that all the channels have the same artifact waveform the documentation says that providing argument artifacts can bypass ms before and ms after but in the implementation of in removeartifactsrecordingsegment in median average mode the slicing of traces still depend on the pad property which is derived from ms before and ms after then it seems to me that we need to manually make sure arguments ms before and ms after need to be aligned with artifacts so that they indicate the same length otherwise there may be mismatch when subtracting the artifacts from the recording traces i am not sure if i understood the logic in source code correctly nor have i tested for my own use case i wanted to mask my stimulation artifacts to zero but different stimulation artifacts have different durations it seems like merely using zeros mode wouldn t accommodate different artifacts if i use median average mode while providing labels it will calculate the artifact waveform template by the label and the subtract from raw data not quite what i wanted to do could you please suggest some efficient ways to zero out the artifacts with varying durations thank you so much for your help i really appreciate your active maintenance and development of this framework it has been making my life a lot easier
| 1
|
155,414
| 24,463,073,836
|
IssuesEvent
|
2022-10-07 12:54:50
|
dccabs/housecallmd
|
https://api.github.com/repos/dccabs/housecallmd
|
closed
|
Contact Us - Redesign with tailwind ui
|
design upgrade
|
Please replace the current contact us page with tailwind ui
https://www.housecallmd.org/contact
- Please replace with this component - https://tailwindui.com/components/marketing/sections/contact-sections#component-dec976a631662b509173ba8b848c49cd
- Make sure form still works
- Show form is submitting indicator
- Please show modal that confirms a message was sent
- Please include some type of error handling - if the request fails we should know there is an error.
- Please replace indigo colors with Rose in tailwind
|
1.0
|
Contact Us - Redesign with tailwind ui - Please replace the current contact us page with tailwind ui
https://www.housecallmd.org/contact
- Please replace with this component - https://tailwindui.com/components/marketing/sections/contact-sections#component-dec976a631662b509173ba8b848c49cd
- Make sure form still works
- Show form is submitting indicator
- Please show modal that confirms a message was sent
- Please include some type of error handling - if the request fails we should know there is an error.
- Please replace indigo colors with Rose in tailwind
|
non_process
|
contact us redesign with tailwind ui please replace the current contact us page with tailwind ui please replace with this component make sure form still works show form is submitting indicator please show modal that confirms a message was sent please include some type of error handling if the request fails we should know there is an error please replace indigo colors with rose in tailwind
| 0
|
68,890
| 7,113,054,759
|
IssuesEvent
|
2018-01-17 19:07:03
|
GalaxyTrail/GFR_bugs
|
https://api.github.com/repos/GalaxyTrail/GFR_bugs
|
closed
|
1-2 eggplant statue house - enemy once got stuck in floor
|
bug maybe fixed - needs more testing
|
Submitter: Patrick Murphy
Email: italiangamer97@gmail.com
In the house of eggplant statues in Course 1-2, I noticed that an enemy managed to get stuck in the floor. Unsure of how I triggered it, but I was able to get the enemy out of the floor and defeat it. Perhaps I did something with one of the statues there?
|
1.0
|
1-2 eggplant statue house - enemy once got stuck in floor - Submitter: Patrick Murphy
Email: italiangamer97@gmail.com
In the house of eggplant statues in Course 1-2, I noticed that an enemy managed to get stuck in the floor. Unsure of how I triggered it, but I was able to get the enemy out of the floor and defeat it. Perhaps I did something with one of the statues there?
|
non_process
|
eggplant statue house enemy once got stuck in floor submitter patrick murphy email gmail com in the house of eggplant statues in course i noticed that an enemy managed to get stuck in the floor unsure of how i triggered it but i was able to get the enemy out of the floor and defeat it perhaps i did something with one of the statues there
| 0
|
559,706
| 16,569,422,307
|
IssuesEvent
|
2021-05-30 04:40:41
|
gitextensions/gitextensions
|
https://api.github.com/repos/gitextensions/gitextensions
|
closed
|
GE hangs when viewing submodules in diff viewer
|
:beetle: type: bug :construction: status: in progress :grey_exclamation: priority: high
|
## Current behaviour
Regression in master in the last days.
If a submodule is selected in worktree (at least - seem OK for HEAD though), GE hangs and stops responding. Status is never changed from the default. Mouse just spinning.
## Expected behaviour
Submodule is displayed.
## Steps to reproduce
Select modified submodule in worktree in the diff viewer.
## Screenshots

## Did this work in previous version of GitExtensions
- Git Extensions 3.6.0.11114
- Build f54a727a95393dcacfbbd8603dfd6daf931618c4 (Dirty)
## Environment
- Git Extensions 3.6.0.11170
- Build 273a0f6fd3e07858f837cdc19d50827871e32319 (Dirty)
- Git 2.30.0.windows.1
- Microsoft Windows NT 10.0.17763.0
- .NET Framework 4.8.4300.0
- DPI 96dpi (no scaling)
|
1.0
|
GE hangs when viewing submodules in diff viewer - ## Current behaviour
Regression in master in the last days.
If a submodule is selected in worktree (at least - seem OK for HEAD though), GE hangs and stops responding. Status is never changed from the default. Mouse just spinning.
## Expected behaviour
Submodule is displayed.
## Steps to reproduce
Select modified submodule in worktree in the diff viewer.
## Screenshots

## Did this work in previous version of GitExtensions
- Git Extensions 3.6.0.11114
- Build f54a727a95393dcacfbbd8603dfd6daf931618c4 (Dirty)
## Environment
- Git Extensions 3.6.0.11170
- Build 273a0f6fd3e07858f837cdc19d50827871e32319 (Dirty)
- Git 2.30.0.windows.1
- Microsoft Windows NT 10.0.17763.0
- .NET Framework 4.8.4300.0
- DPI 96dpi (no scaling)
|
non_process
|
ge hangs when viewing submodules in diff viewer current behaviour regression in master in the last days if a submodule is selected in worktree at least seem ok for head though ge hangs and stops responding status is never changed from the default mouse just spinning expected behaviour submodule is displayed steps to reproduce select modified submodule in worktree in the diff viewer screenshots did this work in previous version of gitextensions git extensions build dirty environment git extensions build dirty git windows microsoft windows nt net framework dpi no scaling
| 0
|
657,658
| 21,799,590,185
|
IssuesEvent
|
2022-05-16 02:30:09
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
closed
|
Add setting to toggle visibility of Bookmark button in the URL bar
|
priority/P4 feature/settings OS/Desktop
|
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
It's easy to accidentally click the bookmark button in the URL bar as it's positioned between the refresh button and the URL input - resulting in unwanted bookmarks. Also, some people don't use the native browser bookmarks at all so this button is not always of use.

Ideally, there should be an Appearance setting for "Show Bookmark button" which works similarly to the "Show Home button" setting.

## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Have poor hand-eye-coordination
2. Accidentally click bookmark button instead of URL input or refresh button
3. Unwanted bookmark saved
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
I do this multiple times per week.
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.18.77 Chromium: 87.0.4280.101 (Official Build) unknown (64-bit)
-- | --
Revision | 9407c80213cda69c2b7abcb4fa8e3f74488f4956-refs/branch-heads/4280@{#1807}
OS | Linux
JavaScript | V8 8.7.220.29
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? Yes
- Can you reproduce this issue with the beta channel? Yes
- Can you reproduce this issue with the nightly channel? Yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? N/A
- Does the issue resolve itself when disabling Brave Rewards? N/A
- Is the issue reproducible on the latest version of Chrome? N/A
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
- Related ticket: https://github.com/brave/brave-browser/issues/774
- Related comment: https://github.com/brave/brave-browser/issues/774#issuecomment-760664719
- [Community thread](https://community.brave.com/t/is-there-a-way-to-move-the-bookmarks-button-to-where-it-is-in-chrome/48785/12)
P.S. Thanks for all the amazing work on Brave!
|
1.0
|
Add setting to toggle visibility of Bookmark button in the URL bar - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
It's easy to accidentally click the bookmark button in the URL bar as it's positioned between the refresh button and the URL input - resulting in unwanted bookmarks. Also, some people don't use the native browser bookmarks at all so this button is not always of use.

Ideally, there should be an Appearance setting for "Show Bookmark button" which works similarly to the "Show Home button" setting.

## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Have poor hand-eye-coordination
2. Accidentally click bookmark button instead of URL input or refresh button
3. Unwanted bookmark saved
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
I do this multiple times per week.
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.18.77 Chromium: 87.0.4280.101 (Official Build) unknown (64-bit)
-- | --
Revision | 9407c80213cda69c2b7abcb4fa8e3f74488f4956-refs/branch-heads/4280@{#1807}
OS | Linux
JavaScript | V8 8.7.220.29
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? Yes
- Can you reproduce this issue with the beta channel? Yes
- Can you reproduce this issue with the nightly channel? Yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? N/A
- Does the issue resolve itself when disabling Brave Rewards? N/A
- Is the issue reproducible on the latest version of Chrome? N/A
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
- Related ticket: https://github.com/brave/brave-browser/issues/774
- Related comment: https://github.com/brave/brave-browser/issues/774#issuecomment-760664719
- [Community thread](https://community.brave.com/t/is-there-a-way-to-move-the-bookmarks-button-to-where-it-is-in-chrome/48785/12)
P.S. Thanks for all the amazing work on Brave!
|
non_process
|
add setting to toggle visibility of bookmark button in the url bar have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description it s easy to accidentally click the bookmark button in the url bar as it s positioned between the refresh button and the url input resulting in unwanted bookmarks also some people don t use the native browser bookmarks at all so this button is not always of use ideally there should be an appearance setting for show bookmark button which works similarly to the show home button setting steps to reproduce have poor hand eye coordination accidentally click bookmark button instead of url input or refresh button unwanted bookmark saved reproduces how often i do this multiple times per week brave version brave version info brave chromium official build unknown bit revision refs branch heads os linux javascript version channel information can you reproduce this issue with the current release yes can you reproduce this issue with the beta channel yes can you reproduce this issue with the nightly channel yes other additional information does the issue resolve itself when disabling brave shields n a does the issue resolve itself when disabling brave rewards n a is the issue reproducible on the latest version of chrome n a miscellaneous information related ticket related comment p s thanks for all the amazing work on brave
| 0
|
731
| 3,214,309,873
|
IssuesEvent
|
2015-10-07 00:42:40
|
broadinstitute/hellbender
|
https://api.github.com/repos/broadinstitute/hellbender
|
closed
|
Add mechanism to generate unique ids for data types (reads, variants, etc.)
|
Dataflow DataflowPreprocessingPipeline
|
Needed for GroupByKey, since Java serialization is not deterministic. Initial idea is to create IDs based on the source of each record (eg., URI + file offset or record number).
|
1.0
|
Add mechanism to generate unique ids for data types (reads, variants, etc.) - Needed for GroupByKey, since Java serialization is not deterministic. Initial idea is to create IDs based on the source of each record (eg., URI + file offset or record number).
|
process
|
add mechanism to generate unique ids for data types reads variants etc needed for groupbykey since java serialization is not deterministic initial idea is to create ids based on the source of each record eg uri file offset or record number
| 1
|
28,127
| 11,590,028,152
|
IssuesEvent
|
2020-02-24 05:03:37
|
aksiksi/vaulty
|
https://api.github.com/repos/aksiksi/vaulty
|
closed
|
Add support for address whitelisting
|
security vaulty-mail
|
Allow users to specify a list of whitelisted sender addresses a Vaulty address under their control.
In addition, it might make sense to allow users to define a "secret word" that **must** be present in the email body or subject to accept the email. The word could be stripped from the email before getting pushed to storage.
|
True
|
Add support for address whitelisting - Allow users to specify a list of whitelisted sender addresses a Vaulty address under their control.
In addition, it might make sense to allow users to define a "secret word" that **must** be present in the email body or subject to accept the email. The word could be stripped from the email before getting pushed to storage.
|
non_process
|
add support for address whitelisting allow users to specify a list of whitelisted sender addresses a vaulty address under their control in addition it might make sense to allow users to define a secret word that must be present in the email body or subject to accept the email the word could be stripped from the email before getting pushed to storage
| 0
|
6,410
| 9,488,866,277
|
IssuesEvent
|
2019-04-22 20:45:24
|
material-components/material-components-ios
|
https://api.github.com/repos/material-components/material-components-ios
|
closed
|
[AppBar] Mark AppBar theming extension ready
|
[AppBar] type:Process
|
This was filed as an internal issue. If you are a Googler, please visit [b/130716217](http://b/130716217) for more details.
<!-- Auto-generated content below, do not modify -->
---
#### Internal data
- Associated internal bug: [b/130716217](http://b/130716217)
|
1.0
|
[AppBar] Mark AppBar theming extension ready - This was filed as an internal issue. If you are a Googler, please visit [b/130716217](http://b/130716217) for more details.
<!-- Auto-generated content below, do not modify -->
---
#### Internal data
- Associated internal bug: [b/130716217](http://b/130716217)
|
process
|
mark appbar theming extension ready this was filed as an internal issue if you are a googler please visit for more details internal data associated internal bug
| 1
|
8,987
| 12,100,407,087
|
IssuesEvent
|
2020-04-20 13:45:31
|
ComposableWeb/poolbase
|
https://api.github.com/repos/ComposableWeb/poolbase
|
opened
|
[💥FEAT] derive color palette from page screenshot
|
enhancement epic: processing
|
**Feature request? Please describe.**
For frontend display of pages: Derive colors from screenshot using AI
**Acceptance Criteria - Describe the solution you'd like**
A clear and concise description of what you want to happen in bullet points:
* ...
**Related issues**
#1
And any other context or screenshots about the feature request here.
|
1.0
|
[💥FEAT] derive color palette from page screenshot - **Feature request? Please describe.**
For frontend display of pages: Derive colors from screenshot using AI
**Acceptance Criteria - Describe the solution you'd like**
A clear and concise description of what you want to happen in bullet points:
* ...
**Related issues**
#1
And any other context or screenshots about the feature request here.
|
process
|
derive color palette from page screenshot feature request please describe for frontend display of pages derive colors from screenshot using ai acceptance criteria describe the solution you d like a clear and concise description of what you want to happen in bullet points related issues and any other context or screenshots about the feature request here
| 1
|
67,843
| 7,065,242,262
|
IssuesEvent
|
2018-01-06 17:38:28
|
pandas-dev/pandas
|
https://api.github.com/repos/pandas-dev/pandas
|
closed
|
`(x + 1) % 1` inside test_multi.py
|
MultiIndex Testing
|
#### Code Sample, a copy-pastable example if possible
https://github.com/pandas-dev/pandas/blob/93033151a8d8aaa650a81df9f41347758bf6c393/pandas/tests/indexes/test_multi.py#L170-L171
#### Problem description
I assume this is supposed to shift the labels cyclically by one. Then `% 4` and `% 2` should be used. Otherwise its at least weird to write `(x + 1) % 1` instead of `0`.
|
1.0
|
`(x + 1) % 1` inside test_multi.py - #### Code Sample, a copy-pastable example if possible
https://github.com/pandas-dev/pandas/blob/93033151a8d8aaa650a81df9f41347758bf6c393/pandas/tests/indexes/test_multi.py#L170-L171
#### Problem description
I assume this is supposed to shift the labels cyclically by one. Then `% 4` and `% 2` should be used. Otherwise its at least weird to write `(x + 1) % 1` instead of `0`.
|
non_process
|
x inside test multi py code sample a copy pastable example if possible problem description i assume this is supposed to shift the labels cyclically by one then and should be used otherwise its at least weird to write x instead of
| 0
|
3,921
| 6,843,043,013
|
IssuesEvent
|
2017-11-12 10:48:24
|
pwittchen/ReactiveNetwork
|
https://api.github.com/repos/pwittchen/ReactiveNetwork
|
opened
|
RxJava1.x: release 0.12.2
|
release process
|
**Initial realese notes**:
- updated project dependencies
- updated Gradle to 3.0.0
**Things to do**:
- [ ] RxJava1.x branch:
- [ ] bump library version
- [ ] upload archives to Maven Central
- [ ] close and release artifact on Maven Central
- [ ] update `CHANGELOG.md` after Maven Sync
- [ ] bump library version in `README.md`
- [ ] create new GitHub release
|
1.0
|
RxJava1.x: release 0.12.2 - **Initial realese notes**:
- updated project dependencies
- updated Gradle to 3.0.0
**Things to do**:
- [ ] RxJava1.x branch:
- [ ] bump library version
- [ ] upload archives to Maven Central
- [ ] close and release artifact on Maven Central
- [ ] update `CHANGELOG.md` after Maven Sync
- [ ] bump library version in `README.md`
- [ ] create new GitHub release
|
process
|
x release initial realese notes updated project dependencies updated gradle to things to do x branch bump library version upload archives to maven central close and release artifact on maven central update changelog md after maven sync bump library version in readme md create new github release
| 1
|
103,843
| 11,383,025,622
|
IssuesEvent
|
2020-01-29 04:20:55
|
car12lin12/InstaBaddies
|
https://api.github.com/repos/car12lin12/InstaBaddies
|
opened
|
Update readme to clarify goals of the project
|
Documentation :memo:
|
Update the readme to show the description / objective / core features of the project.
Core features :
- upload a picture
- follow a user
- receive notifications from that user if they upload a new picture
- comment on a picture
|
1.0
|
Update readme to clarify goals of the project - Update the readme to show the description / objective / core features of the project.
Core features :
- upload a picture
- follow a user
- receive notifications from that user if they upload a new picture
- comment on a picture
|
non_process
|
update readme to clarify goals of the project update the readme to show the description objective core features of the project core features upload a picture follow a user receive notifications from that user if they upload a new picture comment on a picture
| 0
|
238,054
| 19,694,490,725
|
IssuesEvent
|
2022-01-12 10:40:01
|
dotnet/sdk
|
https://api.github.com/repos/dotnet/sdk
|
closed
|
With .NET 7 x86 SDK, Testhost process exited with error: It was not possible to find any compatible framework version
|
Area-DotNet Test untriaged
|
**--Repro Steps---**
1. Install .NET 7 x86 SDK(https://github.com/dotnet/installer) on Win x64 OS
2. Create a UT project and run it
**--Expected Result --**
1. dotnet test works fine
**--Actual Result--**
1. Testhost process exited with error: It was not possible to find any compatible framework version
The framework 'Microsoft.NETCore.App', version '7.0.0-alpha.1.22059.2' (x64) was not found.

**--Note--**
It didnot repro with .NET 6 x86 SDK
**Dotnet info**
c:\6t>dotnet --info
.NET SDK (reflecting any global.json):
Version: 7.0.100-alpha.1.22060.7
Commit: 4a2d57a652
Runtime Environment:
OS Name: Windows
OS Version: 10.0.22000
OS Platform: Windows
RID: win10-x86
Base Path: C:\Program Files (x86)\dotnet\sdk\7.0.100-alpha.1.22060.7\
Host (useful for support):
Version: 7.0.0-alpha.1.22059.2
Commit: a2af629476
.NET SDKs installed:
6.0.200-preview.22055.18 [C:\Program Files (x86)\dotnet\sdk]
7.0.100-alpha.1.22060.7 [C:\Program Files (x86)\dotnet\sdk]
.NET runtimes installed:
Microsoft.AspNetCore.App 6.0.1 [C:\Program Files (x86)\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 7.0.0-alpha.1.22057.17 [C:\Program Files (x86)\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.NETCore.App 6.0.1 [C:\Program Files (x86)\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 7.0.0-alpha.1.22059.2 [C:\Program Files (x86)\dotnet\shared\Microsoft.NETCore.App]
Microsoft.WindowsDesktop.App 6.0.1 [C:\Program Files (x86)\dotnet\shared\Microsoft.WindowsDesktop.App]
Microsoft.WindowsDesktop.App 7.0.0-alpha.1.22055.1 [C:\Program Files (x86)\dotnet\shared\Microsoft.WindowsDesktop.App]
To install additional .NET runtimes or SDKs:
https://aka.ms/dotnet-download
|
1.0
|
With .NET 7 x86 SDK, Testhost process exited with error: It was not possible to find any compatible framework version - **--Repro Steps---**
1. Install .NET 7 x86 SDK(https://github.com/dotnet/installer) on Win x64 OS
2. Create a UT project and run it
**--Expected Result --**
1. dotnet test works fine
**--Actual Result--**
1. Testhost process exited with error: It was not possible to find any compatible framework version
The framework 'Microsoft.NETCore.App', version '7.0.0-alpha.1.22059.2' (x64) was not found.

**--Note--**
It didnot repro with .NET 6 x86 SDK
**Dotnet info**
c:\6t>dotnet --info
.NET SDK (reflecting any global.json):
Version: 7.0.100-alpha.1.22060.7
Commit: 4a2d57a652
Runtime Environment:
OS Name: Windows
OS Version: 10.0.22000
OS Platform: Windows
RID: win10-x86
Base Path: C:\Program Files (x86)\dotnet\sdk\7.0.100-alpha.1.22060.7\
Host (useful for support):
Version: 7.0.0-alpha.1.22059.2
Commit: a2af629476
.NET SDKs installed:
6.0.200-preview.22055.18 [C:\Program Files (x86)\dotnet\sdk]
7.0.100-alpha.1.22060.7 [C:\Program Files (x86)\dotnet\sdk]
.NET runtimes installed:
Microsoft.AspNetCore.App 6.0.1 [C:\Program Files (x86)\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 7.0.0-alpha.1.22057.17 [C:\Program Files (x86)\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.NETCore.App 6.0.1 [C:\Program Files (x86)\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 7.0.0-alpha.1.22059.2 [C:\Program Files (x86)\dotnet\shared\Microsoft.NETCore.App]
Microsoft.WindowsDesktop.App 6.0.1 [C:\Program Files (x86)\dotnet\shared\Microsoft.WindowsDesktop.App]
Microsoft.WindowsDesktop.App 7.0.0-alpha.1.22055.1 [C:\Program Files (x86)\dotnet\shared\Microsoft.WindowsDesktop.App]
To install additional .NET runtimes or SDKs:
https://aka.ms/dotnet-download
|
non_process
|
with net sdk testhost process exited with error it was not possible to find any compatible framework version repro steps install net sdk on win os create a ut project and run it expected result dotnet test works fine actual result testhost process exited with error it was not possible to find any compatible framework version the framework microsoft netcore app version alpha was not found note it didnot repro with net sdk dotnet info c dotnet info net sdk reflecting any global json version alpha commit runtime environment os name windows os version os platform windows rid base path c program files dotnet sdk alpha host useful for support version alpha commit net sdks installed preview alpha net runtimes installed microsoft aspnetcore app microsoft aspnetcore app alpha microsoft netcore app microsoft netcore app alpha microsoft windowsdesktop app microsoft windowsdesktop app alpha to install additional net runtimes or sdks
| 0
|
16,951
| 22,306,072,081
|
IssuesEvent
|
2022-06-13 13:09:39
|
deepset-ai/haystack
|
https://api.github.com/repos/deepset-ai/haystack
|
closed
|
Crawler: can it handle website with contents dynamically loaded by JavaScript?
|
type:feature good first issue Contributions wanted! good second issue topic:preprocessing journey:intermediate
|
**Question**
Per captioned, can crawler handle website with contents dynamically loaded by pressing an "expand" button?
**Additional context**
In today's website designs a lot of them display content dynamically through Javascript. On inspecting Page Source they will be shown as loading in on client side. Some contents are to be loaded when an "expand" button (or eqivalent) is pressed. Tested with the current version of Crawler node and only the statis title/heading can be scraped, but not any of the dynamic text. Is that expected/limitation of Selenium/bug? Any way to scrape from these kind of pages?
E.g. [https://www.sfc.hk/en/Forms/Intermediaries/Licensing-forms](url)
**FAQ Check**
- [X] Have you had a look at [our new FAQ page](https://haystack.deepset.ai/overview/faq)?
|
1.0
|
Crawler: can it handle website with contents dynamically loaded by JavaScript? - **Question**
Per captioned, can crawler handle website with contents dynamically loaded by pressing an "expand" button?
**Additional context**
In today's website designs a lot of them display content dynamically through Javascript. On inspecting Page Source they will be shown as loading in on client side. Some contents are to be loaded when an "expand" button (or eqivalent) is pressed. Tested with the current version of Crawler node and only the statis title/heading can be scraped, but not any of the dynamic text. Is that expected/limitation of Selenium/bug? Any way to scrape from these kind of pages?
E.g. [https://www.sfc.hk/en/Forms/Intermediaries/Licensing-forms](url)
**FAQ Check**
- [X] Have you had a look at [our new FAQ page](https://haystack.deepset.ai/overview/faq)?
|
process
|
crawler can it handle website with contents dynamically loaded by javascript question per captioned can crawler handle website with contents dynamically loaded by pressing an expand button additional context in today s website designs a lot of them display content dynamically through javascript on inspecting page source they will be shown as loading in on client side some contents are to be loaded when an expand button or eqivalent is pressed tested with the current version of crawler node and only the statis title heading can be scraped but not any of the dynamic text is that expected limitation of selenium bug any way to scrape from these kind of pages e g url faq check have you had a look at
| 1
|
11,179
| 13,957,695,350
|
IssuesEvent
|
2020-10-24 08:11:31
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
PT: Harvesting
|
Geoportal Harvesting process PT - Portugal
|
Geoportal team,
Can you please start a harvesting do the Portuguese catalogue?
Thank you!
|
1.0
|
PT: Harvesting - Geoportal team,
Can you please start a harvesting do the Portuguese catalogue?
Thank you!
|
process
|
pt harvesting geoportal team can you please start a harvesting do the portuguese catalogue thank you
| 1
|
8,761
| 11,880,979,031
|
IssuesEvent
|
2020-03-27 11:45:07
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
Test Failed: Process terminated. Assertion failed. "!_fileHandle.IsClosed"
|
area-System.Diagnostics.Process os-windows test bug test-corefx
|
**Detail:**
https://helix.dot.net/api/2019-06-17/jobs/0c15185a-2331-4dcf-978c-cd7c4a4b3189/workitems/System.Diagnostics.Process.Tests/console
**Log:**
```
C:\dotnetbuild\work\0c15185a-2331-4dcf-978c-cd7c4a4b3189\Work\2eeb62f3-774b-43f1-94ac-c844f873d662\Exec>"C:\dotnetbuild\work\0c15185a-2331-4dcf-978c-cd7c4a4b3189\Payload\dotnet.exe" exec --runtimeconfig System.Diagnostics.Process.Tests.runtimeconfig.json xunit.console.dll System.Diagnostics.Process.Tests.dll -xml testResults.xml -nologo -nocolor -notrait category=nonnetcoreapptests -notrait category=nonwindowstests -notrait category=IgnoreForCI -notrait category=OuterLoop -notrait category=failing @C:\dotnetbuild\work\0c15185a-2331-4dcf-978c-cd7c4a4b3189\Payload\CoreFX.issues.rsp
Discovering: System.Diagnostics.Process.Tests (method display = ClassAndMethod, method display options = None)
Discovered: System.Diagnostics.Process.Tests (found 221 of 241 test cases)
Starting: System.Diagnostics.Process.Tests (parallel test collections = on, max threads = 2)
System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_WebPage [SKIP]
Manual test
System.Diagnostics.Tests.ProcessStartInfoTests.ShellExecute_Nano_Fails_Start [SKIP]
Condition(s) not met: "IsWindowsNanoServer"
Process terminated. Assertion failed.
!_fileHandle.IsClosed
at System.IO.FileStream.AssertCanRead() in /_/src/System.Private.CoreLib/shared/System/IO/FileStream.Windows.cs:line 482
at System.IO.FileStream.ReadNative(Span`1 buffer) in /_/src/System.Private.CoreLib/shared/System/IO/FileStream.Windows.cs:line 497
at System.IO.FileStream.ReadSpan(Span`1 destination) in /_/src/System.Private.CoreLib/shared/System/IO/FileStream.Windows.cs:line 429
at System.IO.FileStream.Read(Byte[] array, Int32 offset, Int32 count) in /_/src/System.Private.CoreLib/shared/System/IO/FileStream.cs:line 304
at System.IO.Stream.<>c.<BeginReadInternal>b__43_0(Object <p0>) in /_/src/System.Private.CoreLib/shared/System/IO/Stream.cs:line 303
at System.Threading.Tasks.Task`1.InnerInvoke() in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Future.cs:line 512
at System.Threading.Tasks.Task.<>c.<.cctor>b__274_0(Object obj) in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Task.cs:line 2372
at System.Threading.ExecutionContext.RunFromThreadPoolDispatchLoop(Thread threadPoolThread, ExecutionContext executionContext, ContextCallback callback, Object state) in /_/src/System.Private.CoreLib/shared/System/Threading/ExecutionContext.cs:line 289
at System.Threading.Tasks.Task.ExecuteWithThreadLocal(Task& currentTaskSlot, Thread threadPoolThread) in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Task.cs:line 2333
at System.Threading.Tasks.Task.ExecuteEntryUnsafe(Thread threadPoolThread) in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Task.cs:line 2271
at System.Threading.Tasks.Task.ExecuteFromThreadPool(Thread threadPoolThread) in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Task.cs:line 2256
at System.Threading.ThreadPoolWorkQueue.Dispatch() in /_/src/System.Private.CoreLib/shared/System/Threading/ThreadPool.cs:line 664
at System.Threading._ThreadPoolWaitCallback.PerformWaitCallback() in /_/src/System.Private.CoreLib/src/System/Threading/ThreadPool.CoreCLR.cs:line 29
exit code -2146232797
```
|
1.0
|
Test Failed: Process terminated. Assertion failed. "!_fileHandle.IsClosed" - **Detail:**
https://helix.dot.net/api/2019-06-17/jobs/0c15185a-2331-4dcf-978c-cd7c4a4b3189/workitems/System.Diagnostics.Process.Tests/console
**Log:**
```
C:\dotnetbuild\work\0c15185a-2331-4dcf-978c-cd7c4a4b3189\Work\2eeb62f3-774b-43f1-94ac-c844f873d662\Exec>"C:\dotnetbuild\work\0c15185a-2331-4dcf-978c-cd7c4a4b3189\Payload\dotnet.exe" exec --runtimeconfig System.Diagnostics.Process.Tests.runtimeconfig.json xunit.console.dll System.Diagnostics.Process.Tests.dll -xml testResults.xml -nologo -nocolor -notrait category=nonnetcoreapptests -notrait category=nonwindowstests -notrait category=IgnoreForCI -notrait category=OuterLoop -notrait category=failing @C:\dotnetbuild\work\0c15185a-2331-4dcf-978c-cd7c4a4b3189\Payload\CoreFX.issues.rsp
Discovering: System.Diagnostics.Process.Tests (method display = ClassAndMethod, method display options = None)
Discovered: System.Diagnostics.Process.Tests (found 221 of 241 test cases)
Starting: System.Diagnostics.Process.Tests (parallel test collections = on, max threads = 2)
System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_WebPage [SKIP]
Manual test
System.Diagnostics.Tests.ProcessStartInfoTests.ShellExecute_Nano_Fails_Start [SKIP]
Condition(s) not met: "IsWindowsNanoServer"
Process terminated. Assertion failed.
!_fileHandle.IsClosed
at System.IO.FileStream.AssertCanRead() in /_/src/System.Private.CoreLib/shared/System/IO/FileStream.Windows.cs:line 482
at System.IO.FileStream.ReadNative(Span`1 buffer) in /_/src/System.Private.CoreLib/shared/System/IO/FileStream.Windows.cs:line 497
at System.IO.FileStream.ReadSpan(Span`1 destination) in /_/src/System.Private.CoreLib/shared/System/IO/FileStream.Windows.cs:line 429
at System.IO.FileStream.Read(Byte[] array, Int32 offset, Int32 count) in /_/src/System.Private.CoreLib/shared/System/IO/FileStream.cs:line 304
at System.IO.Stream.<>c.<BeginReadInternal>b__43_0(Object <p0>) in /_/src/System.Private.CoreLib/shared/System/IO/Stream.cs:line 303
at System.Threading.Tasks.Task`1.InnerInvoke() in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Future.cs:line 512
at System.Threading.Tasks.Task.<>c.<.cctor>b__274_0(Object obj) in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Task.cs:line 2372
at System.Threading.ExecutionContext.RunFromThreadPoolDispatchLoop(Thread threadPoolThread, ExecutionContext executionContext, ContextCallback callback, Object state) in /_/src/System.Private.CoreLib/shared/System/Threading/ExecutionContext.cs:line 289
at System.Threading.Tasks.Task.ExecuteWithThreadLocal(Task& currentTaskSlot, Thread threadPoolThread) in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Task.cs:line 2333
at System.Threading.Tasks.Task.ExecuteEntryUnsafe(Thread threadPoolThread) in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Task.cs:line 2271
at System.Threading.Tasks.Task.ExecuteFromThreadPool(Thread threadPoolThread) in /_/src/System.Private.CoreLib/shared/System/Threading/Tasks/Task.cs:line 2256
at System.Threading.ThreadPoolWorkQueue.Dispatch() in /_/src/System.Private.CoreLib/shared/System/Threading/ThreadPool.cs:line 664
at System.Threading._ThreadPoolWaitCallback.PerformWaitCallback() in /_/src/System.Private.CoreLib/src/System/Threading/ThreadPool.CoreCLR.cs:line 29
exit code -2146232797
```
|
process
|
test failed process terminated assertion failed filehandle isclosed detail log c dotnetbuild work work exec c dotnetbuild work payload dotnet exe exec runtimeconfig system diagnostics process tests runtimeconfig json xunit console dll system diagnostics process tests dll xml testresults xml nologo nocolor notrait category nonnetcoreapptests notrait category nonwindowstests notrait category ignoreforci notrait category outerloop notrait category failing c dotnetbuild work payload corefx issues rsp discovering system diagnostics process tests method display classandmethod method display options none discovered system diagnostics process tests found of test cases starting system diagnostics process tests parallel test collections on max threads system diagnostics tests processstartinfotests startinfo webpage manual test system diagnostics tests processstartinfotests shellexecute nano fails start condition s not met iswindowsnanoserver process terminated assertion failed filehandle isclosed at system io filestream assertcanread in src system private corelib shared system io filestream windows cs line at system io filestream readnative span buffer in src system private corelib shared system io filestream windows cs line at system io filestream readspan span destination in src system private corelib shared system io filestream windows cs line at system io filestream read byte array offset count in src system private corelib shared system io filestream cs line at system io stream c b object in src system private corelib shared system io stream cs line at system threading tasks task innerinvoke in src system private corelib shared system threading tasks future cs line at system threading tasks task c b object obj in src system private corelib shared system threading tasks task cs line at system threading executioncontext runfromthreadpooldispatchloop thread threadpoolthread executioncontext executioncontext contextcallback callback object state in src system private corelib shared system threading executioncontext cs line at system threading tasks task executewiththreadlocal task currenttaskslot thread threadpoolthread in src system private corelib shared system threading tasks task cs line at system threading tasks task executeentryunsafe thread threadpoolthread in src system private corelib shared system threading tasks task cs line at system threading tasks task executefromthreadpool thread threadpoolthread in src system private corelib shared system threading tasks task cs line at system threading threadpoolworkqueue dispatch in src system private corelib shared system threading threadpool cs line at system threading threadpoolwaitcallback performwaitcallback in src system private corelib src system threading threadpool coreclr cs line exit code
| 1
|
10,234
| 13,096,025,883
|
IssuesEvent
|
2020-08-03 15:02:04
|
ZbayApp/zbay
|
https://api.github.com/repos/ZbayApp/zbay
|
closed
|
Automate Windows deploy process
|
dev process
|
Automate Windows build, signing with cert if possible. Look into #148 first.
|
1.0
|
Automate Windows deploy process - Automate Windows build, signing with cert if possible. Look into #148 first.
|
process
|
automate windows deploy process automate windows build signing with cert if possible look into first
| 1
|
241,155
| 26,256,663,675
|
IssuesEvent
|
2023-01-06 01:46:12
|
EmpoHQ/empo.im
|
https://api.github.com/repos/EmpoHQ/empo.im
|
opened
|
CVE-2022-24999 (High) detected in qs-6.10.1.tgz
|
security vulnerability
|
## CVE-2022-24999 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-6.10.1.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.10.1.tgz">https://registry.npmjs.org/qs/-/qs-6.10.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- nuxt-2.15.6.tgz (Root Library)
- telemetry-1.3.3.tgz
- git-url-parse-11.4.4.tgz
- git-up-4.0.2.tgz
- parse-url-5.0.2.tgz
- parse-path-4.0.3.tgz
- :x: **qs-6.10.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/EmpoHQ/empo.im/commit/488d7a9b0ad7df016dd006a95c3b06f4883d8561">488d7a9b0ad7df016dd006a95c3b06f4883d8561</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
qs before 6.10.3, as used in Express before 4.17.3 and other products, allows attackers to cause a Node process hang for an Express application because an __ proto__ key can be used. In many typical Express use cases, an unauthenticated remote attacker can place the attack payload in the query string of the URL that is used to visit the application, such as a[__proto__]=b&a[__proto__]&a[length]=100000000. The fix was backported to qs 6.9.7, 6.8.3, 6.7.3, 6.6.1, 6.5.3, 6.4.1, 6.3.3, and 6.2.4 (and therefore Express 4.17.3, which has "deps: qs@6.9.7" in its release description, is not vulnerable).
<p>Publish Date: 2022-11-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24999>CVE-2022-24999</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-24999">https://www.cve.org/CVERecord?id=CVE-2022-24999</a></p>
<p>Release Date: 2022-11-26</p>
<p>Fix Resolution (qs): 6.10.3</p>
<p>Direct dependency fix Resolution (nuxt): 2.15.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-24999 (High) detected in qs-6.10.1.tgz - ## CVE-2022-24999 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-6.10.1.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.10.1.tgz">https://registry.npmjs.org/qs/-/qs-6.10.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- nuxt-2.15.6.tgz (Root Library)
- telemetry-1.3.3.tgz
- git-url-parse-11.4.4.tgz
- git-up-4.0.2.tgz
- parse-url-5.0.2.tgz
- parse-path-4.0.3.tgz
- :x: **qs-6.10.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/EmpoHQ/empo.im/commit/488d7a9b0ad7df016dd006a95c3b06f4883d8561">488d7a9b0ad7df016dd006a95c3b06f4883d8561</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
qs before 6.10.3, as used in Express before 4.17.3 and other products, allows attackers to cause a Node process hang for an Express application because an __ proto__ key can be used. In many typical Express use cases, an unauthenticated remote attacker can place the attack payload in the query string of the URL that is used to visit the application, such as a[__proto__]=b&a[__proto__]&a[length]=100000000. The fix was backported to qs 6.9.7, 6.8.3, 6.7.3, 6.6.1, 6.5.3, 6.4.1, 6.3.3, and 6.2.4 (and therefore Express 4.17.3, which has "deps: qs@6.9.7" in its release description, is not vulnerable).
<p>Publish Date: 2022-11-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24999>CVE-2022-24999</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-24999">https://www.cve.org/CVERecord?id=CVE-2022-24999</a></p>
<p>Release Date: 2022-11-26</p>
<p>Fix Resolution (qs): 6.10.3</p>
<p>Direct dependency fix Resolution (nuxt): 2.15.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in qs tgz cve high severity vulnerability vulnerable library qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href path to dependency file package json path to vulnerable library node modules qs package json dependency hierarchy nuxt tgz root library telemetry tgz git url parse tgz git up tgz parse url tgz parse path tgz x qs tgz vulnerable library found in head commit a href found in base branch main vulnerability details qs before as used in express before and other products allows attackers to cause a node process hang for an express application because an proto key can be used in many typical express use cases an unauthenticated remote attacker can place the attack payload in the query string of the url that is used to visit the application such as a b a a the fix was backported to qs and and therefore express which has deps qs in its release description is not vulnerable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution qs direct dependency fix resolution nuxt step up your open source security game with mend
| 0
|
315,663
| 23,590,886,821
|
IssuesEvent
|
2022-08-23 15:05:59
|
arturo-lang/arturo
|
https://api.github.com/repos/arturo-lang/arturo
|
closed
|
[Reflection\complex?] add documentation example
|
documentation library todo easy
|
[Reflection\complex?] add documentation example
https://github.com/arturo-lang/arturo/blob/3a89302d1b1f33da097ed6e7bd6dfb05fc338482/src/library/Reflection.nim#L292
```text
builtin "color?",
alias = unaliased,
rule = PrefixPrecedence,
description = "checks if given value is of type :color",
args = {
"value" : {Any}
},
attrs = NoAttrs,
returns = {Logical},
# TODO(Reflection\color?) add documentation example
# labels: library, documentation, easy
example = """
""":
##########################################################
push(newLogical(x.kind==Color))
builtin "complex?",
alias = unaliased,
rule = PrefixPrecedence,
description = "checks if given value is of type :complex",
args = {
"value" : {Any}
},
attrs = NoAttrs,
returns = {Logical},
# TODO(Reflection\complex?) add documentation example
# labels: library, documentation, easy
example = """
""":
##########################################################
push(newLogical(x.kind==Complex))
builtin "database?",
alias = unaliased,
rule = PrefixPrecedence,
```
618f5b7315934d100b1650e62313cf8d69700ce6
|
1.0
|
[Reflection\complex?] add documentation example - [Reflection\complex?] add documentation example
https://github.com/arturo-lang/arturo/blob/3a89302d1b1f33da097ed6e7bd6dfb05fc338482/src/library/Reflection.nim#L292
```text
builtin "color?",
alias = unaliased,
rule = PrefixPrecedence,
description = "checks if given value is of type :color",
args = {
"value" : {Any}
},
attrs = NoAttrs,
returns = {Logical},
# TODO(Reflection\color?) add documentation example
# labels: library, documentation, easy
example = """
""":
##########################################################
push(newLogical(x.kind==Color))
builtin "complex?",
alias = unaliased,
rule = PrefixPrecedence,
description = "checks if given value is of type :complex",
args = {
"value" : {Any}
},
attrs = NoAttrs,
returns = {Logical},
# TODO(Reflection\complex?) add documentation example
# labels: library, documentation, easy
example = """
""":
##########################################################
push(newLogical(x.kind==Complex))
builtin "database?",
alias = unaliased,
rule = PrefixPrecedence,
```
618f5b7315934d100b1650e62313cf8d69700ce6
|
non_process
|
add documentation example add documentation example text builtin color alias unaliased rule prefixprecedence description checks if given value is of type color args value any attrs noattrs returns logical todo reflection color add documentation example labels library documentation easy example push newlogical x kind color builtin complex alias unaliased rule prefixprecedence description checks if given value is of type complex args value any attrs noattrs returns logical todo reflection complex add documentation example labels library documentation easy example push newlogical x kind complex builtin database alias unaliased rule prefixprecedence
| 0
|
10,186
| 13,044,162,864
|
IssuesEvent
|
2020-07-29 03:47:37
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `RoundWithFracReal` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `RoundWithFracReal` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @mapleFU
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `RoundWithFracReal` from TiDB -
## Description
Port the scalar function `RoundWithFracReal` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @mapleFU
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function roundwithfracreal from tidb description port the scalar function roundwithfracreal from tidb to coprocessor score mentor s maplefu recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
207,244
| 7,126,399,097
|
IssuesEvent
|
2018-01-20 09:45:55
|
wordpress-mobile/AztecEditor-Android
|
https://api.github.com/repos/wordpress-mobile/AztecEditor-Android
|
reopened
|
Repeating characters in title
|
bug high priority
|
### Expected
If I type "ABCD" in the title box, the title should say "ABCD"
### Observed
I type A, it says A. I type B, the title becomes "AAB". I type C, I get "AABABC", etc.
You can see a video and screenshot of the issue in action here:
https://cloudup.com/cE18S9bfbWY
Also, note if I add a space, it resets, as you can see in the screenshot. Pressing backspace also does wonky things, but I can eventually clear out the title. I can type normally in the content area and paste to the title area.
### Reproduced
1. Make a new post using the beta editor
2. Type a title
Note: We were not able to duplicate this on all phones.
#### Tested
Tested on LG G4, Android 6.0 with alpha 89
|
1.0
|
Repeating characters in title - ### Expected
If I type "ABCD" in the title box, the title should say "ABCD"
### Observed
I type A, it says A. I type B, the title becomes "AAB". I type C, I get "AABABC", etc.
You can see a video and screenshot of the issue in action here:
https://cloudup.com/cE18S9bfbWY
Also, note if I add a space, it resets, as you can see in the screenshot. Pressing backspace also does wonky things, but I can eventually clear out the title. I can type normally in the content area and paste to the title area.
### Reproduced
1. Make a new post using the beta editor
2. Type a title
Note: We were not able to duplicate this on all phones.
#### Tested
Tested on LG G4, Android 6.0 with alpha 89
|
non_process
|
repeating characters in title expected if i type abcd in the title box the title should say abcd observed i type a it says a i type b the title becomes aab i type c i get aababc etc you can see a video and screenshot of the issue in action here also note if i add a space it resets as you can see in the screenshot pressing backspace also does wonky things but i can eventually clear out the title i can type normally in the content area and paste to the title area reproduced make a new post using the beta editor type a title note we were not able to duplicate this on all phones tested tested on lg android with alpha
| 0
|
16,821
| 22,060,943,988
|
IssuesEvent
|
2022-05-30 17:44:09
|
bitPogo/kmock
|
https://api.github.com/repos/bitPogo/kmock
|
closed
|
Relaxation fails for Generics
|
bug kmock-processor
|
## Description
<!--- Provide a detailed introduction to the issue itself, and why you consider it to be a bug -->
Currently Relaxation for Interfaces with generics causes an Compiler Error.
This only applies for inline functions. (The need additional arguments and cannot be inline in this case)
|
1.0
|
Relaxation fails for Generics - ## Description
<!--- Provide a detailed introduction to the issue itself, and why you consider it to be a bug -->
Currently Relaxation for Interfaces with generics causes an Compiler Error.
This only applies for inline functions. (The need additional arguments and cannot be inline in this case)
|
process
|
relaxation fails for generics description currently relaxation for interfaces with generics causes an compiler error this only applies for inline functions the need additional arguments and cannot be inline in this case
| 1
|
345,270
| 30,794,983,040
|
IssuesEvent
|
2023-07-31 19:07:10
|
boltlabs-inc/tss-ecdsa
|
https://api.github.com/repos/boltlabs-inc/tss-ecdsa
|
closed
|
Write tests for PiSchProof
|
tests
|
A `PiSchProof` has two elements and must satisfy one constraint. Some ideas for tests:
- Choose bad secret input `x` (not the dlog of `X`)
- Swap the commitment `A` for a random element
- Swap the response `z` for a random element
- Try bad common input, e.g. where `g` is not a generator of `G`
- Try verifying with different common input that was used to generate the proof
Related issue: #52
|
1.0
|
Write tests for PiSchProof - A `PiSchProof` has two elements and must satisfy one constraint. Some ideas for tests:
- Choose bad secret input `x` (not the dlog of `X`)
- Swap the commitment `A` for a random element
- Swap the response `z` for a random element
- Try bad common input, e.g. where `g` is not a generator of `G`
- Try verifying with different common input that was used to generate the proof
Related issue: #52
|
non_process
|
write tests for pischproof a pischproof has two elements and must satisfy one constraint some ideas for tests choose bad secret input x not the dlog of x swap the commitment a for a random element swap the response z for a random element try bad common input e g where g is not a generator of g try verifying with different common input that was used to generate the proof related issue
| 0
|
57,650
| 7,087,363,575
|
IssuesEvent
|
2018-01-11 17:33:05
|
Nersent/wexond
|
https://api.github.com/repos/Nersent/wexond
|
closed
|
What should be in a new tab page?
|
design question
|
What should be in a new tab page? Please give me your suggestions below. I'm wondering if add bookmarks cards or recently visited websites or something different.
|
1.0
|
What should be in a new tab page? - What should be in a new tab page? Please give me your suggestions below. I'm wondering if add bookmarks cards or recently visited websites or something different.
|
non_process
|
what should be in a new tab page what should be in a new tab page please give me your suggestions below i m wondering if add bookmarks cards or recently visited websites or something different
| 0
|
317,094
| 23,663,707,592
|
IssuesEvent
|
2022-08-26 18:18:30
|
dockstore/dockstore
|
https://api.github.com/repos/dockstore/dockstore
|
closed
|
Remove 1.13 stuff in documentation stable branch
|
bug documentation
|
**Describe the bug**
On our documentation repo, develop got merged/rebased/cherrypicked into the hotfix/1.12.2 branch. There's 99% chance I'm the guy who did this, so I'm going to be the guy who fixes it. The rebase/merge/whatever doesn't break anything, but it does include a few small details about the next release (such as which version of Cromwell 1.13 will be using).
**To Reproduce**
You can see 1.13 referenced in https://docs.dockstore.org/en/stable/advanced-topics/dockstore-cli/advanced-features.html?highlight=cromwell#wdl-launcher-configuration
**Expected behavior**
~Ash is banned from using `git rebase`~ 1.13 should not be referenced on master/stable
**Additional context**
I found this issue this afternoon. Charles and I did some digging, using the addition of images to index.rst (which should be on only develop due to [not being on 1.12's release branch](https://github.com/dockstore/dockstore-documentation/pull/170#issuecomment-1075628831)) as our clue. Charles found the comment "merged commit 9748946" in https://github.com/dockstore/dockstore-documentation/pull/206, so that's probably where the rebase happened. Knowing me, I did it to bring in a linkcheck fix.
┆Issue is synchronized with this [Jira Story](https://ucsc-cgl.atlassian.net/browse/DOCK-2206)
┆fixVersions: Dockstore 1.12.x
┆friendlyId: DOCK-2206
┆sprint: 94- Vampire Pleco
┆taskType: Story
|
1.0
|
Remove 1.13 stuff in documentation stable branch - **Describe the bug**
On our documentation repo, develop got merged/rebased/cherrypicked into the hotfix/1.12.2 branch. There's 99% chance I'm the guy who did this, so I'm going to be the guy who fixes it. The rebase/merge/whatever doesn't break anything, but it does include a few small details about the next release (such as which version of Cromwell 1.13 will be using).
**To Reproduce**
You can see 1.13 referenced in https://docs.dockstore.org/en/stable/advanced-topics/dockstore-cli/advanced-features.html?highlight=cromwell#wdl-launcher-configuration
**Expected behavior**
~Ash is banned from using `git rebase`~ 1.13 should not be referenced on master/stable
**Additional context**
I found this issue this afternoon. Charles and I did some digging, using the addition of images to index.rst (which should be on only develop due to [not being on 1.12's release branch](https://github.com/dockstore/dockstore-documentation/pull/170#issuecomment-1075628831)) as our clue. Charles found the comment "merged commit 9748946" in https://github.com/dockstore/dockstore-documentation/pull/206, so that's probably where the rebase happened. Knowing me, I did it to bring in a linkcheck fix.
┆Issue is synchronized with this [Jira Story](https://ucsc-cgl.atlassian.net/browse/DOCK-2206)
┆fixVersions: Dockstore 1.12.x
┆friendlyId: DOCK-2206
┆sprint: 94- Vampire Pleco
┆taskType: Story
|
non_process
|
remove stuff in documentation stable branch describe the bug on our documentation repo develop got merged rebased cherrypicked into the hotfix branch there s chance i m the guy who did this so i m going to be the guy who fixes it the rebase merge whatever doesn t break anything but it does include a few small details about the next release such as which version of cromwell will be using to reproduce you can see referenced in expected behavior ash is banned from using git rebase should not be referenced on master stable additional context i found this issue this afternoon charles and i did some digging using the addition of images to index rst which should be on only develop due to as our clue charles found the comment merged commit in so that s probably where the rebase happened knowing me i did it to bring in a linkcheck fix ┆issue is synchronized with this ┆fixversions dockstore x ┆friendlyid dock ┆sprint vampire pleco ┆tasktype story
| 0
|
20,148
| 26,695,474,296
|
IssuesEvent
|
2023-01-27 09:59:22
|
ppy/osu-web
|
https://api.github.com/repos/ppy/osu-web
|
closed
|
Error downloading old beatmap
|
area:beatmap-processing
|
This map is unable to download: https://osu.ppy.sh/beatmapsets/13758#osu/50686
If you try download from beatmap page, it returns 500 error.
But if you try doing it form beatmap search page it returns this:
```
Warning: sprintf(): Too few arguments in /var/www/html/S3.php on line 742
Warning: Cannot modify header information - headers already sent by (output started at /var/www/html/S3.php:742) in /var/www/html/d.php on line 109
```
|
1.0
|
Error downloading old beatmap - This map is unable to download: https://osu.ppy.sh/beatmapsets/13758#osu/50686
If you try download from beatmap page, it returns 500 error.
But if you try doing it form beatmap search page it returns this:
```
Warning: sprintf(): Too few arguments in /var/www/html/S3.php on line 742
Warning: Cannot modify header information - headers already sent by (output started at /var/www/html/S3.php:742) in /var/www/html/d.php on line 109
```
|
process
|
error downloading old beatmap this map is unable to download if you try download from beatmap page it returns error but if you try doing it form beatmap search page it returns this warning sprintf too few arguments in var www html php on line warning cannot modify header information headers already sent by output started at var www html php in var www html d php on line
| 1
|
79,228
| 15,166,238,917
|
IssuesEvent
|
2021-02-12 16:06:09
|
sourcegraph/sourcegraph
|
https://api.github.com/repos/sourcegraph/sourcegraph
|
closed
|
Fix pings for lsif WAUs
|
team/code-intelligence
|
We currently aren't getting code intelligence WAUs.
<img width="1482" alt="Screenshot 2021-02-02 at 16 38 10" src="https://user-images.githubusercontent.com/1657213/106653045-99840d00-6596-11eb-97a7-8562c1feb4ab.png">
|
1.0
|
Fix pings for lsif WAUs - We currently aren't getting code intelligence WAUs.
<img width="1482" alt="Screenshot 2021-02-02 at 16 38 10" src="https://user-images.githubusercontent.com/1657213/106653045-99840d00-6596-11eb-97a7-8562c1feb4ab.png">
|
non_process
|
fix pings for lsif waus we currently aren t getting code intelligence waus img width alt screenshot at src
| 0
|
416,243
| 28,074,149,565
|
IssuesEvent
|
2023-03-29 21:35:03
|
sujuka99/dotfiles
|
https://api.github.com/repos/sujuka99/dotfiles
|
reopened
|
Document how to use repo
|
documentation dotfiles
|
Document:
- [x] Clone repo
- [x] Configure git for bare repository
- [ ] List requirements
|
1.0
|
Document how to use repo - Document:
- [x] Clone repo
- [x] Configure git for bare repository
- [ ] List requirements
|
non_process
|
document how to use repo document clone repo configure git for bare repository list requirements
| 0
|
682,863
| 23,360,272,544
|
IssuesEvent
|
2022-08-10 11:03:33
|
microsoft/fluentui
|
https://api.github.com/repos/microsoft/fluentui
|
opened
|
[Feature]: Fluent React v9 Preview Banner 🎉
|
Type: Feature Priority 1: High Area: Website Partner Ask
|
### Library
React Components / v9 (@fluentui/react-components)
### Describe the feature that you would like added
In mid to late August the Microsite is planned to go live to the public. To let our customers know of the new site we would like to have a top banner added to the existing [Fluent Design System](https://www.microsoft.com/design/fluent/) to direct customers to the Microsite.
### Solution

### Asset Resources
- For more information into the styling and design, please, view the `website banner` pages in the [Figma file](https://www.figma.com/file/Jm7YiqB9W4lyMA3NxPlK9R/V9-Build-website?node-id=958%3A1908)
- The `Go To Fluent React v9` button in the banner should direct to https://react.fluentui.dev/
### Have you discussed this feature with our team
yes (Design and Website Team)
### Additional context
_No response_
### Validations
- [X] Check that there isn't already an issue that request the same feature to avoid creating a duplicate.
|
1.0
|
[Feature]: Fluent React v9 Preview Banner 🎉 - ### Library
React Components / v9 (@fluentui/react-components)
### Describe the feature that you would like added
In mid to late August the Microsite is planned to go live to the public. To let our customers know of the new site we would like to have a top banner added to the existing [Fluent Design System](https://www.microsoft.com/design/fluent/) to direct customers to the Microsite.
### Solution

### Asset Resources
- For more information into the styling and design, please, view the `website banner` pages in the [Figma file](https://www.figma.com/file/Jm7YiqB9W4lyMA3NxPlK9R/V9-Build-website?node-id=958%3A1908)
- The `Go To Fluent React v9` button in the banner should direct to https://react.fluentui.dev/
### Have you discussed this feature with our team
yes (Design and Website Team)
### Additional context
_No response_
### Validations
- [X] Check that there isn't already an issue that request the same feature to avoid creating a duplicate.
|
non_process
|
fluent react preview banner 🎉 library react components fluentui react components describe the feature that you would like added in mid to late august the microsite is planned to go live to the public to let our customers know of the new site we would like to have a top banner added to the existing to direct customers to the microsite solution asset resources for more information into the styling and design please view the website banner pages in the the go to fluent react button in the banner should direct to have you discussed this feature with our team yes design and website team additional context no response validations check that there isn t already an issue that request the same feature to avoid creating a duplicate
| 0
|
654,114
| 21,638,101,550
|
IssuesEvent
|
2022-05-05 15:55:14
|
internetarchive/openlibrary
|
https://api.github.com/repos/internetarchive/openlibrary
|
opened
|
Enriching Reading Log data exports (more fields)
|
Theme: Reading Log export Module: Data dumps Priority: 2 Affects: Data Lead: @mekarpeles
|
* Currently Reading Log dumps are limited to IDs because large Reading Logs (500+ items) can cause performance issues aggregating results in memory and and serving as large files. Currently, Lists impose an export limit if there are more than ~1k items
* The manual work around is to look over each Reading Log entry (e.g. OL16801714W) and to fetch the json for the corresponding item from Open Library API or the data dumps: https://openlibrary.org/works/OL16801714W.json
* 2 possible solutions to this problem are:
* Give an option to download all work data but limit to e.g. 500 results (and show a notice, similar to lists page)
* Offer a js option which allows one (from the OL website) to download pages of ~500 results from Reading Log (calculated based on total number of patron's books on Reading Log)
Questions for community:
* What format of data would you prefer/expect? CSV? JSON?
|
1.0
|
Enriching Reading Log data exports (more fields) - * Currently Reading Log dumps are limited to IDs because large Reading Logs (500+ items) can cause performance issues aggregating results in memory and and serving as large files. Currently, Lists impose an export limit if there are more than ~1k items
* The manual work around is to look over each Reading Log entry (e.g. OL16801714W) and to fetch the json for the corresponding item from Open Library API or the data dumps: https://openlibrary.org/works/OL16801714W.json
* 2 possible solutions to this problem are:
* Give an option to download all work data but limit to e.g. 500 results (and show a notice, similar to lists page)
* Offer a js option which allows one (from the OL website) to download pages of ~500 results from Reading Log (calculated based on total number of patron's books on Reading Log)
Questions for community:
* What format of data would you prefer/expect? CSV? JSON?
|
non_process
|
enriching reading log data exports more fields currently reading log dumps are limited to ids because large reading logs items can cause performance issues aggregating results in memory and and serving as large files currently lists impose an export limit if there are more than items the manual work around is to look over each reading log entry e g and to fetch the json for the corresponding item from open library api or the data dumps possible solutions to this problem are give an option to download all work data but limit to e g results and show a notice similar to lists page offer a js option which allows one from the ol website to download pages of results from reading log calculated based on total number of patron s books on reading log questions for community what format of data would you prefer expect csv json
| 0
|
10,482
| 13,252,913,227
|
IssuesEvent
|
2020-08-20 06:33:10
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
Deadline check in table / index scan executor
|
sig/coprocessor status/discussion
|
## Feature Request
Currently we check deadline in DAG next(). However there are many scenarios that this doesn't work, i.e. there is a selection executor that filtered a lot of lines, or an aggregation executor that emits a row only after collecting all rows.
By checking deadline in the innermost executor, i.e. table scan or index scan executor, we can avoid this issue and make deadline check work correctly in all scenarios.
|
1.0
|
Deadline check in table / index scan executor - ## Feature Request
Currently we check deadline in DAG next(). However there are many scenarios that this doesn't work, i.e. there is a selection executor that filtered a lot of lines, or an aggregation executor that emits a row only after collecting all rows.
By checking deadline in the innermost executor, i.e. table scan or index scan executor, we can avoid this issue and make deadline check work correctly in all scenarios.
|
process
|
deadline check in table index scan executor feature request currently we check deadline in dag next however there are many scenarios that this doesn t work i e there is a selection executor that filtered a lot of lines or an aggregation executor that emits a row only after collecting all rows by checking deadline in the innermost executor i e table scan or index scan executor we can avoid this issue and make deadline check work correctly in all scenarios
| 1
|
6,835
| 9,978,102,343
|
IssuesEvent
|
2019-07-09 18:59:30
|
TileDB-Inc/TileDB
|
https://api.github.com/repos/TileDB-Inc/TileDB
|
closed
|
Lock release with multiple contexts
|
bug process safety vfs
|
The following program demonstrates a failure to release the lockfile when a local URI is opened from multiple contexts.
<details>
```
import tiledb
import numpy as np
import shutil, os
if os.path.isdir("C:/tmp/foo"):
shutil.rmtree("C:/tmp/foo")
ctx = tiledb.Ctx()
ctx2= tiledb.Ctx()
dom = tiledb.Domain(tiledb.Dim(name="rows", domain=(1, 4), tile=4, dtype=np.int32, ctx=ctx),
tiledb.Dim(name="cols", domain=(1, 4), tile=4, dtype=np.int32, ctx=ctx),
ctx=ctx)
schema = tiledb.ArraySchema(domain=dom, sparse=False,
attrs=[tiledb.Attr(name="a", dtype=np.int32)], ctx=ctx)
uri = "C:/tmp/foo"
tiledb.Array.create("C:/tmp/foo", schema)
a = tiledb.DenseArray(uri, 'w', ctx=ctx)
a[:] = np.random.rand(4,4)
b = tiledb.DenseArray(uri, 'r', ctx=ctx2)
a.close()
b.close()
# the call below causes a PermissionDenied error because the handle for
# C:/tmp/foo/__lock.tdb is still open.
shutil.rmtree(uri)
```
</details>
The issue arises due to the following sequence:
- array 1, ctx 1, the storage manager calls `OpenArray::file_lock`, which calls `vfs->filelock_lock` w/ a reference to the `filelock_` member:
https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/storage_manager/open_array.cc#L93-L99
- ctx 1 increments the global counter (URI count ==1) and calls the platform lock, which sets the `OpenArray::filelock_` member passed above to the FD:
https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/filesystem/vfs.cc#L365-L372
- array 2, ctx 2, increments the counter in `VFS::filelock_lock` (now URI count == 2) but *does not* call the platform file lock, so the `OpenArray::filelock_` member for this instance continues to have the value `INVALID_FILELOCK`.
https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/filesystem/vfs.cc#L367
- when `tiledb_array_close` is called the first time, the storage manager calls `OpenArray::file_unlock`
https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/storage_manager/open_array.cc#L101-L108
- which calls `VFS::file_unlock` and decrements the counter -- but doesn't unlock because the URI global count == 1.
https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/filesystem/vfs.cc#L408-L412
- when `tiledb_array_close` is called the 2nd time, we reach `OpenArray::file_unlock` again, but the `filelock_` member for this `OpenArray` instance was not initialized previously, so the call to `vfs->file_unlock` is skipped, leaving the handle open:
https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/storage_manager/open_array.cc#L101-L108
|
1.0
|
Lock release with multiple contexts - The following program demonstrates a failure to release the lockfile when a local URI is opened from multiple contexts.
<details>
```
import tiledb
import numpy as np
import shutil, os
if os.path.isdir("C:/tmp/foo"):
shutil.rmtree("C:/tmp/foo")
ctx = tiledb.Ctx()
ctx2= tiledb.Ctx()
dom = tiledb.Domain(tiledb.Dim(name="rows", domain=(1, 4), tile=4, dtype=np.int32, ctx=ctx),
tiledb.Dim(name="cols", domain=(1, 4), tile=4, dtype=np.int32, ctx=ctx),
ctx=ctx)
schema = tiledb.ArraySchema(domain=dom, sparse=False,
attrs=[tiledb.Attr(name="a", dtype=np.int32)], ctx=ctx)
uri = "C:/tmp/foo"
tiledb.Array.create("C:/tmp/foo", schema)
a = tiledb.DenseArray(uri, 'w', ctx=ctx)
a[:] = np.random.rand(4,4)
b = tiledb.DenseArray(uri, 'r', ctx=ctx2)
a.close()
b.close()
# the call below causes a PermissionDenied error because the handle for
# C:/tmp/foo/__lock.tdb is still open.
shutil.rmtree(uri)
```
</details>
The issue arises due to the following sequence:
- array 1, ctx 1, the storage manager calls `OpenArray::file_lock`, which calls `vfs->filelock_lock` w/ a reference to the `filelock_` member:
https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/storage_manager/open_array.cc#L93-L99
- ctx 1 increments the global counter (URI count ==1) and calls the platform lock, which sets the `OpenArray::filelock_` member passed above to the FD:
https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/filesystem/vfs.cc#L365-L372
- array 2, ctx 2, increments the counter in `VFS::filelock_lock` (now URI count == 2) but *does not* call the platform file lock, so the `OpenArray::filelock_` member for this instance continues to have the value `INVALID_FILELOCK`.
https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/filesystem/vfs.cc#L367
- when `tiledb_array_close` is called the first time, the storage manager calls `OpenArray::file_unlock`
https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/storage_manager/open_array.cc#L101-L108
- which calls `VFS::file_unlock` and decrements the counter -- but doesn't unlock because the URI global count == 1.
https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/filesystem/vfs.cc#L408-L412
- when `tiledb_array_close` is called the 2nd time, we reach `OpenArray::file_unlock` again, but the `filelock_` member for this `OpenArray` instance was not initialized previously, so the call to `vfs->file_unlock` is skipped, leaving the handle open:
https://github.com/TileDB-Inc/TileDB/blob/24417f5d1c1448f5ad59e8aa68625bace61b1ae0/tiledb/sm/storage_manager/open_array.cc#L101-L108
|
process
|
lock release with multiple contexts the following program demonstrates a failure to release the lockfile when a local uri is opened from multiple contexts import tiledb import numpy as np import shutil os if os path isdir c tmp foo shutil rmtree c tmp foo ctx tiledb ctx tiledb ctx dom tiledb domain tiledb dim name rows domain tile dtype np ctx ctx tiledb dim name cols domain tile dtype np ctx ctx ctx ctx schema tiledb arrayschema domain dom sparse false attrs ctx ctx uri c tmp foo tiledb array create c tmp foo schema a tiledb densearray uri w ctx ctx a np random rand b tiledb densearray uri r ctx a close b close the call below causes a permissiondenied error because the handle for c tmp foo lock tdb is still open shutil rmtree uri the issue arises due to the following sequence array ctx the storage manager calls openarray file lock which calls vfs filelock lock w a reference to the filelock member ctx increments the global counter uri count and calls the platform lock which sets the openarray filelock member passed above to the fd array ctx increments the counter in vfs filelock lock now uri count but does not call the platform file lock so the openarray filelock member for this instance continues to have the value invalid filelock when tiledb array close is called the first time the storage manager calls openarray file unlock which calls vfs file unlock and decrements the counter but doesn t unlock because the uri global count when tiledb array close is called the time we reach openarray file unlock again but the filelock member for this openarray instance was not initialized previously so the call to vfs file unlock is skipped leaving the handle open
| 1
|
99,060
| 30,268,069,549
|
IssuesEvent
|
2023-07-07 13:23:20
|
cms-sw/cmssw
|
https://api.github.com/repos/cms-sw/cmssw
|
closed
|
Build CMSSW_13_0_10
|
release-notes-requested release-announced release-build-request slc7_amd64_gcc11-finished el8_amd64_gcc11-finished el8_aarch64_gcc11-finished el8_ppc64le_gcc11-finished el9_amd64_gcc11-finished
|
To start the MC production campaign for Run3 2023
The build will go in parallel with the IB tests in CMSSW_13_0_X_2023-07-05-1100, to speed up the procedure: the release will get uploaded only if those tests show no issues.
|
1.0
|
Build CMSSW_13_0_10 - To start the MC production campaign for Run3 2023
The build will go in parallel with the IB tests in CMSSW_13_0_X_2023-07-05-1100, to speed up the procedure: the release will get uploaded only if those tests show no issues.
|
non_process
|
build cmssw to start the mc production campaign for the build will go in parallel with the ib tests in cmssw x to speed up the procedure the release will get uploaded only if those tests show no issues
| 0
|
16,624
| 21,678,126,955
|
IssuesEvent
|
2022-05-09 01:23:15
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
Add Super in America
|
suggested title in process
|
Please add as much of the following info as you can:
Title: Super in America
Type (film/tv show): reality TV show
Film or show in which it appears: The Boys
Is the parent film/show streaming anywhere? Yes - Amazon Prime
About when in the parent film/show does it appear? Throughout ep. 1x06; "The Innocents." Timestamps: 0:15 - 0:52, 4:24- 5:53 (peppered with behind the scenes footage), 12:30 - 14:30 (peppered with behind the scenes footage), 34:57 - 35:29, 37:54 - 39:33 (peppered with behind the scenes footage).
Actual footage of the film/show can be seen (yes/no)? Yes
Production Company: Vought Studios
|
1.0
|
Add Super in America - Please add as much of the following info as you can:
Title: Super in America
Type (film/tv show): reality TV show
Film or show in which it appears: The Boys
Is the parent film/show streaming anywhere? Yes - Amazon Prime
About when in the parent film/show does it appear? Throughout ep. 1x06; "The Innocents." Timestamps: 0:15 - 0:52, 4:24- 5:53 (peppered with behind the scenes footage), 12:30 - 14:30 (peppered with behind the scenes footage), 34:57 - 35:29, 37:54 - 39:33 (peppered with behind the scenes footage).
Actual footage of the film/show can be seen (yes/no)? Yes
Production Company: Vought Studios
|
process
|
add super in america please add as much of the following info as you can title super in america type film tv show reality tv show film or show in which it appears the boys is the parent film show streaming anywhere yes amazon prime about when in the parent film show does it appear throughout ep the innocents timestamps peppered with behind the scenes footage peppered with behind the scenes footage peppered with behind the scenes footage actual footage of the film show can be seen yes no yes production company vought studios
| 1
|
64,212
| 18,279,840,558
|
IssuesEvent
|
2021-10-05 00:47:17
|
microsoft/STL
|
https://api.github.com/repos/microsoft/STL
|
opened
|
P2418R2 Add Support For `std::generator`-like Types To `std::format`
|
cxx20 format defect report
|
[P2418R2](https://wg21.link/P2418R2) Add Support For `std::generator`-like Types To `std::format`
This increases the value of `__cpp_lib_format`, also increased by [P2372R3](https://wg21.link/P2372R3) (see #2237).
|
1.0
|
P2418R2 Add Support For `std::generator`-like Types To `std::format` - [P2418R2](https://wg21.link/P2418R2) Add Support For `std::generator`-like Types To `std::format`
This increases the value of `__cpp_lib_format`, also increased by [P2372R3](https://wg21.link/P2372R3) (see #2237).
|
non_process
|
add support for std generator like types to std format add support for std generator like types to std format this increases the value of cpp lib format also increased by see
| 0
|
348,012
| 31,392,433,350
|
IssuesEvent
|
2023-08-26 14:06:54
|
void-linux/void-packages
|
https://api.github.com/repos/void-linux/void-packages
|
opened
|
`docker build` fails with "ERROR: http: invalid Host header" with docker-buildx-0.10.3_3
|
bug needs-testing
|
### Is this a new report?
Yes
### System Info
Void 6.3.13_1 x86_64 AuthenticAMD notuptodate DFFF
### Package(s) Affected
docker-buildx-0.10.3_3
### Does a report exist for this bug with the project's home (upstream) and/or another distro?
https://github.com/moby/moby/issues/45935
### Expected behaviour
work with
```
docker-24.0.5_2
docker-cli-24.0.5_2
moby-24.0.5_2
```
and lower versions (tested 24.0.5_1)
### Actual behaviour
"ERROR: http: invalid Host header" when building
downgrade to docker-buildx-0.10.3_2 solves the issue. (docker-buildx-0.10.3_2 works with docker-24.0.5_2 and moby-24.0.5_2, but docker-buildx-0.10.3_3 doesn't)
### Steps to reproduce
1. `xbps-install -S docker-buildx-0.10.3_3`
2. `docker build .` fails
3. `xdowngrade /var/cache/xbps/docker-buildx-0.10.3_2.x86_64.xbps`
4. `docker build .` is fine
"FROM node:18-alpine" is the only line in Dockerfile
|
1.0
|
`docker build` fails with "ERROR: http: invalid Host header" with docker-buildx-0.10.3_3 - ### Is this a new report?
Yes
### System Info
Void 6.3.13_1 x86_64 AuthenticAMD notuptodate DFFF
### Package(s) Affected
docker-buildx-0.10.3_3
### Does a report exist for this bug with the project's home (upstream) and/or another distro?
https://github.com/moby/moby/issues/45935
### Expected behaviour
work with
```
docker-24.0.5_2
docker-cli-24.0.5_2
moby-24.0.5_2
```
and lower versions (tested 24.0.5_1)
### Actual behaviour
"ERROR: http: invalid Host header" when building
downgrade to docker-buildx-0.10.3_2 solves the issue. (docker-buildx-0.10.3_2 works with docker-24.0.5_2 and moby-24.0.5_2, but docker-buildx-0.10.3_3 doesn't)
### Steps to reproduce
1. `xbps-install -S docker-buildx-0.10.3_3`
2. `docker build .` fails
3. `xdowngrade /var/cache/xbps/docker-buildx-0.10.3_2.x86_64.xbps`
4. `docker build .` is fine
"FROM node:18-alpine" is the only line in Dockerfile
|
non_process
|
docker build fails with error http invalid host header with docker buildx is this a new report yes system info void authenticamd notuptodate dfff package s affected docker buildx does a report exist for this bug with the project s home upstream and or another distro expected behaviour work with docker docker cli moby and lower versions tested actual behaviour error http invalid host header when building downgrade to docker buildx solves the issue docker buildx works with docker and moby but docker buildx doesn t steps to reproduce xbps install s docker buildx docker build fails xdowngrade var cache xbps docker buildx xbps docker build is fine from node alpine is the only line in dockerfile
| 0
|
1,561
| 2,645,217,029
|
IssuesEvent
|
2015-03-12 21:20:23
|
editorconfig/editorconfig
|
https://api.github.com/repos/editorconfig/editorconfig
|
closed
|
Add support for existing `**` expansion and bash-like `**` expansion
|
core code core code (C) core code (python) feature request
|
Tests have been added for this in the [v0.11.0-development](https://github.com/editorconfig/editorconfig-core-test/tree/v0.11.0-development) branch.
In reference to editorconfig/editorconfig-core-js#1
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/381233-add-support-for-existing-expansion-and-bash-like-expansion?utm_campaign=plugin&utm_content=tracker%2F76099&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F76099&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
3.0
|
Add support for existing `**` expansion and bash-like `**` expansion - Tests have been added for this in the [v0.11.0-development](https://github.com/editorconfig/editorconfig-core-test/tree/v0.11.0-development) branch.
In reference to editorconfig/editorconfig-core-js#1
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/381233-add-support-for-existing-expansion-and-bash-like-expansion?utm_campaign=plugin&utm_content=tracker%2F76099&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F76099&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
non_process
|
add support for existing expansion and bash like expansion tests have been added for this in the branch in reference to editorconfig editorconfig core js want to back this issue we accept bounties via
| 0
|
21,420
| 17,038,387,673
|
IssuesEvent
|
2021-07-05 10:07:18
|
indico/indico
|
https://api.github.com/repos/indico/indico
|
closed
|
Timetable: add warning in case of draft mode
|
enhancement usability
|
We have a "Draft" switch in the list of contributions, which decides whether the Timetable is published or not.

This is at times confusing for users. One nice improvement would be to add a warning at the top of the "Timetable management" page, something like "The timetable is in draft move". (followed by a "publish it" button)
|
True
|
Timetable: add warning in case of draft mode - We have a "Draft" switch in the list of contributions, which decides whether the Timetable is published or not.

This is at times confusing for users. One nice improvement would be to add a warning at the top of the "Timetable management" page, something like "The timetable is in draft move". (followed by a "publish it" button)
|
non_process
|
timetable add warning in case of draft mode we have a draft switch in the list of contributions which decides whether the timetable is published or not this is at times confusing for users one nice improvement would be to add a warning at the top of the timetable management page something like the timetable is in draft move followed by a publish it button
| 0
|
76,667
| 15,496,164,254
|
IssuesEvent
|
2021-03-11 02:10:43
|
jinuem/Shopping-Cart-POC
|
https://api.github.com/repos/jinuem/Shopping-Cart-POC
|
opened
|
CVE-2021-23337 (High) detected in lodash-4.17.11.tgz
|
security vulnerability
|
## CVE-2021-23337 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.11.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p>
<p>Path to dependency file: /Shopping-Cart-POC/rejsx/package.json</p>
<p>Path to vulnerable library: Shopping-Cart-POC/rejsx/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-2.1.5.tgz (Root Library)
- core-7.2.2.tgz
- :x: **lodash-4.17.11.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package lodash; all versions of package org.fujion.webjars:lodash are vulnerable to Command Injection via template.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash - 4.17.21</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-23337 (High) detected in lodash-4.17.11.tgz - ## CVE-2021-23337 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.11.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p>
<p>Path to dependency file: /Shopping-Cart-POC/rejsx/package.json</p>
<p>Path to vulnerable library: Shopping-Cart-POC/rejsx/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-2.1.5.tgz (Root Library)
- core-7.2.2.tgz
- :x: **lodash-4.17.11.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package lodash; all versions of package org.fujion.webjars:lodash are vulnerable to Command Injection via template.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash - 4.17.21</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in lodash tgz cve high severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file shopping cart poc rejsx package json path to vulnerable library shopping cart poc rejsx node modules lodash package json dependency hierarchy react scripts tgz root library core tgz x lodash tgz vulnerable library vulnerability details all versions of package lodash all versions of package org fujion webjars lodash are vulnerable to command injection via template publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash step up your open source security game with whitesource
| 0
|
5,129
| 7,896,221,382
|
IssuesEvent
|
2018-06-29 07:47:46
|
decidim/decidim
|
https://api.github.com/repos/decidim/decidim
|
closed
|
Special content can be added to the process or superprocess home (2nd iteration)
|
Fridge lot: mods space: processes status: Needs-definition
|
- [ ] Active surveys (we are missing)
- [ ] Active sortitions
* Related issues:
#2273
|
1.0
|
Special content can be added to the process or superprocess home (2nd iteration) - - [ ] Active surveys (we are missing)
- [ ] Active sortitions
* Related issues:
#2273
|
process
|
special content can be added to the process or superprocess home iteration active surveys we are missing active sortitions related issues
| 1
|
249,894
| 18,858,247,870
|
IssuesEvent
|
2021-11-12 09:33:02
|
Lemonsr/pe
|
https://api.github.com/repos/Lemonsr/pe
|
opened
|
"Lifeline should end at the destroy marker (X)" is mentioned for only one specific diagram
|
severity.Low type.DocumentationBug
|

I believe it is unclear to readers if you only mention that for that specific sequence diagram, the lifeline should be destroyed. For every other sequence diagram, it should be mentioned as well. Otherwise, put this warning at the very top to encompass all sequence diagrams in one go.

<!--session: 1636703785006-31579939-b7a4-4e4a-ad80-94aa4ad34348-->
<!--Version: Web v3.4.1-->
|
1.0
|
"Lifeline should end at the destroy marker (X)" is mentioned for only one specific diagram - 
I believe it is unclear to readers if you only mention that for that specific sequence diagram, the lifeline should be destroyed. For every other sequence diagram, it should be mentioned as well. Otherwise, put this warning at the very top to encompass all sequence diagrams in one go.

<!--session: 1636703785006-31579939-b7a4-4e4a-ad80-94aa4ad34348-->
<!--Version: Web v3.4.1-->
|
non_process
|
lifeline should end at the destroy marker x is mentioned for only one specific diagram i believe it is unclear to readers if you only mention that for that specific sequence diagram the lifeline should be destroyed for every other sequence diagram it should be mentioned as well otherwise put this warning at the very top to encompass all sequence diagrams in one go
| 0
|
21,545
| 29,865,284,736
|
IssuesEvent
|
2023-06-20 02:53:33
|
cncf/tag-security
|
https://api.github.com/repos/cncf/tag-security
|
closed
|
[Suggestion] assesment outline: does "operation" belong in "intended use"?
|
assessment-process suggestion inactive
|
Description: I was reading the https://github.com/cncf/sig-security/blob/master/assessments/guide/outline.md doc, and I saw that under "Intended Use", which describes the use cases and personas in high level, we ask for "Operation" which describes operational considerations of the application. This is a bit confusing because at this place in the document the system design hadn't yet introduced so diving into operational aspects, which will include internal terminology and familiarity with the design, might be confusing for the reader. (I'm also not sure semantic the relation to the use cases section). I think that the "Operation" part could belongs under "Configuration and Set-Up" which comes right after the "Design" section, or possibly in the design section.
Impact: Describe your hopes for how this would reduce risk for the cloud native ecosystem. Who will this help? How will it help them?
Make the security assessment doc more readable and understandable
Scope: How much effort will this take? ok to provide a range of options if or "not yet determined"
easy fix
|
1.0
|
[Suggestion] assesment outline: does "operation" belong in "intended use"? - Description: I was reading the https://github.com/cncf/sig-security/blob/master/assessments/guide/outline.md doc, and I saw that under "Intended Use", which describes the use cases and personas in high level, we ask for "Operation" which describes operational considerations of the application. This is a bit confusing because at this place in the document the system design hadn't yet introduced so diving into operational aspects, which will include internal terminology and familiarity with the design, might be confusing for the reader. (I'm also not sure semantic the relation to the use cases section). I think that the "Operation" part could belongs under "Configuration and Set-Up" which comes right after the "Design" section, or possibly in the design section.
Impact: Describe your hopes for how this would reduce risk for the cloud native ecosystem. Who will this help? How will it help them?
Make the security assessment doc more readable and understandable
Scope: How much effort will this take? ok to provide a range of options if or "not yet determined"
easy fix
|
process
|
assesment outline does operation belong in intended use description i was reading the doc and i saw that under intended use which describes the use cases and personas in high level we ask for operation which describes operational considerations of the application this is a bit confusing because at this place in the document the system design hadn t yet introduced so diving into operational aspects which will include internal terminology and familiarity with the design might be confusing for the reader i m also not sure semantic the relation to the use cases section i think that the operation part could belongs under configuration and set up which comes right after the design section or possibly in the design section impact describe your hopes for how this would reduce risk for the cloud native ecosystem who will this help how will it help them make the security assessment doc more readable and understandable scope how much effort will this take ok to provide a range of options if or not yet determined easy fix
| 1
|
2,258
| 5,089,817,186
|
IssuesEvent
|
2017-01-01 22:04:46
|
NJDaeger/EssentialCommands
|
https://api.github.com/repos/NJDaeger/EssentialCommands
|
opened
|
Add messages.yml
|
feature in process
|
I want messages to be able to be customized. Also create a parser that will find keywords and replace them. This may take some time.
|
1.0
|
Add messages.yml - I want messages to be able to be customized. Also create a parser that will find keywords and replace them. This may take some time.
|
process
|
add messages yml i want messages to be able to be customized also create a parser that will find keywords and replace them this may take some time
| 1
|
394,998
| 27,049,809,527
|
IssuesEvent
|
2023-02-13 12:27:39
|
grace287/uxui-issue-practice
|
https://api.github.com/repos/grace287/uxui-issue-practice
|
opened
|
유저 모델링 작업
|
documentation enhancement
|
## 설명
우리 서비스에 가입할 유저의 속성을 정의한다.
## 할 일
- [ ] 유저 모델 초기화
- [ ] 유저 모델 정의
- [ ] 유저 모델 검증
## 참조
[Go to google](https://www.google.com)
|
1.0
|
유저 모델링 작업 - ## 설명
우리 서비스에 가입할 유저의 속성을 정의한다.
## 할 일
- [ ] 유저 모델 초기화
- [ ] 유저 모델 정의
- [ ] 유저 모델 검증
## 참조
[Go to google](https://www.google.com)
|
non_process
|
유저 모델링 작업 설명 우리 서비스에 가입할 유저의 속성을 정의한다 할 일 유저 모델 초기화 유저 모델 정의 유저 모델 검증 참조
| 0
|
397,997
| 27,183,412,602
|
IssuesEvent
|
2023-02-18 23:00:20
|
Vurv78/WebAudio
|
https://api.github.com/repos/Vurv78/WebAudio
|
closed
|
Document `wa_sc_compat` or enable by default
|
documentation enhancement Complexity: Low
|
**Is your feature request related to a problem? Please describe.**
Need to document our convars
**Describe the solution you'd like**
Document it
**Describe alternatives you've considered**
Enabling it by default (not good)
**Additional context**
A bunch of people don't like using webaudio and prefer streamcore's api, `wa_sc_compat` adds backwards compatibility, exposing streamcore's api but using webaudio internally.
|
1.0
|
Document `wa_sc_compat` or enable by default - **Is your feature request related to a problem? Please describe.**
Need to document our convars
**Describe the solution you'd like**
Document it
**Describe alternatives you've considered**
Enabling it by default (not good)
**Additional context**
A bunch of people don't like using webaudio and prefer streamcore's api, `wa_sc_compat` adds backwards compatibility, exposing streamcore's api but using webaudio internally.
|
non_process
|
document wa sc compat or enable by default is your feature request related to a problem please describe need to document our convars describe the solution you d like document it describe alternatives you ve considered enabling it by default not good additional context a bunch of people don t like using webaudio and prefer streamcore s api wa sc compat adds backwards compatibility exposing streamcore s api but using webaudio internally
| 0
|
4,967
| 7,806,309,630
|
IssuesEvent
|
2018-06-11 13:44:09
|
sysown/proxysql
|
https://api.github.com/repos/sysown/proxysql
|
closed
|
Query_Processor_time_nsec per rules
|
QUERY PROCESSOR STATISTICS enhancement
|
Hi,
as we already discussed measuring the process time per rule would handy when you want to find a "bad" rules or you just want to see which one is the slowest.
Of course this could have some overhead so this should be optional.
Thans.
|
1.0
|
Query_Processor_time_nsec per rules - Hi,
as we already discussed measuring the process time per rule would handy when you want to find a "bad" rules or you just want to see which one is the slowest.
Of course this could have some overhead so this should be optional.
Thans.
|
process
|
query processor time nsec per rules hi as we already discussed measuring the process time per rule would handy when you want to find a bad rules or you just want to see which one is the slowest of course this could have some overhead so this should be optional thans
| 1
|
20,445
| 27,102,418,666
|
IssuesEvent
|
2023-02-15 09:39:06
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
Add integration tests for java annotation plugins.
|
P2 type: process team-Rules-Java stale
|
Currently we don't test the `plugins` attribute on the java rules. Testing this feature would have caught https://github.com/bazelbuild/intellij/issues/845 at presubmit time and would have avoided a bazel patch release.
|
1.0
|
Add integration tests for java annotation plugins. - Currently we don't test the `plugins` attribute on the java rules. Testing this feature would have caught https://github.com/bazelbuild/intellij/issues/845 at presubmit time and would have avoided a bazel patch release.
|
process
|
add integration tests for java annotation plugins currently we don t test the plugins attribute on the java rules testing this feature would have caught at presubmit time and would have avoided a bazel patch release
| 1
|
18,646
| 24,580,930,703
|
IssuesEvent
|
2022-10-13 15:33:26
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[FHIR] JSON > Text response type > Some other values are getting displayed
|
Bug P1 Response datastore Process: Fixed Process: Tested dev
|
JSON > Text response type > Some other values are getting displayed
1. Configured values in SB

2. JSON values

|
2.0
|
[FHIR] JSON > Text response type > Some other values are getting displayed - JSON > Text response type > Some other values are getting displayed
1. Configured values in SB

2. JSON values

|
process
|
json text response type some other values are getting displayed json text response type some other values are getting displayed configured values in sb json values
| 1
|
417,096
| 12,155,777,366
|
IssuesEvent
|
2020-04-25 14:36:57
|
FMS-Cat/automaton-with-gui
|
https://api.github.com/repos/FMS-Cat/automaton-with-gui
|
closed
|
Create Channel -> Create Item -> Undo -> Undo -> Redo -> Redo causes an error
|
category: 🐛 Bug priority: 0-LIT
|
same goes to Curves
|
1.0
|
Create Channel -> Create Item -> Undo -> Undo -> Redo -> Redo causes an error - same goes to Curves
|
non_process
|
create channel create item undo undo redo redo causes an error same goes to curves
| 0
|
19,252
| 25,447,718,654
|
IssuesEvent
|
2022-11-24 07:57:26
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Having "Split with lines" as in-place editing.
|
Editing Processing Feature Request
|
### Feature description
It would be good to have "Split by lines" algorithm as in-place editing too. It will allow users carry out complex feature splitting without needing to create other layers.
### Additional context
_No response_
|
1.0
|
Having "Split with lines" as in-place editing. - ### Feature description
It would be good to have "Split by lines" algorithm as in-place editing too. It will allow users carry out complex feature splitting without needing to create other layers.
### Additional context
_No response_
|
process
|
having split with lines as in place editing feature description it would be good to have split by lines algorithm as in place editing too it will allow users carry out complex feature splitting without needing to create other layers additional context no response
| 1
|
85,690
| 16,719,034,165
|
IssuesEvent
|
2021-06-10 03:44:24
|
ArctosDB/arctos
|
https://api.github.com/repos/ArctosDB/arctos
|
closed
|
Code Table Request - sort code table categories alphabetically
|
Curatorial Search Display/Interface Enhancement Function-CodeTables Priority-Normal
|
We just organized the code tables by category and added descriptions, which is great, but the categories are not alphabetical as I would have expected. Can we list these alphabetically so that all of the code tables for the same category (e.g., locality) are together? The current (random?) sorting makes it harder to find the table you're looking for.
|
1.0
|
Code Table Request - sort code table categories alphabetically - We just organized the code tables by category and added descriptions, which is great, but the categories are not alphabetical as I would have expected. Can we list these alphabetically so that all of the code tables for the same category (e.g., locality) are together? The current (random?) sorting makes it harder to find the table you're looking for.
|
non_process
|
code table request sort code table categories alphabetically we just organized the code tables by category and added descriptions which is great but the categories are not alphabetical as i would have expected can we list these alphabetically so that all of the code tables for the same category e g locality are together the current random sorting makes it harder to find the table you re looking for
| 0
|
5,313
| 8,128,775,538
|
IssuesEvent
|
2018-08-17 13:04:59
|
w3c/transitions
|
https://api.github.com/repos/w3c/transitions
|
closed
|
Automatically Superseding previous versions
|
Process Issue
|
Add for PR transition:
Does the Working Group believes this specification should supersede a previous Recommendation?
Add a note for WBS to highlight this.
|
1.0
|
Automatically Superseding previous versions - Add for PR transition:
Does the Working Group believes this specification should supersede a previous Recommendation?
Add a note for WBS to highlight this.
|
process
|
automatically superseding previous versions add for pr transition does the working group believes this specification should supersede a previous recommendation add a note for wbs to highlight this
| 1
|
86,232
| 24,796,315,002
|
IssuesEvent
|
2022-10-24 17:35:17
|
LLNL/serac
|
https://api.github.com/repos/LLNL/serac
|
closed
|
Fix device initialization call on Lassen
|
CS build
|
The lassen builds are currently broken due to a device initialization call. We need to figure out the root cause and re-enable this.
_Originally posted by @jamiebramwell in https://github.com/LLNL/serac/pull/565#r721618590_
|
1.0
|
Fix device initialization call on Lassen - The lassen builds are currently broken due to a device initialization call. We need to figure out the root cause and re-enable this.
_Originally posted by @jamiebramwell in https://github.com/LLNL/serac/pull/565#r721618590_
|
non_process
|
fix device initialization call on lassen the lassen builds are currently broken due to a device initialization call we need to figure out the root cause and re enable this originally posted by jamiebramwell in
| 0
|
15,227
| 19,099,316,082
|
IssuesEvent
|
2021-11-29 20:23:47
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Reprojecting -180 to 180 dg raster to polar projection bilinear/cubic blurry
|
Feedback Processing Bug
|
### What is the bug or the crash?
Reprojecting a raster with a -180 to 180 degrees longitudal extent, to a polar projection, using the Warp tool, and with e.g. bilinear or cubic rather than nearest neighbour resampling method, creates blurry/stretched segments around the 180dg meridian. In the attached screenshots, the upper result is 3995 nearest neighbour, which is all good, followed by buggy 3995 bilinear, both viewed in a 3995 project CRS. The third result is 3413 bilinear viewed in a 3413 project CRS. The bug seems to affect only the warp tool, while when setting bilinear resampling on zoomed in/out rendering of the original 4326 raster dataset in the Symbology properties, it looks all good.
### Steps to reproduce the issue
1. Extract a -180 to 180 dg northern section (65 to 70 dg N) of e.g. ETOPO1 GeoTIFF.
2. Select Raster / Projections / Warp
3. Set output CRS to e.g. 3995, resampling method Bilinear, and resolution 1000 m, and run.
### Versions
<!--StartFragment--><!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd">
<html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /><style type="text/css">
p, li { white-space: pre-wrap; }
</style></head><body>
QGIS version | 3.22.0-Białowieża | QGIS code revision | d9022691f1
-- | -- | -- | --
Qt version | 5.15.2
Python version | 3.9.5
GDAL/OGR version | 3.4.0
PROJ version | 8.2.0
EPSG Registry database version | v10.038 (2021-10-21)
GEOS version | 3.10.0-CAPI-1.16.0
SQLite version | 3.35.2
PDAL version | 2.3.0
PostgreSQL client version | 13.0
SpatiaLite version | 5.0.1
QWT version | 6.1.3
QScintilla2 version | 2.11.5
OS version | Windows 10 Version 2009
| | |
Active Python plugins
latlontools | 3.5.9
grassprovider | 2.12.99
processing | 2.12.99
sagaprovider | 2.12.99

### Supported QGIS version
- [ ] I'm running a supported QGIS version according to the roadmap.
### New profile
- [ ] I tried with a new QGIS profile
### Additional context
_No response_
|
1.0
|
Reprojecting -180 to 180 dg raster to polar projection bilinear/cubic blurry - ### What is the bug or the crash?
Reprojecting a raster with a -180 to 180 degrees longitudal extent, to a polar projection, using the Warp tool, and with e.g. bilinear or cubic rather than nearest neighbour resampling method, creates blurry/stretched segments around the 180dg meridian. In the attached screenshots, the upper result is 3995 nearest neighbour, which is all good, followed by buggy 3995 bilinear, both viewed in a 3995 project CRS. The third result is 3413 bilinear viewed in a 3413 project CRS. The bug seems to affect only the warp tool, while when setting bilinear resampling on zoomed in/out rendering of the original 4326 raster dataset in the Symbology properties, it looks all good.
### Steps to reproduce the issue
1. Extract a -180 to 180 dg northern section (65 to 70 dg N) of e.g. ETOPO1 GeoTIFF.
2. Select Raster / Projections / Warp
3. Set output CRS to e.g. 3995, resampling method Bilinear, and resolution 1000 m, and run.
### Versions
<!--StartFragment--><!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd">
<html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /><style type="text/css">
p, li { white-space: pre-wrap; }
</style></head><body>
QGIS version | 3.22.0-Białowieża | QGIS code revision | d9022691f1
-- | -- | -- | --
Qt version | 5.15.2
Python version | 3.9.5
GDAL/OGR version | 3.4.0
PROJ version | 8.2.0
EPSG Registry database version | v10.038 (2021-10-21)
GEOS version | 3.10.0-CAPI-1.16.0
SQLite version | 3.35.2
PDAL version | 2.3.0
PostgreSQL client version | 13.0
SpatiaLite version | 5.0.1
QWT version | 6.1.3
QScintilla2 version | 2.11.5
OS version | Windows 10 Version 2009
| | |
Active Python plugins
latlontools | 3.5.9
grassprovider | 2.12.99
processing | 2.12.99
sagaprovider | 2.12.99

### Supported QGIS version
- [ ] I'm running a supported QGIS version according to the roadmap.
### New profile
- [ ] I tried with a new QGIS profile
### Additional context
_No response_
|
process
|
reprojecting to dg raster to polar projection bilinear cubic blurry what is the bug or the crash reprojecting a raster with a to degrees longitudal extent to a polar projection using the warp tool and with e g bilinear or cubic rather than nearest neighbour resampling method creates blurry stretched segments around the meridian in the attached screenshots the upper result is nearest neighbour which is all good followed by buggy bilinear both viewed in a project crs the third result is bilinear viewed in a project crs the bug seems to affect only the warp tool while when setting bilinear resampling on zoomed in out rendering of the original raster dataset in the symbology properties it looks all good steps to reproduce the issue extract a to dg northern section to dg n of e g geotiff select raster projections warp set output crs to e g resampling method bilinear and resolution m and run versions doctype html public dtd html en p li white space pre wrap qgis version białowieża qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version spatialite version qwt version version os version windows version active python plugins latlontools grassprovider processing sagaprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
| 1
|
17,112
| 22,634,729,217
|
IssuesEvent
|
2022-06-30 17:45:54
|
hashgraph/hedera-json-rpc-relay
|
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
|
closed
|
Deploy Integration relay instance using Helm
|
enhancement P1 process
|
### Problem
We currently lack a flow to easily deploy to integration using the new Helm Chart flow.
### Solution
2 phase approach
Phase 1 : From a local repo, deploy to an integration cluster in a fashion that updates the existing instances.
Phase 2: Automate so CI flow automatically does this on every merge to main, potentially using [release integration Github workflow](https://github.com/hashgraph/hedera-json-rpc-relay/blob/main/.github/workflows/release-integration.yml)
### Alternatives
_No response_
|
1.0
|
Deploy Integration relay instance using Helm - ### Problem
We currently lack a flow to easily deploy to integration using the new Helm Chart flow.
### Solution
2 phase approach
Phase 1 : From a local repo, deploy to an integration cluster in a fashion that updates the existing instances.
Phase 2: Automate so CI flow automatically does this on every merge to main, potentially using [release integration Github workflow](https://github.com/hashgraph/hedera-json-rpc-relay/blob/main/.github/workflows/release-integration.yml)
### Alternatives
_No response_
|
process
|
deploy integration relay instance using helm problem we currently lack a flow to easily deploy to integration using the new helm chart flow solution phase approach phase from a local repo deploy to an integration cluster in a fashion that updates the existing instances phase automate so ci flow automatically does this on every merge to main potentially using alternatives no response
| 1
|
9,709
| 12,705,222,941
|
IssuesEvent
|
2020-06-23 03:55:20
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
Deadlock with shared CUDA tensors and multiprocessing (spawn)
|
high priority module: cuda module: multiprocessing triaged
|
## 🐛 Bug
Reported by iffiX in https://discuss.pytorch.org/t/freezing-problem-while-using-cuda-tensor-in-multiprocessing-environment/80000.
(I have moved the code into a `main()` function to rule out some other multiprocessing bugs)
```python3
import itertools as it
import torch as t
import torch.multiprocessing as mp
def infer(id, tensor):
print(id)
print(tensor)
# del tensor immediately doesn't solve the problem
del tensor
def main():
# some global tensor
g_tensor = t.full([1000, 1000], 2, device="cuda:0")
g_tensor.share_memory_()
ctx = mp.get_context("spawn")
pool = ctx.Pool(2)
for i in range(10000000):
print("start")
pool.starmap(infer, zip(range(5), it.repeat(g_tensor)))
# cpu tensors work just fine
# for cuda tensors:
# if I delete the global tensor, reassign it with a new cuda tensor
# or if I use a tensor created dynamically in each iteration
# the program freezes after 2 iterations.
# Comment out the following lines and everything will work fine.
del g_tensor
g_tensor = t.full([1000, 1000], 2, device="cuda:0")
g_tensor.share_memory_()
if __name__ == "__main__":
main()
```
The backtrace for the deadlocked thread is:
```
#0 __lll_lock_wait () at ../sysdeps/unix/sysv/linux/x86_64/lowlevellock.S:135
#1 0x00007ffff7bc0023 in __GI___pthread_mutex_lock (mutex=0x7fffd105aaa8 <torch::(anonymous namespace)::cuda_ipc_global_entities+136>) at ../nptl/pthread_mutex_lock.c:78
#2 0x00007fffd04135b7 in torch::(anonymous namespace)::CudaIPCSentDataLimbo::collect() () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so
#3 0x00007fffd04138ec in torch::(anonymous namespace)::CudaIPCSentDataDelete(void*) () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so
#4 0x00007fffd0410aae in torch::CudaIPCSentData::~CudaIPCSentData() () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so
#5 0x00007fffd0413615 in torch::(anonymous namespace)::CudaIPCSentDataLimbo::collect() () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so
#6 0x00007fffd04138c9 in torch::(anonymous namespace)::CudaIPCSentDataDelete(void*) () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so
#7 0x00007fffb67b40bd in c10::TensorImpl::release_resources() [clone .localalias.204] () from /scratch/sgross/pytorch/torch/lib/libc10.so
#8 0x00007fffd0215649 in c10::intrusive_ptr<c10::TensorImpl, c10::UndefinedTensorImpl>::reset_() () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so
#9 0x00007fffd0519437 in THPVariable_dealloc(THPVariable*) () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so
#10 0x00005555556f584f in subtype_dealloc () at /tmp/build/80754af9/python_1565725737370/work/Objects/typeobject.c:1278
#11 0x0000555555725cbe in _PyEval_EvalFrameDefault () at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:2156
#12 0x00005555556b7ccb in function_code_fastcall (globals=<optimized out>, nargs=0, args=<optimized out>, co=<optimized out>)
at /tmp/build/80754af9/python_1565725737370/work/Objects/call.c:283
#13 _PyFunction_FastCallKeywords () at /tmp/build/80754af9/python_1565725737370/work/Objects/call.c:408
#14 0x000055555571f806 in call_function (kwnames=0x0, oparg=<optimized out>, pp_stack=<synthetic pointer>) at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:4616
#15 _PyEval_EvalFrameDefault () at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:3124
#16 0x0000555555668539 in _PyEval_EvalCodeWithName () at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:3930
#17 0x0000555555669424 in PyEval_EvalCodeEx () at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:3959
#18 0x000055555566944c in PyEval_EvalCode (co=<optimized out>, globals=<optimized out>, locals=<optimized out>) at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:524
#19 0x000055555577eb74 in run_mod () at /tmp/build/80754af9/python_1565725737370/work/Python/pythonrun.c:1035
#20 0x0000555555788eb1 in PyRun_FileExFlags () at /tmp/build/80754af9/python_1565725737370/work/Python/pythonrun.c:988
#21 0x00005555557890a3 in PyRun_SimpleFileExFlags () at /tmp/build/80754af9/python_1565725737370/work/Python/pythonrun.c:429
#22 0x000055555578a195 in pymain_run_file (p_cf=0x7fffffffd310, filename=0x5555558c06a0 L"prog.py", fp=0x555555907310) at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:433
#23 pymain_run_filename (cf=0x7fffffffd310, pymain=0x7fffffffd420) at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:1612
#24 pymain_run_python (pymain=0x7fffffffd420) at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:2873
#25 pymain_main () at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:3413
#26 0x000055555578a2bc in _Py_UnixMain () at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:3448
#27 0x00007ffff77e6b97 in __libc_start_main (main=0x5555556496c0 <main>, argc=2, argv=0x7fffffffd578, init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized out>,
stack_end=0x7fffffffd568) at ../csu/libc-start.c:310
#28 0x000055555572f062 in _start () at ../sysdeps/x86_64/elf/start.S:103
```
The original report was for PyTorch 1.1.0, reproduced in 1.5.0 as well. I see it in 1.6.0a0+324dc16.
cc @ezyang @gchanan @zou3519 @ngimel @VitalyFedyunin since this is in `CudaIPCSentData`
|
1.0
|
Deadlock with shared CUDA tensors and multiprocessing (spawn) - ## 🐛 Bug
Reported by iffiX in https://discuss.pytorch.org/t/freezing-problem-while-using-cuda-tensor-in-multiprocessing-environment/80000.
(I have moved the code into a `main()` function to rule out some other multiprocessing bugs)
```python3
import itertools as it
import torch as t
import torch.multiprocessing as mp
def infer(id, tensor):
print(id)
print(tensor)
# del tensor immediately doesn't solve the problem
del tensor
def main():
# some global tensor
g_tensor = t.full([1000, 1000], 2, device="cuda:0")
g_tensor.share_memory_()
ctx = mp.get_context("spawn")
pool = ctx.Pool(2)
for i in range(10000000):
print("start")
pool.starmap(infer, zip(range(5), it.repeat(g_tensor)))
# cpu tensors work just fine
# for cuda tensors:
# if I delete the global tensor, reassign it with a new cuda tensor
# or if I use a tensor created dynamically in each iteration
# the program freezes after 2 iterations.
# Comment out the following lines and everything will work fine.
del g_tensor
g_tensor = t.full([1000, 1000], 2, device="cuda:0")
g_tensor.share_memory_()
if __name__ == "__main__":
main()
```
The backtrace for the deadlocked thread is:
```
#0 __lll_lock_wait () at ../sysdeps/unix/sysv/linux/x86_64/lowlevellock.S:135
#1 0x00007ffff7bc0023 in __GI___pthread_mutex_lock (mutex=0x7fffd105aaa8 <torch::(anonymous namespace)::cuda_ipc_global_entities+136>) at ../nptl/pthread_mutex_lock.c:78
#2 0x00007fffd04135b7 in torch::(anonymous namespace)::CudaIPCSentDataLimbo::collect() () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so
#3 0x00007fffd04138ec in torch::(anonymous namespace)::CudaIPCSentDataDelete(void*) () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so
#4 0x00007fffd0410aae in torch::CudaIPCSentData::~CudaIPCSentData() () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so
#5 0x00007fffd0413615 in torch::(anonymous namespace)::CudaIPCSentDataLimbo::collect() () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so
#6 0x00007fffd04138c9 in torch::(anonymous namespace)::CudaIPCSentDataDelete(void*) () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so
#7 0x00007fffb67b40bd in c10::TensorImpl::release_resources() [clone .localalias.204] () from /scratch/sgross/pytorch/torch/lib/libc10.so
#8 0x00007fffd0215649 in c10::intrusive_ptr<c10::TensorImpl, c10::UndefinedTensorImpl>::reset_() () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so
#9 0x00007fffd0519437 in THPVariable_dealloc(THPVariable*) () from /scratch/sgross/pytorch/torch/lib/libtorch_python.so
#10 0x00005555556f584f in subtype_dealloc () at /tmp/build/80754af9/python_1565725737370/work/Objects/typeobject.c:1278
#11 0x0000555555725cbe in _PyEval_EvalFrameDefault () at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:2156
#12 0x00005555556b7ccb in function_code_fastcall (globals=<optimized out>, nargs=0, args=<optimized out>, co=<optimized out>)
at /tmp/build/80754af9/python_1565725737370/work/Objects/call.c:283
#13 _PyFunction_FastCallKeywords () at /tmp/build/80754af9/python_1565725737370/work/Objects/call.c:408
#14 0x000055555571f806 in call_function (kwnames=0x0, oparg=<optimized out>, pp_stack=<synthetic pointer>) at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:4616
#15 _PyEval_EvalFrameDefault () at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:3124
#16 0x0000555555668539 in _PyEval_EvalCodeWithName () at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:3930
#17 0x0000555555669424 in PyEval_EvalCodeEx () at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:3959
#18 0x000055555566944c in PyEval_EvalCode (co=<optimized out>, globals=<optimized out>, locals=<optimized out>) at /tmp/build/80754af9/python_1565725737370/work/Python/ceval.c:524
#19 0x000055555577eb74 in run_mod () at /tmp/build/80754af9/python_1565725737370/work/Python/pythonrun.c:1035
#20 0x0000555555788eb1 in PyRun_FileExFlags () at /tmp/build/80754af9/python_1565725737370/work/Python/pythonrun.c:988
#21 0x00005555557890a3 in PyRun_SimpleFileExFlags () at /tmp/build/80754af9/python_1565725737370/work/Python/pythonrun.c:429
#22 0x000055555578a195 in pymain_run_file (p_cf=0x7fffffffd310, filename=0x5555558c06a0 L"prog.py", fp=0x555555907310) at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:433
#23 pymain_run_filename (cf=0x7fffffffd310, pymain=0x7fffffffd420) at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:1612
#24 pymain_run_python (pymain=0x7fffffffd420) at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:2873
#25 pymain_main () at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:3413
#26 0x000055555578a2bc in _Py_UnixMain () at /tmp/build/80754af9/python_1565725737370/work/Modules/main.c:3448
#27 0x00007ffff77e6b97 in __libc_start_main (main=0x5555556496c0 <main>, argc=2, argv=0x7fffffffd578, init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized out>,
stack_end=0x7fffffffd568) at ../csu/libc-start.c:310
#28 0x000055555572f062 in _start () at ../sysdeps/x86_64/elf/start.S:103
```
The original report was for PyTorch 1.1.0, reproduced in 1.5.0 as well. I see it in 1.6.0a0+324dc16.
cc @ezyang @gchanan @zou3519 @ngimel @VitalyFedyunin since this is in `CudaIPCSentData`
|
process
|
deadlock with shared cuda tensors and multiprocessing spawn 🐛 bug reported by iffix in i have moved the code into a main function to rule out some other multiprocessing bugs import itertools as it import torch as t import torch multiprocessing as mp def infer id tensor print id print tensor del tensor immediately doesn t solve the problem del tensor def main some global tensor g tensor t full device cuda g tensor share memory ctx mp get context spawn pool ctx pool for i in range print start pool starmap infer zip range it repeat g tensor cpu tensors work just fine for cuda tensors if i delete the global tensor reassign it with a new cuda tensor or if i use a tensor created dynamically in each iteration the program freezes after iterations comment out the following lines and everything will work fine del g tensor g tensor t full device cuda g tensor share memory if name main main the backtrace for the deadlocked thread is lll lock wait at sysdeps unix sysv linux lowlevellock s in gi pthread mutex lock mutex at nptl pthread mutex lock c in torch anonymous namespace cudaipcsentdatalimbo collect from scratch sgross pytorch torch lib libtorch python so in torch anonymous namespace cudaipcsentdatadelete void from scratch sgross pytorch torch lib libtorch python so in torch cudaipcsentdata cudaipcsentdata from scratch sgross pytorch torch lib libtorch python so in torch anonymous namespace cudaipcsentdatalimbo collect from scratch sgross pytorch torch lib libtorch python so in torch anonymous namespace cudaipcsentdatadelete void from scratch sgross pytorch torch lib libtorch python so in tensorimpl release resources from scratch sgross pytorch torch lib so in intrusive ptr reset from scratch sgross pytorch torch lib libtorch python so in thpvariable dealloc thpvariable from scratch sgross pytorch torch lib libtorch python so in subtype dealloc at tmp build python work objects typeobject c in pyeval evalframedefault at tmp build python work python ceval c in function code fastcall globals nargs args co at tmp build python work objects call c pyfunction fastcallkeywords at tmp build python work objects call c in call function kwnames oparg pp stack at tmp build python work python ceval c pyeval evalframedefault at tmp build python work python ceval c in pyeval evalcodewithname at tmp build python work python ceval c in pyeval evalcodeex at tmp build python work python ceval c in pyeval evalcode co globals locals at tmp build python work python ceval c in run mod at tmp build python work python pythonrun c in pyrun fileexflags at tmp build python work python pythonrun c in pyrun simplefileexflags at tmp build python work python pythonrun c in pymain run file p cf filename l prog py fp at tmp build python work modules main c pymain run filename cf pymain at tmp build python work modules main c pymain run python pymain at tmp build python work modules main c pymain main at tmp build python work modules main c in py unixmain at tmp build python work modules main c in libc start main main argc argv init fini rtld fini stack end at csu libc start c in start at sysdeps elf start s the original report was for pytorch reproduced in as well i see it in cc ezyang gchanan ngimel vitalyfedyunin since this is in cudaipcsentdata
| 1
|
264,104
| 28,106,214,316
|
IssuesEvent
|
2023-03-31 01:03:56
|
Nivaskumark/CVE-2020-0097-frameworks_base_after
|
https://api.github.com/repos/Nivaskumark/CVE-2020-0097-frameworks_base_after
|
reopened
|
CVE-2019-9377 (Low) detected in baseandroid-10.0.0_r14
|
Mend: dependency security vulnerability
|
## CVE-2019-9377 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-10.0.0_r14</b></p></summary>
<p>
<p>Android framework classes and services</p>
<p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/CVE-2020-0097-frameworks_base/commit/ab450da4821e38b32262bb7c728210e89468ce0e">ab450da4821e38b32262bb7c728210e89468ce0e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/core/java/com/android/server/biometrics/fingerprint/FingerprintService.java</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In FingerprintService, there is a possible bypass for operating system protections that isolate user profiles from each other due to a missing permission check. This could lead to a local information disclosure of metadata about the biometrics of another user on the device with no additional execution privileges needed. User interaction is not needed for exploitation. Product: AndroidVersions: Android-10Android ID: A-128599663
<p>Publish Date: 2019-09-27
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-9377>CVE-2019-9377</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-9377">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-9377</a></p>
<p>Release Date: 2019-09-27</p>
<p>Fix Resolution: android-10.0.0_r30</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-9377 (Low) detected in baseandroid-10.0.0_r14 - ## CVE-2019-9377 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-10.0.0_r14</b></p></summary>
<p>
<p>Android framework classes and services</p>
<p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Nivaskumark/CVE-2020-0097-frameworks_base/commit/ab450da4821e38b32262bb7c728210e89468ce0e">ab450da4821e38b32262bb7c728210e89468ce0e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/core/java/com/android/server/biometrics/fingerprint/FingerprintService.java</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In FingerprintService, there is a possible bypass for operating system protections that isolate user profiles from each other due to a missing permission check. This could lead to a local information disclosure of metadata about the biometrics of another user on the device with no additional execution privileges needed. User interaction is not needed for exploitation. Product: AndroidVersions: Android-10Android ID: A-128599663
<p>Publish Date: 2019-09-27
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-9377>CVE-2019-9377</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-9377">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-9377</a></p>
<p>Release Date: 2019-09-27</p>
<p>Fix Resolution: android-10.0.0_r30</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve low detected in baseandroid cve low severity vulnerability vulnerable library baseandroid android framework classes and services library home page a href found in head commit a href found in base branch master vulnerable source files core java com android server biometrics fingerprint fingerprintservice java vulnerability details in fingerprintservice there is a possible bypass for operating system protections that isolate user profiles from each other due to a missing permission check this could lead to a local information disclosure of metadata about the biometrics of another user on the device with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android step up your open source security game with mend
| 0
|
15,163
| 18,913,017,122
|
IssuesEvent
|
2021-11-16 15:49:26
|
googleapis/sphinx-docfx-yaml
|
https://api.github.com/repos/googleapis/sphinx-docfx-yaml
|
closed
|
Suppress blib2to3.pgen2.driver output...
|
type: process priority: p3
|
After adding `format_code` support in #144, there's a lot of debugging output like `blib2to3.pgen2.driver > Stop.` that is littered over the logs, cluttering that doesn't help with debugging through the logs. Will investigate to suppress this 😢
|
1.0
|
Suppress blib2to3.pgen2.driver output... - After adding `format_code` support in #144, there's a lot of debugging output like `blib2to3.pgen2.driver > Stop.` that is littered over the logs, cluttering that doesn't help with debugging through the logs. Will investigate to suppress this 😢
|
process
|
suppress driver output after adding format code support in there s a lot of debugging output like driver stop that is littered over the logs cluttering that doesn t help with debugging through the logs will investigate to suppress this 😢
| 1
|
231,364
| 7,631,589,272
|
IssuesEvent
|
2018-05-05 03:55:01
|
jahirfiquitiva/Blueprint
|
https://api.github.com/repos/jahirfiquitiva/Blueprint
|
closed
|
Please update FAQs to be include bottom navigation
|
accepted enhancement low priority
|
<!--
Any HTML comment will be stripped when the markdown is rendered, so you don't need to delete them.
Put an x inside the [] like this: [x] to mark the checkbox.
-->
- [x] I have verified there are no duplicate active or recent bugs, questions, or requests
- [x] I have verified that I am using the latest version of Blueprint.
### Device/dashboard info:
- Blueprint version: `?`
- Android version: `?`
- Device Manufacturer: `?`
- Device Name: `?`
<!--
Please wrap code with correct syntax highlighting. You can remove it if you think it isn't necessary.
-->
```kotlin
println("Hello, world!")
```
<!--
If you are getting an error in the LogCat, paste here the stack trace.
Please wrap logs with Gradle syntax highlighting (it makes them look better).
-->
```Gradle
java.lang.RuntimeException: This is an example Exception log
at jahirfiquitiva.libs.blueprint.HelloWorld
at jahirfiquitiva.libs.blueprint.HelloWorld$ThisIsNotARealLog
at android.app.Instrumentation.callActivityOnResume(Instrumentation.kt)
```
<!--
The required steps to reproduce it.
-->
### Reproduction Steps
1.
2.
3.
### Expected Result
### Actual Result
<!--
Adding pictures/screenshots/videos of the expected/actual result is always helpful :)
-->
|
1.0
|
Please update FAQs to be include bottom navigation - <!--
Any HTML comment will be stripped when the markdown is rendered, so you don't need to delete them.
Put an x inside the [] like this: [x] to mark the checkbox.
-->
- [x] I have verified there are no duplicate active or recent bugs, questions, or requests
- [x] I have verified that I am using the latest version of Blueprint.
### Device/dashboard info:
- Blueprint version: `?`
- Android version: `?`
- Device Manufacturer: `?`
- Device Name: `?`
<!--
Please wrap code with correct syntax highlighting. You can remove it if you think it isn't necessary.
-->
```kotlin
println("Hello, world!")
```
<!--
If you are getting an error in the LogCat, paste here the stack trace.
Please wrap logs with Gradle syntax highlighting (it makes them look better).
-->
```Gradle
java.lang.RuntimeException: This is an example Exception log
at jahirfiquitiva.libs.blueprint.HelloWorld
at jahirfiquitiva.libs.blueprint.HelloWorld$ThisIsNotARealLog
at android.app.Instrumentation.callActivityOnResume(Instrumentation.kt)
```
<!--
The required steps to reproduce it.
-->
### Reproduction Steps
1.
2.
3.
### Expected Result
### Actual Result
<!--
Adding pictures/screenshots/videos of the expected/actual result is always helpful :)
-->
|
non_process
|
please update faqs to be include bottom navigation any html comment will be stripped when the markdown is rendered so you don t need to delete them put an x inside the like this to mark the checkbox i have verified there are no duplicate active or recent bugs questions or requests i have verified that i am using the latest version of blueprint device dashboard info blueprint version android version device manufacturer device name please wrap code with correct syntax highlighting you can remove it if you think it isn t necessary kotlin println hello world if you are getting an error in the logcat paste here the stack trace please wrap logs with gradle syntax highlighting it makes them look better gradle java lang runtimeexception this is an example exception log at jahirfiquitiva libs blueprint helloworld at jahirfiquitiva libs blueprint helloworld thisisnotareallog at android app instrumentation callactivityonresume instrumentation kt the required steps to reproduce it reproduction steps expected result actual result adding pictures screenshots videos of the expected actual result is always helpful
| 0
|
118,964
| 15,383,340,552
|
IssuesEvent
|
2021-03-03 02:32:19
|
openrewrite/rewrite
|
https://api.github.com/repos/openrewrite/rewrite
|
closed
|
Unable to resolve dependencies from internal repository for unknown reason
|
bug design-partner-1
|
All this says is "`MavenPomDownloader.download()` returned null, which it does if the dependency cannot be downloaded for any reason". First step is to capture more information about why this could have failed.
```
org.openrewrite.maven.internal.MavenParsingException: Unable to download commons-logging:commons-logging:1.2. Including POM is at pom.xml
at org.openrewrite.maven.internal.RawMavenResolver.lambda$processDependencies$4 (RawMavenResolver.java:300)
at java.util.stream.ReferencePipeline$3$1.accept (ReferencePipeline.java:193)
at java.util.stream.ReferencePipeline$2$1.accept (ReferencePipeline.java:175)
at java.util.ArrayList$ArrayListSpliterator.forEachRemaining (ArrayList.java:1382)
at java.util.stream.AbstractPipeline.copyInto (AbstractPipeline.java:481)
at java.util.stream.AbstractPipeline.wrapAndCopyInto (AbstractPipeline.java:471)
at java.util.stream.ReduceOps$ReduceOp.evaluateSequential (ReduceOps.java:708)
at java.util.stream.AbstractPipeline.evaluate (AbstractPipeline.java:234)
at java.util.stream.ReferencePipeline.collect (ReferencePipeline.java:499)
at org.openrewrite.maven.internal.RawMavenResolver.processDependencies (RawMavenResolver.java:335)
at org.openrewrite.maven.internal.RawMavenResolver.processTask (RawMavenResolver.java:122)
at org.openrewrite.maven.internal.RawMavenResolver.resolve (RawMavenResolver.java:103)
at org.openrewrite.maven.internal.RawMavenResolver.resolve (RawMavenResolver.java:91)
at org.openrewrite.maven.internal.RawMavenResolver.resolve (RawMavenResolver.java:74)
at org.openrewrite.maven.MavenParser.parseInputs (MavenParser.java:88)
at org.openrewrite.Parser.parse (Parser.java:39)
at org.openrewrite.maven.AbstractRewriteMojo.parseMaven (AbstractRewriteMojo.java:148)
at org.openrewrite.maven.AbstractRewriteMojo.listResults (AbstractRewriteMojo.java:252)
at org.openrewrite.maven.RewriteFixMojo.execute (RewriteFixMojo.java:36)
at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo (DefaultBuildPluginManager.java:137)
at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:208)
at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:154)
at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:146)
at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117)
at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:81)
at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build (SingleThreadedBuilder.java:56)
at org.apache.maven.lifecycle.internal.LifecycleStarter.execute (LifecycleStarter.java:128)
at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:305)
at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:192)
at org.apache.maven.DefaultMaven.execute (DefaultMaven.java:105)
at org.apache.maven.cli.MavenCli.execute (MavenCli.java:954)
at org.apache.maven.cli.MavenCli.doMain (MavenCli.java:288)
at org.apache.maven.cli.MavenCli.main (MavenCli.java:192)
at sun.reflect.NativeMethodAccessorImpl.invoke0 (Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke (NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke (DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke (Method.java:498)
at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced (Launcher.java:289)
at org.codehaus.plexus.classworlds.launcher.Launcher.launch (Launcher.java:229)
at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode (Launcher.java:415)
at org.codehaus.plexus.classworlds.launcher.Launcher.main (Launcher.java:356)
```
|
1.0
|
Unable to resolve dependencies from internal repository for unknown reason - All this says is "`MavenPomDownloader.download()` returned null, which it does if the dependency cannot be downloaded for any reason". First step is to capture more information about why this could have failed.
```
org.openrewrite.maven.internal.MavenParsingException: Unable to download commons-logging:commons-logging:1.2. Including POM is at pom.xml
at org.openrewrite.maven.internal.RawMavenResolver.lambda$processDependencies$4 (RawMavenResolver.java:300)
at java.util.stream.ReferencePipeline$3$1.accept (ReferencePipeline.java:193)
at java.util.stream.ReferencePipeline$2$1.accept (ReferencePipeline.java:175)
at java.util.ArrayList$ArrayListSpliterator.forEachRemaining (ArrayList.java:1382)
at java.util.stream.AbstractPipeline.copyInto (AbstractPipeline.java:481)
at java.util.stream.AbstractPipeline.wrapAndCopyInto (AbstractPipeline.java:471)
at java.util.stream.ReduceOps$ReduceOp.evaluateSequential (ReduceOps.java:708)
at java.util.stream.AbstractPipeline.evaluate (AbstractPipeline.java:234)
at java.util.stream.ReferencePipeline.collect (ReferencePipeline.java:499)
at org.openrewrite.maven.internal.RawMavenResolver.processDependencies (RawMavenResolver.java:335)
at org.openrewrite.maven.internal.RawMavenResolver.processTask (RawMavenResolver.java:122)
at org.openrewrite.maven.internal.RawMavenResolver.resolve (RawMavenResolver.java:103)
at org.openrewrite.maven.internal.RawMavenResolver.resolve (RawMavenResolver.java:91)
at org.openrewrite.maven.internal.RawMavenResolver.resolve (RawMavenResolver.java:74)
at org.openrewrite.maven.MavenParser.parseInputs (MavenParser.java:88)
at org.openrewrite.Parser.parse (Parser.java:39)
at org.openrewrite.maven.AbstractRewriteMojo.parseMaven (AbstractRewriteMojo.java:148)
at org.openrewrite.maven.AbstractRewriteMojo.listResults (AbstractRewriteMojo.java:252)
at org.openrewrite.maven.RewriteFixMojo.execute (RewriteFixMojo.java:36)
at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo (DefaultBuildPluginManager.java:137)
at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:208)
at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:154)
at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:146)
at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117)
at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:81)
at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build (SingleThreadedBuilder.java:56)
at org.apache.maven.lifecycle.internal.LifecycleStarter.execute (LifecycleStarter.java:128)
at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:305)
at org.apache.maven.DefaultMaven.doExecute (DefaultMaven.java:192)
at org.apache.maven.DefaultMaven.execute (DefaultMaven.java:105)
at org.apache.maven.cli.MavenCli.execute (MavenCli.java:954)
at org.apache.maven.cli.MavenCli.doMain (MavenCli.java:288)
at org.apache.maven.cli.MavenCli.main (MavenCli.java:192)
at sun.reflect.NativeMethodAccessorImpl.invoke0 (Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke (NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke (DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke (Method.java:498)
at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced (Launcher.java:289)
at org.codehaus.plexus.classworlds.launcher.Launcher.launch (Launcher.java:229)
at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode (Launcher.java:415)
at org.codehaus.plexus.classworlds.launcher.Launcher.main (Launcher.java:356)
```
|
non_process
|
unable to resolve dependencies from internal repository for unknown reason all this says is mavenpomdownloader download returned null which it does if the dependency cannot be downloaded for any reason first step is to capture more information about why this could have failed org openrewrite maven internal mavenparsingexception unable to download commons logging commons logging including pom is at pom xml at org openrewrite maven internal rawmavenresolver lambda processdependencies rawmavenresolver java at java util stream referencepipeline accept referencepipeline java at java util stream referencepipeline accept referencepipeline java at java util arraylist arraylistspliterator foreachremaining arraylist java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream reduceops reduceop evaluatesequential reduceops java at java util stream abstractpipeline evaluate abstractpipeline java at java util stream referencepipeline collect referencepipeline java at org openrewrite maven internal rawmavenresolver processdependencies rawmavenresolver java at org openrewrite maven internal rawmavenresolver processtask rawmavenresolver java at org openrewrite maven internal rawmavenresolver resolve rawmavenresolver java at org openrewrite maven internal rawmavenresolver resolve rawmavenresolver java at org openrewrite maven internal rawmavenresolver resolve rawmavenresolver java at org openrewrite maven mavenparser parseinputs mavenparser java at org openrewrite parser parse parser java at org openrewrite maven abstractrewritemojo parsemaven abstractrewritemojo java at org openrewrite maven abstractrewritemojo listresults abstractrewritemojo java at org openrewrite maven rewritefixmojo execute rewritefixmojo java at org apache maven plugin defaultbuildpluginmanager executemojo defaultbuildpluginmanager java at org apache maven lifecycle internal mojoexecutor execute mojoexecutor java at org apache maven lifecycle internal mojoexecutor execute mojoexecutor java at org apache maven lifecycle internal mojoexecutor execute mojoexecutor java at org apache maven lifecycle internal lifecyclemodulebuilder buildproject lifecyclemodulebuilder java at org apache maven lifecycle internal lifecyclemodulebuilder buildproject lifecyclemodulebuilder java at org apache maven lifecycle internal builder singlethreaded singlethreadedbuilder build singlethreadedbuilder java at org apache maven lifecycle internal lifecyclestarter execute lifecyclestarter java at org apache maven defaultmaven doexecute defaultmaven java at org apache maven defaultmaven doexecute defaultmaven java at org apache maven defaultmaven execute defaultmaven java at org apache maven cli mavencli execute mavencli java at org apache maven cli mavencli domain mavencli java at org apache maven cli mavencli main mavencli java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org codehaus plexus classworlds launcher launcher launchenhanced launcher java at org codehaus plexus classworlds launcher launcher launch launcher java at org codehaus plexus classworlds launcher launcher mainwithexitcode launcher java at org codehaus plexus classworlds launcher launcher main launcher java
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.