Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
22,254
| 30,802,636,546
|
IssuesEvent
|
2023-08-01 03:34:55
|
emily-writes-poems/emily-writes-poems-processing
|
https://api.github.com/repos/emily-writes-poems/emily-writes-poems-processing
|
closed
|
display linked poems
|
processing refinement
|
in poems table. maybe as a modal? can also redesign the link poem option to be in this modal...?
|
1.0
|
display linked poems - in poems table. maybe as a modal? can also redesign the link poem option to be in this modal...?
|
process
|
display linked poems in poems table maybe as a modal can also redesign the link poem option to be in this modal
| 1
|
4,324
| 7,229,376,901
|
IssuesEvent
|
2018-02-11 19:14:49
|
pwittchen/ReactiveNetwork
|
https://api.github.com/repos/pwittchen/ReactiveNetwork
|
closed
|
Release 0.12.3 (RxJava1.x)
|
RxJava1.x release process
|
**Initial release notes**:
- bumped RxJava: 1.3.3 -> 1.3.5
- updated Gradle Build Tools: 3.0.0 -> 3.0.1
- added script for publishing JavaDoc on gh-pages
- added script for publishing documentation on gh-pages
**Things to do**:
- [x] update JavaDoc on `gh-pages`
- [x] update documentation on `gh-pages`
- [x] bump library version
- [x] upload archives to Maven Central
- [x] close and release artifact on Maven Central
- [x] update `CHANGELOG.md` after Maven Sync
- [x] bump library version in `README.md`
- [x] update docs on gh-pages after updating `README.md`
- [x] create new GitHub release
|
1.0
|
Release 0.12.3 (RxJava1.x) - **Initial release notes**:
- bumped RxJava: 1.3.3 -> 1.3.5
- updated Gradle Build Tools: 3.0.0 -> 3.0.1
- added script for publishing JavaDoc on gh-pages
- added script for publishing documentation on gh-pages
**Things to do**:
- [x] update JavaDoc on `gh-pages`
- [x] update documentation on `gh-pages`
- [x] bump library version
- [x] upload archives to Maven Central
- [x] close and release artifact on Maven Central
- [x] update `CHANGELOG.md` after Maven Sync
- [x] bump library version in `README.md`
- [x] update docs on gh-pages after updating `README.md`
- [x] create new GitHub release
|
process
|
release x initial release notes bumped rxjava updated gradle build tools added script for publishing javadoc on gh pages added script for publishing documentation on gh pages things to do update javadoc on gh pages update documentation on gh pages bump library version upload archives to maven central close and release artifact on maven central update changelog md after maven sync bump library version in readme md update docs on gh pages after updating readme md create new github release
| 1
|
182,036
| 30,782,689,347
|
IssuesEvent
|
2023-07-31 11:08:42
|
DeveloperAcademy-POSTECH/MC3-G5T15-DreamEgg
|
https://api.github.com/repos/DeveloperAcademy-POSTECH/MC3-G5T15-DreamEgg
|
opened
|
[Feat] UI Design 다듬기
|
🎨 Design
|
### 📝 작업 목적 / Miro / Figma
<!-- 기능 혹은 버그에 대한 설명 -->
디벨롭하거나 디자인 변경이 안된 부분을 반영하기
---
### 🛠️ Tasks
* [ ] line 1
* [ ] line 2
|
1.0
|
[Feat] UI Design 다듬기 - ### 📝 작업 목적 / Miro / Figma
<!-- 기능 혹은 버그에 대한 설명 -->
디벨롭하거나 디자인 변경이 안된 부분을 반영하기
---
### 🛠️ Tasks
* [ ] line 1
* [ ] line 2
|
non_process
|
ui design 다듬기 📝 작업 목적 miro figma 디벨롭하거나 디자인 변경이 안된 부분을 반영하기 🛠️ tasks line line
| 0
|
17,411
| 23,228,628,390
|
IssuesEvent
|
2022-08-03 04:44:37
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
Allow copying equivalent Python command directly from Processing algorithm dialog, without running algorithm first (Request in QGIS)
|
Processing 3.24
|
### Request for documentation
From pull request QGIS/qgis#46499
Author: @nyalldawson
QGIS version: 3.24
**Allow copying equivalent Python command directly from Processing algorithm dialog, without running algorithm first**
### PR Description:
Adds an advanced button to the toolbox algorithm dialog, which contains an option to copy the equivalent Python command as the parameters define in the dialog.
While this command is also available from the history dialog, the advanced button provides a way for users to generate these
commands WITHOUT actually having to run the algorithm in advance.
This menu will be extended with additional items (such as copy parameter as json, paste parameters, copy as qgis_process command) in follow up PRs.
Sponsored by the Research Institute for Nature and Forest, Flemish Govt
### Commits tagged with [need-docs] or [FEATURE]
"[feature][processing] Add an advanced button to the toolbox algorithm\ndialog, which contains an option to copy the equivalent Python\ncommand as the parameters define in the dialog\n\nWhile this command is also available from the history dialog, the\nadvanced button provides a way for users to generate these\ncommands WITHOUT actually having to run the algorithm in advance.\n\nSponsored by the Research Institute for Nature and Forest, Flemish Govt"
|
1.0
|
Allow copying equivalent Python command directly from Processing algorithm dialog, without running algorithm first (Request in QGIS) - ### Request for documentation
From pull request QGIS/qgis#46499
Author: @nyalldawson
QGIS version: 3.24
**Allow copying equivalent Python command directly from Processing algorithm dialog, without running algorithm first**
### PR Description:
Adds an advanced button to the toolbox algorithm dialog, which contains an option to copy the equivalent Python command as the parameters define in the dialog.
While this command is also available from the history dialog, the advanced button provides a way for users to generate these
commands WITHOUT actually having to run the algorithm in advance.
This menu will be extended with additional items (such as copy parameter as json, paste parameters, copy as qgis_process command) in follow up PRs.
Sponsored by the Research Institute for Nature and Forest, Flemish Govt
### Commits tagged with [need-docs] or [FEATURE]
"[feature][processing] Add an advanced button to the toolbox algorithm\ndialog, which contains an option to copy the equivalent Python\ncommand as the parameters define in the dialog\n\nWhile this command is also available from the history dialog, the\nadvanced button provides a way for users to generate these\ncommands WITHOUT actually having to run the algorithm in advance.\n\nSponsored by the Research Institute for Nature and Forest, Flemish Govt"
|
process
|
allow copying equivalent python command directly from processing algorithm dialog without running algorithm first request in qgis request for documentation from pull request qgis qgis author nyalldawson qgis version allow copying equivalent python command directly from processing algorithm dialog without running algorithm first pr description adds an advanced button to the toolbox algorithm dialog which contains an option to copy the equivalent python command as the parameters define in the dialog while this command is also available from the history dialog the advanced button provides a way for users to generate these commands without actually having to run the algorithm in advance this menu will be extended with additional items such as copy parameter as json paste parameters copy as qgis process command in follow up prs sponsored by the research institute for nature and forest flemish govt commits tagged with or add an advanced button to the toolbox algorithm ndialog which contains an option to copy the equivalent python ncommand as the parameters define in the dialog n nwhile this command is also available from the history dialog the nadvanced button provides a way for users to generate these ncommands without actually having to run the algorithm in advance n nsponsored by the research institute for nature and forest flemish govt
| 1
|
221,914
| 17,377,057,227
|
IssuesEvent
|
2021-07-31 00:10:23
|
microsoft/AzureStorageExplorer
|
https://api.github.com/repos/microsoft/AzureStorageExplorer
|
closed
|
The blob which has virtual directory doesn't display under the Versions/Snapshots view when managing versions/snapshots from the root under flat list mode
|
:gear: blobs :heavy_check_mark: merged 🧪 testing
|
**Storage Explorer Version**: 1.21.0-dev
**Build Number**: 20210727.2
**Branch**: main
**Platform/OS**: Windows 10
**Architecture**: ia32
**How Found**: Exploratory testing
**Regression From**: Not a regression
## Steps to Reproduce ##
1. Expand one Non-ADLS Gen2 storage account -> Blob Containers.
2. Create a blob container -> Create a folder then upload one blob to it -> Click 'Back'.
3. Click 'Show View Options' -> Select 'Flat'.
4. Hide view options panel -> Right click the blob.
5. Click 'Manage History -> Manage Versions'.
6. Check whether the blob displays under the Versions view.
## Expected Experience ##
The blob displays under the Versions view.
## Actual Experience ##
The blob does not display under the Versions view.
|
1.0
|
The blob which has virtual directory doesn't display under the Versions/Snapshots view when managing versions/snapshots from the root under flat list mode - **Storage Explorer Version**: 1.21.0-dev
**Build Number**: 20210727.2
**Branch**: main
**Platform/OS**: Windows 10
**Architecture**: ia32
**How Found**: Exploratory testing
**Regression From**: Not a regression
## Steps to Reproduce ##
1. Expand one Non-ADLS Gen2 storage account -> Blob Containers.
2. Create a blob container -> Create a folder then upload one blob to it -> Click 'Back'.
3. Click 'Show View Options' -> Select 'Flat'.
4. Hide view options panel -> Right click the blob.
5. Click 'Manage History -> Manage Versions'.
6. Check whether the blob displays under the Versions view.
## Expected Experience ##
The blob displays under the Versions view.
## Actual Experience ##
The blob does not display under the Versions view.
|
non_process
|
the blob which has virtual directory doesn t display under the versions snapshots view when managing versions snapshots from the root under flat list mode storage explorer version dev build number branch main platform os windows architecture how found exploratory testing regression from not a regression steps to reproduce expand one non adls storage account blob containers create a blob container create a folder then upload one blob to it click back click show view options select flat hide view options panel right click the blob click manage history manage versions check whether the blob displays under the versions view expected experience the blob displays under the versions view actual experience the blob does not display under the versions view
| 0
|
18,108
| 24,134,131,721
|
IssuesEvent
|
2022-09-21 09:50:19
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Taxon constraint: GO:0032922
|
taxon constraints cellular processes
|
Hello,
GO:0032922 (circadian regulation of gene expression) should not be excluded from cyanobacteria, and maybe not from archaea either.
The circadian clock is highly studied in cyanobacteria, especially Synechococcus elongatus PCC 7942 and Thermosynechococcus elongatus BP-1 (now Thermosynechococcus vestitus), see review PMID:33419320. This paper (https://bmcecolevol.biomedcentral.com/articles/10.1186/s12862-017-0999-7) discusses a possible clock in archaea, although it is not yet clear if they really exist. There may be a clock in a few other (non-oxygenic) photosynthetic bacteria as well based simply on gene conservation.
I think you should remove this taxonomic restraint. If you want it could be limited to cyanobacteria, but I wonder if we would have to revisit this issue one day if you chose to limit it to that taxonomic level.
Thanks, Andrea
|
1.0
|
Taxon constraint: GO:0032922 - Hello,
GO:0032922 (circadian regulation of gene expression) should not be excluded from cyanobacteria, and maybe not from archaea either.
The circadian clock is highly studied in cyanobacteria, especially Synechococcus elongatus PCC 7942 and Thermosynechococcus elongatus BP-1 (now Thermosynechococcus vestitus), see review PMID:33419320. This paper (https://bmcecolevol.biomedcentral.com/articles/10.1186/s12862-017-0999-7) discusses a possible clock in archaea, although it is not yet clear if they really exist. There may be a clock in a few other (non-oxygenic) photosynthetic bacteria as well based simply on gene conservation.
I think you should remove this taxonomic restraint. If you want it could be limited to cyanobacteria, but I wonder if we would have to revisit this issue one day if you chose to limit it to that taxonomic level.
Thanks, Andrea
|
process
|
taxon constraint go hello go circadian regulation of gene expression should not be excluded from cyanobacteria and maybe not from archaea either the circadian clock is highly studied in cyanobacteria especially synechococcus elongatus pcc and thermosynechococcus elongatus bp now thermosynechococcus vestitus see review pmid this paper discusses a possible clock in archaea although it is not yet clear if they really exist there may be a clock in a few other non oxygenic photosynthetic bacteria as well based simply on gene conservation i think you should remove this taxonomic restraint if you want it could be limited to cyanobacteria but i wonder if we would have to revisit this issue one day if you chose to limit it to that taxonomic level thanks andrea
| 1
|
14,796
| 18,072,553,905
|
IssuesEvent
|
2021-09-21 05:41:39
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Obsoletion notice: GO:1902186 regulation of viral release from host cell & children
|
obsoletion multi-species process
|
Dear all,
The proposal has been made to obsolete :
GO:1902186 regulation of viral release from host cell
GO:0044789 modulation by host of viral release from host cell
GO:1902188 positive regulation of viral release from host cell
GO:0044791 positive regulation by host of viral release from host cell
The reason for obsoletion is that there is no evidence that this process is regulated. Annotations are to mutants, and should be moved to 'GO:0019076 viral release from host cell' or a child.
There are 23 EXP annotations by UniProt (18) and AgBase (5). No action is needed, since annotations will be replaced by GO:0019076 viral release from host cell' .
There are no mappings to these terms; these terms are not present in any subsets.
You may comment on the ticket:
Thanks, Pascale
|
1.0
|
Obsoletion notice: GO:1902186 regulation of viral release from host cell & children - Dear all,
The proposal has been made to obsolete :
GO:1902186 regulation of viral release from host cell
GO:0044789 modulation by host of viral release from host cell
GO:1902188 positive regulation of viral release from host cell
GO:0044791 positive regulation by host of viral release from host cell
The reason for obsoletion is that there is no evidence that this process is regulated. Annotations are to mutants, and should be moved to 'GO:0019076 viral release from host cell' or a child.
There are 23 EXP annotations by UniProt (18) and AgBase (5). No action is needed, since annotations will be replaced by GO:0019076 viral release from host cell' .
There are no mappings to these terms; these terms are not present in any subsets.
You may comment on the ticket:
Thanks, Pascale
|
process
|
obsoletion notice go regulation of viral release from host cell children dear all the proposal has been made to obsolete go regulation of viral release from host cell go modulation by host of viral release from host cell go positive regulation of viral release from host cell go positive regulation by host of viral release from host cell the reason for obsoletion is that there is no evidence that this process is regulated annotations are to mutants and should be moved to go viral release from host cell or a child there are exp annotations by uniprot and agbase no action is needed since annotations will be replaced by go viral release from host cell there are no mappings to these terms these terms are not present in any subsets you may comment on the ticket thanks pascale
| 1
|
6,278
| 9,255,975,562
|
IssuesEvent
|
2019-03-16 15:14:47
|
chuminh712/BookStorage---Group-2
|
https://api.github.com/repos/chuminh712/BookStorage---Group-2
|
closed
|
Detail Design
|
In Process
|
Design class diagram for Use Case Manage Customer
Design class diagram for Use Case Manage Supplier
|
1.0
|
Detail Design - Design class diagram for Use Case Manage Customer
Design class diagram for Use Case Manage Supplier
|
process
|
detail design design class diagram for use case manage customer design class diagram for use case manage supplier
| 1
|
306,508
| 9,395,829,872
|
IssuesEvent
|
2019-04-08 04:42:32
|
wso2/product-ei
|
https://api.github.com/repos/wso2/product-ei
|
opened
|
Proxy service results in a errorneous state if multiple transports are defined with incorrect configurations
|
Priority/Low Severity/Minor
|
**Description:**
Assume a proxy is defined with mutiple transports vfs and http. If the proxy service is not configured for the vfs transport properly, the service will be marked as faulty with the following warn message
```
WARN - VFSTransportListener Unable to configure the service SplitAggregateProxy for the VFS transport: Service doesn't have configuration information for transport vfs. This service is being marked as faulty and will not be available over the VFS transport.
```
and will not be listed by the admin service (or in the management console). However, it will be acessible via http. This is an inconsistent state.
sample proxy:
```
<proxy name="StockQuoteProxy">
<target>
<inSequence>
<filter source="$trp:priority" regex="1">
<then>
<enqueue priority="1" sequence="priority_sequence" executor="exec"/>
</then>
<else>
<enqueue priority="10" sequence="priority_sequence" executor="exec"/>
</else>
</filter>
</inSequence>
<outSequence>
<send/>
</outSequence>
</target>
<publishWSDL uri="file:samples/service-bus/resources/proxy/sample_proxy_1.wsdl"/>
</proxy>
```
**Affected Product Version:**
All EI versions
**Related Issues:**
https://github.com/wso2/product-ei/issues/3631
|
1.0
|
Proxy service results in a errorneous state if multiple transports are defined with incorrect configurations - **Description:**
Assume a proxy is defined with mutiple transports vfs and http. If the proxy service is not configured for the vfs transport properly, the service will be marked as faulty with the following warn message
```
WARN - VFSTransportListener Unable to configure the service SplitAggregateProxy for the VFS transport: Service doesn't have configuration information for transport vfs. This service is being marked as faulty and will not be available over the VFS transport.
```
and will not be listed by the admin service (or in the management console). However, it will be acessible via http. This is an inconsistent state.
sample proxy:
```
<proxy name="StockQuoteProxy">
<target>
<inSequence>
<filter source="$trp:priority" regex="1">
<then>
<enqueue priority="1" sequence="priority_sequence" executor="exec"/>
</then>
<else>
<enqueue priority="10" sequence="priority_sequence" executor="exec"/>
</else>
</filter>
</inSequence>
<outSequence>
<send/>
</outSequence>
</target>
<publishWSDL uri="file:samples/service-bus/resources/proxy/sample_proxy_1.wsdl"/>
</proxy>
```
**Affected Product Version:**
All EI versions
**Related Issues:**
https://github.com/wso2/product-ei/issues/3631
|
non_process
|
proxy service results in a errorneous state if multiple transports are defined with incorrect configurations description assume a proxy is defined with mutiple transports vfs and http if the proxy service is not configured for the vfs transport properly the service will be marked as faulty with the following warn message warn vfstransportlistener unable to configure the service splitaggregateproxy for the vfs transport service doesn t have configuration information for transport vfs this service is being marked as faulty and will not be available over the vfs transport and will not be listed by the admin service or in the management console however it will be acessible via http this is an inconsistent state sample proxy affected product version all ei versions related issues
| 0
|
5,792
| 8,638,867,707
|
IssuesEvent
|
2018-11-23 16:13:13
|
Dweepa/DataAnalytics-
|
https://api.github.com/repos/Dweepa/DataAnalytics-
|
closed
|
Filling in missing values using clusters
|
Cleaning and pre-processing
|
When dealing with missing values, for continuous variables, chose to group by fields that are correlated with those variables, and then take the mean of the grouped values to fill the missing variable
|
1.0
|
Filling in missing values using clusters - When dealing with missing values, for continuous variables, chose to group by fields that are correlated with those variables, and then take the mean of the grouped values to fill the missing variable
|
process
|
filling in missing values using clusters when dealing with missing values for continuous variables chose to group by fields that are correlated with those variables and then take the mean of the grouped values to fill the missing variable
| 1
|
414,175
| 27,980,163,256
|
IssuesEvent
|
2023-03-26 03:12:12
|
ahg3/mywebclass-simulation
|
https://api.github.com/repos/ahg3/mywebclass-simulation
|
opened
|
Imaging and File Optimization
|
documentation good first issue
|
Test that the website's images and files are optimized for faster loading times.
|
1.0
|
Imaging and File Optimization - Test that the website's images and files are optimized for faster loading times.
|
non_process
|
imaging and file optimization test that the website s images and files are optimized for faster loading times
| 0
|
16,009
| 10,423,688,746
|
IssuesEvent
|
2019-09-16 12:04:05
|
kyma-project/kyma
|
https://api.github.com/repos/kyma-project/kyma
|
closed
|
Release process should work both for the master branch and dedicated api-server branch
|
area/service-catalog
|
**Description**
The release process should work both for the master and dedicated api-server branch.
The release is done by the Travis job. The current release process is documented here: https://github.com/kubernetes-sigs/service-catalog/wiki/Release-Process
**AC**
- release process for CRD only from master is working
- release process for the dedicated api-server branch is working
|
1.0
|
Release process should work both for the master branch and dedicated api-server branch - **Description**
The release process should work both for the master and dedicated api-server branch.
The release is done by the Travis job. The current release process is documented here: https://github.com/kubernetes-sigs/service-catalog/wiki/Release-Process
**AC**
- release process for CRD only from master is working
- release process for the dedicated api-server branch is working
|
non_process
|
release process should work both for the master branch and dedicated api server branch description the release process should work both for the master and dedicated api server branch the release is done by the travis job the current release process is documented here ac release process for crd only from master is working release process for the dedicated api server branch is working
| 0
|
339,163
| 10,243,460,719
|
IssuesEvent
|
2019-08-20 08:13:45
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.estv.admin.ch - design is broken
|
browser-firefox engine-gecko priority-normal
|
<!-- @browser: Firefox 69.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:69.0) Gecko/20100101 Firefox/69.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html
**Browser / Version**: Firefox 69.0
**Operating System**: Windows 7
**Tested Another Browser**: No
**Problem type**: Design is broken
**Description**: no access to login
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/8/3f045f63-c65e-4b6b-9708-ee580cf21921.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190815163925</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Warning: "Request to access cookie or storage on https://googleads.g.doubleclick.net/pagead/id was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://static.doubleclick.net/instream/ad_status.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://googleads.g.doubleclick.net/pagead/id was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://googleads.g.doubleclick.net/pagead/id?slf_rd=1 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://googleads.g.doubleclick.net/pagead/id?slf_rd=1 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "MouseEvent.mozPressure is deprecated. Use PointerEvent.pressure instead." {file: "https://www.youtube.com/yts/jsbin/player_ias-vflubst9M/en_US/base.js" line: 1135}]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.estv.admin.ch - design is broken - <!-- @browser: Firefox 69.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:69.0) Gecko/20100101 Firefox/69.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html
**Browser / Version**: Firefox 69.0
**Operating System**: Windows 7
**Tested Another Browser**: No
**Problem type**: Design is broken
**Description**: no access to login
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/8/3f045f63-c65e-4b6b-9708-ee580cf21921.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190815163925</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Warning: "Request to access cookie or storage on https://googleads.g.doubleclick.net/pagead/id was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://static.doubleclick.net/instream/ad_status.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://googleads.g.doubleclick.net/pagead/id was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://googleads.g.doubleclick.net/pagead/id?slf_rd=1 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://googleads.g.doubleclick.net/pagead/id?slf_rd=1 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "MouseEvent.mozPressure is deprecated. Use PointerEvent.pressure instead." {file: "https://www.youtube.com/yts/jsbin/player_ias-vflubst9M/en_US/base.js" line: 1135}]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
design is broken url browser version firefox operating system windows tested another browser no problem type design is broken description no access to login steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta console messages u u u u u from with ❤️
| 0
|
8,319
| 11,486,378,628
|
IssuesEvent
|
2020-02-11 09:50:14
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Multi-organism processes tidying up: feeding behavior
|
multi-species process
|
Hello,
Following up on the discussion we had on the ontology call regarding #14956
I will merge:
GO:0044366 feeding on or from other organism
GO:0044367 feeding from tissue of other organism
(no annotations)
into GO:0042755 eating behavior
and move GO:0002120 predatory behavior
and GO:0060756 foraging behavior
as children of
GO:0007631 feeding behavior
Thanks, Pascale
|
1.0
|
Multi-organism processes tidying up: feeding behavior - Hello,
Following up on the discussion we had on the ontology call regarding #14956
I will merge:
GO:0044366 feeding on or from other organism
GO:0044367 feeding from tissue of other organism
(no annotations)
into GO:0042755 eating behavior
and move GO:0002120 predatory behavior
and GO:0060756 foraging behavior
as children of
GO:0007631 feeding behavior
Thanks, Pascale
|
process
|
multi organism processes tidying up feeding behavior hello following up on the discussion we had on the ontology call regarding i will merge go feeding on or from other organism go feeding from tissue of other organism no annotations into go eating behavior and move go predatory behavior and go foraging behavior as children of go feeding behavior thanks pascale
| 1
|
68,746
| 14,948,852,750
|
IssuesEvent
|
2021-01-26 10:39:58
|
symfony/symfony
|
https://api.github.com/repos/symfony/symfony
|
closed
|
[Security] TypeError: Argument 1 passed to Symfony\Component\Security\Core\Exception\AuthenticationException::__unserialize() must be of the type array, null given
|
Bug Security Status: Needs Review
|
**Symfony version(s) affected**: 4.4.17
**Description**
In some specific circumstances, PHP 7.4 session could not be unserialized in PHP 7.3. I asusme it's related to https://bugs.php.net/bug.php?id=79759.
This is a common scenario in multi-server environment, when PHP is upgraded server by server. It's strange that there are no previous bug reports on this subject.
Errors I collected:
```
TypeError: Argument 1 passed to Symfony\Component\Security\Core\Exception\AuthenticationException::__unserialize() must be of the type array, null given, called in vendor/symfony/symfony/src/Symfony/Component/Security/Core/Exception/AuthenticationException.php on line 137
#24 /vendor/symfony/symfony/src/Symfony/Component/Security/Core/Exception/AuthenticationException.php(96): Symfony\Component\Security\Core\Exception\AuthenticationException::__unserialize
#23 /vendor/symfony/symfony/src/Symfony/Component/Security/Core/Exception/AuthenticationException.php(137): Symfony\Component\Security\Core\Exception\AuthenticationException::__wakeup
#22 [internal](0): session_start
#21 /vendor/symfony/symfony/src/Symfony/Component/HttpFoundation/Session/Storage/NativeSessionStorage.php(156): Symfony\Component\HttpFoundation\Session\Storage\NativeSessionStorage::start
#20 /vendor/symfony/symfony/src/Symfony/Component/HttpFoundation/Session/Storage/NativeSessionStorage.php(323): Symfony\Component\HttpFoundation\Session\Storage\NativeSessionStorage::getBag
```
How does `session_start()` knows about `AuthenticationException`?
```
"PHP message: PHP Warning: session_start(): Session object destruction failed in vendor/symfony/symfony/src/Symfony/Component/HttpFoundation/Session/Storage/NativeSessionStorage.php on line 156"
```
**How to reproduce**
Save session in PHP 7.3, and load it in PHP 7.4. but bug only affects small portion of users, so it's not easy to reproduce.
**Possible Improvement**
One minor improvement for better DX and error handling is to check if `$this->serialized` is `null` before trying to unserialize and fail with a better error.
|
True
|
[Security] TypeError: Argument 1 passed to Symfony\Component\Security\Core\Exception\AuthenticationException::__unserialize() must be of the type array, null given - **Symfony version(s) affected**: 4.4.17
**Description**
In some specific circumstances, PHP 7.4 session could not be unserialized in PHP 7.3. I asusme it's related to https://bugs.php.net/bug.php?id=79759.
This is a common scenario in multi-server environment, when PHP is upgraded server by server. It's strange that there are no previous bug reports on this subject.
Errors I collected:
```
TypeError: Argument 1 passed to Symfony\Component\Security\Core\Exception\AuthenticationException::__unserialize() must be of the type array, null given, called in vendor/symfony/symfony/src/Symfony/Component/Security/Core/Exception/AuthenticationException.php on line 137
#24 /vendor/symfony/symfony/src/Symfony/Component/Security/Core/Exception/AuthenticationException.php(96): Symfony\Component\Security\Core\Exception\AuthenticationException::__unserialize
#23 /vendor/symfony/symfony/src/Symfony/Component/Security/Core/Exception/AuthenticationException.php(137): Symfony\Component\Security\Core\Exception\AuthenticationException::__wakeup
#22 [internal](0): session_start
#21 /vendor/symfony/symfony/src/Symfony/Component/HttpFoundation/Session/Storage/NativeSessionStorage.php(156): Symfony\Component\HttpFoundation\Session\Storage\NativeSessionStorage::start
#20 /vendor/symfony/symfony/src/Symfony/Component/HttpFoundation/Session/Storage/NativeSessionStorage.php(323): Symfony\Component\HttpFoundation\Session\Storage\NativeSessionStorage::getBag
```
How does `session_start()` knows about `AuthenticationException`?
```
"PHP message: PHP Warning: session_start(): Session object destruction failed in vendor/symfony/symfony/src/Symfony/Component/HttpFoundation/Session/Storage/NativeSessionStorage.php on line 156"
```
**How to reproduce**
Save session in PHP 7.3, and load it in PHP 7.4. but bug only affects small portion of users, so it's not easy to reproduce.
**Possible Improvement**
One minor improvement for better DX and error handling is to check if `$this->serialized` is `null` before trying to unserialize and fail with a better error.
|
non_process
|
typeerror argument passed to symfony component security core exception authenticationexception unserialize must be of the type array null given symfony version s affected description in some specific circumstances php session could not be unserialized in php i asusme it s related to this is a common scenario in multi server environment when php is upgraded server by server it s strange that there are no previous bug reports on this subject errors i collected typeerror argument passed to symfony component security core exception authenticationexception unserialize must be of the type array null given called in vendor symfony symfony src symfony component security core exception authenticationexception php on line vendor symfony symfony src symfony component security core exception authenticationexception php symfony component security core exception authenticationexception unserialize vendor symfony symfony src symfony component security core exception authenticationexception php symfony component security core exception authenticationexception wakeup session start vendor symfony symfony src symfony component httpfoundation session storage nativesessionstorage php symfony component httpfoundation session storage nativesessionstorage start vendor symfony symfony src symfony component httpfoundation session storage nativesessionstorage php symfony component httpfoundation session storage nativesessionstorage getbag how does session start knows about authenticationexception php message php warning session start session object destruction failed in vendor symfony symfony src symfony component httpfoundation session storage nativesessionstorage php on line how to reproduce save session in php and load it in php but bug only affects small portion of users so it s not easy to reproduce possible improvement one minor improvement for better dx and error handling is to check if this serialized is null before trying to unserialize and fail with a better error
| 0
|
735,616
| 25,406,705,155
|
IssuesEvent
|
2022-11-22 15:47:18
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
closed
|
Unable to install grpcio 1.51.0 for Python on Mac OS Apple M1
|
kind/bug lang/Python priority/P2 disposition/requires reporter action
|
<!--
PLEASE DO NOT POST A QUESTION HERE.
This form is for bug reports and feature requests ONLY!
For general questions and troubleshooting, please ask/look for answers at StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc
For questions that specifically need to be answered by gRPC team members, please ask/look for answers at grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io
Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new)
-->
### What version of gRPC and what language are you using?
grpcio==1.51.0
grpcio-tools=1.51.0
### What operating system (Linux, Windows,...) and version?
Mac OS Ventura (Apple M1) 13.0.1
### What runtime / compiler are you using (e.g. python version or version of gcc)
Python 3.11.0 / 3.10.8 / 3.9.15
### What did you do?
Please provide either 1) A unit test for reproducing the bug or 2) Specific steps for us to follow to reproduce the bug. If there’s not enough information to debug the problem, gRPC team may close the issue at their discretion. You’re welcome to re-open the issue once you have a reproduction.
I wanted to upgrade a Poetry project depending on `grpcio`, and I'm facing a `setuptools` errors while trying to install the new wheel.
I've tried to install the project manually through `pip` with this command line:
```sh
$ pip install --upgrade grpcio grpcio-tools
```
And I was still unable to update my project dependency.
### What did you expect to see?
A successful installation.
### What did you see instead?
```sh
$ pip install --upgrade grpcio grpcio-tools
Collecting grpcio
Using cached grpcio-1.51.0.tar.gz (22.1 MB)
Preparing metadata (setup.py): started
Preparing metadata (setup.py): finished with status 'done'
Collecting grpcio-tools
Downloading grpcio-tools-1.51.0.tar.gz (2.3 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.3/2.3 MB 24.5 MB/s eta 0:00:00
Preparing metadata (setup.py): started
Preparing metadata (setup.py): finished with status 'done'
Requirement already satisfied: protobuf<5.0dev,>=4.21.6 in ./.venv/lib/python3.9/site-packages (from grpcio-tools) (4.21.9)
Requirement already satisfied: setuptools in ./.venv/lib/python3.9/site-packages (from grpcio-tools) (65.6.0)
Building wheels for collected packages: grpcio, grpcio-tools
Building wheel for grpcio (setup.py): started
Building wheel for grpcio (setup.py): finished with status 'error'
error: subprocess-exited-with-error
× python setup.py bdist_wheel did not run successfully.
│ exit code: 1
╰─> [2106 lines of output]
Boringssl currently does not support macOS arm64, so we'll try to use the system installation of 'openssl' to build by default, make sure you have 'openssl' installed in this case
Found cython-generated files...
running bdist_wheel
running build
running build_py
running build_project_metadata
... Truncated mkdir and file copy ...
clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX12.sdk -D_WIN32_WINNT=1536 -DGRPC_XDS_USER_AGENT_NAME_SUFFIX=\"Python\" -DGRPC_XDS_USER_AGENT_VERSION_SUFFIX=\"1.51.0\" -DOPENSSL_NO_ASM=1 -DGPR_BACKWARDS_COMPATIBILITY_MODE=1 -DHAVE_CONFIG_H=1 -DGRPC_ENABLE_FORK_SUPPORT=1 "-DPyMODINIT_FUNC=extern \"C\" __attribute__((visibility (\"default\"))) PyObject*" -DGRPC_POSIX_FORK_ALLOW_PTHREAD_ATFORK=1 -Isrc/python/grpcio -Iinclude -I. -Ithird_party/abseil-cpp -Ithird_party/address_sorting/include -Ithird_party/cares/cares/include -Ithird_party/cares -Ithird_party/cares/cares -Ithird_party/cares/config_darwin -Ithird_party/re2 -I/usr/include/openssl -Ithird_party/upb -Isrc/core/ext/upb-generated -Isrc/core/ext/upbdefs-generated -Ithird_party/xxhash -Ithird_party/zlib -I/path/to/.venv/include -I/opt/homebrew/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/include/python3.9 -c src/core/ext/filters/census/grpc_context.cc -o python_build/temp.macosx-12-arm64-cpython-39/src/core/ext/filters/census/grpc_context.o -std=c++14 -stdlib=libc++ -fvisibility=hidden -fno-wrapv -fno-exceptions -DHAVE_UNISTD_H -pthread
clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX12.sdk -D_WIN32_WINNT=1536 -DGRPC_XDS_USER_AGENT_NAME_SUFFIX=\"Python\" -DGRPC_XDS_USER_AGENT_VERSION_SUFFIX=\"1.51.0\" -DOPENSSL_NO_ASM=1 -DGPR_BACKWARDS_COMPATIBILITY_MODE=1 -DHAVE_CONFIG_H=1 -DGRPC_ENABLE_FORK_SUPPORT=1 "-DPyMODINIT_FUNC=extern \"C\" __attribute__((visibility (\"default\"))) PyObject*" -DGRPC_POSIX_FORK_ALLOW_PTHREAD_ATFORK=1 -Isrc/python/grpcio -Iinclude -I. -Ithird_party/abseil-cpp -Ithird_party/address_sorting/include -Ithird_party/cares/cares/include -Ithird_party/cares -Ithird_party/cares/cares -Ithird_party/cares/config_darwin -Ithird_party/re2 -I/usr/include/openssl -Ithird_party/upb -Isrc/core/ext/upb-generated -Isrc/core/ext/upbdefs-generated -Ithird_party/xxhash -Ithird_party/zlib -I/path/to/.venv/include -I/opt/homebrew/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/include/python3.9 -c src/core/ext/filters/client_channel/lb_policy/weighted_target/weighted_target.cc -o python_build/temp.macosx-12-arm64-cpython-39/src/core/ext/filters/client_channel/lb_policy/weighted_target/weighted_target.o -std=c++14 -stdlib=libc++ -fvisibility=hidden -fno-wrapv -fno-exceptions -DHAVE_UNISTD_H -pthread
... Truncated clang invokations ...
clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX12.sdk -D_WIN32_WINNT=1536 -DGRPC_XDS_USER_AGENT_NAME_SUFFIX=\"Python\" -DGRPC_XDS_USER_AGENT_VERSION_SUFFIX=\"1.51.0\" -DOPENSSL_NO_ASM=1 -DGPR_BACKWARDS_COMPATIBILITY_MODE=1 -DHAVE_CONFIG_H=1 -DGRPC_ENABLE_FORK_SUPPORT=1 "-DPyMODINIT_FUNC=extern \"C\" __attribute__((visibility (\"default\"))) PyObject*" -DGRPC_POSIX_FORK_ALLOW_PTHREAD_ATFORK=1 -Isrc/python/grpcio -Iinclude -I. -Ithird_party/abseil-cpp -Ithird_party/address_sorting/include -Ithird_party/cares/cares/include -Ithird_party/cares -Ithird_party/cares/cares -Ithird_party/cares/config_darwin -Ithird_party/re2 -I/usr/include/openssl -Ithird_party/upb -Isrc/core/ext/upb-generated -Isrc/core/ext/upbdefs-generated -Ithird_party/xxhash -Ithird_party/zlib -I/path/to/.venv/include -I/opt/homebrew/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/include/python3.9 -c third_party/upb/upb/array.c -o python_build/temp.macosx-12-arm64-cpython-39/third_party/upb/upb/array.o -stdlib=libc++ -fvisibility=hidden -fno-wrapv -fno-exceptions -DHAVE_UNISTD_H -pthread
creating None
creating None/var
creating None/var/folders
creating None/var/folders/9g
creating None/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn
creating None/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T
creating None/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/tmp6oq83wvk
clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX12.sdk -I/path/to/.venv/include -I/opt/homebrew/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/include/python3.9 -c /var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/tmp6oq83wvk/a.c -o None/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/tmp6oq83wvk/a.o
Traceback (most recent call last):
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/unixccompiler.py", line 186, in _compile
self.spawn(compiler_so + cc_args + [src, '-o', obj] + extra_postargs)
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_spawn_patch.py", line 54, in _commandfile_spawn
_classic_spawn(self, command)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/ccompiler.py", line 1007, in spawn
spawn(cmd, dry_run=self.dry_run, **kwargs)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/spawn.py", line 70, in spawn
raise DistutilsExecError(
distutils.errors.DistutilsExecError: command '/usr/bin/clang' failed with exit code 1
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 280, in build_extensions
build_ext.build_ext.build_extensions(self)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 468, in build_extensions
self._build_extensions_serial()
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 494, in _build_extensions_serial
self.build_extension(ext)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/command/build_ext.py", line 246, in build_extension
_build_ext.build_extension(self, ext)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 549, in build_extension
objects = self.compiler.compile(
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_parallel_compile_patch.py", line 58, in _parallel_compile
multiprocessing.pool.ThreadPool(BUILD_EXT_COMPILER_JOBS).map(
File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 364, in map
return self._map_async(func, iterable, mapstar, chunksize).get()
File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 771, in get
raise self._value
File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 125, in worker
result = (True, func(*args, **kwds))
File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 48, in mapstar
return list(map(*args))
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_parallel_compile_patch.py", line 54, in _compile_single_file
self._compile(obj, src, ext, cc_args, extra_postargs, pp_opts)
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 263, in new_compile
return old_compile(obj, src, ext, cc_args, extra_postargs,
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/unixccompiler.py", line 188, in _compile
raise CompileError(msg)
distutils.errors.CompileError: command '/usr/bin/clang' failed with exit code 1
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "<string>", line 2, in <module>
File "<pip-setuptools-caller>", line 34, in <module>
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/setup.py", line 546, in <module>
setuptools.setup(
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/__init__.py", line 87, in setup
return distutils.core.setup(**attrs)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/core.py", line 185, in setup
return run_commands(dist)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/core.py", line 201, in run_commands
dist.run_commands()
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 969, in run_commands
self.run_command(cmd)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/dist.py", line 1217, in run_command
super().run_command(command)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/path/to/.venv/lib/python3.9/site-packages/wheel/bdist_wheel.py", line 299, in run
self.run_command('build')
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command
self.distribution.run_command(command)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/dist.py", line 1217, in run_command
super().run_command(command)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build.py", line 132, in run
self.run_command(cmd_name)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command
self.distribution.run_command(command)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/dist.py", line 1217, in run_command
super().run_command(command)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/command/build_ext.py", line 84, in run
_build_ext.run(self)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 346, in run
self.build_extensions()
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 284, in build_extensions
raise CommandError(
commands.CommandError: Failed `build_ext` step:
Traceback (most recent call last):
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/unixccompiler.py", line 186, in _compile
self.spawn(compiler_so + cc_args + [src, '-o', obj] + extra_postargs)
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_spawn_patch.py", line 54, in _commandfile_spawn
_classic_spawn(self, command)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/ccompiler.py", line 1007, in spawn
spawn(cmd, dry_run=self.dry_run, **kwargs)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/spawn.py", line 70, in spawn
raise DistutilsExecError(
distutils.errors.DistutilsExecError: command '/usr/bin/clang' failed with exit code 1
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 280, in build_extensions
build_ext.build_ext.build_extensions(self)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 468, in build_extensions
self._build_extensions_serial()
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 494, in _build_extensions_serial
self.build_extension(ext)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/command/build_ext.py", line 246, in build_extension
_build_ext.build_extension(self, ext)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 549, in build_extension
objects = self.compiler.compile(
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_parallel_compile_patch.py", line 58, in _parallel_compile
multiprocessing.pool.ThreadPool(BUILD_EXT_COMPILER_JOBS).map(
File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 364, in map
return self._map_async(func, iterable, mapstar, chunksize).get()
File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 771, in get
raise self._value
File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 125, in worker
result = (True, func(*args, **kwds))
File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 48, in mapstar
return list(map(*args))
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_parallel_compile_patch.py", line 54, in _compile_single_file
self._compile(obj, src, ext, cc_args, extra_postargs, pp_opts)
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 263, in new_compile
return old_compile(obj, src, ext, cc_args, extra_postargs,
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/unixccompiler.py", line 188, in _compile
raise CompileError(msg)
distutils.errors.CompileError: command '/usr/bin/clang' failed with exit code 1
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
ERROR: Failed building wheel for grpcio
Running setup.py clean for grpcio
Building wheel for grpcio-tools (setup.py): started
Building wheel for grpcio-tools (setup.py): finished with status 'done'
Created wheel for grpcio-tools: filename=grpcio_tools-1.51.0-cp39-cp39-macosx_12_0_arm64.whl size=1889028 sha256=01a33f1fdde7f11a87ad82211aa73c9f79cc9c85ba735aec144d0a857b81d351
Stored in directory: /Users/rigma/Library/Caches/pip/wheels/17/1d/7a/1d007fdff1149d47af524dbb20479c5ee32a81636b302a6fbd
Successfully built grpcio-tools
Failed to build grpcio
... Same errors with legacy pip installation ...
× Encountered error while trying to install package.
╰─> grpcio
note: This is an issue with the package mentioned above, not pip.
hint: See above for output from the failure.
```
I've got the same output for Python 3.10 and 3.11.
### Anything else we should know about your project / environment?
My project is setup thanks to Poetry with the following `pyproject.toml` subset:
```toml
[build-system]
requires = ["poetry-core"]
build-backend = "poetry.core.masonry.api"
[tool.poetry]
name = "example-project"
version = "0.0.0"
readme = "README.md"
packages = [
{ include = "app" },
]
[tool.poetry.dependencies]
python = "^3.9"
dependency-injector = { version = "4.40.0", extras = ["pydantic"] }
fastapi = "0.87.0"
google-cloud-storage = "2.6.0"
orjson = "3.8.2"
pydantic = { version = "1.10.2", extras = ["email"] }
typing-extensions = "4.4.0"
[tool.poetry.group.dev.dependencies]
alembic = "1.8.1"
black = "*"
pyproject-parser = "0.7.0"
pytest = "7.2.0"
pytest-asyncio = "0.20.2"
pytest-cov = "4.0.0"
uvicorn = "0.19.0"
[tool.pytest.ini_options]
addopts = "--cov app --cov-report term --cov-report xml:cov.xml --disable-warnings --strict-markers"
asyncio_mode = "strict"
testpaths = ["tests"]
markers = []
filterwarnings = []
[tool.coverage.run]
branch = true
source = ["app"]
omit = ["app/__main__.py"]
```
|
1.0
|
Unable to install grpcio 1.51.0 for Python on Mac OS Apple M1 - <!--
PLEASE DO NOT POST A QUESTION HERE.
This form is for bug reports and feature requests ONLY!
For general questions and troubleshooting, please ask/look for answers at StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc
For questions that specifically need to be answered by gRPC team members, please ask/look for answers at grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io
Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new)
-->
### What version of gRPC and what language are you using?
grpcio==1.51.0
grpcio-tools=1.51.0
### What operating system (Linux, Windows,...) and version?
Mac OS Ventura (Apple M1) 13.0.1
### What runtime / compiler are you using (e.g. python version or version of gcc)
Python 3.11.0 / 3.10.8 / 3.9.15
### What did you do?
Please provide either 1) A unit test for reproducing the bug or 2) Specific steps for us to follow to reproduce the bug. If there’s not enough information to debug the problem, gRPC team may close the issue at their discretion. You’re welcome to re-open the issue once you have a reproduction.
I wanted to upgrade a Poetry project depending on `grpcio`, and I'm facing a `setuptools` errors while trying to install the new wheel.
I've tried to install the project manually through `pip` with this command line:
```sh
$ pip install --upgrade grpcio grpcio-tools
```
And I was still unable to update my project dependency.
### What did you expect to see?
A successful installation.
### What did you see instead?
```sh
$ pip install --upgrade grpcio grpcio-tools
Collecting grpcio
Using cached grpcio-1.51.0.tar.gz (22.1 MB)
Preparing metadata (setup.py): started
Preparing metadata (setup.py): finished with status 'done'
Collecting grpcio-tools
Downloading grpcio-tools-1.51.0.tar.gz (2.3 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.3/2.3 MB 24.5 MB/s eta 0:00:00
Preparing metadata (setup.py): started
Preparing metadata (setup.py): finished with status 'done'
Requirement already satisfied: protobuf<5.0dev,>=4.21.6 in ./.venv/lib/python3.9/site-packages (from grpcio-tools) (4.21.9)
Requirement already satisfied: setuptools in ./.venv/lib/python3.9/site-packages (from grpcio-tools) (65.6.0)
Building wheels for collected packages: grpcio, grpcio-tools
Building wheel for grpcio (setup.py): started
Building wheel for grpcio (setup.py): finished with status 'error'
error: subprocess-exited-with-error
× python setup.py bdist_wheel did not run successfully.
│ exit code: 1
╰─> [2106 lines of output]
Boringssl currently does not support macOS arm64, so we'll try to use the system installation of 'openssl' to build by default, make sure you have 'openssl' installed in this case
Found cython-generated files...
running bdist_wheel
running build
running build_py
running build_project_metadata
... Truncated mkdir and file copy ...
clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX12.sdk -D_WIN32_WINNT=1536 -DGRPC_XDS_USER_AGENT_NAME_SUFFIX=\"Python\" -DGRPC_XDS_USER_AGENT_VERSION_SUFFIX=\"1.51.0\" -DOPENSSL_NO_ASM=1 -DGPR_BACKWARDS_COMPATIBILITY_MODE=1 -DHAVE_CONFIG_H=1 -DGRPC_ENABLE_FORK_SUPPORT=1 "-DPyMODINIT_FUNC=extern \"C\" __attribute__((visibility (\"default\"))) PyObject*" -DGRPC_POSIX_FORK_ALLOW_PTHREAD_ATFORK=1 -Isrc/python/grpcio -Iinclude -I. -Ithird_party/abseil-cpp -Ithird_party/address_sorting/include -Ithird_party/cares/cares/include -Ithird_party/cares -Ithird_party/cares/cares -Ithird_party/cares/config_darwin -Ithird_party/re2 -I/usr/include/openssl -Ithird_party/upb -Isrc/core/ext/upb-generated -Isrc/core/ext/upbdefs-generated -Ithird_party/xxhash -Ithird_party/zlib -I/path/to/.venv/include -I/opt/homebrew/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/include/python3.9 -c src/core/ext/filters/census/grpc_context.cc -o python_build/temp.macosx-12-arm64-cpython-39/src/core/ext/filters/census/grpc_context.o -std=c++14 -stdlib=libc++ -fvisibility=hidden -fno-wrapv -fno-exceptions -DHAVE_UNISTD_H -pthread
clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX12.sdk -D_WIN32_WINNT=1536 -DGRPC_XDS_USER_AGENT_NAME_SUFFIX=\"Python\" -DGRPC_XDS_USER_AGENT_VERSION_SUFFIX=\"1.51.0\" -DOPENSSL_NO_ASM=1 -DGPR_BACKWARDS_COMPATIBILITY_MODE=1 -DHAVE_CONFIG_H=1 -DGRPC_ENABLE_FORK_SUPPORT=1 "-DPyMODINIT_FUNC=extern \"C\" __attribute__((visibility (\"default\"))) PyObject*" -DGRPC_POSIX_FORK_ALLOW_PTHREAD_ATFORK=1 -Isrc/python/grpcio -Iinclude -I. -Ithird_party/abseil-cpp -Ithird_party/address_sorting/include -Ithird_party/cares/cares/include -Ithird_party/cares -Ithird_party/cares/cares -Ithird_party/cares/config_darwin -Ithird_party/re2 -I/usr/include/openssl -Ithird_party/upb -Isrc/core/ext/upb-generated -Isrc/core/ext/upbdefs-generated -Ithird_party/xxhash -Ithird_party/zlib -I/path/to/.venv/include -I/opt/homebrew/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/include/python3.9 -c src/core/ext/filters/client_channel/lb_policy/weighted_target/weighted_target.cc -o python_build/temp.macosx-12-arm64-cpython-39/src/core/ext/filters/client_channel/lb_policy/weighted_target/weighted_target.o -std=c++14 -stdlib=libc++ -fvisibility=hidden -fno-wrapv -fno-exceptions -DHAVE_UNISTD_H -pthread
... Truncated clang invokations ...
clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX12.sdk -D_WIN32_WINNT=1536 -DGRPC_XDS_USER_AGENT_NAME_SUFFIX=\"Python\" -DGRPC_XDS_USER_AGENT_VERSION_SUFFIX=\"1.51.0\" -DOPENSSL_NO_ASM=1 -DGPR_BACKWARDS_COMPATIBILITY_MODE=1 -DHAVE_CONFIG_H=1 -DGRPC_ENABLE_FORK_SUPPORT=1 "-DPyMODINIT_FUNC=extern \"C\" __attribute__((visibility (\"default\"))) PyObject*" -DGRPC_POSIX_FORK_ALLOW_PTHREAD_ATFORK=1 -Isrc/python/grpcio -Iinclude -I. -Ithird_party/abseil-cpp -Ithird_party/address_sorting/include -Ithird_party/cares/cares/include -Ithird_party/cares -Ithird_party/cares/cares -Ithird_party/cares/config_darwin -Ithird_party/re2 -I/usr/include/openssl -Ithird_party/upb -Isrc/core/ext/upb-generated -Isrc/core/ext/upbdefs-generated -Ithird_party/xxhash -Ithird_party/zlib -I/path/to/.venv/include -I/opt/homebrew/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/include/python3.9 -c third_party/upb/upb/array.c -o python_build/temp.macosx-12-arm64-cpython-39/third_party/upb/upb/array.o -stdlib=libc++ -fvisibility=hidden -fno-wrapv -fno-exceptions -DHAVE_UNISTD_H -pthread
creating None
creating None/var
creating None/var/folders
creating None/var/folders/9g
creating None/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn
creating None/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T
creating None/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/tmp6oq83wvk
clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX12.sdk -I/path/to/.venv/include -I/opt/homebrew/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/include/python3.9 -c /var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/tmp6oq83wvk/a.c -o None/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/tmp6oq83wvk/a.o
Traceback (most recent call last):
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/unixccompiler.py", line 186, in _compile
self.spawn(compiler_so + cc_args + [src, '-o', obj] + extra_postargs)
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_spawn_patch.py", line 54, in _commandfile_spawn
_classic_spawn(self, command)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/ccompiler.py", line 1007, in spawn
spawn(cmd, dry_run=self.dry_run, **kwargs)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/spawn.py", line 70, in spawn
raise DistutilsExecError(
distutils.errors.DistutilsExecError: command '/usr/bin/clang' failed with exit code 1
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 280, in build_extensions
build_ext.build_ext.build_extensions(self)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 468, in build_extensions
self._build_extensions_serial()
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 494, in _build_extensions_serial
self.build_extension(ext)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/command/build_ext.py", line 246, in build_extension
_build_ext.build_extension(self, ext)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 549, in build_extension
objects = self.compiler.compile(
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_parallel_compile_patch.py", line 58, in _parallel_compile
multiprocessing.pool.ThreadPool(BUILD_EXT_COMPILER_JOBS).map(
File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 364, in map
return self._map_async(func, iterable, mapstar, chunksize).get()
File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 771, in get
raise self._value
File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 125, in worker
result = (True, func(*args, **kwds))
File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 48, in mapstar
return list(map(*args))
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_parallel_compile_patch.py", line 54, in _compile_single_file
self._compile(obj, src, ext, cc_args, extra_postargs, pp_opts)
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 263, in new_compile
return old_compile(obj, src, ext, cc_args, extra_postargs,
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/unixccompiler.py", line 188, in _compile
raise CompileError(msg)
distutils.errors.CompileError: command '/usr/bin/clang' failed with exit code 1
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "<string>", line 2, in <module>
File "<pip-setuptools-caller>", line 34, in <module>
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/setup.py", line 546, in <module>
setuptools.setup(
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/__init__.py", line 87, in setup
return distutils.core.setup(**attrs)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/core.py", line 185, in setup
return run_commands(dist)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/core.py", line 201, in run_commands
dist.run_commands()
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 969, in run_commands
self.run_command(cmd)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/dist.py", line 1217, in run_command
super().run_command(command)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/path/to/.venv/lib/python3.9/site-packages/wheel/bdist_wheel.py", line 299, in run
self.run_command('build')
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command
self.distribution.run_command(command)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/dist.py", line 1217, in run_command
super().run_command(command)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build.py", line 132, in run
self.run_command(cmd_name)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command
self.distribution.run_command(command)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/dist.py", line 1217, in run_command
super().run_command(command)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/command/build_ext.py", line 84, in run
_build_ext.run(self)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 346, in run
self.build_extensions()
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 284, in build_extensions
raise CommandError(
commands.CommandError: Failed `build_ext` step:
Traceback (most recent call last):
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/unixccompiler.py", line 186, in _compile
self.spawn(compiler_so + cc_args + [src, '-o', obj] + extra_postargs)
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_spawn_patch.py", line 54, in _commandfile_spawn
_classic_spawn(self, command)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/ccompiler.py", line 1007, in spawn
spawn(cmd, dry_run=self.dry_run, **kwargs)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/spawn.py", line 70, in spawn
raise DistutilsExecError(
distutils.errors.DistutilsExecError: command '/usr/bin/clang' failed with exit code 1
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 280, in build_extensions
build_ext.build_ext.build_extensions(self)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 468, in build_extensions
self._build_extensions_serial()
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 494, in _build_extensions_serial
self.build_extension(ext)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/command/build_ext.py", line 246, in build_extension
_build_ext.build_extension(self, ext)
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 549, in build_extension
objects = self.compiler.compile(
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_parallel_compile_patch.py", line 58, in _parallel_compile
multiprocessing.pool.ThreadPool(BUILD_EXT_COMPILER_JOBS).map(
File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 364, in map
return self._map_async(func, iterable, mapstar, chunksize).get()
File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 771, in get
raise self._value
File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 125, in worker
result = (True, func(*args, **kwds))
File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 48, in mapstar
return list(map(*args))
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_parallel_compile_patch.py", line 54, in _compile_single_file
self._compile(obj, src, ext, cc_args, extra_postargs, pp_opts)
File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 263, in new_compile
return old_compile(obj, src, ext, cc_args, extra_postargs,
File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/unixccompiler.py", line 188, in _compile
raise CompileError(msg)
distutils.errors.CompileError: command '/usr/bin/clang' failed with exit code 1
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
ERROR: Failed building wheel for grpcio
Running setup.py clean for grpcio
Building wheel for grpcio-tools (setup.py): started
Building wheel for grpcio-tools (setup.py): finished with status 'done'
Created wheel for grpcio-tools: filename=grpcio_tools-1.51.0-cp39-cp39-macosx_12_0_arm64.whl size=1889028 sha256=01a33f1fdde7f11a87ad82211aa73c9f79cc9c85ba735aec144d0a857b81d351
Stored in directory: /Users/rigma/Library/Caches/pip/wheels/17/1d/7a/1d007fdff1149d47af524dbb20479c5ee32a81636b302a6fbd
Successfully built grpcio-tools
Failed to build grpcio
... Same errors with legacy pip installation ...
× Encountered error while trying to install package.
╰─> grpcio
note: This is an issue with the package mentioned above, not pip.
hint: See above for output from the failure.
```
I've got the same output for Python 3.10 and 3.11.
### Anything else we should know about your project / environment?
My project is setup thanks to Poetry with the following `pyproject.toml` subset:
```toml
[build-system]
requires = ["poetry-core"]
build-backend = "poetry.core.masonry.api"
[tool.poetry]
name = "example-project"
version = "0.0.0"
readme = "README.md"
packages = [
{ include = "app" },
]
[tool.poetry.dependencies]
python = "^3.9"
dependency-injector = { version = "4.40.0", extras = ["pydantic"] }
fastapi = "0.87.0"
google-cloud-storage = "2.6.0"
orjson = "3.8.2"
pydantic = { version = "1.10.2", extras = ["email"] }
typing-extensions = "4.4.0"
[tool.poetry.group.dev.dependencies]
alembic = "1.8.1"
black = "*"
pyproject-parser = "0.7.0"
pytest = "7.2.0"
pytest-asyncio = "0.20.2"
pytest-cov = "4.0.0"
uvicorn = "0.19.0"
[tool.pytest.ini_options]
addopts = "--cov app --cov-report term --cov-report xml:cov.xml --disable-warnings --strict-markers"
asyncio_mode = "strict"
testpaths = ["tests"]
markers = []
filterwarnings = []
[tool.coverage.run]
branch = true
source = ["app"]
omit = ["app/__main__.py"]
```
|
non_process
|
unable to install grpcio for python on mac os apple please do not post a question here this form is for bug reports and feature requests only for general questions and troubleshooting please ask look for answers at stackoverflow with grpc tag for questions that specifically need to be answered by grpc team members please ask look for answers at grpc io mailing list issues specific to grpc java grpc go grpc node grpc dart grpc web should be created in the repository they belong to e g what version of grpc and what language are you using grpcio grpcio tools what operating system linux windows and version mac os ventura apple what runtime compiler are you using e g python version or version of gcc python what did you do please provide either a unit test for reproducing the bug or specific steps for us to follow to reproduce the bug if there’s not enough information to debug the problem grpc team may close the issue at their discretion you’re welcome to re open the issue once you have a reproduction i wanted to upgrade a poetry project depending on grpcio and i m facing a setuptools errors while trying to install the new wheel i ve tried to install the project manually through pip with this command line sh pip install upgrade grpcio grpcio tools and i was still unable to update my project dependency what did you expect to see a successful installation what did you see instead sh pip install upgrade grpcio grpcio tools collecting grpcio using cached grpcio tar gz mb preparing metadata setup py started preparing metadata setup py finished with status done collecting grpcio tools downloading grpcio tools tar gz mb ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ mb mb s eta preparing metadata setup py started preparing metadata setup py finished with status done requirement already satisfied protobuf in venv lib site packages from grpcio tools requirement already satisfied setuptools in venv lib site packages from grpcio tools building wheels for collected packages grpcio grpcio tools building wheel for grpcio setup py started building wheel for grpcio setup py finished with status error error subprocess exited with error × python setup py bdist wheel did not run successfully │ exit code ╰─ boringssl currently does not support macos so we ll try to use the system installation of openssl to build by default make sure you have openssl installed in this case found cython generated files running bdist wheel running build running build py running build project metadata truncated mkdir and file copy clang wno unused result wsign compare wunreachable code fno common dynamic dndebug g fwrapv wall isysroot library developer commandlinetools sdks sdk d winnt dgrpc xds user agent name suffix python dgrpc xds user agent version suffix dopenssl no asm dgpr backwards compatibility mode dhave config h dgrpc enable fork support dpymodinit func extern c attribute visibility default pyobject dgrpc posix fork allow pthread atfork isrc python grpcio iinclude i ithird party abseil cpp ithird party address sorting include ithird party cares cares include ithird party cares ithird party cares cares ithird party cares config darwin ithird party i usr include openssl ithird party upb isrc core ext upb generated isrc core ext upbdefs generated ithird party xxhash ithird party zlib i path to venv include i opt homebrew opt python frameworks python framework versions include c src core ext filters census grpc context cc o python build temp macosx cpython src core ext filters census grpc context o std c stdlib libc fvisibility hidden fno wrapv fno exceptions dhave unistd h pthread clang wno unused result wsign compare wunreachable code fno common dynamic dndebug g fwrapv wall isysroot library developer commandlinetools sdks sdk d winnt dgrpc xds user agent name suffix python dgrpc xds user agent version suffix dopenssl no asm dgpr backwards compatibility mode dhave config h dgrpc enable fork support dpymodinit func extern c attribute visibility default pyobject dgrpc posix fork allow pthread atfork isrc python grpcio iinclude i ithird party abseil cpp ithird party address sorting include ithird party cares cares include ithird party cares ithird party cares cares ithird party cares config darwin ithird party i usr include openssl ithird party upb isrc core ext upb generated isrc core ext upbdefs generated ithird party xxhash ithird party zlib i path to venv include i opt homebrew opt python frameworks python framework versions include c src core ext filters client channel lb policy weighted target weighted target cc o python build temp macosx cpython src core ext filters client channel lb policy weighted target weighted target o std c stdlib libc fvisibility hidden fno wrapv fno exceptions dhave unistd h pthread truncated clang invokations clang wno unused result wsign compare wunreachable code fno common dynamic dndebug g fwrapv wall isysroot library developer commandlinetools sdks sdk d winnt dgrpc xds user agent name suffix python dgrpc xds user agent version suffix dopenssl no asm dgpr backwards compatibility mode dhave config h dgrpc enable fork support dpymodinit func extern c attribute visibility default pyobject dgrpc posix fork allow pthread atfork isrc python grpcio iinclude i ithird party abseil cpp ithird party address sorting include ithird party cares cares include ithird party cares ithird party cares cares ithird party cares config darwin ithird party i usr include openssl ithird party upb isrc core ext upb generated isrc core ext upbdefs generated ithird party xxhash ithird party zlib i path to venv include i opt homebrew opt python frameworks python framework versions include c third party upb upb array c o python build temp macosx cpython third party upb upb array o stdlib libc fvisibility hidden fno wrapv fno exceptions dhave unistd h pthread creating none creating none var creating none var folders creating none var folders creating none var folders c creating none var folders c t creating none var folders c t clang wno unused result wsign compare wunreachable code fno common dynamic dndebug g fwrapv wall isysroot library developer commandlinetools sdks sdk i path to venv include i opt homebrew opt python frameworks python framework versions include c var folders c t a c o none var folders c t a o traceback most recent call last file path to venv lib site packages setuptools distutils unixccompiler py line in compile self spawn compiler so cc args extra postargs file private var folders c t pip install grpcio src python grpcio spawn patch py line in commandfile spawn classic spawn self command file path to venv lib site packages setuptools distutils ccompiler py line in spawn spawn cmd dry run self dry run kwargs file path to venv lib site packages setuptools distutils spawn py line in spawn raise distutilsexecerror distutils errors distutilsexecerror command usr bin clang failed with exit code during handling of the above exception another exception occurred traceback most recent call last file private var folders c t pip install grpcio src python grpcio commands py line in build extensions build ext build ext build extensions self file path to venv lib site packages setuptools distutils command build ext py line in build extensions self build extensions serial file path to venv lib site packages setuptools distutils command build ext py line in build extensions serial self build extension ext file path to venv lib site packages setuptools command build ext py line in build extension build ext build extension self ext file path to venv lib site packages setuptools distutils command build ext py line in build extension objects self compiler compile file private var folders c t pip install grpcio src python grpcio parallel compile patch py line in parallel compile multiprocessing pool threadpool build ext compiler jobs map file opt homebrew cellar python frameworks python framework versions lib multiprocessing pool py line in map return self map async func iterable mapstar chunksize get file opt homebrew cellar python frameworks python framework versions lib multiprocessing pool py line in get raise self value file opt homebrew cellar python frameworks python framework versions lib multiprocessing pool py line in worker result true func args kwds file opt homebrew cellar python frameworks python framework versions lib multiprocessing pool py line in mapstar return list map args file private var folders c t pip install grpcio src python grpcio parallel compile patch py line in compile single file self compile obj src ext cc args extra postargs pp opts file private var folders c t pip install grpcio src python grpcio commands py line in new compile return old compile obj src ext cc args extra postargs file path to venv lib site packages setuptools distutils unixccompiler py line in compile raise compileerror msg distutils errors compileerror command usr bin clang failed with exit code during handling of the above exception another exception occurred traceback most recent call last file line in file line in file private var folders c t pip install grpcio setup py line in setuptools setup file path to venv lib site packages setuptools init py line in setup return distutils core setup attrs file path to venv lib site packages setuptools distutils core py line in setup return run commands dist file path to venv lib site packages setuptools distutils core py line in run commands dist run commands file path to venv lib site packages setuptools distutils dist py line in run commands self run command cmd file path to venv lib site packages setuptools dist py line in run command super run command command file path to venv lib site packages setuptools distutils dist py line in run command cmd obj run file path to venv lib site packages wheel bdist wheel py line in run self run command build file path to venv lib site packages setuptools distutils cmd py line in run command self distribution run command command file path to venv lib site packages setuptools dist py line in run command super run command command file path to venv lib site packages setuptools distutils dist py line in run command cmd obj run file path to venv lib site packages setuptools distutils command build py line in run self run command cmd name file path to venv lib site packages setuptools distutils cmd py line in run command self distribution run command command file path to venv lib site packages setuptools dist py line in run command super run command command file path to venv lib site packages setuptools distutils dist py line in run command cmd obj run file path to venv lib site packages setuptools command build ext py line in run build ext run self file path to venv lib site packages setuptools distutils command build ext py line in run self build extensions file private var folders c t pip install grpcio src python grpcio commands py line in build extensions raise commanderror commands commanderror failed build ext step traceback most recent call last file path to venv lib site packages setuptools distutils unixccompiler py line in compile self spawn compiler so cc args extra postargs file private var folders c t pip install grpcio src python grpcio spawn patch py line in commandfile spawn classic spawn self command file path to venv lib site packages setuptools distutils ccompiler py line in spawn spawn cmd dry run self dry run kwargs file path to venv lib site packages setuptools distutils spawn py line in spawn raise distutilsexecerror distutils errors distutilsexecerror command usr bin clang failed with exit code during handling of the above exception another exception occurred traceback most recent call last file private var folders c t pip install grpcio src python grpcio commands py line in build extensions build ext build ext build extensions self file path to venv lib site packages setuptools distutils command build ext py line in build extensions self build extensions serial file path to venv lib site packages setuptools distutils command build ext py line in build extensions serial self build extension ext file path to venv lib site packages setuptools command build ext py line in build extension build ext build extension self ext file path to venv lib site packages setuptools distutils command build ext py line in build extension objects self compiler compile file private var folders c t pip install grpcio src python grpcio parallel compile patch py line in parallel compile multiprocessing pool threadpool build ext compiler jobs map file opt homebrew cellar python frameworks python framework versions lib multiprocessing pool py line in map return self map async func iterable mapstar chunksize get file opt homebrew cellar python frameworks python framework versions lib multiprocessing pool py line in get raise self value file opt homebrew cellar python frameworks python framework versions lib multiprocessing pool py line in worker result true func args kwds file opt homebrew cellar python frameworks python framework versions lib multiprocessing pool py line in mapstar return list map args file private var folders c t pip install grpcio src python grpcio parallel compile patch py line in compile single file self compile obj src ext cc args extra postargs pp opts file private var folders c t pip install grpcio src python grpcio commands py line in new compile return old compile obj src ext cc args extra postargs file path to venv lib site packages setuptools distutils unixccompiler py line in compile raise compileerror msg distutils errors compileerror command usr bin clang failed with exit code note this error originates from a subprocess and is likely not a problem with pip error failed building wheel for grpcio running setup py clean for grpcio building wheel for grpcio tools setup py started building wheel for grpcio tools setup py finished with status done created wheel for grpcio tools filename grpcio tools macosx whl size stored in directory users rigma library caches pip wheels successfully built grpcio tools failed to build grpcio same errors with legacy pip installation × encountered error while trying to install package ╰─ grpcio note this is an issue with the package mentioned above not pip hint see above for output from the failure i ve got the same output for python and anything else we should know about your project environment my project is setup thanks to poetry with the following pyproject toml subset toml requires build backend poetry core masonry api name example project version readme readme md packages include app python dependency injector version extras fastapi google cloud storage orjson pydantic version extras typing extensions alembic black pyproject parser pytest pytest asyncio pytest cov uvicorn addopts cov app cov report term cov report xml cov xml disable warnings strict markers asyncio mode strict testpaths markers filterwarnings branch true source omit
| 0
|
336
| 2,792,238,767
|
IssuesEvent
|
2015-05-10 20:42:43
|
t3kt/vjzual
|
https://api.github.com/repos/t3kt/vjzual
|
opened
|
add more blend modes
|
enhancement video processing
|
possible modes:
* tima-based time distortion
* masking using an external mask source
|
1.0
|
add more blend modes - possible modes:
* tima-based time distortion
* masking using an external mask source
|
process
|
add more blend modes possible modes tima based time distortion masking using an external mask source
| 1
|
163,779
| 12,744,477,720
|
IssuesEvent
|
2020-06-26 12:35:21
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
certified-conformance : ./sonobuoy run --mode certified-conformance --wait -v 5 failed with timeout error
|
kind/failing-test needs-sig
|
certified-conformance : ./sonobuoy run --mode certified-conformance --wait -v 5 failed with below error
level=error msg="Result processing error (409): result systemd-logs/cluster already received"
level=error msg="Result processing error (409): result systemd-logs/cluster already received"
level=error msg="Result processing error (409): result systemd-logs/cluster already received"
level=error msg="Result processing error (409): result systemd-logs/cluster already received"
level=error msg="Timeout waiting for plugin e2e"
level=error msg="error running plugins: timeout occurred when waiting for plugin results"
level=error msg="Timeout occurred when running plugins. Inspect logs further for details."
|
1.0
|
certified-conformance : ./sonobuoy run --mode certified-conformance --wait -v 5 failed with timeout error -
certified-conformance : ./sonobuoy run --mode certified-conformance --wait -v 5 failed with below error
level=error msg="Result processing error (409): result systemd-logs/cluster already received"
level=error msg="Result processing error (409): result systemd-logs/cluster already received"
level=error msg="Result processing error (409): result systemd-logs/cluster already received"
level=error msg="Result processing error (409): result systemd-logs/cluster already received"
level=error msg="Timeout waiting for plugin e2e"
level=error msg="error running plugins: timeout occurred when waiting for plugin results"
level=error msg="Timeout occurred when running plugins. Inspect logs further for details."
|
non_process
|
certified conformance sonobuoy run mode certified conformance wait v failed with timeout error certified conformance sonobuoy run mode certified conformance wait v failed with below error level error msg result processing error result systemd logs cluster already received level error msg result processing error result systemd logs cluster already received level error msg result processing error result systemd logs cluster already received level error msg result processing error result systemd logs cluster already received level error msg timeout waiting for plugin level error msg error running plugins timeout occurred when waiting for plugin results level error msg timeout occurred when running plugins inspect logs further for details
| 0
|
151,416
| 19,650,296,376
|
IssuesEvent
|
2022-01-10 05:50:32
|
panasalap/linux-4.19.72
|
https://api.github.com/repos/panasalap/linux-4.19.72
|
opened
|
CVE-2021-26932 (Medium) detected in multiple libraries
|
security vulnerability
|
## CVE-2021-26932 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-yoctov5.4.51</b>, <b>linuxlinux-4.19.224</b>, <b>linux-yoctov5.4.51</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel 3.2 through 5.10.16, as used by Xen. Grant mapping operations often occur in batch hypercalls, where a number of operations are done in a single hypercall, the success or failure of each one is reported to the backend driver, and the backend driver then loops over the results, performing follow-up actions based on the success or failure of each operation. Unfortunately, when running in PV mode, the Linux backend drivers mishandle this: Some errors are ignored, effectively implying their success from the success of related batch elements. In other cases, errors resulting from one batch element lead to further batch elements not being inspected, and hence successful ones to not be possible to properly unmap upon error recovery. Only systems with Linux backends running in PV mode are vulnerable. Linux backends run in HVM / PVH modes are not vulnerable. This affects arch/*/xen/p2m.c and drivers/xen/gntdev.c.
<p>Publish Date: 2021-02-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-26932>CVE-2021-26932</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-26932 (Medium) detected in multiple libraries - ## CVE-2021-26932 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-yoctov5.4.51</b>, <b>linuxlinux-4.19.224</b>, <b>linux-yoctov5.4.51</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel 3.2 through 5.10.16, as used by Xen. Grant mapping operations often occur in batch hypercalls, where a number of operations are done in a single hypercall, the success or failure of each one is reported to the backend driver, and the backend driver then loops over the results, performing follow-up actions based on the success or failure of each operation. Unfortunately, when running in PV mode, the Linux backend drivers mishandle this: Some errors are ignored, effectively implying their success from the success of related batch elements. In other cases, errors resulting from one batch element lead to further batch elements not being inspected, and hence successful ones to not be possible to properly unmap upon error recovery. Only systems with Linux backends running in PV mode are vulnerable. Linux backends run in HVM / PVH modes are not vulnerable. This affects arch/*/xen/p2m.c and drivers/xen/gntdev.c.
<p>Publish Date: 2021-02-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-26932>CVE-2021-26932</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries linux linuxlinux linux vulnerability details an issue was discovered in the linux kernel through as used by xen grant mapping operations often occur in batch hypercalls where a number of operations are done in a single hypercall the success or failure of each one is reported to the backend driver and the backend driver then loops over the results performing follow up actions based on the success or failure of each operation unfortunately when running in pv mode the linux backend drivers mishandle this some errors are ignored effectively implying their success from the success of related batch elements in other cases errors resulting from one batch element lead to further batch elements not being inspected and hence successful ones to not be possible to properly unmap upon error recovery only systems with linux backends running in pv mode are vulnerable linux backends run in hvm pvh modes are not vulnerable this affects arch xen c and drivers xen gntdev c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource
| 0
|
57,646
| 6,552,434,475
|
IssuesEvent
|
2017-09-05 18:20:34
|
Aubron/scoreshots-templates
|
https://api.github.com/repos/Aubron/scoreshots-templates
|
closed
|
Modify Football Statshot, XML Integration
|
Priority: High Status: Testing Needed
|

### Requested by:
Internal
## Dynamic Considerations:
Adding XML integration to template 381, pictured above. No other edits should be needed.
|
1.0
|
Modify Football Statshot, XML Integration -

### Requested by:
Internal
## Dynamic Considerations:
Adding XML integration to template 381, pictured above. No other edits should be needed.
|
non_process
|
modify football statshot xml integration requested by internal dynamic considerations adding xml integration to template pictured above no other edits should be needed
| 0
|
22,341
| 31,018,532,597
|
IssuesEvent
|
2023-08-10 02:00:07
|
lizhihao6/get-daily-arxiv-noti
|
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
|
opened
|
New submissions for Thu, 10 Aug 23
|
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
|
## Keyword: events
### Constructing Holistic Spatio-Temporal Scene Graph for Video Semantic Role Labeling
- **Authors:** Yu Zhao, Hao Fei, Yixin Cao, Bobo Li, Meishan Zhang, Jianguo Wei, Min Zhang, Tat-Seng Chua
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL)
- **Arxiv link:** https://arxiv.org/abs/2308.05081
- **Pdf link:** https://arxiv.org/pdf/2308.05081
- **Abstract**
Video Semantic Role Labeling (VidSRL) aims to detect the salient events from given videos, by recognizing the predict-argument event structures and the interrelationships between events. While recent endeavors have put forth methods for VidSRL, they can be mostly subject to two key drawbacks, including the lack of fine-grained spatial scene perception and the insufficiently modeling of video temporality. Towards this end, this work explores a novel holistic spatio-temporal scene graph (namely HostSG) representation based on the existing dynamic scene graph structures, which well model both the fine-grained spatial semantics and temporal dynamics of videos for VidSRL. Built upon the HostSG, we present a nichetargeting VidSRL framework. A scene-event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high-level event semantic structure, resulting in an overall hierarchical scene-event (termed ICE) graph structure. We further perform iterative structure refinement to optimize the ICE graph, such that the overall structure representation can best coincide with end task demand. Finally, three subtask predictions of VidSRL are jointly decoded, where the end-to-end paradigm effectively avoids error propagation. On the benchmark dataset, our framework boosts significantly over the current best-performing model. Further analyses are shown for a better understanding of the advances of our methods.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
### Rendering Humans from Object-Occluded Monocular Videos
- **Authors:** Tiange Xiang, Adam Sun, Jiajun Wu, Ehsan Adeli, Li Fei-Fei
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2308.04622
- **Pdf link:** https://arxiv.org/pdf/2308.04622
- **Abstract**
3D understanding and rendering of moving humans from monocular videos is a challenging task. Despite recent progress, the task remains difficult in real-world scenarios, where obstacles may block the camera view and cause partial occlusions in the captured videos. Existing methods cannot handle such defects due to two reasons. First, the standard rendering strategy relies on point-point mapping, which could lead to dramatic disparities between the visible and occluded areas of the body. Second, the naive direct regression approach does not consider any feasibility criteria (ie, prior information) for rendering under occlusions. To tackle the above drawbacks, we present OccNeRF, a neural rendering method that achieves better rendering of humans in severely occluded scenes. As direct solutions to the two drawbacks, we propose surface-based rendering by integrating geometry and visibility priors. We validate our method on both simulated and real-world occlusions and demonstrate our method's superiority.
### Constructing Holistic Spatio-Temporal Scene Graph for Video Semantic Role Labeling
- **Authors:** Yu Zhao, Hao Fei, Yixin Cao, Bobo Li, Meishan Zhang, Jianguo Wei, Min Zhang, Tat-Seng Chua
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL)
- **Arxiv link:** https://arxiv.org/abs/2308.05081
- **Pdf link:** https://arxiv.org/pdf/2308.05081
- **Abstract**
Video Semantic Role Labeling (VidSRL) aims to detect the salient events from given videos, by recognizing the predict-argument event structures and the interrelationships between events. While recent endeavors have put forth methods for VidSRL, they can be mostly subject to two key drawbacks, including the lack of fine-grained spatial scene perception and the insufficiently modeling of video temporality. Towards this end, this work explores a novel holistic spatio-temporal scene graph (namely HostSG) representation based on the existing dynamic scene graph structures, which well model both the fine-grained spatial semantics and temporal dynamics of videos for VidSRL. Built upon the HostSG, we present a nichetargeting VidSRL framework. A scene-event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high-level event semantic structure, resulting in an overall hierarchical scene-event (termed ICE) graph structure. We further perform iterative structure refinement to optimize the ICE graph, such that the overall structure representation can best coincide with end task demand. Finally, three subtask predictions of VidSRL are jointly decoded, where the end-to-end paradigm effectively avoids error propagation. On the benchmark dataset, our framework boosts significantly over the current best-performing model. Further analyses are shown for a better understanding of the advances of our methods.
## Keyword: ISP
### Rendering Humans from Object-Occluded Monocular Videos
- **Authors:** Tiange Xiang, Adam Sun, Jiajun Wu, Ehsan Adeli, Li Fei-Fei
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2308.04622
- **Pdf link:** https://arxiv.org/pdf/2308.04622
- **Abstract**
3D understanding and rendering of moving humans from monocular videos is a challenging task. Despite recent progress, the task remains difficult in real-world scenarios, where obstacles may block the camera view and cause partial occlusions in the captured videos. Existing methods cannot handle such defects due to two reasons. First, the standard rendering strategy relies on point-point mapping, which could lead to dramatic disparities between the visible and occluded areas of the body. Second, the naive direct regression approach does not consider any feasibility criteria (ie, prior information) for rendering under occlusions. To tackle the above drawbacks, we present OccNeRF, a neural rendering method that achieves better rendering of humans in severely occluded scenes. As direct solutions to the two drawbacks, we propose surface-based rendering by integrating geometry and visibility priors. We validate our method on both simulated and real-world occlusions and demonstrate our method's superiority.
### Addressing Racial Bias in Facial Emotion Recognition
- **Authors:** Alex Fan, Xingshuo Xiao, Peter Washington
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Computers and Society (cs.CY)
- **Arxiv link:** https://arxiv.org/abs/2308.04674
- **Pdf link:** https://arxiv.org/pdf/2308.04674
- **Abstract**
Fairness in deep learning models trained with high-dimensional inputs and subjective labels remains a complex and understudied area. Facial emotion recognition, a domain where datasets are often racially imbalanced, can lead to models that yield disparate outcomes across racial groups. This study focuses on analyzing racial bias by sub-sampling training sets with varied racial distributions and assessing test performance across these simulations. Our findings indicate that smaller datasets with posed faces improve on both fairness and performance metrics as the simulations approach racial balance. Notably, the F1-score increases by $27.2\%$ points, and demographic parity increases by $15.7\%$ points on average across the simulations. However, in larger datasets with greater facial variation, fairness metrics generally remain constant, suggesting that racial balance by itself is insufficient to achieve parity in test performance across different racial groups.
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### Resource Constrained Model Compression via Minimax Optimization for Spiking Neural Networks
- **Authors:** Jue Chen, Huan Yuan, Jianchao Tan, Bin Chen, Chengru Song, Di Zhang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Multimedia (cs.MM)
- **Arxiv link:** https://arxiv.org/abs/2308.04672
- **Pdf link:** https://arxiv.org/pdf/2308.04672
- **Abstract**
Brain-inspired Spiking Neural Networks (SNNs) have the characteristics of event-driven and high energy-efficient, which are different from traditional Artificial Neural Networks (ANNs) when deployed on edge devices such as neuromorphic chips. Most previous work focuses on SNNs training strategies to improve model performance and brings larger and deeper network architectures. It is difficult to deploy these complex networks on resource-limited edge devices directly. To meet such demand, people compress SNNs very cautiously to balance the performance and the computation efficiency. Existing compression methods either iteratively pruned SNNs using weights norm magnitude or formulated the problem as a sparse learning optimization. We propose an improved end-to-end Minimax optimization method for this sparse learning problem to better balance the model performance and the computation efficiency. We also demonstrate that jointly applying compression and finetuning on SNNs is better than sequentially, especially for extreme compression ratios. The compressed SNN models achieved state-of-the-art (SOTA) performance on various benchmark datasets and architectures. Our code is available at https://github.com/chenjallen/Resource-Constrained-Compression-on-SNN.
## Keyword: RAW
### Rendering Humans from Object-Occluded Monocular Videos
- **Authors:** Tiange Xiang, Adam Sun, Jiajun Wu, Ehsan Adeli, Li Fei-Fei
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2308.04622
- **Pdf link:** https://arxiv.org/pdf/2308.04622
- **Abstract**
3D understanding and rendering of moving humans from monocular videos is a challenging task. Despite recent progress, the task remains difficult in real-world scenarios, where obstacles may block the camera view and cause partial occlusions in the captured videos. Existing methods cannot handle such defects due to two reasons. First, the standard rendering strategy relies on point-point mapping, which could lead to dramatic disparities between the visible and occluded areas of the body. Second, the naive direct regression approach does not consider any feasibility criteria (ie, prior information) for rendering under occlusions. To tackle the above drawbacks, we present OccNeRF, a neural rendering method that achieves better rendering of humans in severely occluded scenes. As direct solutions to the two drawbacks, we propose surface-based rendering by integrating geometry and visibility priors. We validate our method on both simulated and real-world occlusions and demonstrate our method's superiority.
### SAfER: Layer-Level Sensitivity Assessment for Efficient and Robust Neural Network Inference
- **Authors:** Edouard Yvinec, Arnaud Dapogny, Kevin Bailly
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2308.04753
- **Pdf link:** https://arxiv.org/pdf/2308.04753
- **Abstract**
Deep neural networks (DNNs) demonstrate outstanding performance across most computer vision tasks. Some critical applications, such as autonomous driving or medical imaging, also require investigation into their behavior and the reasons behind the decisions they make. In this vein, DNN attribution consists in studying the relationship between the predictions of a DNN and its inputs. Attribution methods have been adapted to highlight the most relevant weights or neurons in a DNN, allowing to more efficiently select which weights or neurons can be pruned. However, a limitation of these approaches is that weights are typically compared within each layer separately, while some layers might appear as more critical than others. In this work, we propose to investigate DNN layer importance, i.e. to estimate the sensitivity of the accuracy w.r.t. perturbations applied at the layer level. To do so, we propose a novel dataset to evaluate our method as well as future works. We benchmark a number of criteria and draw conclusions regarding how to assess DNN layer importance and, consequently, how to budgetize layers for increased DNN efficiency (with applications for DNN pruning and quantization), as well as robustness to hardware failure (e.g. bit swaps).
### View while Moving: Efficient Video Recognition in Long-untrimmed Videos
- **Authors:** Ye Tian, Mengyu Yang, Lanshan Zhang, Zhizhen Zhang, Yang Liu, Xiaohui Xie, Xirong Que, Wendong Wang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2308.04834
- **Pdf link:** https://arxiv.org/pdf/2308.04834
- **Abstract**
Recent adaptive methods for efficient video recognition mostly follow the two-stage paradigm of "preview-then-recognition" and have achieved great success on multiple video benchmarks. However, this two-stage paradigm involves two visits of raw frames from coarse-grained to fine-grained during inference (cannot be parallelized), and the captured spatiotemporal features cannot be reused in the second stage (due to varying granularity), being not friendly to efficiency and computation optimization. To this end, inspired by human cognition, we propose a novel recognition paradigm of "View while Moving" for efficient long-untrimmed video recognition. In contrast to the two-stage paradigm, our paradigm only needs to access the raw frame once. The two phases of coarse-grained sampling and fine-grained recognition are combined into unified spatiotemporal modeling, showing great performance. Moreover, we investigate the properties of semantic units in video and propose a hierarchical mechanism to efficiently capture and reason about the unit-level and video-level temporal semantics in long-untrimmed videos respectively. Extensive experiments on both long-untrimmed and short-trimmed videos demonstrate that our approach outperforms state-of-the-art methods in terms of accuracy as well as efficiency, yielding new efficiency and accuracy trade-offs for video spatiotemporal modeling.
### Constructing Holistic Spatio-Temporal Scene Graph for Video Semantic Role Labeling
- **Authors:** Yu Zhao, Hao Fei, Yixin Cao, Bobo Li, Meishan Zhang, Jianguo Wei, Min Zhang, Tat-Seng Chua
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL)
- **Arxiv link:** https://arxiv.org/abs/2308.05081
- **Pdf link:** https://arxiv.org/pdf/2308.05081
- **Abstract**
Video Semantic Role Labeling (VidSRL) aims to detect the salient events from given videos, by recognizing the predict-argument event structures and the interrelationships between events. While recent endeavors have put forth methods for VidSRL, they can be mostly subject to two key drawbacks, including the lack of fine-grained spatial scene perception and the insufficiently modeling of video temporality. Towards this end, this work explores a novel holistic spatio-temporal scene graph (namely HostSG) representation based on the existing dynamic scene graph structures, which well model both the fine-grained spatial semantics and temporal dynamics of videos for VidSRL. Built upon the HostSG, we present a nichetargeting VidSRL framework. A scene-event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high-level event semantic structure, resulting in an overall hierarchical scene-event (termed ICE) graph structure. We further perform iterative structure refinement to optimize the ICE graph, such that the overall structure representation can best coincide with end task demand. Finally, three subtask predictions of VidSRL are jointly decoded, where the end-to-end paradigm effectively avoids error propagation. On the benchmark dataset, our framework boosts significantly over the current best-performing model. Further analyses are shown for a better understanding of the advances of our methods.
## Keyword: raw image
There is no result
|
2.0
|
New submissions for Thu, 10 Aug 23 - ## Keyword: events
### Constructing Holistic Spatio-Temporal Scene Graph for Video Semantic Role Labeling
- **Authors:** Yu Zhao, Hao Fei, Yixin Cao, Bobo Li, Meishan Zhang, Jianguo Wei, Min Zhang, Tat-Seng Chua
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL)
- **Arxiv link:** https://arxiv.org/abs/2308.05081
- **Pdf link:** https://arxiv.org/pdf/2308.05081
- **Abstract**
Video Semantic Role Labeling (VidSRL) aims to detect the salient events from given videos, by recognizing the predict-argument event structures and the interrelationships between events. While recent endeavors have put forth methods for VidSRL, they can be mostly subject to two key drawbacks, including the lack of fine-grained spatial scene perception and the insufficiently modeling of video temporality. Towards this end, this work explores a novel holistic spatio-temporal scene graph (namely HostSG) representation based on the existing dynamic scene graph structures, which well model both the fine-grained spatial semantics and temporal dynamics of videos for VidSRL. Built upon the HostSG, we present a nichetargeting VidSRL framework. A scene-event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high-level event semantic structure, resulting in an overall hierarchical scene-event (termed ICE) graph structure. We further perform iterative structure refinement to optimize the ICE graph, such that the overall structure representation can best coincide with end task demand. Finally, three subtask predictions of VidSRL are jointly decoded, where the end-to-end paradigm effectively avoids error propagation. On the benchmark dataset, our framework boosts significantly over the current best-performing model. Further analyses are shown for a better understanding of the advances of our methods.
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
### Rendering Humans from Object-Occluded Monocular Videos
- **Authors:** Tiange Xiang, Adam Sun, Jiajun Wu, Ehsan Adeli, Li Fei-Fei
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2308.04622
- **Pdf link:** https://arxiv.org/pdf/2308.04622
- **Abstract**
3D understanding and rendering of moving humans from monocular videos is a challenging task. Despite recent progress, the task remains difficult in real-world scenarios, where obstacles may block the camera view and cause partial occlusions in the captured videos. Existing methods cannot handle such defects due to two reasons. First, the standard rendering strategy relies on point-point mapping, which could lead to dramatic disparities between the visible and occluded areas of the body. Second, the naive direct regression approach does not consider any feasibility criteria (ie, prior information) for rendering under occlusions. To tackle the above drawbacks, we present OccNeRF, a neural rendering method that achieves better rendering of humans in severely occluded scenes. As direct solutions to the two drawbacks, we propose surface-based rendering by integrating geometry and visibility priors. We validate our method on both simulated and real-world occlusions and demonstrate our method's superiority.
### Constructing Holistic Spatio-Temporal Scene Graph for Video Semantic Role Labeling
- **Authors:** Yu Zhao, Hao Fei, Yixin Cao, Bobo Li, Meishan Zhang, Jianguo Wei, Min Zhang, Tat-Seng Chua
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL)
- **Arxiv link:** https://arxiv.org/abs/2308.05081
- **Pdf link:** https://arxiv.org/pdf/2308.05081
- **Abstract**
Video Semantic Role Labeling (VidSRL) aims to detect the salient events from given videos, by recognizing the predict-argument event structures and the interrelationships between events. While recent endeavors have put forth methods for VidSRL, they can be mostly subject to two key drawbacks, including the lack of fine-grained spatial scene perception and the insufficiently modeling of video temporality. Towards this end, this work explores a novel holistic spatio-temporal scene graph (namely HostSG) representation based on the existing dynamic scene graph structures, which well model both the fine-grained spatial semantics and temporal dynamics of videos for VidSRL. Built upon the HostSG, we present a nichetargeting VidSRL framework. A scene-event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high-level event semantic structure, resulting in an overall hierarchical scene-event (termed ICE) graph structure. We further perform iterative structure refinement to optimize the ICE graph, such that the overall structure representation can best coincide with end task demand. Finally, three subtask predictions of VidSRL are jointly decoded, where the end-to-end paradigm effectively avoids error propagation. On the benchmark dataset, our framework boosts significantly over the current best-performing model. Further analyses are shown for a better understanding of the advances of our methods.
## Keyword: ISP
### Rendering Humans from Object-Occluded Monocular Videos
- **Authors:** Tiange Xiang, Adam Sun, Jiajun Wu, Ehsan Adeli, Li Fei-Fei
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2308.04622
- **Pdf link:** https://arxiv.org/pdf/2308.04622
- **Abstract**
3D understanding and rendering of moving humans from monocular videos is a challenging task. Despite recent progress, the task remains difficult in real-world scenarios, where obstacles may block the camera view and cause partial occlusions in the captured videos. Existing methods cannot handle such defects due to two reasons. First, the standard rendering strategy relies on point-point mapping, which could lead to dramatic disparities between the visible and occluded areas of the body. Second, the naive direct regression approach does not consider any feasibility criteria (ie, prior information) for rendering under occlusions. To tackle the above drawbacks, we present OccNeRF, a neural rendering method that achieves better rendering of humans in severely occluded scenes. As direct solutions to the two drawbacks, we propose surface-based rendering by integrating geometry and visibility priors. We validate our method on both simulated and real-world occlusions and demonstrate our method's superiority.
### Addressing Racial Bias in Facial Emotion Recognition
- **Authors:** Alex Fan, Xingshuo Xiao, Peter Washington
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Computers and Society (cs.CY)
- **Arxiv link:** https://arxiv.org/abs/2308.04674
- **Pdf link:** https://arxiv.org/pdf/2308.04674
- **Abstract**
Fairness in deep learning models trained with high-dimensional inputs and subjective labels remains a complex and understudied area. Facial emotion recognition, a domain where datasets are often racially imbalanced, can lead to models that yield disparate outcomes across racial groups. This study focuses on analyzing racial bias by sub-sampling training sets with varied racial distributions and assessing test performance across these simulations. Our findings indicate that smaller datasets with posed faces improve on both fairness and performance metrics as the simulations approach racial balance. Notably, the F1-score increases by $27.2\%$ points, and demographic parity increases by $15.7\%$ points on average across the simulations. However, in larger datasets with greater facial variation, fairness metrics generally remain constant, suggesting that racial balance by itself is insufficient to achieve parity in test performance across different racial groups.
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### Resource Constrained Model Compression via Minimax Optimization for Spiking Neural Networks
- **Authors:** Jue Chen, Huan Yuan, Jianchao Tan, Bin Chen, Chengru Song, Di Zhang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Multimedia (cs.MM)
- **Arxiv link:** https://arxiv.org/abs/2308.04672
- **Pdf link:** https://arxiv.org/pdf/2308.04672
- **Abstract**
Brain-inspired Spiking Neural Networks (SNNs) have the characteristics of event-driven and high energy-efficient, which are different from traditional Artificial Neural Networks (ANNs) when deployed on edge devices such as neuromorphic chips. Most previous work focuses on SNNs training strategies to improve model performance and brings larger and deeper network architectures. It is difficult to deploy these complex networks on resource-limited edge devices directly. To meet such demand, people compress SNNs very cautiously to balance the performance and the computation efficiency. Existing compression methods either iteratively pruned SNNs using weights norm magnitude or formulated the problem as a sparse learning optimization. We propose an improved end-to-end Minimax optimization method for this sparse learning problem to better balance the model performance and the computation efficiency. We also demonstrate that jointly applying compression and finetuning on SNNs is better than sequentially, especially for extreme compression ratios. The compressed SNN models achieved state-of-the-art (SOTA) performance on various benchmark datasets and architectures. Our code is available at https://github.com/chenjallen/Resource-Constrained-Compression-on-SNN.
## Keyword: RAW
### Rendering Humans from Object-Occluded Monocular Videos
- **Authors:** Tiange Xiang, Adam Sun, Jiajun Wu, Ehsan Adeli, Li Fei-Fei
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2308.04622
- **Pdf link:** https://arxiv.org/pdf/2308.04622
- **Abstract**
3D understanding and rendering of moving humans from monocular videos is a challenging task. Despite recent progress, the task remains difficult in real-world scenarios, where obstacles may block the camera view and cause partial occlusions in the captured videos. Existing methods cannot handle such defects due to two reasons. First, the standard rendering strategy relies on point-point mapping, which could lead to dramatic disparities between the visible and occluded areas of the body. Second, the naive direct regression approach does not consider any feasibility criteria (ie, prior information) for rendering under occlusions. To tackle the above drawbacks, we present OccNeRF, a neural rendering method that achieves better rendering of humans in severely occluded scenes. As direct solutions to the two drawbacks, we propose surface-based rendering by integrating geometry and visibility priors. We validate our method on both simulated and real-world occlusions and demonstrate our method's superiority.
### SAfER: Layer-Level Sensitivity Assessment for Efficient and Robust Neural Network Inference
- **Authors:** Edouard Yvinec, Arnaud Dapogny, Kevin Bailly
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2308.04753
- **Pdf link:** https://arxiv.org/pdf/2308.04753
- **Abstract**
Deep neural networks (DNNs) demonstrate outstanding performance across most computer vision tasks. Some critical applications, such as autonomous driving or medical imaging, also require investigation into their behavior and the reasons behind the decisions they make. In this vein, DNN attribution consists in studying the relationship between the predictions of a DNN and its inputs. Attribution methods have been adapted to highlight the most relevant weights or neurons in a DNN, allowing to more efficiently select which weights or neurons can be pruned. However, a limitation of these approaches is that weights are typically compared within each layer separately, while some layers might appear as more critical than others. In this work, we propose to investigate DNN layer importance, i.e. to estimate the sensitivity of the accuracy w.r.t. perturbations applied at the layer level. To do so, we propose a novel dataset to evaluate our method as well as future works. We benchmark a number of criteria and draw conclusions regarding how to assess DNN layer importance and, consequently, how to budgetize layers for increased DNN efficiency (with applications for DNN pruning and quantization), as well as robustness to hardware failure (e.g. bit swaps).
### View while Moving: Efficient Video Recognition in Long-untrimmed Videos
- **Authors:** Ye Tian, Mengyu Yang, Lanshan Zhang, Zhizhen Zhang, Yang Liu, Xiaohui Xie, Xirong Que, Wendong Wang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2308.04834
- **Pdf link:** https://arxiv.org/pdf/2308.04834
- **Abstract**
Recent adaptive methods for efficient video recognition mostly follow the two-stage paradigm of "preview-then-recognition" and have achieved great success on multiple video benchmarks. However, this two-stage paradigm involves two visits of raw frames from coarse-grained to fine-grained during inference (cannot be parallelized), and the captured spatiotemporal features cannot be reused in the second stage (due to varying granularity), being not friendly to efficiency and computation optimization. To this end, inspired by human cognition, we propose a novel recognition paradigm of "View while Moving" for efficient long-untrimmed video recognition. In contrast to the two-stage paradigm, our paradigm only needs to access the raw frame once. The two phases of coarse-grained sampling and fine-grained recognition are combined into unified spatiotemporal modeling, showing great performance. Moreover, we investigate the properties of semantic units in video and propose a hierarchical mechanism to efficiently capture and reason about the unit-level and video-level temporal semantics in long-untrimmed videos respectively. Extensive experiments on both long-untrimmed and short-trimmed videos demonstrate that our approach outperforms state-of-the-art methods in terms of accuracy as well as efficiency, yielding new efficiency and accuracy trade-offs for video spatiotemporal modeling.
### Constructing Holistic Spatio-Temporal Scene Graph for Video Semantic Role Labeling
- **Authors:** Yu Zhao, Hao Fei, Yixin Cao, Bobo Li, Meishan Zhang, Jianguo Wei, Min Zhang, Tat-Seng Chua
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL)
- **Arxiv link:** https://arxiv.org/abs/2308.05081
- **Pdf link:** https://arxiv.org/pdf/2308.05081
- **Abstract**
Video Semantic Role Labeling (VidSRL) aims to detect the salient events from given videos, by recognizing the predict-argument event structures and the interrelationships between events. While recent endeavors have put forth methods for VidSRL, they can be mostly subject to two key drawbacks, including the lack of fine-grained spatial scene perception and the insufficiently modeling of video temporality. Towards this end, this work explores a novel holistic spatio-temporal scene graph (namely HostSG) representation based on the existing dynamic scene graph structures, which well model both the fine-grained spatial semantics and temporal dynamics of videos for VidSRL. Built upon the HostSG, we present a nichetargeting VidSRL framework. A scene-event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high-level event semantic structure, resulting in an overall hierarchical scene-event (termed ICE) graph structure. We further perform iterative structure refinement to optimize the ICE graph, such that the overall structure representation can best coincide with end task demand. Finally, three subtask predictions of VidSRL are jointly decoded, where the end-to-end paradigm effectively avoids error propagation. On the benchmark dataset, our framework boosts significantly over the current best-performing model. Further analyses are shown for a better understanding of the advances of our methods.
## Keyword: raw image
There is no result
|
process
|
new submissions for thu aug keyword events constructing holistic spatio temporal scene graph for video semantic role labeling authors yu zhao hao fei yixin cao bobo li meishan zhang jianguo wei min zhang tat seng chua subjects computer vision and pattern recognition cs cv computation and language cs cl arxiv link pdf link abstract video semantic role labeling vidsrl aims to detect the salient events from given videos by recognizing the predict argument event structures and the interrelationships between events while recent endeavors have put forth methods for vidsrl they can be mostly subject to two key drawbacks including the lack of fine grained spatial scene perception and the insufficiently modeling of video temporality towards this end this work explores a novel holistic spatio temporal scene graph namely hostsg representation based on the existing dynamic scene graph structures which well model both the fine grained spatial semantics and temporal dynamics of videos for vidsrl built upon the hostsg we present a nichetargeting vidsrl framework a scene event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high level event semantic structure resulting in an overall hierarchical scene event termed ice graph structure we further perform iterative structure refinement to optimize the ice graph such that the overall structure representation can best coincide with end task demand finally three subtask predictions of vidsrl are jointly decoded where the end to end paradigm effectively avoids error propagation on the benchmark dataset our framework boosts significantly over the current best performing model further analyses are shown for a better understanding of the advances of our methods keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb rendering humans from object occluded monocular videos authors tiange xiang adam sun jiajun wu ehsan adeli li fei fei subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract understanding and rendering of moving humans from monocular videos is a challenging task despite recent progress the task remains difficult in real world scenarios where obstacles may block the camera view and cause partial occlusions in the captured videos existing methods cannot handle such defects due to two reasons first the standard rendering strategy relies on point point mapping which could lead to dramatic disparities between the visible and occluded areas of the body second the naive direct regression approach does not consider any feasibility criteria ie prior information for rendering under occlusions to tackle the above drawbacks we present occnerf a neural rendering method that achieves better rendering of humans in severely occluded scenes as direct solutions to the two drawbacks we propose surface based rendering by integrating geometry and visibility priors we validate our method on both simulated and real world occlusions and demonstrate our method s superiority constructing holistic spatio temporal scene graph for video semantic role labeling authors yu zhao hao fei yixin cao bobo li meishan zhang jianguo wei min zhang tat seng chua subjects computer vision and pattern recognition cs cv computation and language cs cl arxiv link pdf link abstract video semantic role labeling vidsrl aims to detect the salient events from given videos by recognizing the predict argument event structures and the interrelationships between events while recent endeavors have put forth methods for vidsrl they can be mostly subject to two key drawbacks including the lack of fine grained spatial scene perception and the insufficiently modeling of video temporality towards this end this work explores a novel holistic spatio temporal scene graph namely hostsg representation based on the existing dynamic scene graph structures which well model both the fine grained spatial semantics and temporal dynamics of videos for vidsrl built upon the hostsg we present a nichetargeting vidsrl framework a scene event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high level event semantic structure resulting in an overall hierarchical scene event termed ice graph structure we further perform iterative structure refinement to optimize the ice graph such that the overall structure representation can best coincide with end task demand finally three subtask predictions of vidsrl are jointly decoded where the end to end paradigm effectively avoids error propagation on the benchmark dataset our framework boosts significantly over the current best performing model further analyses are shown for a better understanding of the advances of our methods keyword isp rendering humans from object occluded monocular videos authors tiange xiang adam sun jiajun wu ehsan adeli li fei fei subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract understanding and rendering of moving humans from monocular videos is a challenging task despite recent progress the task remains difficult in real world scenarios where obstacles may block the camera view and cause partial occlusions in the captured videos existing methods cannot handle such defects due to two reasons first the standard rendering strategy relies on point point mapping which could lead to dramatic disparities between the visible and occluded areas of the body second the naive direct regression approach does not consider any feasibility criteria ie prior information for rendering under occlusions to tackle the above drawbacks we present occnerf a neural rendering method that achieves better rendering of humans in severely occluded scenes as direct solutions to the two drawbacks we propose surface based rendering by integrating geometry and visibility priors we validate our method on both simulated and real world occlusions and demonstrate our method s superiority addressing racial bias in facial emotion recognition authors alex fan xingshuo xiao peter washington subjects computer vision and pattern recognition cs cv artificial intelligence cs ai computers and society cs cy arxiv link pdf link abstract fairness in deep learning models trained with high dimensional inputs and subjective labels remains a complex and understudied area facial emotion recognition a domain where datasets are often racially imbalanced can lead to models that yield disparate outcomes across racial groups this study focuses on analyzing racial bias by sub sampling training sets with varied racial distributions and assessing test performance across these simulations our findings indicate that smaller datasets with posed faces improve on both fairness and performance metrics as the simulations approach racial balance notably the score increases by points and demographic parity increases by points on average across the simulations however in larger datasets with greater facial variation fairness metrics generally remain constant suggesting that racial balance by itself is insufficient to achieve parity in test performance across different racial groups keyword image signal processing there is no result keyword image signal process there is no result keyword compression resource constrained model compression via minimax optimization for spiking neural networks authors jue chen huan yuan jianchao tan bin chen chengru song di zhang subjects computer vision and pattern recognition cs cv artificial intelligence cs ai multimedia cs mm arxiv link pdf link abstract brain inspired spiking neural networks snns have the characteristics of event driven and high energy efficient which are different from traditional artificial neural networks anns when deployed on edge devices such as neuromorphic chips most previous work focuses on snns training strategies to improve model performance and brings larger and deeper network architectures it is difficult to deploy these complex networks on resource limited edge devices directly to meet such demand people compress snns very cautiously to balance the performance and the computation efficiency existing compression methods either iteratively pruned snns using weights norm magnitude or formulated the problem as a sparse learning optimization we propose an improved end to end minimax optimization method for this sparse learning problem to better balance the model performance and the computation efficiency we also demonstrate that jointly applying compression and finetuning on snns is better than sequentially especially for extreme compression ratios the compressed snn models achieved state of the art sota performance on various benchmark datasets and architectures our code is available at keyword raw rendering humans from object occluded monocular videos authors tiange xiang adam sun jiajun wu ehsan adeli li fei fei subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract understanding and rendering of moving humans from monocular videos is a challenging task despite recent progress the task remains difficult in real world scenarios where obstacles may block the camera view and cause partial occlusions in the captured videos existing methods cannot handle such defects due to two reasons first the standard rendering strategy relies on point point mapping which could lead to dramatic disparities between the visible and occluded areas of the body second the naive direct regression approach does not consider any feasibility criteria ie prior information for rendering under occlusions to tackle the above drawbacks we present occnerf a neural rendering method that achieves better rendering of humans in severely occluded scenes as direct solutions to the two drawbacks we propose surface based rendering by integrating geometry and visibility priors we validate our method on both simulated and real world occlusions and demonstrate our method s superiority safer layer level sensitivity assessment for efficient and robust neural network inference authors edouard yvinec arnaud dapogny kevin bailly subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract deep neural networks dnns demonstrate outstanding performance across most computer vision tasks some critical applications such as autonomous driving or medical imaging also require investigation into their behavior and the reasons behind the decisions they make in this vein dnn attribution consists in studying the relationship between the predictions of a dnn and its inputs attribution methods have been adapted to highlight the most relevant weights or neurons in a dnn allowing to more efficiently select which weights or neurons can be pruned however a limitation of these approaches is that weights are typically compared within each layer separately while some layers might appear as more critical than others in this work we propose to investigate dnn layer importance i e to estimate the sensitivity of the accuracy w r t perturbations applied at the layer level to do so we propose a novel dataset to evaluate our method as well as future works we benchmark a number of criteria and draw conclusions regarding how to assess dnn layer importance and consequently how to budgetize layers for increased dnn efficiency with applications for dnn pruning and quantization as well as robustness to hardware failure e g bit swaps view while moving efficient video recognition in long untrimmed videos authors ye tian mengyu yang lanshan zhang zhizhen zhang yang liu xiaohui xie xirong que wendong wang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract recent adaptive methods for efficient video recognition mostly follow the two stage paradigm of preview then recognition and have achieved great success on multiple video benchmarks however this two stage paradigm involves two visits of raw frames from coarse grained to fine grained during inference cannot be parallelized and the captured spatiotemporal features cannot be reused in the second stage due to varying granularity being not friendly to efficiency and computation optimization to this end inspired by human cognition we propose a novel recognition paradigm of view while moving for efficient long untrimmed video recognition in contrast to the two stage paradigm our paradigm only needs to access the raw frame once the two phases of coarse grained sampling and fine grained recognition are combined into unified spatiotemporal modeling showing great performance moreover we investigate the properties of semantic units in video and propose a hierarchical mechanism to efficiently capture and reason about the unit level and video level temporal semantics in long untrimmed videos respectively extensive experiments on both long untrimmed and short trimmed videos demonstrate that our approach outperforms state of the art methods in terms of accuracy as well as efficiency yielding new efficiency and accuracy trade offs for video spatiotemporal modeling constructing holistic spatio temporal scene graph for video semantic role labeling authors yu zhao hao fei yixin cao bobo li meishan zhang jianguo wei min zhang tat seng chua subjects computer vision and pattern recognition cs cv computation and language cs cl arxiv link pdf link abstract video semantic role labeling vidsrl aims to detect the salient events from given videos by recognizing the predict argument event structures and the interrelationships between events while recent endeavors have put forth methods for vidsrl they can be mostly subject to two key drawbacks including the lack of fine grained spatial scene perception and the insufficiently modeling of video temporality towards this end this work explores a novel holistic spatio temporal scene graph namely hostsg representation based on the existing dynamic scene graph structures which well model both the fine grained spatial semantics and temporal dynamics of videos for vidsrl built upon the hostsg we present a nichetargeting vidsrl framework a scene event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high level event semantic structure resulting in an overall hierarchical scene event termed ice graph structure we further perform iterative structure refinement to optimize the ice graph such that the overall structure representation can best coincide with end task demand finally three subtask predictions of vidsrl are jointly decoded where the end to end paradigm effectively avoids error propagation on the benchmark dataset our framework boosts significantly over the current best performing model further analyses are shown for a better understanding of the advances of our methods keyword raw image there is no result
| 1
|
521
| 2,994,323,890
|
IssuesEvent
|
2015-07-22 11:00:03
|
genomizer/genomizer-server
|
https://api.github.com/repos/genomizer/genomizer-server
|
closed
|
genomizer-server-tester fails nondeterministically on Travis
|
bug Medium priority Processing
|
Sometimes I see nondeterministic single-test failures in `genomizer-server-tester` tests on Travis ([example](https://travis-ci.org/genomizer/genomizer-server/builds/65365650)) that go away after restarting the build. May be a race condition somewhere - if the issue persists, it can be temporarily papered over by ignoring `genomizer-server-tester`'s exit code. Won't have time to investigate, so it's just FYI.
|
1.0
|
genomizer-server-tester fails nondeterministically on Travis - Sometimes I see nondeterministic single-test failures in `genomizer-server-tester` tests on Travis ([example](https://travis-ci.org/genomizer/genomizer-server/builds/65365650)) that go away after restarting the build. May be a race condition somewhere - if the issue persists, it can be temporarily papered over by ignoring `genomizer-server-tester`'s exit code. Won't have time to investigate, so it's just FYI.
|
process
|
genomizer server tester fails nondeterministically on travis sometimes i see nondeterministic single test failures in genomizer server tester tests on travis that go away after restarting the build may be a race condition somewhere if the issue persists it can be temporarily papered over by ignoring genomizer server tester s exit code won t have time to investigate so it s just fyi
| 1
|
7,753
| 10,866,673,790
|
IssuesEvent
|
2019-11-14 21:48:14
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[FEATURE] Add feature filter algorithm
|
3.2 Automatic new feature Processing Alg
|
Original commit: https://github.com/qgis/QGIS/commit/147f6e71c46a5f1fd3d5b0da97f0a7028568552f by m-kuhn
Adds a new "feature filter" algorithm. This algorithm takes a list of expressions and
creates an output for each of them. Matching features from the input layer are sent
to the outputs.
This allows for an algorithm to only handle a subset of incoming features or to send
different features to different outputs.
The algorithm is only available in the modeler.
|
1.0
|
[FEATURE] Add feature filter algorithm - Original commit: https://github.com/qgis/QGIS/commit/147f6e71c46a5f1fd3d5b0da97f0a7028568552f by m-kuhn
Adds a new "feature filter" algorithm. This algorithm takes a list of expressions and
creates an output for each of them. Matching features from the input layer are sent
to the outputs.
This allows for an algorithm to only handle a subset of incoming features or to send
different features to different outputs.
The algorithm is only available in the modeler.
|
process
|
add feature filter algorithm original commit by m kuhn adds a new feature filter algorithm this algorithm takes a list of expressions and creates an output for each of them matching features from the input layer are sent to the outputs this allows for an algorithm to only handle a subset of incoming features or to send different features to different outputs the algorithm is only available in the modeler
| 1
|
13,814
| 16,576,213,168
|
IssuesEvent
|
2021-05-31 05:24:56
|
arcus-azure/arcus.messaging
|
https://api.github.com/repos/arcus-azure/arcus.messaging
|
closed
|
Add ContentType to AzureServiceBusSystemProperties (Arcus.Messaging.Abstractions.ServiceBus)
|
area:message-processing enhancement
|
**Is your feature request related to a problem? Please describe.**
Currently we are trying to make a distinction between messages based on the contentType but unfortunately, arcus doesn't support this property.
**Describe the solution you'd like**
In AzureServiceBusSystemProperties add the following property : ContentType. This is already a property from ServiceBusReceivedMessage but is currently not getting used

|
1.0
|
Add ContentType to AzureServiceBusSystemProperties (Arcus.Messaging.Abstractions.ServiceBus) - **Is your feature request related to a problem? Please describe.**
Currently we are trying to make a distinction between messages based on the contentType but unfortunately, arcus doesn't support this property.
**Describe the solution you'd like**
In AzureServiceBusSystemProperties add the following property : ContentType. This is already a property from ServiceBusReceivedMessage but is currently not getting used

|
process
|
add contenttype to azureservicebussystemproperties arcus messaging abstractions servicebus is your feature request related to a problem please describe currently we are trying to make a distinction between messages based on the contenttype but unfortunately arcus doesn t support this property describe the solution you d like in azureservicebussystemproperties add the following property contenttype this is already a property from servicebusreceivedmessage but is currently not getting used
| 1
|
371
| 2,499,901,565
|
IssuesEvent
|
2015-01-08 07:42:41
|
fossology/fossology
|
https://api.github.com/repos/fossology/fossology
|
opened
|
Binary analysis for open source discovery
|
Component: Rank Component: Tester Priority: Low Status: New Tracker: Feature
|
---
Author Name: **Paul Holland**
Original Redmine Issue: 2457, http://www.fossology.org/issues/2457
Original Date: 2012/06/15
---
None
|
1.0
|
Binary analysis for open source discovery - ---
Author Name: **Paul Holland**
Original Redmine Issue: 2457, http://www.fossology.org/issues/2457
Original Date: 2012/06/15
---
None
|
non_process
|
binary analysis for open source discovery author name paul holland original redmine issue original date none
| 0
|
102,105
| 12,745,233,297
|
IssuesEvent
|
2020-06-26 13:55:21
|
magento/magento2
|
https://api.github.com/repos/magento/magento2
|
closed
|
Shipping Title translation is not being translated when you switch the langauge in Checkout Cart Page
|
Area: Design/Frontend Component: Checkout Component: Shipping Issue: Confirmed Issue: Format is valid Issue: Ready for Work Progress: dev in progress Reproduced on 2.4.x
|
<!---
Please review our guidelines before adding a new issue: https://github.com/magento/magento2/wiki/Issue-reporting-guidelines
Fields marked with (*) are required. Please don't remove the template.
-->
### Preconditions (*)
<!---
Provide the exact Magento version (example: 2.3.2) and any important information on the environment where bug is reproducible.
-->
1. Magneot 2.3.3
### Steps to reproduce (*)
<!---
Important: Provide a set of clear steps to reproduce this bug. We can not provide support without clear instructions on how to reproduce.
-->
1. Add Two Store views
2. Change the Label for Free Shipping method in the second store
3. add a new product to be able to order it
4. disable multi shipping address
5. add the product to cart
6. go to the Checkout cart page
7. switch the languages
8. look at the shipping label in the right sidebar
### Expected result (*)
<!--- Tell us what do you expect to happen. -->
1. when we switch the language we have to see the correct label per store view
### Actual result (*)
<!--- Tell us what happened instead. Include error messages and issues. -->
1. we see don't see the correct language and we see the other language , when we change any information in the estimation shipping cost we got the correct label
|
1.0
|
Shipping Title translation is not being translated when you switch the langauge in Checkout Cart Page - <!---
Please review our guidelines before adding a new issue: https://github.com/magento/magento2/wiki/Issue-reporting-guidelines
Fields marked with (*) are required. Please don't remove the template.
-->
### Preconditions (*)
<!---
Provide the exact Magento version (example: 2.3.2) and any important information on the environment where bug is reproducible.
-->
1. Magneot 2.3.3
### Steps to reproduce (*)
<!---
Important: Provide a set of clear steps to reproduce this bug. We can not provide support without clear instructions on how to reproduce.
-->
1. Add Two Store views
2. Change the Label for Free Shipping method in the second store
3. add a new product to be able to order it
4. disable multi shipping address
5. add the product to cart
6. go to the Checkout cart page
7. switch the languages
8. look at the shipping label in the right sidebar
### Expected result (*)
<!--- Tell us what do you expect to happen. -->
1. when we switch the language we have to see the correct label per store view
### Actual result (*)
<!--- Tell us what happened instead. Include error messages and issues. -->
1. we see don't see the correct language and we see the other language , when we change any information in the estimation shipping cost we got the correct label
|
non_process
|
shipping title translation is not being translated when you switch the langauge in checkout cart page please review our guidelines before adding a new issue fields marked with are required please don t remove the template preconditions provide the exact magento version example and any important information on the environment where bug is reproducible magneot steps to reproduce important provide a set of clear steps to reproduce this bug we can not provide support without clear instructions on how to reproduce add two store views change the label for free shipping method in the second store add a new product to be able to order it disable multi shipping address add the product to cart go to the checkout cart page switch the languages look at the shipping label in the right sidebar expected result when we switch the language we have to see the correct label per store view actual result we see don t see the correct language and we see the other language when we change any information in the estimation shipping cost we got the correct label
| 0
|
83,633
| 24,107,857,969
|
IssuesEvent
|
2022-09-20 08:56:47
|
openvinotoolkit/openvino
|
https://api.github.com/repos/openvinotoolkit/openvino
|
closed
|
Build error when build python api from resource
|
bug category: build support_request
|
##### System information (version)
<!-- Example
- OpenVINO => 2022.1
- Operating System / Platform => Windows 64 Bit
- Compiler => Visual Studio 2017
- Problem classification: Model Conversion
- Framework: TensorFlow (if applicable)
- Model name: ResNet50 (if applicable)
-->
- OpenVINO=> 2022.1
- Operating System / Platform => windows11
- Compiler => vs2022
- Problem classification =>build error
##### Detailed description
Hi, when I want to build python api from resource follow https://github.com/openvinotoolkit/openvino/wiki/BuildingForWindows.
I used `` cmake -DENABLE_TESTS=ON -DENABLE_FUNCTIONAL_TESTS=ON -DENABLE_PYTHON=ON -DPYTHON_EXECUTABLE="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\python.exe" -DPYTHON_LIBRARY="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\libs\python38.lib" -DPYTHON_INCLUDE_DIR="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\include" -DENABLE_OV_ONNX_FRONTEND=ON .. `` then `` cmake --build . --config Release -j 8 ``
but there were some wrong :
```
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(10,33): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.
vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(613,19): message : see previous definition of 'private: static FeStat FrontEndMockPy::m_stat' [C:\openvino\build\src\bindings\pyth
on\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(10,33): error C2491: 'FrontEndMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mock
\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(11,38): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.
vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(429,22): message : see previous definition of 'private: static ModelStat InputModelMockPy::m_stat' [C:\openvino\build\src\bindings
\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(11,38): error C2491: 'InputModelMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mo
ck\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(12,33): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.
vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(91,22): message : see previous definition of 'private: static PlaceStat PlaceMockPy::m_stat' [C:\openvino\build\src\bindings\pytho
n\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(12,33): error C2491: 'PlaceMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mock\mo
ck_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(17,53): error C2491: 'GetAPIVersion': definition of dllimport function not allowed [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\o
penvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(21,45): error C2491: 'GetFrontEndData': definition of dllimport function not allowed [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend
\openvino_mock_py_frontend.vcxproj]
```
Any help? Thanks!
##### Steps to reproduce
<!--
Describe your problem and steps you've done before you got to this point.
to add code example fence it with triple backticks and optional file extension
```.cpp
// C++ code example
```
or attach as .txt or .zip file
-->
##### Issue submission checklist
- [ ] I report the issue, it's not a question
<!--
OpenVINO team works with support forum, Stack Overflow and other communities
to discuss problems. Tickets with question without real issue statement will be
closed.
-->
- [ ] I checked the problem with documentation, FAQ, open issues, Stack Overflow, etc and have not found solution
<!--
Places to check:
* OpenVINO documentation: https://docs.openvino.ai/
* OpenVINO forum: https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/bd-p/distribution-openvino-toolkit
* OpenVINO issue tracker: https://github.com/openvinotoolkit/openvino/issues?q=is%3Aissue
* Stack Overflow branch: https://stackoverflow.com/questions/tagged/openvino
-->
- [ ] There is reproducer code and related data files: images, videos, models, etc.
<!--
The best reproducer -- test case for OpenVINO that we can add to the library.
-->
|
1.0
|
Build error when build python api from resource - ##### System information (version)
<!-- Example
- OpenVINO => 2022.1
- Operating System / Platform => Windows 64 Bit
- Compiler => Visual Studio 2017
- Problem classification: Model Conversion
- Framework: TensorFlow (if applicable)
- Model name: ResNet50 (if applicable)
-->
- OpenVINO=> 2022.1
- Operating System / Platform => windows11
- Compiler => vs2022
- Problem classification =>build error
##### Detailed description
Hi, when I want to build python api from resource follow https://github.com/openvinotoolkit/openvino/wiki/BuildingForWindows.
I used `` cmake -DENABLE_TESTS=ON -DENABLE_FUNCTIONAL_TESTS=ON -DENABLE_PYTHON=ON -DPYTHON_EXECUTABLE="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\python.exe" -DPYTHON_LIBRARY="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\libs\python38.lib" -DPYTHON_INCLUDE_DIR="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\include" -DENABLE_OV_ONNX_FRONTEND=ON .. `` then `` cmake --build . --config Release -j 8 ``
but there were some wrong :
```
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(10,33): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.
vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(613,19): message : see previous definition of 'private: static FeStat FrontEndMockPy::m_stat' [C:\openvino\build\src\bindings\pyth
on\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(10,33): error C2491: 'FrontEndMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mock
\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(11,38): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.
vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(429,22): message : see previous definition of 'private: static ModelStat InputModelMockPy::m_stat' [C:\openvino\build\src\bindings
\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(11,38): error C2491: 'InputModelMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mo
ck\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(12,33): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.
vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(91,22): message : see previous definition of 'private: static PlaceStat PlaceMockPy::m_stat' [C:\openvino\build\src\bindings\pytho
n\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(12,33): error C2491: 'PlaceMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mock\mo
ck_py_frontend\openvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(17,53): error C2491: 'GetAPIVersion': definition of dllimport function not allowed [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\o
penvino_mock_py_frontend.vcxproj]
C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(21,45): error C2491: 'GetFrontEndData': definition of dllimport function not allowed [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend
\openvino_mock_py_frontend.vcxproj]
```
Any help? Thanks!
##### Steps to reproduce
<!--
Describe your problem and steps you've done before you got to this point.
to add code example fence it with triple backticks and optional file extension
```.cpp
// C++ code example
```
or attach as .txt or .zip file
-->
##### Issue submission checklist
- [ ] I report the issue, it's not a question
<!--
OpenVINO team works with support forum, Stack Overflow and other communities
to discuss problems. Tickets with question without real issue statement will be
closed.
-->
- [ ] I checked the problem with documentation, FAQ, open issues, Stack Overflow, etc and have not found solution
<!--
Places to check:
* OpenVINO documentation: https://docs.openvino.ai/
* OpenVINO forum: https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/bd-p/distribution-openvino-toolkit
* OpenVINO issue tracker: https://github.com/openvinotoolkit/openvino/issues?q=is%3Aissue
* Stack Overflow branch: https://stackoverflow.com/questions/tagged/openvino
-->
- [ ] There is reproducer code and related data files: images, videos, models, etc.
<!--
The best reproducer -- test case for OpenVINO that we can add to the library.
-->
|
non_process
|
build error when build python api from resource system information version example openvino operating system platform windows bit compiler visual studio problem classification model conversion framework tensorflow if applicable model name if applicable openvino operating system platform compiler problem classification build error detailed description hi when i want to build python api from resource follow i used cmake denable tests on denable functional tests on denable python on dpython executable c users local admin appdata local programs python python exe dpython library c users local admin appdata local programs python libs lib dpython include dir c users local admin appdata local programs python include denable ov onnx frontend on then cmake build config release j but there were some wrong c openvino src bindings python tests mock mock py frontend src mock py frontend cpp warning m stat inconsistent dll linkage c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend include mock py frontend mock py frontend hpp message see previous definition of private static festat frontendmockpy m stat c openvino build src bindings pyth on tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp error frontendmockpy m stat definition of dllimport static data member not allowed c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp warning m stat inconsistent dll linkage c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend include mock py frontend mock py frontend hpp message see previous definition of private static modelstat inputmodelmockpy m stat c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp error inputmodelmockpy m stat definition of dllimport static data member not allowed c openvino build src bindings python tests mo ck mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp warning m stat inconsistent dll linkage c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend include mock py frontend mock py frontend hpp message see previous definition of private static placestat placemockpy m stat c openvino build src bindings pytho n tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp error placemockpy m stat definition of dllimport static data member not allowed c openvino build src bindings python tests mock mo ck py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp error getapiversion definition of dllimport function not allowed c openvino build src bindings python tests mock mock py frontend o penvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp error getfrontenddata definition of dllimport function not allowed c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj any help thanks steps to reproduce describe your problem and steps you ve done before you got to this point to add code example fence it with triple backticks and optional file extension cpp c code example or attach as txt or zip file issue submission checklist i report the issue it s not a question openvino team works with support forum stack overflow and other communities to discuss problems tickets with question without real issue statement will be closed i checked the problem with documentation faq open issues stack overflow etc and have not found solution places to check openvino documentation openvino forum openvino issue tracker stack overflow branch there is reproducer code and related data files images videos models etc the best reproducer test case for openvino that we can add to the library
| 0
|
31,512
| 7,375,323,920
|
IssuesEvent
|
2018-03-13 23:49:40
|
MicrosoftDocs/live-share
|
https://api.github.com/repos/MicrosoftDocs/live-share
|
closed
|
[VS Code] Sign-in failed.
|
area: identity and sign-in duplicate needs more info vscode
|
## Error:
Sign-in failed.
## Steps to Reproduce:
1. Connect to NetMotion (VPN)
2. Open Visual Studio or Visual Studio Code.
3. Attempt to start a new sharing session (Visual Studio) or sign into Live Share (VS Code)
4. Visual Studio will immediately fail with the error:
> [2018-03-09 22:02:37.396 VSIX I] Acquired an access token for the VSO account '<14:eb047f03>'
> [2018-03-09 22:02:37.420 VSIX E] Failed to create a collaboration session. An internal error occurred. Please contact Microsoft Product Support Services.
VS Code will get to 'Ready to Collaborate' page on browser, but will display 'Sign-in Failed' on actual instance of VS Code.
Note: If I disable VPNing, which probably also disables use of my corporate proxy, everything works as expected.
Note: My system currently works (on the VPN + Proxy) with Spotify, MS Teams, NuGet in Visual Studio, and Docker, which either natively support IE proxy settings or have their own setup for proxy details.
Note: When I use bash shell on windows and CURL https://download.microsoft.com, it fails with `(56) Proxy CONNECT aborted`.
Note: I have the HTTP_PROXY and HTTPS_PROXY environment variables set up with my corporate proxy url (url only, no credentials).
||Version Data|
|-:|:-|
|**extensionName**|VSLS|
|**extensionVersion**|0.2.206|
|**protocolVersion**|2.1|
|**applicationName**|VSCode|
|**applicationVersion**|1.21.0|
|**platformName**|Windows|
|**platformVersion**|10.0.14393|
|
1.0
|
[VS Code] Sign-in failed. -
## Error:
Sign-in failed.
## Steps to Reproduce:
1. Connect to NetMotion (VPN)
2. Open Visual Studio or Visual Studio Code.
3. Attempt to start a new sharing session (Visual Studio) or sign into Live Share (VS Code)
4. Visual Studio will immediately fail with the error:
> [2018-03-09 22:02:37.396 VSIX I] Acquired an access token for the VSO account '<14:eb047f03>'
> [2018-03-09 22:02:37.420 VSIX E] Failed to create a collaboration session. An internal error occurred. Please contact Microsoft Product Support Services.
VS Code will get to 'Ready to Collaborate' page on browser, but will display 'Sign-in Failed' on actual instance of VS Code.
Note: If I disable VPNing, which probably also disables use of my corporate proxy, everything works as expected.
Note: My system currently works (on the VPN + Proxy) with Spotify, MS Teams, NuGet in Visual Studio, and Docker, which either natively support IE proxy settings or have their own setup for proxy details.
Note: When I use bash shell on windows and CURL https://download.microsoft.com, it fails with `(56) Proxy CONNECT aborted`.
Note: I have the HTTP_PROXY and HTTPS_PROXY environment variables set up with my corporate proxy url (url only, no credentials).
||Version Data|
|-:|:-|
|**extensionName**|VSLS|
|**extensionVersion**|0.2.206|
|**protocolVersion**|2.1|
|**applicationName**|VSCode|
|**applicationVersion**|1.21.0|
|**platformName**|Windows|
|**platformVersion**|10.0.14393|
|
non_process
|
sign in failed error sign in failed steps to reproduce connect to netmotion vpn open visual studio or visual studio code attempt to start a new sharing session visual studio or sign into live share vs code visual studio will immediately fail with the error acquired an access token for the vso account failed to create a collaboration session an internal error occurred please contact microsoft product support services vs code will get to ready to collaborate page on browser but will display sign in failed on actual instance of vs code note if i disable vpning which probably also disables use of my corporate proxy everything works as expected note my system currently works on the vpn proxy with spotify ms teams nuget in visual studio and docker which either natively support ie proxy settings or have their own setup for proxy details note when i use bash shell on windows and curl it fails with proxy connect aborted note i have the http proxy and https proxy environment variables set up with my corporate proxy url url only no credentials version data extensionname vsls extensionversion protocolversion applicationname vscode applicationversion platformname windows platformversion
| 0
|
20,542
| 27,192,655,179
|
IssuesEvent
|
2023-02-20 00:05:19
|
cse442-at-ub/project_s23-team-infinity
|
https://api.github.com/repos/cse442-at-ub/project_s23-team-infinity
|
opened
|
Syntax Taskes
|
Processing Task
|
Task 1:
Print Hello World in php
Test 1:
Check if Hello World was printed:
Task 2:
Loop and print number 1 to 10 in php
Test 2:
Check if all numbers have been printed
Task 3:
Created a function that when called with parameters of 2 numbers will add then compare if it is greater than 10 and print out yes or no
Test 3.1:
Input: 3 and 4
Output: no
Test 3.2:
Input: 5 and 6
Output: yes
|
1.0
|
Syntax Taskes - Task 1:
Print Hello World in php
Test 1:
Check if Hello World was printed:
Task 2:
Loop and print number 1 to 10 in php
Test 2:
Check if all numbers have been printed
Task 3:
Created a function that when called with parameters of 2 numbers will add then compare if it is greater than 10 and print out yes or no
Test 3.1:
Input: 3 and 4
Output: no
Test 3.2:
Input: 5 and 6
Output: yes
|
process
|
syntax taskes task print hello world in php test check if hello world was printed task loop and print number to in php test check if all numbers have been printed task created a function that when called with parameters of numbers will add then compare if it is greater than and print out yes or no test input and output no test input and output yes
| 1
|
158,293
| 12,412,181,869
|
IssuesEvent
|
2020-05-22 10:02:37
|
aliasrobotics/RVD
|
https://api.github.com/repos/aliasrobotics/RVD
|
opened
|
(warning) %d in format string (no. 7) requires 'int' but the argument type is 'unsigned int'.
|
bug cppcheck static analysis testing triage
|
```yaml
{
"id": 1,
"title": "(warning) %d in format string (no. 7) requires 'int' but the argument type is 'unsigned int'.",
"type": "bug",
"description": "[src/opencv3/3rdparty/libjasper/jpc_cs.c:568]: (warning) %d in format string (no. 7) requires 'int' but the argument type is 'unsigned int'.",
"cwe": "None",
"cve": "None",
"keywords": [
"cppcheck",
"static analysis",
"testing",
"triage",
"bug"
],
"system": "src/opencv3/3rdparty/libjasper/jpc_cs.c",
"vendor": null,
"severity": {
"rvss-score": 0,
"rvss-vector": "",
"severity-description": "",
"cvss-score": 0,
"cvss-vector": ""
},
"links": "",
"flaw": {
"phase": "testing",
"specificity": "N/A",
"architectural-location": "N/A",
"application": "N/A",
"subsystem": "N/A",
"package": "N/A",
"languages": "None",
"date-detected": "2020-05-22 (10:02)",
"detected-by": "Alias Robotics",
"detected-by-method": "testing static",
"date-reported": "2020-05-22 (10:02)",
"reported-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"issue": "",
"reproducibility": "always",
"trace": "",
"reproduction": "See artifacts below (if available)",
"reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_kinetic/-/jobs/563367426/artifacts/download"
},
"exploitation": {
"description": "",
"exploitation-image": "",
"exploitation-vector": ""
},
"mitigation": {
"description": "",
"pull-request": "",
"date-mitigation": ""
}
}
```
|
1.0
|
(warning) %d in format string (no. 7) requires 'int' but the argument type is 'unsigned int'. - ```yaml
{
"id": 1,
"title": "(warning) %d in format string (no. 7) requires 'int' but the argument type is 'unsigned int'.",
"type": "bug",
"description": "[src/opencv3/3rdparty/libjasper/jpc_cs.c:568]: (warning) %d in format string (no. 7) requires 'int' but the argument type is 'unsigned int'.",
"cwe": "None",
"cve": "None",
"keywords": [
"cppcheck",
"static analysis",
"testing",
"triage",
"bug"
],
"system": "src/opencv3/3rdparty/libjasper/jpc_cs.c",
"vendor": null,
"severity": {
"rvss-score": 0,
"rvss-vector": "",
"severity-description": "",
"cvss-score": 0,
"cvss-vector": ""
},
"links": "",
"flaw": {
"phase": "testing",
"specificity": "N/A",
"architectural-location": "N/A",
"application": "N/A",
"subsystem": "N/A",
"package": "N/A",
"languages": "None",
"date-detected": "2020-05-22 (10:02)",
"detected-by": "Alias Robotics",
"detected-by-method": "testing static",
"date-reported": "2020-05-22 (10:02)",
"reported-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"issue": "",
"reproducibility": "always",
"trace": "",
"reproduction": "See artifacts below (if available)",
"reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_kinetic/-/jobs/563367426/artifacts/download"
},
"exploitation": {
"description": "",
"exploitation-image": "",
"exploitation-vector": ""
},
"mitigation": {
"description": "",
"pull-request": "",
"date-mitigation": ""
}
}
```
|
non_process
|
warning d in format string no requires int but the argument type is unsigned int yaml id title warning d in format string no requires int but the argument type is unsigned int type bug description warning d in format string no requires int but the argument type is unsigned int cwe none cve none keywords cppcheck static analysis testing triage bug system src libjasper jpc cs c vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity n a architectural location n a application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace reproduction see artifacts below if available reproduction image gitlab com aliasrobotics offensive alurity pipelines active pipeline ros kinetic jobs artifacts download exploitation description exploitation image exploitation vector mitigation description pull request date mitigation
| 0
|
6,778
| 9,915,064,612
|
IssuesEvent
|
2019-06-28 15:50:50
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Unable to edit the variables after deploying the Solution
|
assigned-to-author automation/svc process-automation/subsvc product-question triaged
|
Hi,
I need to exclude the VM's for this solution and trying to edit External_ExcludeVMNames variable , im not getting any option to edit .
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 225c9d05-83dd-b006-0025-3753f5ab25bf
* Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096
* Content: [Start/Stop VMs during off-hours solution](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management#solution-components)
* Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @georgewallace
* Microsoft Alias: **gwallace**
|
1.0
|
Unable to edit the variables after deploying the Solution - Hi,
I need to exclude the VM's for this solution and trying to edit External_ExcludeVMNames variable , im not getting any option to edit .
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 225c9d05-83dd-b006-0025-3753f5ab25bf
* Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096
* Content: [Start/Stop VMs during off-hours solution](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management#solution-components)
* Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @georgewallace
* Microsoft Alias: **gwallace**
|
process
|
unable to edit the variables after deploying the solution hi i need to exclude the vm s for this solution and trying to edit external excludevmnames variable im not getting any option to edit document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login georgewallace microsoft alias gwallace
| 1
|
773,025
| 27,143,097,000
|
IssuesEvent
|
2023-02-16 17:44:16
|
containrrr/watchtower
|
https://api.github.com/repos/containrrr/watchtower
|
opened
|
Non-label method for excluding containers
|
Type: Enhancement Priority: Low Status: Available
|
### Is your feature request related to a problem? Please describe.
On Synology (and I assume other environments - UniFi comes to mind), there are often "private" containers that we don't control the creation of. As a consequence, watchtower tries to update those but fails.
Since there's no way to update labels for those containers after they were created, and we don't create them, there's no good way to exclude them, short of
### Describe the solution you'd like
I'd be happy with a `--exclude` or similar flag (possibly taking a regex?) to tell watchtower what to exclude - that's easy to update.
### Describe alternatives you've considered
* Using `--label-enable` is harder to manage (requires managing labels for all the other containers)
* Creating derived containers just to add the label is not really an option for a lot of these private containers, especially as some of them have their own update mechanism.
* On Synology specifically, updating environment variables is easier (doesn't require recreating the container), so that's definitely a good option there
* There could simply be a flag to silently ignore any containers that aren't on the registry (would solve the Synology docviewer case, might not solve the UniFi case where they create common containers too)
* A file with the list of exclusions might also work
### Additional context
Related to https://github.com/containrrr/watchtower/issues/923 .
|
1.0
|
Non-label method for excluding containers - ### Is your feature request related to a problem? Please describe.
On Synology (and I assume other environments - UniFi comes to mind), there are often "private" containers that we don't control the creation of. As a consequence, watchtower tries to update those but fails.
Since there's no way to update labels for those containers after they were created, and we don't create them, there's no good way to exclude them, short of
### Describe the solution you'd like
I'd be happy with a `--exclude` or similar flag (possibly taking a regex?) to tell watchtower what to exclude - that's easy to update.
### Describe alternatives you've considered
* Using `--label-enable` is harder to manage (requires managing labels for all the other containers)
* Creating derived containers just to add the label is not really an option for a lot of these private containers, especially as some of them have their own update mechanism.
* On Synology specifically, updating environment variables is easier (doesn't require recreating the container), so that's definitely a good option there
* There could simply be a flag to silently ignore any containers that aren't on the registry (would solve the Synology docviewer case, might not solve the UniFi case where they create common containers too)
* A file with the list of exclusions might also work
### Additional context
Related to https://github.com/containrrr/watchtower/issues/923 .
|
non_process
|
non label method for excluding containers is your feature request related to a problem please describe on synology and i assume other environments unifi comes to mind there are often private containers that we don t control the creation of as a consequence watchtower tries to update those but fails since there s no way to update labels for those containers after they were created and we don t create them there s no good way to exclude them short of describe the solution you d like i d be happy with a exclude or similar flag possibly taking a regex to tell watchtower what to exclude that s easy to update describe alternatives you ve considered using label enable is harder to manage requires managing labels for all the other containers creating derived containers just to add the label is not really an option for a lot of these private containers especially as some of them have their own update mechanism on synology specifically updating environment variables is easier doesn t require recreating the container so that s definitely a good option there there could simply be a flag to silently ignore any containers that aren t on the registry would solve the synology docviewer case might not solve the unifi case where they create common containers too a file with the list of exclusions might also work additional context related to
| 0
|
78,861
| 3,518,417,793
|
IssuesEvent
|
2016-01-12 12:49:04
|
blue-systems/pm-img
|
https://api.github.com/repos/blue-systems/pm-img
|
closed
|
[XWayland]: Provide XWayland 1.18
|
high priority ready for QA
|
Please provide the latest XWayland server (1.18) in the image. It got released last week and brings important improvements for X11 applications like touch support.
|
1.0
|
[XWayland]: Provide XWayland 1.18 - Please provide the latest XWayland server (1.18) in the image. It got released last week and brings important improvements for X11 applications like touch support.
|
non_process
|
provide xwayland please provide the latest xwayland server in the image it got released last week and brings important improvements for applications like touch support
| 0
|
4,275
| 7,189,966,494
|
IssuesEvent
|
2018-02-02 15:44:40
|
parcel-bundler/parcel
|
https://api.github.com/repos/parcel-bundler/parcel
|
closed
|
🐛 Using type="text" on input in HTML gets stripped
|
#Question HTML Preprocessing
|
<!---
Thanks for filing an issue 😄 ! Before you submit, please read the following:
Search open/closed issues before submitting since someone might have asked the same thing before!
-->
**Choose one:** is this a 🐛 bug report or 🙋 feature request?
🐛
<!--- Provide a general summary of the issue in the title above -->
### 🎛 Configuration (.babelrc, package.json, cli command)
<!--- If describing a bug, tell us what your babel configuration looks like -->
No config. Just a HTML5 Boilerplate with Parcel.
### 🤔 Expected Behavior
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
When running ```parcel build index.html --public-url ./```, any HTML <input type="text"> should not be altered.
### 😯 Current Behavior
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If you are seeing an error, please include the full error message and stack trace -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
When running ```parcel build index.html --public-url ./```:
```<input type="text" name="FirstName" class="form-control" id="FirstName" placeholder="*Firstname" required>```
changes to
```<input name="FirstName" class="form-control" id="FirstName" placeholder="*Firstname" required="">```
### 💁 Possible Solution
Not sure.
### 🔦 Context
Need type="text" to stay, especially when using CSS [type="text"].
### 🌍 Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
| Software | Version(s) |
| ---------------- | ---------- |
| Parcel | 1.4.1
| Node | 8.5.0
| npm/Yarn | 5.3.0
| Operating System | macOS 10.13.2
<!-- Love parcel? Please consider supporting our collective:
👉 https://opencollective.com/parcel/donate -->
|
1.0
|
🐛 Using type="text" on input in HTML gets stripped - <!---
Thanks for filing an issue 😄 ! Before you submit, please read the following:
Search open/closed issues before submitting since someone might have asked the same thing before!
-->
**Choose one:** is this a 🐛 bug report or 🙋 feature request?
🐛
<!--- Provide a general summary of the issue in the title above -->
### 🎛 Configuration (.babelrc, package.json, cli command)
<!--- If describing a bug, tell us what your babel configuration looks like -->
No config. Just a HTML5 Boilerplate with Parcel.
### 🤔 Expected Behavior
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
When running ```parcel build index.html --public-url ./```, any HTML <input type="text"> should not be altered.
### 😯 Current Behavior
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If you are seeing an error, please include the full error message and stack trace -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
When running ```parcel build index.html --public-url ./```:
```<input type="text" name="FirstName" class="form-control" id="FirstName" placeholder="*Firstname" required>```
changes to
```<input name="FirstName" class="form-control" id="FirstName" placeholder="*Firstname" required="">```
### 💁 Possible Solution
Not sure.
### 🔦 Context
Need type="text" to stay, especially when using CSS [type="text"].
### 🌍 Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
| Software | Version(s) |
| ---------------- | ---------- |
| Parcel | 1.4.1
| Node | 8.5.0
| npm/Yarn | 5.3.0
| Operating System | macOS 10.13.2
<!-- Love parcel? Please consider supporting our collective:
👉 https://opencollective.com/parcel/donate -->
|
process
|
🐛 using type text on input in html gets stripped thanks for filing an issue 😄 before you submit please read the following search open closed issues before submitting since someone might have asked the same thing before choose one is this a 🐛 bug report or 🙋 feature request 🐛 🎛 configuration babelrc package json cli command no config just a boilerplate with parcel 🤔 expected behavior when running parcel build index html public url any html should not be altered 😯 current behavior when running parcel build index html public url changes to 💁 possible solution not sure 🔦 context need type text to stay especially when using css 🌍 your environment software version s parcel node npm yarn operating system macos love parcel please consider supporting our collective 👉
| 1
|
20,632
| 27,314,544,652
|
IssuesEvent
|
2023-02-24 14:45:35
|
ESMValGroup/ESMValCore
|
https://api.github.com/repos/ESMValGroup/ESMValCore
|
closed
|
Possibility to provide a '*' to the definition of a dataset in a recipe
|
enhancement preprocessor
|
Hello everyone,
this issue follows our discussion during the Is-ENES3 GA with @valeriupredoi
With CliMAF we found very interesting to be able to specify a wildcard '*' to our dataset definitions, like for instance model='*', realization='*' to work on all the models or realizations available.
In the same way, we implemented the possibility to specify period='last_XXY', 'first_XXY' or '*', with XX being a number of years, to retrieve the last XX, first XX years available, or the full period.
Do you guys think you could consider adding this functionality?
Cheers,
J.
|
1.0
|
Possibility to provide a '*' to the definition of a dataset in a recipe - Hello everyone,
this issue follows our discussion during the Is-ENES3 GA with @valeriupredoi
With CliMAF we found very interesting to be able to specify a wildcard '*' to our dataset definitions, like for instance model='*', realization='*' to work on all the models or realizations available.
In the same way, we implemented the possibility to specify period='last_XXY', 'first_XXY' or '*', with XX being a number of years, to retrieve the last XX, first XX years available, or the full period.
Do you guys think you could consider adding this functionality?
Cheers,
J.
|
process
|
possibility to provide a to the definition of a dataset in a recipe hello everyone this issue follows our discussion during the is ga with valeriupredoi with climaf we found very interesting to be able to specify a wildcard to our dataset definitions like for instance model realization to work on all the models or realizations available in the same way we implemented the possibility to specify period last xxy first xxy or with xx being a number of years to retrieve the last xx first xx years available or the full period do you guys think you could consider adding this functionality cheers j
| 1
|
22,025
| 30,541,186,737
|
IssuesEvent
|
2023-07-19 21:34:11
|
tradingstrategy-ai/frontend
|
https://api.github.com/repos/tradingstrategy-ai/frontend
|
closed
|
Aave: Add lending reserve pages to sitemap.xml
|
enhancement process size: XS
|
# Goal
As a Trading Strategy marketer, I want Aave lending reserve pages to be discoverable / indexable by search engines so consumers can find them via web search.
# Details
Add `/trading-view/lending-reserves` and all `/trading-view/polygon/lending/[protocol]/[reserve]` pages to `sitemap.xml`
|
1.0
|
Aave: Add lending reserve pages to sitemap.xml - # Goal
As a Trading Strategy marketer, I want Aave lending reserve pages to be discoverable / indexable by search engines so consumers can find them via web search.
# Details
Add `/trading-view/lending-reserves` and all `/trading-view/polygon/lending/[protocol]/[reserve]` pages to `sitemap.xml`
|
process
|
aave add lending reserve pages to sitemap xml goal as a trading strategy marketer i want aave lending reserve pages to be discoverable indexable by search engines so consumers can find them via web search details add trading view lending reserves and all trading view polygon lending pages to sitemap xml
| 1
|
15,058
| 18,763,401,040
|
IssuesEvent
|
2021-11-05 19:27:01
|
ORNL-AMO/AMO-Tools-Desktop
|
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
|
closed
|
Available Heat Algorithm
|
Process Heating
|
Will need to do something things on the Desktop side to account for Suite Issue #623
|
1.0
|
Available Heat Algorithm - Will need to do something things on the Desktop side to account for Suite Issue #623
|
process
|
available heat algorithm will need to do something things on the desktop side to account for suite issue
| 1
|
137,610
| 18,755,107,429
|
IssuesEvent
|
2021-11-05 09:43:49
|
Dima2022/node-kms
|
https://api.github.com/repos/Dima2022/node-kms
|
opened
|
CVE-2020-7656 (Medium) detected in jquery-1.7.1.min.js
|
security vulnerability
|
## CVE-2020-7656 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: node-kms/node_modules/vm-browserify/example/run/index.html</p>
<p>Path to vulnerable library: /node_modules/vm-browserify/example/run/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Dima2022/node-kms/commit/0174d52c9bbe1d27d5a8e0359134e9ffdbe6db6f">0174d52c9bbe1d27d5a8e0359134e9ffdbe6db6f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed.
<p>Publish Date: 2020-05-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-q4m3-2j7h-f7xw">https://github.com/advisories/GHSA-q4m3-2j7h-f7xw</a></p>
<p>Release Date: 2020-05-28</p>
<p>Fix Resolution: jquery - 1.9.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.7.1","packageFilePaths":["/node_modules/vm-browserify/example/run/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jquery - 1.9.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-7656","vulnerabilityDetails":"jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove \"\u003cscript\u003e\" HTML tags that contain a whitespace character, i.e: \"\u003c/script \u003e\", which results in the enclosed script logic to be executed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-7656 (Medium) detected in jquery-1.7.1.min.js - ## CVE-2020-7656 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: node-kms/node_modules/vm-browserify/example/run/index.html</p>
<p>Path to vulnerable library: /node_modules/vm-browserify/example/run/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Dima2022/node-kms/commit/0174d52c9bbe1d27d5a8e0359134e9ffdbe6db6f">0174d52c9bbe1d27d5a8e0359134e9ffdbe6db6f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed.
<p>Publish Date: 2020-05-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-q4m3-2j7h-f7xw">https://github.com/advisories/GHSA-q4m3-2j7h-f7xw</a></p>
<p>Release Date: 2020-05-28</p>
<p>Fix Resolution: jquery - 1.9.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.7.1","packageFilePaths":["/node_modules/vm-browserify/example/run/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jquery - 1.9.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-7656","vulnerabilityDetails":"jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove \"\u003cscript\u003e\" HTML tags that contain a whitespace character, i.e: \"\u003c/script \u003e\", which results in the enclosed script logic to be executed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file node kms node modules vm browserify example run index html path to vulnerable library node modules vm browserify example run index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery prior to allows cross site scripting attacks via the load method the load method fails to recognize and remove html tags that contain a whitespace character i e which results in the enclosed script logic to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion jquery basebranches vulnerabilityidentifier cve vulnerabilitydetails jquery prior to allows cross site scripting attacks via the load method the load method fails to recognize and remove html tags that contain a whitespace character i e script which results in the enclosed script logic to be executed vulnerabilityurl
| 0
|
12,414
| 14,920,392,045
|
IssuesEvent
|
2021-01-23 04:29:03
|
e4exp/paper_manager_abstract
|
https://api.github.com/repos/e4exp/paper_manager_abstract
|
opened
|
The Pile: An 800GB Dataset of Diverse Text for Language Modeling
|
2020 Dataset Natural Language Processing
|
* https://arxiv.org/abs/2101.00027
* 2020
最近の研究では、学習データセットの多様性を高めることで、大規模言語モデルの一般的なクロスドメイン知識と下流の一般化能力が向上することが実証されている。
このことを念頭に置いて,我々は大規模言語モデルの学習を目的とした825GiBの英文コーパスであるtextit{the Pile}を提示する.
このコーパスは,既存のものと新たに構築されたものの両方を含む22の多様で質の高いサブセットから構成されており,その多くは学術的・専門的な情報源に由来するものである.
GPT-2とGPT-3のPile上での未調整の性能を評価したところ、これらのモデルはアカデミックライティングなどの多くの構成要素で苦戦していることがわかりました。
逆に、Pile上で訓練されたモデルは、Pileのすべての構成要素においてRaw CCとCC-100を大幅に上回り、下流の評価でのパフォーマンスを向上させました。
徹底的な探索的分析により、将来のユーザーのためにデータの潜在的な側面を文書化しています。その構築に使用されたコードを公開しています。
|
1.0
|
The Pile: An 800GB Dataset of Diverse Text for Language Modeling - * https://arxiv.org/abs/2101.00027
* 2020
最近の研究では、学習データセットの多様性を高めることで、大規模言語モデルの一般的なクロスドメイン知識と下流の一般化能力が向上することが実証されている。
このことを念頭に置いて,我々は大規模言語モデルの学習を目的とした825GiBの英文コーパスであるtextit{the Pile}を提示する.
このコーパスは,既存のものと新たに構築されたものの両方を含む22の多様で質の高いサブセットから構成されており,その多くは学術的・専門的な情報源に由来するものである.
GPT-2とGPT-3のPile上での未調整の性能を評価したところ、これらのモデルはアカデミックライティングなどの多くの構成要素で苦戦していることがわかりました。
逆に、Pile上で訓練されたモデルは、Pileのすべての構成要素においてRaw CCとCC-100を大幅に上回り、下流の評価でのパフォーマンスを向上させました。
徹底的な探索的分析により、将来のユーザーのためにデータの潜在的な側面を文書化しています。その構築に使用されたコードを公開しています。
|
process
|
the pile an dataset of diverse text for language modeling 最近の研究では、学習データセットの多様性を高めることで、大規模言語モデルの一般的なクロスドメイン知識と下流の一般化能力が向上することが実証されている。 このことを念頭に置いて, the pile を提示する. このコーパスは, ,その多くは学術的・専門的な情報源に由来するものである. gpt 、これらのモデルはアカデミックライティングなどの多くの構成要素で苦戦していることがわかりました。 逆に、pile上で訓練されたモデルは、pileのすべての構成要素においてraw ccとcc 、下流の評価でのパフォーマンスを向上させました。 徹底的な探索的分析により、将来のユーザーのためにデータの潜在的な側面を文書化しています。その構築に使用されたコードを公開しています。
| 1
|
273,833
| 29,831,099,222
|
IssuesEvent
|
2023-06-18 09:31:46
|
RG4421/ampere-centos-kernel
|
https://api.github.com/repos/RG4421/ampere-centos-kernel
|
closed
|
CVE-2020-15437 (Medium) detected in linuxv5.2 - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2020-15437 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/serial/8250/8250_core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/serial/8250/8250_core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/serial/8250/8250_core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The Linux kernel before version 5.8 is vulnerable to a NULL pointer dereference in drivers/tty/serial/8250/8250_core.c:serial8250_isa_init_ports() that allows local users to cause a denial of service by using the p->serial_in pointer which uninitialized.
Mend Note: After conducting further research, Mend has determined that versions v3.9-rc5 through v4.4.231, v4.5-rc1 through v4.9.231, v4.10-rc1 through v4.14.189, v4.15-rc1 through v4.19.134, v5.0-rc1 through v5.4.53, v5.5-rc1 through v5.7.10 and v5.8-rc1 through v5.8-rc6 of Linux Kernel are vulnerable to CVE-2020-15437.
<p>Publish Date: 2020-11-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-15437>CVE-2020-15437</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-15437">https://www.linuxkernelcves.com/cves/CVE-2020-15437</a></p>
<p>Release Date: 2020-11-23</p>
<p>Fix Resolution: v4.4.232, v4.9.232, v4.14.190, v4.19.135, v5.4.54, v5.7.11, v5.8-rc7</p>
</p>
</details>
<p></p>
|
True
|
CVE-2020-15437 (Medium) detected in linuxv5.2 - autoclosed - ## CVE-2020-15437 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/serial/8250/8250_core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/serial/8250/8250_core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/serial/8250/8250_core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The Linux kernel before version 5.8 is vulnerable to a NULL pointer dereference in drivers/tty/serial/8250/8250_core.c:serial8250_isa_init_ports() that allows local users to cause a denial of service by using the p->serial_in pointer which uninitialized.
Mend Note: After conducting further research, Mend has determined that versions v3.9-rc5 through v4.4.231, v4.5-rc1 through v4.9.231, v4.10-rc1 through v4.14.189, v4.15-rc1 through v4.19.134, v5.0-rc1 through v5.4.53, v5.5-rc1 through v5.7.10 and v5.8-rc1 through v5.8-rc6 of Linux Kernel are vulnerable to CVE-2020-15437.
<p>Publish Date: 2020-11-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-15437>CVE-2020-15437</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-15437">https://www.linuxkernelcves.com/cves/CVE-2020-15437</a></p>
<p>Release Date: 2020-11-23</p>
<p>Fix Resolution: v4.4.232, v4.9.232, v4.14.190, v4.19.135, v5.4.54, v5.7.11, v5.8-rc7</p>
</p>
</details>
<p></p>
|
non_process
|
cve medium detected in autoclosed cve medium severity vulnerability vulnerable library linux kernel source tree library home page a href found in base branch amp centos kernel vulnerable source files drivers tty serial core c drivers tty serial core c drivers tty serial core c vulnerability details the linux kernel before version is vulnerable to a null pointer dereference in drivers tty serial core c isa init ports that allows local users to cause a denial of service by using the p serial in pointer which uninitialized mend note after conducting further research mend has determined that versions through through through through through through and through of linux kernel are vulnerable to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
| 0
|
316,378
| 23,628,475,513
|
IssuesEvent
|
2022-08-25 07:14:15
|
iotaledger/inx-chronicle
|
https://api.github.com/repos/iotaledger/inx-chronicle
|
closed
|
Document MongoDB `transactionLifetimeLimitSeconds` configuration
|
documentation
|
When we call `ReadUnspentOutputs` the resulting transaction can take a really long time, depending on the amount of outputs. In some cases, the default timeout from MongoDB kills the transaction before completion. To work around this, users should configure their MongoDB instance with a larger timeout using the `transactionLifetimeLimitSeconds` option.
We should document that behavior, provide an example on how to set the flag, and adapt our `docker-compose.yml` file accordingly.
|
1.0
|
Document MongoDB `transactionLifetimeLimitSeconds` configuration - When we call `ReadUnspentOutputs` the resulting transaction can take a really long time, depending on the amount of outputs. In some cases, the default timeout from MongoDB kills the transaction before completion. To work around this, users should configure their MongoDB instance with a larger timeout using the `transactionLifetimeLimitSeconds` option.
We should document that behavior, provide an example on how to set the flag, and adapt our `docker-compose.yml` file accordingly.
|
non_process
|
document mongodb transactionlifetimelimitseconds configuration when we call readunspentoutputs the resulting transaction can take a really long time depending on the amount of outputs in some cases the default timeout from mongodb kills the transaction before completion to work around this users should configure their mongodb instance with a larger timeout using the transactionlifetimelimitseconds option we should document that behavior provide an example on how to set the flag and adapt our docker compose yml file accordingly
| 0
|
20,157
| 26,710,314,262
|
IssuesEvent
|
2023-01-27 22:47:33
|
openxla/stablehlo
|
https://api.github.com/repos/openxla/stablehlo
|
opened
|
Consider standardizing test files
|
Process
|
Currently, several of our tests do not follow a consistent style in the way the tests are written.
Though not an exhaustive list some issues are:
- No standard test naming convention
- Inconsistent placement of CHECK labels
- Inconsistent string we are comparing CHECK-NEXT directives on
- Ordering of the tests
- Non-pretty printed ops
- Using default SSA values (e.g. %0, %1, etc.)
- etc.
Although #951 tackles this for future tests, this is not the case for already existing tests that is not part of this workflow. This ticket should track the progress of how consistent the tests are written.
|
1.0
|
Consider standardizing test files - Currently, several of our tests do not follow a consistent style in the way the tests are written.
Though not an exhaustive list some issues are:
- No standard test naming convention
- Inconsistent placement of CHECK labels
- Inconsistent string we are comparing CHECK-NEXT directives on
- Ordering of the tests
- Non-pretty printed ops
- Using default SSA values (e.g. %0, %1, etc.)
- etc.
Although #951 tackles this for future tests, this is not the case for already existing tests that is not part of this workflow. This ticket should track the progress of how consistent the tests are written.
|
process
|
consider standardizing test files currently several of our tests do not follow a consistent style in the way the tests are written though not an exhaustive list some issues are no standard test naming convention inconsistent placement of check labels inconsistent string we are comparing check next directives on ordering of the tests non pretty printed ops using default ssa values e g etc etc although tackles this for future tests this is not the case for already existing tests that is not part of this workflow this ticket should track the progress of how consistent the tests are written
| 1
|
7,144
| 10,288,573,369
|
IssuesEvent
|
2019-08-27 10:43:30
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
The Runbook will Add the AZ modules but cannot remove the old AzureRM
|
Pri2 automation/svc cxp process-automation/subsvc product-question triaged
|
When a new Automation account is created it comes by pre-installed with AzureRM modules.
This runbook will add the AZ modules no worries. But it does not remove the AzureRM. Neither you can delete the old AzureRm modules
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 55d92628-d657-759d-d301-06029cd61138
* Version Independent ID: ca381b85-2be5-24af-be18-00c13a7e53df
* Content: [Update Azure modules in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/automation-update-azure-modules#feedback)
* Content Source: [articles/automation/automation-update-azure-modules.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-update-azure-modules.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @bobbytreed
* Microsoft Alias: **robreed**
|
1.0
|
The Runbook will Add the AZ modules but cannot remove the old AzureRM - When a new Automation account is created it comes by pre-installed with AzureRM modules.
This runbook will add the AZ modules no worries. But it does not remove the AzureRM. Neither you can delete the old AzureRm modules
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 55d92628-d657-759d-d301-06029cd61138
* Version Independent ID: ca381b85-2be5-24af-be18-00c13a7e53df
* Content: [Update Azure modules in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/automation-update-azure-modules#feedback)
* Content Source: [articles/automation/automation-update-azure-modules.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-update-azure-modules.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @bobbytreed
* Microsoft Alias: **robreed**
|
process
|
the runbook will add the az modules but cannot remove the old azurerm when a new automation account is created it comes by pre installed with azurerm modules this runbook will add the az modules no worries but it does not remove the azurerm neither you can delete the old azurerm modules document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login bobbytreed microsoft alias robreed
| 1
|
205,149
| 15,964,633,018
|
IssuesEvent
|
2021-04-16 06:34:25
|
Icelenaugust/pe
|
https://api.github.com/repos/Icelenaugust/pe
|
opened
|
Mismatch of diagram in UG findp
|
severity.VeryLow type.DocumentationBug
|
in UG the displayed message is 2 persons listed!

while in actual product, the displayed message is 2 person(s) listed!
It might be a good habit to update UG!

<!--session: 1618551311074-1a994e8c-6641-4e86-94f8-3737348b1ccc-->
|
1.0
|
Mismatch of diagram in UG findp - in UG the displayed message is 2 persons listed!

while in actual product, the displayed message is 2 person(s) listed!
It might be a good habit to update UG!

<!--session: 1618551311074-1a994e8c-6641-4e86-94f8-3737348b1ccc-->
|
non_process
|
mismatch of diagram in ug findp in ug the displayed message is persons listed while in actual product the displayed message is person s listed it might be a good habit to update ug
| 0
|
20,669
| 27,335,048,724
|
IssuesEvent
|
2023-02-26 04:35:00
|
cse442-at-ub/project_s23-cinco
|
https://api.github.com/repos/cse442-at-ub/project_s23-cinco
|
closed
|
Create a sign up frame as well as interactivity to go to and from the new frame
|
Processing Task Sprint 1
|
Test 1:
go to figma page: https://www.figma.com/file/5qJUyXFUAdbtiobIQYqH20/Project-Prototype?node-id=151%3A3&t=QH31e7QFg894LkNj-0
click sign up in the top right corner
confirm that it takes you to a sign up page
confirm that clicking sign up brings you back to the login screen where you can log in with your new credentials
|
1.0
|
Create a sign up frame as well as interactivity to go to and from the new frame - Test 1:
go to figma page: https://www.figma.com/file/5qJUyXFUAdbtiobIQYqH20/Project-Prototype?node-id=151%3A3&t=QH31e7QFg894LkNj-0
click sign up in the top right corner
confirm that it takes you to a sign up page
confirm that clicking sign up brings you back to the login screen where you can log in with your new credentials
|
process
|
create a sign up frame as well as interactivity to go to and from the new frame test go to figma page click sign up in the top right corner confirm that it takes you to a sign up page confirm that clicking sign up brings you back to the login screen where you can log in with your new credentials
| 1
|
2,362
| 5,166,321,914
|
IssuesEvent
|
2017-01-17 15:57:46
|
inasafe/inasafe
|
https://api.github.com/repos/inasafe/inasafe
|
closed
|
Communicate the limitations of running with big datasets to users
|
Feature request Needs review Preprocessing User Interface
|
Either by docuemntation or the GUI, when users select detailed datasets with a big viewport we should set their expectations instead of them having to wonder wether or not it is going to finish successfully.
Perhaps warnings when the data is too big? Like OSM does. (Ask Liana for more info).
|
1.0
|
Communicate the limitations of running with big datasets to users - Either by docuemntation or the GUI, when users select detailed datasets with a big viewport we should set their expectations instead of them having to wonder wether or not it is going to finish successfully.
Perhaps warnings when the data is too big? Like OSM does. (Ask Liana for more info).
|
process
|
communicate the limitations of running with big datasets to users either by docuemntation or the gui when users select detailed datasets with a big viewport we should set their expectations instead of them having to wonder wether or not it is going to finish successfully perhaps warnings when the data is too big like osm does ask liana for more info
| 1
|
444,624
| 31,081,006,601
|
IssuesEvent
|
2023-08-13 04:13:11
|
aws/aws-sdk-ruby
|
https://api.github.com/repos/aws/aws-sdk-ruby
|
opened
|
Stub response_target
|
documentation needs-triage
|
### Describe the issue
I am trying to stub the `response_target` in the `get_object` function call.
I currently have this setup in my test suite
```
Aws.config[:s3] = {
stub_responses: {
list_buckets: {
buckets: [name: 'bucket_name'
},
list_objects_v2: {
contents: [{key: "some_key"}]
},
get_object: {
body: "lorem ipsum",
response_target: "trying_to_stub/this"
}
}
}
```
I get a `ArgumentError: unexpected value at params[:response_target]`
I don't see any documentation on this.
If I remove the `response_target` key, it will try to create the file(which i don't want)
### Links
https://docs.aws.amazon.com/sdk-for-ruby/v3/api/Aws/ClientStubs.html
|
1.0
|
Stub response_target - ### Describe the issue
I am trying to stub the `response_target` in the `get_object` function call.
I currently have this setup in my test suite
```
Aws.config[:s3] = {
stub_responses: {
list_buckets: {
buckets: [name: 'bucket_name'
},
list_objects_v2: {
contents: [{key: "some_key"}]
},
get_object: {
body: "lorem ipsum",
response_target: "trying_to_stub/this"
}
}
}
```
I get a `ArgumentError: unexpected value at params[:response_target]`
I don't see any documentation on this.
If I remove the `response_target` key, it will try to create the file(which i don't want)
### Links
https://docs.aws.amazon.com/sdk-for-ruby/v3/api/Aws/ClientStubs.html
|
non_process
|
stub response target describe the issue i am trying to stub the response target in the get object function call i currently have this setup in my test suite aws config stub responses list buckets buckets name bucket name list objects contents get object body lorem ipsum response target trying to stub this i get a argumenterror unexpected value at params i don t see any documentation on this if i remove the response target key it will try to create the file which i don t want links
| 0
|
540
| 3,001,553,233
|
IssuesEvent
|
2015-07-24 12:14:25
|
genomizer/genomizer-server
|
https://api.github.com/repos/genomizer/genomizer-server
|
closed
|
Smooth file truncated by IGB, should be investigated
|
bug High priority Processing
|
Check result in Integrated Genome Browser.
Smoothing should use a moving window and compute the mean or median of the window and assign to the rounded (any direction) center coordinate. This can be rewritten in Java with relative ease.
|
1.0
|
Smooth file truncated by IGB, should be investigated - Check result in Integrated Genome Browser.
Smoothing should use a moving window and compute the mean or median of the window and assign to the rounded (any direction) center coordinate. This can be rewritten in Java with relative ease.
|
process
|
smooth file truncated by igb should be investigated check result in integrated genome browser smoothing should use a moving window and compute the mean or median of the window and assign to the rounded any direction center coordinate this can be rewritten in java with relative ease
| 1
|
98,444
| 11,083,446,007
|
IssuesEvent
|
2019-12-13 14:29:37
|
esdc-devcop/esdc-devcop.github.io
|
https://api.github.com/repos/esdc-devcop/esdc-devcop.github.io
|
closed
|
Update Strategy
|
documentation
|
The strategy is changing to include streams and hasn't had the time to focus on management support. So the documentation on this should be updated.
|
1.0
|
Update Strategy - The strategy is changing to include streams and hasn't had the time to focus on management support. So the documentation on this should be updated.
|
non_process
|
update strategy the strategy is changing to include streams and hasn t had the time to focus on management support so the documentation on this should be updated
| 0
|
327,818
| 9,981,678,467
|
IssuesEvent
|
2019-07-10 08:06:02
|
Sinapse-Energia/IoT-Hub-Home
|
https://api.github.com/repos/Sinapse-Energia/IoT-Hub-Home
|
opened
|
[BUG] External Metering is not working with channel 2 & 3
|
Client Priority: high Size: 2 Status: new Type: bug
|
We connect a load to each 3 phases of the EASTRON but is only working the first one.
The phase 2 return negative values and the phase 3 return 0.
`L3_READ_EXT_METERING_R;855022;1;2;0.0;0.0;0.0;0.0;0.0;-21.81;-1;-186.1;0.0;780.2;-21.8;1562742958;`
`L3_READ_EXT_METERING_R;855022;1;3;0.0;0.0;0.0;0.0;0.0;0.00;1;0.0;0.0;0.0;0.0;1562742985;`
Also, when there is not load on the channel 2, we get also strange values.
The values displayed in the EASTRON screen are OK
|
1.0
|
[BUG] External Metering is not working with channel 2 & 3 - We connect a load to each 3 phases of the EASTRON but is only working the first one.
The phase 2 return negative values and the phase 3 return 0.
`L3_READ_EXT_METERING_R;855022;1;2;0.0;0.0;0.0;0.0;0.0;-21.81;-1;-186.1;0.0;780.2;-21.8;1562742958;`
`L3_READ_EXT_METERING_R;855022;1;3;0.0;0.0;0.0;0.0;0.0;0.00;1;0.0;0.0;0.0;0.0;1562742985;`
Also, when there is not load on the channel 2, we get also strange values.
The values displayed in the EASTRON screen are OK
|
non_process
|
external metering is not working with channel we connect a load to each phases of the eastron but is only working the first one the phase return negative values and the phase return read ext metering r read ext metering r also when there is not load on the channel we get also strange values the values displayed in the eastron screen are ok
| 0
|
20,720
| 27,419,639,087
|
IssuesEvent
|
2023-03-01 15:54:35
|
xataio/xata-py
|
https://api.github.com/repos/xataio/xata-py
|
closed
|
`throw_exception` option in BulkProcessor
|
bulk-processor
|
Currently, every exception in the bulk processor throws an exception and terminates the thread. Add an option `throw_exception` that allows a bool flag to trigger throw or not to throw. Default: `False`.
|
1.0
|
`throw_exception` option in BulkProcessor - Currently, every exception in the bulk processor throws an exception and terminates the thread. Add an option `throw_exception` that allows a bool flag to trigger throw or not to throw. Default: `False`.
|
process
|
throw exception option in bulkprocessor currently every exception in the bulk processor throws an exception and terminates the thread add an option throw exception that allows a bool flag to trigger throw or not to throw default false
| 1
|
83,289
| 10,324,745,366
|
IssuesEvent
|
2019-09-01 11:56:47
|
coredns/coredns
|
https://api.github.com/repos/coredns/coredns
|
closed
|
Need document on how to set up gRPC server
|
documentation plugin/grpc
|
Hi,
I'm trying to setup CoreDNS client and server to communicate via gRPC, but totally cannot figure out how to. I tried this config for server:
`.grpc://test.com:4343 {
forward . 8.8.8.8
}`
and this for client on another machine:
`.:4343 {
grpc . 192.168.1.19:4343 {
tls_servername test.com
}
}`
But looks it's not working and no any error printed. For gRPC server, I think we need http2 and https enabled? If so I'd like to use self-signed cert, but don't know where to specify the certs, and how to let client don't check the cert.
I could neither find related article on Google, so could you add some basic instruction on this? Thanks.
|
1.0
|
Need document on how to set up gRPC server - Hi,
I'm trying to setup CoreDNS client and server to communicate via gRPC, but totally cannot figure out how to. I tried this config for server:
`.grpc://test.com:4343 {
forward . 8.8.8.8
}`
and this for client on another machine:
`.:4343 {
grpc . 192.168.1.19:4343 {
tls_servername test.com
}
}`
But looks it's not working and no any error printed. For gRPC server, I think we need http2 and https enabled? If so I'd like to use self-signed cert, but don't know where to specify the certs, and how to let client don't check the cert.
I could neither find related article on Google, so could you add some basic instruction on this? Thanks.
|
non_process
|
need document on how to set up grpc server hi i m trying to setup coredns client and server to communicate via grpc but totally cannot figure out how to i tried this config for server grpc test com forward and this for client on another machine grpc tls servername test com but looks it s not working and no any error printed for grpc server i think we need and https enabled if so i d like to use self signed cert but don t know where to specify the certs and how to let client don t check the cert i could neither find related article on google so could you add some basic instruction on this thanks
| 0
|
417,531
| 12,167,037,057
|
IssuesEvent
|
2020-04-27 10:13:12
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.chase.com - site is not usable
|
browser-firefox engine-gecko priority-important
|
<!-- @browser: Firefox 75.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:75.0) Gecko/20100101 Firefox/75.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/52130 -->
**URL**: http://www.chase.com
**Browser / Version**: Firefox 75.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Browser unsupported
**Steps to Reproduce**:
site does not load images or stylesheets, only test including, "Please update your browser. We don't support this browser anymore. Using another one will help protect your accounts and provide a better experience. " despite claiming to support firefox.
Console:
Loading failed for the <script> with source “blob:https://www.chase.com/51046d69-62b6-4c5b-bb2e-e4f6407d4948”. www.chase.com:1:1
Loading failed for the <script> with source “https://www.chase.com/c/041820/etc/designs/chase-ux/clientlibs/jquery/js/jquery.min.js”. www.chase.com:32:1
Loading failed for the <script> with source “https://www.chase.com/c/041820/etc/designs/chase-ux/clientlibs/require/js/require.min.js”. www.chase.com:33:1
Loading failed for the <script> with source “https://www.chase.com/c/041820/apps/chase/clientlibs/foundation/scripts/Reporting.js”. www.chase.com:41:1
ReferenceError: requirejs is not defined
www.chase.com:49:15
Loading failed for the <script> with source “https://www.chase.com/c/041820/etc/designs/chase-ux/clientlibs/chase-ux/js/dist/home.min.js”. www.chase.com:64:1
ReferenceError: require is not defined
www.chase.com:66:7
ReferenceError: require is not defined
www.chase.com:76:11
ReferenceError: require is not defined
www.chase.com:166:5
ReferenceError: require is not defined
www.chase.com:546:7
ReferenceError: require is not defined
www.chase.com:627:31
Content Security Policy: Ignoring ‘x-frame-options’ because of ‘frame-ancestors’ directive.
ReferenceError: CHASE is not defined
<details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/4/2ec9b877-4e62-4abf-b916-764181138c3f.jpg'></details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.chase.com - site is not usable - <!-- @browser: Firefox 75.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:75.0) Gecko/20100101 Firefox/75.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/52130 -->
**URL**: http://www.chase.com
**Browser / Version**: Firefox 75.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Browser unsupported
**Steps to Reproduce**:
site does not load images or stylesheets, only test including, "Please update your browser. We don't support this browser anymore. Using another one will help protect your accounts and provide a better experience. " despite claiming to support firefox.
Console:
Loading failed for the <script> with source “blob:https://www.chase.com/51046d69-62b6-4c5b-bb2e-e4f6407d4948”. www.chase.com:1:1
Loading failed for the <script> with source “https://www.chase.com/c/041820/etc/designs/chase-ux/clientlibs/jquery/js/jquery.min.js”. www.chase.com:32:1
Loading failed for the <script> with source “https://www.chase.com/c/041820/etc/designs/chase-ux/clientlibs/require/js/require.min.js”. www.chase.com:33:1
Loading failed for the <script> with source “https://www.chase.com/c/041820/apps/chase/clientlibs/foundation/scripts/Reporting.js”. www.chase.com:41:1
ReferenceError: requirejs is not defined
www.chase.com:49:15
Loading failed for the <script> with source “https://www.chase.com/c/041820/etc/designs/chase-ux/clientlibs/chase-ux/js/dist/home.min.js”. www.chase.com:64:1
ReferenceError: require is not defined
www.chase.com:66:7
ReferenceError: require is not defined
www.chase.com:76:11
ReferenceError: require is not defined
www.chase.com:166:5
ReferenceError: require is not defined
www.chase.com:546:7
ReferenceError: require is not defined
www.chase.com:627:31
Content Security Policy: Ignoring ‘x-frame-options’ because of ‘frame-ancestors’ directive.
ReferenceError: CHASE is not defined
<details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/4/2ec9b877-4e62-4abf-b916-764181138c3f.jpg'></details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
site is not usable url browser version firefox operating system windows tested another browser yes chrome problem type site is not usable description browser unsupported steps to reproduce site does not load images or stylesheets only test including please update your browser we don t support this browser anymore using another one will help protect your accounts and provide a better experience despite claiming to support firefox console loading failed for the with source “blob loading failed for the with source “ loading failed for the with source “ loading failed for the with source “ referenceerror requirejs is not defined loading failed for the with source “ referenceerror require is not defined referenceerror require is not defined referenceerror require is not defined referenceerror require is not defined referenceerror require is not defined content security policy ignoring ‘x frame options’ because of ‘frame ancestors’ directive referenceerror chase is not defined view the screenshot img alt screenshot src browser configuration none from with ❤️
| 0
|
54,832
| 7,926,704,136
|
IssuesEvent
|
2018-07-06 03:55:02
|
pkulchenko/ZeroBraneStudio
|
https://api.github.com/repos/pkulchenko/ZeroBraneStudio
|
closed
|
Suggetion - Start debug then run directly until break point
|
Documentation
|
Currently, start debug will begin step at the first line.
I think this is not necessary, because user could put break point at first line if they need.
As the title said, maybe 'Start debug then run directly until break point' is more handy.
|
1.0
|
Suggetion - Start debug then run directly until break point - Currently, start debug will begin step at the first line.
I think this is not necessary, because user could put break point at first line if they need.
As the title said, maybe 'Start debug then run directly until break point' is more handy.
|
non_process
|
suggetion start debug then run directly until break point currently start debug will begin step at the first line i think this is not necessary because user could put break point at first line if they need as the title said maybe start debug then run directly until break point is more handy
| 0
|
288,504
| 31,861,420,698
|
IssuesEvent
|
2023-09-15 11:12:31
|
nidhi7598/linux-v4.19.72_CVE-2022-3564
|
https://api.github.com/repos/nidhi7598/linux-v4.19.72_CVE-2022-3564
|
opened
|
CVE-2021-3347 (High) detected in linuxlinux-4.19.294
|
Mend: dependency security vulnerability
|
## CVE-2021-3347 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-v4.19.72_CVE-2022-3564/commit/9ffee08efa44c7887e2babb8f304df0fa1094efb">9ffee08efa44c7887e2babb8f304df0fa1094efb</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/futex.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel through 5.10.11. PI futexes have a kernel stack use-after-free during fault handling, allowing local users to execute code in the kernel, aka CID-34b1a1ce1458.
<p>Publish Date: 2021-01-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3347>CVE-2021-3347</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-3347">https://www.linuxkernelcves.com/cves/CVE-2021-3347</a></p>
<p>Release Date: 2021-01-29</p>
<p>Fix Resolution: v4.14.218,v4.19.172,v5.10.12,v5.4.94</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-3347 (High) detected in linuxlinux-4.19.294 - ## CVE-2021-3347 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-v4.19.72_CVE-2022-3564/commit/9ffee08efa44c7887e2babb8f304df0fa1094efb">9ffee08efa44c7887e2babb8f304df0fa1094efb</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/futex.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel through 5.10.11. PI futexes have a kernel stack use-after-free during fault handling, allowing local users to execute code in the kernel, aka CID-34b1a1ce1458.
<p>Publish Date: 2021-01-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3347>CVE-2021-3347</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-3347">https://www.linuxkernelcves.com/cves/CVE-2021-3347</a></p>
<p>Release Date: 2021-01-29</p>
<p>Fix Resolution: v4.14.218,v4.19.172,v5.10.12,v5.4.94</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files kernel futex c vulnerability details an issue was discovered in the linux kernel through pi futexes have a kernel stack use after free during fault handling allowing local users to execute code in the kernel aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
16,334
| 20,990,704,947
|
IssuesEvent
|
2022-03-29 09:02:26
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Move GO:1990109 rejection of pollen from other species
|
multi-species process
|
Hi
GO:1990109 rejection of pollen from other species is a child of 'GO:0044419 biological process involved in interspecies interaction between organisms'
In the multiorg call we concluded it should be under 'GO:0048544 recognition of pollen'
@tberardini is that OK with you?
Thanks, Pascale
|
1.0
|
Move GO:1990109 rejection of pollen from other species - Hi
GO:1990109 rejection of pollen from other species is a child of 'GO:0044419 biological process involved in interspecies interaction between organisms'
In the multiorg call we concluded it should be under 'GO:0048544 recognition of pollen'
@tberardini is that OK with you?
Thanks, Pascale
|
process
|
move go rejection of pollen from other species hi go rejection of pollen from other species is a child of go biological process involved in interspecies interaction between organisms in the multiorg call we concluded it should be under go recognition of pollen tberardini is that ok with you thanks pascale
| 1
|
418,075
| 12,192,906,405
|
IssuesEvent
|
2020-04-29 13:39:40
|
semanticarts/gist
|
https://api.github.com/repos/semanticarts/gist
|
closed
|
Clean up documentation and tools
|
effort: small impact: patch priority: should have status: triaged
|
- bundle.bat can be deleted - now replaced by ontology-toolkit. Instructions in README.md should be updated to reflect the use of ontology-toolkit.
- versionize scripts can be deleted
- serialize scripts can be deleted
- version.txt - delete
- curl_requests.txt - do we need this?
- Only rdf-toolkit is needed in tools directory
|
1.0
|
Clean up documentation and tools - - bundle.bat can be deleted - now replaced by ontology-toolkit. Instructions in README.md should be updated to reflect the use of ontology-toolkit.
- versionize scripts can be deleted
- serialize scripts can be deleted
- version.txt - delete
- curl_requests.txt - do we need this?
- Only rdf-toolkit is needed in tools directory
|
non_process
|
clean up documentation and tools bundle bat can be deleted now replaced by ontology toolkit instructions in readme md should be updated to reflect the use of ontology toolkit versionize scripts can be deleted serialize scripts can be deleted version txt delete curl requests txt do we need this only rdf toolkit is needed in tools directory
| 0
|
168,904
| 13,106,875,378
|
IssuesEvent
|
2020-08-04 14:28:22
|
foodcoopshop/foodcoopshop
|
https://api.github.com/repos/foodcoopshop/foodcoopshop
|
closed
|
Implement integration tests with IntegrationTestTrait
|
cleanup 🧹 tests
|
Since v3.7 CakePHP provides IntegrationTestTrait which is really great!
This refactoring
* should make the tests quicker
* reduces code - as methods (eg. assert methods) from the cake core are used
**Todos**
**done**
* [x] netzwerk modul testen
* [x] remove html output in SendShells (unit tests break on travis)
* [x] rechnngs-download geht nicht: https://demo-de.foodcoopshop.com/admin/action-logs?types=&types%5B%5D=cronjob_send_invoices&customerId=&productId=&dateFrom=06.05.2020&dateTo=03.08.2020
* [x] open order lists inline in browser and do not force download
* [x] PaymentsController: file upload not yet working
* [x] Changelog: speed up travis integration testing by 50%!
* [x] ListsController: link with path in order_lists.php (wrong), tests not yet working
* [x] as the new LoginTrait just fakes a login (to be quick), the "real" login needs to get an own test
* [x] Wrap up 'headers' => [ 'X_REQUESTED_WITH' => 'XMLHttpRequest', 'ACCEPT' => application/json' ],
* [x] do not show stack trace: ProductsControllerTest.php:257
* [x] InstantOrderTests are not working (no AppSession in AppAuth)
* [x] AppCakeTestCase should contain much less assertMethods. Use the built-in assert-methods in IntegrationTestCase.
* [x] remove usage of httpClient in all tests
* [x] HttpClient removed: BlogPostControllerTest (#529), ManufacturersFrontendController, ProductsFrontendController (#531)
* [x] CartControllerTest::testAddLoggedOut fails due to missing exit on autorization
* [x] CartController authorize does not work any more
* [x] ProductsControllerTest::testDeleteProduct fails
* [x] Refactor CartsControllerTest
* [x] Refactor OrderDetailsControllerAddFeedbackTest
* [x] Refactor OrderDetailsControllerCancellationTest.php
* [x] Refactor OrderDetailsControllerEditAmountTest.php
* [x] Refactor OrderDetailsControllerEdittCustomerTest.php
* [x] Refactor OrderDetailsControllerEditPickupDayTest
* [x] Refactor OrderDetailsControllerEditPriceTest
* [x] Refactor OrderDetailsControllerEditQuantityTest
* [x] addProductToCart() and finishCart without httpClient
* [x] enable testCancellationWithTimebasedCurrency
* [x] if (!$_SERVER['REQUEST_METHOD'] == 'POST') { is not set with IntegrationTest
* [x] testNewPasswordRequestWithValidEmail test login with new password
|
1.0
|
Implement integration tests with IntegrationTestTrait - Since v3.7 CakePHP provides IntegrationTestTrait which is really great!
This refactoring
* should make the tests quicker
* reduces code - as methods (eg. assert methods) from the cake core are used
**Todos**
**done**
* [x] netzwerk modul testen
* [x] remove html output in SendShells (unit tests break on travis)
* [x] rechnngs-download geht nicht: https://demo-de.foodcoopshop.com/admin/action-logs?types=&types%5B%5D=cronjob_send_invoices&customerId=&productId=&dateFrom=06.05.2020&dateTo=03.08.2020
* [x] open order lists inline in browser and do not force download
* [x] PaymentsController: file upload not yet working
* [x] Changelog: speed up travis integration testing by 50%!
* [x] ListsController: link with path in order_lists.php (wrong), tests not yet working
* [x] as the new LoginTrait just fakes a login (to be quick), the "real" login needs to get an own test
* [x] Wrap up 'headers' => [ 'X_REQUESTED_WITH' => 'XMLHttpRequest', 'ACCEPT' => application/json' ],
* [x] do not show stack trace: ProductsControllerTest.php:257
* [x] InstantOrderTests are not working (no AppSession in AppAuth)
* [x] AppCakeTestCase should contain much less assertMethods. Use the built-in assert-methods in IntegrationTestCase.
* [x] remove usage of httpClient in all tests
* [x] HttpClient removed: BlogPostControllerTest (#529), ManufacturersFrontendController, ProductsFrontendController (#531)
* [x] CartControllerTest::testAddLoggedOut fails due to missing exit on autorization
* [x] CartController authorize does not work any more
* [x] ProductsControllerTest::testDeleteProduct fails
* [x] Refactor CartsControllerTest
* [x] Refactor OrderDetailsControllerAddFeedbackTest
* [x] Refactor OrderDetailsControllerCancellationTest.php
* [x] Refactor OrderDetailsControllerEditAmountTest.php
* [x] Refactor OrderDetailsControllerEdittCustomerTest.php
* [x] Refactor OrderDetailsControllerEditPickupDayTest
* [x] Refactor OrderDetailsControllerEditPriceTest
* [x] Refactor OrderDetailsControllerEditQuantityTest
* [x] addProductToCart() and finishCart without httpClient
* [x] enable testCancellationWithTimebasedCurrency
* [x] if (!$_SERVER['REQUEST_METHOD'] == 'POST') { is not set with IntegrationTest
* [x] testNewPasswordRequestWithValidEmail test login with new password
|
non_process
|
implement integration tests with integrationtesttrait since cakephp provides integrationtesttrait which is really great this refactoring should make the tests quicker reduces code as methods eg assert methods from the cake core are used todos done netzwerk modul testen remove html output in sendshells unit tests break on travis rechnngs download geht nicht open order lists inline in browser and do not force download paymentscontroller file upload not yet working changelog speed up travis integration testing by listscontroller link with path in order lists php wrong tests not yet working as the new logintrait just fakes a login to be quick the real login needs to get an own test wrap up headers do not show stack trace productscontrollertest php instantordertests are not working no appsession in appauth appcaketestcase should contain much less assertmethods use the built in assert methods in integrationtestcase remove usage of httpclient in all tests httpclient removed blogpostcontrollertest manufacturersfrontendcontroller productsfrontendcontroller cartcontrollertest testaddloggedout fails due to missing exit on autorization cartcontroller authorize does not work any more productscontrollertest testdeleteproduct fails refactor cartscontrollertest refactor orderdetailscontrolleraddfeedbacktest refactor orderdetailscontrollercancellationtest php refactor orderdetailscontrollereditamounttest php refactor orderdetailscontrolleredittcustomertest php refactor orderdetailscontrollereditpickupdaytest refactor orderdetailscontrollereditpricetest refactor orderdetailscontrollereditquantitytest addproducttocart and finishcart without httpclient enable testcancellationwithtimebasedcurrency if server post is not set with integrationtest testnewpasswordrequestwithvalidemail test login with new password
| 0
|
223,600
| 17,117,010,832
|
IssuesEvent
|
2021-07-11 15:10:03
|
WebDevsOrg/calculator
|
https://api.github.com/repos/WebDevsOrg/calculator
|
opened
|
Capture standard mode requirements
|
documentation
|
This should have been done earlier, but we were following MS calculator and try to replicate some part of it. But before we start user interface its good to capture these requirements and pick good features, UI layout etc. from other calculator projects too.
Its also important to understand the use case
1. Is it targeted for mobile first customer
2. Is it targeted for desktop users
3. Or both
Its also important to consider who is going to use this calculator, so we should think about accessibility issues, like themes, colors, button sizes, labels, font size, contrast

We should think mobile first design but initially we may target for desktop and later on mobile as it may require different setup.
|
1.0
|
Capture standard mode requirements - This should have been done earlier, but we were following MS calculator and try to replicate some part of it. But before we start user interface its good to capture these requirements and pick good features, UI layout etc. from other calculator projects too.
Its also important to understand the use case
1. Is it targeted for mobile first customer
2. Is it targeted for desktop users
3. Or both
Its also important to consider who is going to use this calculator, so we should think about accessibility issues, like themes, colors, button sizes, labels, font size, contrast

We should think mobile first design but initially we may target for desktop and later on mobile as it may require different setup.
|
non_process
|
capture standard mode requirements this should have been done earlier but we were following ms calculator and try to replicate some part of it but before we start user interface its good to capture these requirements and pick good features ui layout etc from other calculator projects too its also important to understand the use case is it targeted for mobile first customer is it targeted for desktop users or both its also important to consider who is going to use this calculator so we should think about accessibility issues like themes colors button sizes labels font size contrast we should think mobile first design but initially we may target for desktop and later on mobile as it may require different setup
| 0
|
18,920
| 24,867,914,700
|
IssuesEvent
|
2022-10-27 13:20:02
|
OpenDataScotland/the_od_bods
|
https://api.github.com/repos/OpenDataScotland/the_od_bods
|
opened
|
Scottish OpenStreetMap repository
|
data processing back end new source
|
**Is your feature request related to a problem? Please describe.**
OpenStreetMap holds a lot of great GIS data for Scotland that we could scrape and store in a GitHub repository to then be surfaced on OpenData.Scot
**Describe the solution you'd like**
A script for making queries to the OSM Overpass API to then save in various formats (CSV, JSON, GEOJSON) in a GitHub repo. The script could run on a variable scheduled basis depending on how often we reckon the data would be updated (e.g. weekly, monthly, annually)
**Describe alternatives you've considered**
We could just link directly to overpass queries but that's fairly basic, can have slow API response times and requires more work for the user vs. serving up ready-made and curated datasets
**Additional context**
N/A
|
1.0
|
Scottish OpenStreetMap repository - **Is your feature request related to a problem? Please describe.**
OpenStreetMap holds a lot of great GIS data for Scotland that we could scrape and store in a GitHub repository to then be surfaced on OpenData.Scot
**Describe the solution you'd like**
A script for making queries to the OSM Overpass API to then save in various formats (CSV, JSON, GEOJSON) in a GitHub repo. The script could run on a variable scheduled basis depending on how often we reckon the data would be updated (e.g. weekly, monthly, annually)
**Describe alternatives you've considered**
We could just link directly to overpass queries but that's fairly basic, can have slow API response times and requires more work for the user vs. serving up ready-made and curated datasets
**Additional context**
N/A
|
process
|
scottish openstreetmap repository is your feature request related to a problem please describe openstreetmap holds a lot of great gis data for scotland that we could scrape and store in a github repository to then be surfaced on opendata scot describe the solution you d like a script for making queries to the osm overpass api to then save in various formats csv json geojson in a github repo the script could run on a variable scheduled basis depending on how often we reckon the data would be updated e g weekly monthly annually describe alternatives you ve considered we could just link directly to overpass queries but that s fairly basic can have slow api response times and requires more work for the user vs serving up ready made and curated datasets additional context n a
| 1
|
334,004
| 24,399,523,080
|
IssuesEvent
|
2022-10-04 23:08:54
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
closed
|
better description of solveh_banded limitations
|
scipy.linalg Documentation
|
Related to the discussion in the original pull request for linalg.solveh_banded, #4373, the solver requires additional constraints on the matrix compared to linalg.solve and linalg.solve_banded.
The routine solveh_banded requires leading minors of the matrix A in the equation A * x = b be positive definite. I do not understand this requirement at all, so I can not begin to delve into the linear algebra discussion. This requirement is very confusing as it stands, however, because it is not raised in very similar solvers that it is meant to supersede.
In an ideal world, I would like to help relax this requirement to be more in line with the other solvers. I do not know how to go about this as it stands. At a minimum, could we work on a more informative error message that mentions the other solvers may still work in situations where solveh_banded fails?
Here is an example of an error that is raised in an application finding eigenvectors from eigenvalues by inverse iteration
```
import numpy as np
from scipy.linalg import solve, solveh_banded, solve_banded
numz = 105
d = np.full(numz, -1.9964864348874727)
C = np.diag(d) + np.diag(np.ones(numz - 1), k=1) + np.diag(np.ones(numz - 1), k=-1)
# first iteration to find eigen-vector
phi_test = np.ones(numz)
phi1 = solve(C, phi_test)
C_b = np.ones((3, numz))
C_b[1, :] = d
phi2 = solve_banded((1, 1), C_b, phi_test)
C_h = np.ones((2, numz))
C_h[1, :] = d
phi3 = solveh_banded(C_h, phi_test)
```
|
1.0
|
better description of solveh_banded limitations - Related to the discussion in the original pull request for linalg.solveh_banded, #4373, the solver requires additional constraints on the matrix compared to linalg.solve and linalg.solve_banded.
The routine solveh_banded requires leading minors of the matrix A in the equation A * x = b be positive definite. I do not understand this requirement at all, so I can not begin to delve into the linear algebra discussion. This requirement is very confusing as it stands, however, because it is not raised in very similar solvers that it is meant to supersede.
In an ideal world, I would like to help relax this requirement to be more in line with the other solvers. I do not know how to go about this as it stands. At a minimum, could we work on a more informative error message that mentions the other solvers may still work in situations where solveh_banded fails?
Here is an example of an error that is raised in an application finding eigenvectors from eigenvalues by inverse iteration
```
import numpy as np
from scipy.linalg import solve, solveh_banded, solve_banded
numz = 105
d = np.full(numz, -1.9964864348874727)
C = np.diag(d) + np.diag(np.ones(numz - 1), k=1) + np.diag(np.ones(numz - 1), k=-1)
# first iteration to find eigen-vector
phi_test = np.ones(numz)
phi1 = solve(C, phi_test)
C_b = np.ones((3, numz))
C_b[1, :] = d
phi2 = solve_banded((1, 1), C_b, phi_test)
C_h = np.ones((2, numz))
C_h[1, :] = d
phi3 = solveh_banded(C_h, phi_test)
```
|
non_process
|
better description of solveh banded limitations related to the discussion in the original pull request for linalg solveh banded the solver requires additional constraints on the matrix compared to linalg solve and linalg solve banded the routine solveh banded requires leading minors of the matrix a in the equation a x b be positive definite i do not understand this requirement at all so i can not begin to delve into the linear algebra discussion this requirement is very confusing as it stands however because it is not raised in very similar solvers that it is meant to supersede in an ideal world i would like to help relax this requirement to be more in line with the other solvers i do not know how to go about this as it stands at a minimum could we work on a more informative error message that mentions the other solvers may still work in situations where solveh banded fails here is an example of an error that is raised in an application finding eigenvectors from eigenvalues by inverse iteration import numpy as np from scipy linalg import solve solveh banded solve banded numz d np full numz c np diag d np diag np ones numz k np diag np ones numz k first iteration to find eigen vector phi test np ones numz solve c phi test c b np ones numz c b d solve banded c b phi test c h np ones numz c h d solveh banded c h phi test
| 0
|
61,954
| 7,533,917,929
|
IssuesEvent
|
2018-04-16 05:42:59
|
ParabolInc/action
|
https://api.github.com/repos/ParabolInc/action
|
closed
|
Design: duplicate/clone cards
|
design enhancement user request
|
## Issue - Enhancement
User `google-oauth2|103150684875644068831` requested the following:
> [I] need to duplicate cards
This is an alternative to [#1071 Assign to All](https://github.com/ParabolInc/action/issues/1071), and probably easier to implement. We'd need to design where we'd put this control.
### Acceptance Criteria (optional)
A design in which users can:
- Quickly access a control that allows them to clone a card
- When the control is pressed, it duplicates the card in its current context...
- ...cloning the card within its column if pressed on the dashboard or _Project Updates_
- ...cloning the card when processing an agenda item
- **Estimated effort:** 13 points ([see CONTRIBUTING.md](https://github.com/ParabolInc/action/blob/master/CONTRIBUTING.md#points-and-sizes))
|
1.0
|
Design: duplicate/clone cards - ## Issue - Enhancement
User `google-oauth2|103150684875644068831` requested the following:
> [I] need to duplicate cards
This is an alternative to [#1071 Assign to All](https://github.com/ParabolInc/action/issues/1071), and probably easier to implement. We'd need to design where we'd put this control.
### Acceptance Criteria (optional)
A design in which users can:
- Quickly access a control that allows them to clone a card
- When the control is pressed, it duplicates the card in its current context...
- ...cloning the card within its column if pressed on the dashboard or _Project Updates_
- ...cloning the card when processing an agenda item
- **Estimated effort:** 13 points ([see CONTRIBUTING.md](https://github.com/ParabolInc/action/blob/master/CONTRIBUTING.md#points-and-sizes))
|
non_process
|
design duplicate clone cards issue enhancement user google requested the following need to duplicate cards this is an alternative to and probably easier to implement we d need to design where we d put this control acceptance criteria optional a design in which users can quickly access a control that allows them to clone a card when the control is pressed it duplicates the card in its current context cloning the card within its column if pressed on the dashboard or project updates cloning the card when processing an agenda item estimated effort points
| 0
|
10,271
| 13,125,342,255
|
IssuesEvent
|
2020-08-06 06:27:35
|
linuxdeepin/developer-center
|
https://api.github.com/repos/linuxdeepin/developer-center
|
closed
|
Change keyboard layout shortcut doesn't work on lock screen
|
Solved | resolved bug | functional behavior other | delay processing place | WH
|
I'm running Deepin Desktop 15.11. In the lock screen, changing keyboard layout using shortcuts `Alt + Shift` or `Win + Space` to enter password doesn't work.
|
1.0
|
Change keyboard layout shortcut doesn't work on lock screen - I'm running Deepin Desktop 15.11. In the lock screen, changing keyboard layout using shortcuts `Alt + Shift` or `Win + Space` to enter password doesn't work.
|
process
|
change keyboard layout shortcut doesn t work on lock screen i m running deepin desktop in the lock screen changing keyboard layout using shortcuts alt shift or win space to enter password doesn t work
| 1
|
15,279
| 19,269,704,153
|
IssuesEvent
|
2021-12-10 02:41:14
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Merge Vector Layer missing in Saga
|
Feedback stale Processing Bug
|
### What is the bug or the crash?
The merge vector layer tool is not available in the SAGA processing toolbox. I am running ubuntu 20.04 and the latest version of qgis.
### Steps to reproduce the issue
Go to QGIS Processing Toolbox
Open the Saga drop down
select vector general
merge vector layer missing
does not show up in search either.
### Versions
<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd">
<html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /><style type="text/css">
p, li { white-space: pre-wrap; }
</style></head><body>
QGIS version | 3.20.3-Odense | QGIS code revision | 495fbaecaf
-- | -- | -- | --
Qt version | 5.12.8
Python version | 3.8.10
GDAL/OGR version | 3.0.4
PROJ version | 6.3.1
EPSG Registry database version | v9.8.6 (2020-01-22)
Compiled against GEOS | 3.8.0-CAPI-1.13.1 | Running against GEOS | 3.8.0-CAPI-1.13.1
SQLite version | 3.31.1
PDAL version | 2.0.1
PostgreSQL client version | 12.8 (Ubuntu 12.8-0ubuntu0.20.04.1)
SpatiaLite version | 4.3.0a
QWT version | 6.1.4
QScintilla2 version | 2.11.2
OS version | Ubuntu 20.04.3 LTS
| | |
Active Python plugins | quick_map_servicescartography_toolsQuickOSMAutomaticBackup-masterMetaSearchdb_managerprocessing
</body></html>
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
_No response_
|
1.0
|
Merge Vector Layer missing in Saga - ### What is the bug or the crash?
The merge vector layer tool is not available in the SAGA processing toolbox. I am running ubuntu 20.04 and the latest version of qgis.
### Steps to reproduce the issue
Go to QGIS Processing Toolbox
Open the Saga drop down
select vector general
merge vector layer missing
does not show up in search either.
### Versions
<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd">
<html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /><style type="text/css">
p, li { white-space: pre-wrap; }
</style></head><body>
QGIS version | 3.20.3-Odense | QGIS code revision | 495fbaecaf
-- | -- | -- | --
Qt version | 5.12.8
Python version | 3.8.10
GDAL/OGR version | 3.0.4
PROJ version | 6.3.1
EPSG Registry database version | v9.8.6 (2020-01-22)
Compiled against GEOS | 3.8.0-CAPI-1.13.1 | Running against GEOS | 3.8.0-CAPI-1.13.1
SQLite version | 3.31.1
PDAL version | 2.0.1
PostgreSQL client version | 12.8 (Ubuntu 12.8-0ubuntu0.20.04.1)
SpatiaLite version | 4.3.0a
QWT version | 6.1.4
QScintilla2 version | 2.11.2
OS version | Ubuntu 20.04.3 LTS
| | |
Active Python plugins | quick_map_servicescartography_toolsQuickOSMAutomaticBackup-masterMetaSearchdb_managerprocessing
</body></html>
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
_No response_
|
process
|
merge vector layer missing in saga what is the bug or the crash the merge vector layer tool is not available in the saga processing toolbox i am running ubuntu and the latest version of qgis steps to reproduce the issue go to qgis processing toolbox open the saga drop down select vector general merge vector layer missing does not show up in search either versions doctype html public dtd html en p li white space pre wrap qgis version odense qgis code revision qt version python version gdal ogr version proj version epsg registry database version compiled against geos capi running against geos capi sqlite version pdal version postgresql client version ubuntu spatialite version qwt version version os version ubuntu lts active python plugins quick map servicescartography toolsquickosmautomaticbackup mastermetasearchdb managerprocessing supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
| 1
|
9,126
| 8,528,049,705
|
IssuesEvent
|
2018-11-02 21:46:07
|
goharbor/harbor
|
https://api.github.com/repos/goharbor/harbor
|
closed
|
Dead loop redis event after fired demo job
|
area/job-services kind/bug
|
**Expected behavior and actual behavior:**
expect job service could receive cancel job event, and do CancelJob at once.
**Steps to reproduce the problem:**
```
POST /api/v1/jobs/a5fe27325fbfb9a5df7f9e2a
{
"action": "stop"
}
```
**Versions:**
- job service: [v.1.6.1]
**Additional context:**
I'm running the raw job service, and want to use it as back service just sth like celery, and I found that there is dead loop events while do canceljob or stopjob.
steps are:
1. we fired job operations, and redis pool received at [here](https://github.com/goharbor/harbor/blob/v1.6.1/src/jobservice/pool/redis_pool.go#L135);
2. gcwp do [`SendCommand`](https://github.com/goharbor/harbor/blob/v1.6.1/src/jobservice/pool/redis_pool.go#L606);
3. at redis stats mgr, we `fired` msg at [here](https://github.com/goharbor/harbor/blob/v1.6.1/src/jobservice/opm/redis_job_stats_mgr.go#L221), and it [published to redis channel](https://github.com/goharbor/harbor/blob/v1.6.1/src/jobservice/opm/op_commands.go#L90)
4. and redis pool received msg again...
so why we need to fire this again and again? or there is something which i misunderstand?
Thanks.
|
1.0
|
Dead loop redis event after fired demo job - **Expected behavior and actual behavior:**
expect job service could receive cancel job event, and do CancelJob at once.
**Steps to reproduce the problem:**
```
POST /api/v1/jobs/a5fe27325fbfb9a5df7f9e2a
{
"action": "stop"
}
```
**Versions:**
- job service: [v.1.6.1]
**Additional context:**
I'm running the raw job service, and want to use it as back service just sth like celery, and I found that there is dead loop events while do canceljob or stopjob.
steps are:
1. we fired job operations, and redis pool received at [here](https://github.com/goharbor/harbor/blob/v1.6.1/src/jobservice/pool/redis_pool.go#L135);
2. gcwp do [`SendCommand`](https://github.com/goharbor/harbor/blob/v1.6.1/src/jobservice/pool/redis_pool.go#L606);
3. at redis stats mgr, we `fired` msg at [here](https://github.com/goharbor/harbor/blob/v1.6.1/src/jobservice/opm/redis_job_stats_mgr.go#L221), and it [published to redis channel](https://github.com/goharbor/harbor/blob/v1.6.1/src/jobservice/opm/op_commands.go#L90)
4. and redis pool received msg again...
so why we need to fire this again and again? or there is something which i misunderstand?
Thanks.
|
non_process
|
dead loop redis event after fired demo job expected behavior and actual behavior expect job service could receive cancel job event and do canceljob at once steps to reproduce the problem post api jobs action stop versions job service additional context i m running the raw job service and want to use it as back service just sth like celery and i found that there is dead loop events while do canceljob or stopjob steps are we fired job operations and redis pool received at gcwp do at redis stats mgr we fired msg at and it and redis pool received msg again so why we need to fire this again and again or there is something which i misunderstand thanks
| 0
|
596
| 3,071,598,424
|
IssuesEvent
|
2015-08-19 13:06:22
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
Cannot publish remote HTTP DITA Map to XHTML [DOT 2.x develop branch]
|
bug P2 preprocess
|
I'm publishing a DITA Map from a webdav HTTP server without authentication.
I have images referenced in my topics, at some point the publishing breaks with this error:
```shell
BUILD FAILED
D:\projects\eXml\frameworks\dita\DITA-OT2.x\build.xml:41: The following error occurred while executing this line:
D:\projects\eXml\frameworks\dita\DITA-OT2.x\plugins\org.dita.xhtml\build_general.xml:112: java.lang.IllegalArgumentException: URI scheme is not "file"
at java.io.File.<init>(Unknown Source)
at org.dita.dost.writer.ImageMetadataFilter.getImageFile(ImageMetadataFilter.java:172)
at org.dita.dost.writer.ImageMetadataFilter.startElement(ImageMetadataFilter.java:84)
at org.apache.xerces.parsers.AbstractSAXParser.startElement(Unknown Source)
at org.apache.xerces.parsers.AbstractXMLDocumentParser.emptyElement(Unknown Source)
at org.ditang.relaxng.defaults.RelaxNGDefaultsComponent.emptyElement(RelaxNGDefaultsComponent.java:635)
at org.apache.xerces.impl.XMLNSDocumentScannerImpl.scanStartElement(XMLNSDocumentScannerImpl.java:260)
at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl$FragmentContentDispatcher.dispatch(XMLDocumentFragmentScannerImpl.java:1655)
at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl.scanDocument(XMLDocumentFragmentScannerImpl.java:325)
at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
at org.ditang.relaxng.defaults.RelaxDefaultsParserConfiguration.parse(RelaxDefaultsParserConfiguration.java:150)
at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
at org.apache.xerces.parsers.XMLParser.parse(Unknown Source)
at org.apache.xerces.parsers.AbstractSAXParser.parse(Unknown Source)
at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source)
at net.sf.saxon.event.Sender.sendSAXSource(Sender.java:441)
at net.sf.saxon.event.Sender.send(Sender.java:146)
at net.sf.saxon.jaxp.IdentityTransformer.transform(IdentityTransformer.java:363)
at org.dita.dost.util.XMLUtils.transform(XMLUtils.java:244)
at org.dita.dost.util.XMLUtils.transform(XMLUtils.java:204)
at org.dita.dost.writer.AbstractXMLFilter.write(AbstractXMLFilter.java:32)
at org.dita.dost.writer.ImageMetadataFilter.write(ImageMetadataFilter.java:73)
at org.dita.dost.module.ImageMetadataModule.execute(ImageMetadataModule.java:46)
at org.dita.dost.pipeline.PipelineFacade.execute(PipelineFacade.java:68)
at org.dita.dost.invoker.ExtensibleAntInvoker.execute(ExtensibleAntInvoker.java:193)
at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292)
at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
```
|
1.0
|
Cannot publish remote HTTP DITA Map to XHTML [DOT 2.x develop branch] - I'm publishing a DITA Map from a webdav HTTP server without authentication.
I have images referenced in my topics, at some point the publishing breaks with this error:
```shell
BUILD FAILED
D:\projects\eXml\frameworks\dita\DITA-OT2.x\build.xml:41: The following error occurred while executing this line:
D:\projects\eXml\frameworks\dita\DITA-OT2.x\plugins\org.dita.xhtml\build_general.xml:112: java.lang.IllegalArgumentException: URI scheme is not "file"
at java.io.File.<init>(Unknown Source)
at org.dita.dost.writer.ImageMetadataFilter.getImageFile(ImageMetadataFilter.java:172)
at org.dita.dost.writer.ImageMetadataFilter.startElement(ImageMetadataFilter.java:84)
at org.apache.xerces.parsers.AbstractSAXParser.startElement(Unknown Source)
at org.apache.xerces.parsers.AbstractXMLDocumentParser.emptyElement(Unknown Source)
at org.ditang.relaxng.defaults.RelaxNGDefaultsComponent.emptyElement(RelaxNGDefaultsComponent.java:635)
at org.apache.xerces.impl.XMLNSDocumentScannerImpl.scanStartElement(XMLNSDocumentScannerImpl.java:260)
at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl$FragmentContentDispatcher.dispatch(XMLDocumentFragmentScannerImpl.java:1655)
at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl.scanDocument(XMLDocumentFragmentScannerImpl.java:325)
at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
at org.ditang.relaxng.defaults.RelaxDefaultsParserConfiguration.parse(RelaxDefaultsParserConfiguration.java:150)
at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
at org.apache.xerces.parsers.XMLParser.parse(Unknown Source)
at org.apache.xerces.parsers.AbstractSAXParser.parse(Unknown Source)
at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source)
at net.sf.saxon.event.Sender.sendSAXSource(Sender.java:441)
at net.sf.saxon.event.Sender.send(Sender.java:146)
at net.sf.saxon.jaxp.IdentityTransformer.transform(IdentityTransformer.java:363)
at org.dita.dost.util.XMLUtils.transform(XMLUtils.java:244)
at org.dita.dost.util.XMLUtils.transform(XMLUtils.java:204)
at org.dita.dost.writer.AbstractXMLFilter.write(AbstractXMLFilter.java:32)
at org.dita.dost.writer.ImageMetadataFilter.write(ImageMetadataFilter.java:73)
at org.dita.dost.module.ImageMetadataModule.execute(ImageMetadataModule.java:46)
at org.dita.dost.pipeline.PipelineFacade.execute(PipelineFacade.java:68)
at org.dita.dost.invoker.ExtensibleAntInvoker.execute(ExtensibleAntInvoker.java:193)
at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292)
at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
```
|
process
|
cannot publish remote http dita map to xhtml i m publishing a dita map from a webdav http server without authentication i have images referenced in my topics at some point the publishing breaks with this error shell build failed d projects exml frameworks dita dita x build xml the following error occurred while executing this line d projects exml frameworks dita dita x plugins org dita xhtml build general xml java lang illegalargumentexception uri scheme is not file at java io file unknown source at org dita dost writer imagemetadatafilter getimagefile imagemetadatafilter java at org dita dost writer imagemetadatafilter startelement imagemetadatafilter java at org apache xerces parsers abstractsaxparser startelement unknown source at org apache xerces parsers abstractxmldocumentparser emptyelement unknown source at org ditang relaxng defaults relaxngdefaultscomponent emptyelement relaxngdefaultscomponent java at org apache xerces impl xmlnsdocumentscannerimpl scanstartelement xmlnsdocumentscannerimpl java at org apache xerces impl xmldocumentfragmentscannerimpl fragmentcontentdispatcher dispatch xmldocumentfragmentscannerimpl java at org apache xerces impl xmldocumentfragmentscannerimpl scandocument xmldocumentfragmentscannerimpl java at org apache xerces parsers parse unknown source at org ditang relaxng defaults relaxdefaultsparserconfiguration parse relaxdefaultsparserconfiguration java at org apache xerces parsers parse unknown source at org apache xerces parsers xmlparser parse unknown source at org apache xerces parsers abstractsaxparser parse unknown source at org xml sax helpers xmlfilterimpl parse unknown source at net sf saxon event sender sendsaxsource sender java at net sf saxon event sender send sender java at net sf saxon jaxp identitytransformer transform identitytransformer java at org dita dost util xmlutils transform xmlutils java at org dita dost util xmlutils transform xmlutils java at org dita dost writer abstractxmlfilter write abstractxmlfilter java at org dita dost writer imagemetadatafilter write imagemetadatafilter java at org dita dost module imagemetadatamodule execute imagemetadatamodule java at org dita dost pipeline pipelinefacade execute pipelinefacade java at org dita dost invoker extensibleantinvoker execute extensibleantinvoker java at org apache tools ant unknownelement execute unknownelement java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke unknown source
| 1
|
21,214
| 28,292,127,141
|
IssuesEvent
|
2023-04-09 10:44:57
|
firebase/firebase-cpp-sdk
|
https://api.github.com/repos/firebase/firebase-cpp-sdk
|
closed
|
[C++] Nightly Integration Testing Report
|
type: process nightly-testing
|
Note: This report excludes firestore. Please also check **[the report for firestore](https://github.com/firebase/firebase-cpp-sdk/issues/1178)**
***
<hidden value="integration-test-status-comment"></hidden>
### [build against repo] Integration test with FLAKINESS (succeeded after retry)
Requested by @DellaBitta on commit 8b4226383959ef9558c4bf13e3b40012b4f75d08
Last updated: Sat Apr 8 03:50 PDT 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4644379829)**
| Failures | Configs |
|----------|---------|
| messaging | [TEST] [FLAKINESS] [Android] [1/3 os: windows] [1/4 android_device: android_target]<details><summary>(1 failed tests)</summary> CRASH/TIMEOUT</details> |
| storage | [TEST] [FLAKINESS] [iOS] [macos] [1/6 ios_device: ios_target]<details><summary>(1 failed tests)</summary> FirebaseStorageTest.TestWriteAndReadFileWithCustomMetadata</details> |
Add flaky tests to **[go/fpl-cpp-flake-tracker](http://go/fpl-cpp-flake-tracker)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against SDK] Integration test succeeded!
Requested by @firebase-workflow-trigger[bot] on commit 8b4226383959ef9558c4bf13e3b40012b4f75d08
Last updated: Sat Apr 8 11:44 PDT 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4645148803)**
<hidden value="integration-test-status-comment"></hidden>
|
1.0
|
[C++] Nightly Integration Testing Report - Note: This report excludes firestore. Please also check **[the report for firestore](https://github.com/firebase/firebase-cpp-sdk/issues/1178)**
***
<hidden value="integration-test-status-comment"></hidden>
### [build against repo] Integration test with FLAKINESS (succeeded after retry)
Requested by @DellaBitta on commit 8b4226383959ef9558c4bf13e3b40012b4f75d08
Last updated: Sat Apr 8 03:50 PDT 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4644379829)**
| Failures | Configs |
|----------|---------|
| messaging | [TEST] [FLAKINESS] [Android] [1/3 os: windows] [1/4 android_device: android_target]<details><summary>(1 failed tests)</summary> CRASH/TIMEOUT</details> |
| storage | [TEST] [FLAKINESS] [iOS] [macos] [1/6 ios_device: ios_target]<details><summary>(1 failed tests)</summary> FirebaseStorageTest.TestWriteAndReadFileWithCustomMetadata</details> |
Add flaky tests to **[go/fpl-cpp-flake-tracker](http://go/fpl-cpp-flake-tracker)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against SDK] Integration test succeeded!
Requested by @firebase-workflow-trigger[bot] on commit 8b4226383959ef9558c4bf13e3b40012b4f75d08
Last updated: Sat Apr 8 11:44 PDT 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4645148803)**
<hidden value="integration-test-status-comment"></hidden>
|
process
|
nightly integration testing report note this report excludes firestore please also check integration test with flakiness succeeded after retry requested by dellabitta on commit last updated sat apr pdt failures configs messaging failed tests nbsp nbsp crash timeout storage failed tests nbsp nbsp firebasestoragetest testwriteandreadfilewithcustommetadata add flaky tests to ✅ nbsp integration test succeeded requested by firebase workflow trigger on commit last updated sat apr pdt
| 1
|
1,264
| 3,797,110,332
|
IssuesEvent
|
2016-03-23 05:26:15
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
No link present in TOC (DITA OT 2.2.2)
|
bug P1 preprocess/keyref
|
I'm attaching a sample DITA Map:
http://www.oxygenxml.com/forum/files/noLinkInTOC.zip
When published to XHTML, the TOC has no link in it. It should have one because inside the subordinate DITA Map there is a topicref with keyref:
<topicref keyref="topicref_intro"/>
This works with DITA OT 1.8.
I tested on the 2.2.3 branch and does not seem to work. This also does not seem to work on the "develop" branch.
|
1.0
|
No link present in TOC (DITA OT 2.2.2) - I'm attaching a sample DITA Map:
http://www.oxygenxml.com/forum/files/noLinkInTOC.zip
When published to XHTML, the TOC has no link in it. It should have one because inside the subordinate DITA Map there is a topicref with keyref:
<topicref keyref="topicref_intro"/>
This works with DITA OT 1.8.
I tested on the 2.2.3 branch and does not seem to work. This also does not seem to work on the "develop" branch.
|
process
|
no link present in toc dita ot i m attaching a sample dita map when published to xhtml the toc has no link in it it should have one because inside the subordinate dita map there is a topicref with keyref this works with dita ot i tested on the branch and does not seem to work this also does not seem to work on the develop branch
| 1
|
327,617
| 28,073,959,400
|
IssuesEvent
|
2023-03-29 21:24:13
|
microsoft/vscode
|
https://api.github.com/repos/microsoft/vscode
|
closed
|
Go to Test Alt+Enter doesn't work on Linux
|
bug testing
|
Testing microsoft/vscode-python#17039
testing on Linux Debian build
Select a test in the Test Explorer and press Alt+Enter (called out in the context menu). Focus goes to the File menu and the cursor does not go to the selected test in the editor.
|
1.0
|
Go to Test Alt+Enter doesn't work on Linux - Testing microsoft/vscode-python#17039
testing on Linux Debian build
Select a test in the Test Explorer and press Alt+Enter (called out in the context menu). Focus goes to the File menu and the cursor does not go to the selected test in the editor.
|
non_process
|
go to test alt enter doesn t work on linux testing microsoft vscode python testing on linux debian build select a test in the test explorer and press alt enter called out in the context menu focus goes to the file menu and the cursor does not go to the selected test in the editor
| 0
|
85,635
| 3,697,061,759
|
IssuesEvent
|
2016-02-27 11:57:17
|
TwidereProject/Twidere-Android
|
https://api.github.com/repos/TwidereProject/Twidere-Android
|
closed
|
"Like"/fav star is not saved between timeline and tweet view
|
category:ui priority:high type:bug
|
When I "like" a tweet in my timeline and then go to the tweet detail view the "like" is not shown? Seems like a recent regression.
|
1.0
|
"Like"/fav star is not saved between timeline and tweet view - When I "like" a tweet in my timeline and then go to the tweet detail view the "like" is not shown? Seems like a recent regression.
|
non_process
|
like fav star is not saved between timeline and tweet view when i like a tweet in my timeline and then go to the tweet detail view the like is not shown seems like a recent regression
| 0
|
13,918
| 2,789,755,504
|
IssuesEvent
|
2015-05-08 21:17:26
|
google/google-visualization-api-issues
|
https://api.github.com/repos/google/google-visualization-api-issues
|
closed
|
Gauges API fails to display in IE8
|
Priority-Medium Type-Defect
|
Original [issue 58](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=58) created by orwant on 2009-09-18T04:57:11.000Z:
<b>What steps will reproduce the problem? Please provide a link to a</b>
<b>demonstration page if at all possible, or attach code.</b>
1.Make sure your IE8 is in IE8 mode, not Compatibility View or IE7
2.Load the following URL:
http://code.google.com/apis/visualization/documentation/gallery/gauge.html#Example
The gauges do not display.
<b>What component is this issue related to (PieChart, LineChart, DataTable,</b>
<b>Query, etc)?</b>
Gauges
<b>Are you using the test environment (version 1.1)?</b>
<b>(If you are not sure, answer NO)</b>
NO
<b>What operating system and browser are you using?</b>
Windows 7 64 bit RC and IE8.0.7100
For those who need a workaround add the following meta tag to your head
section (make sure it's before any CSS or other includes.
<meta http-equiv="X-UA-Compatible" content="IE=7" />
or
<meta http-equiv="X-UA-Compatible" content="IE=EmulateIE7" />
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
|
1.0
|
Gauges API fails to display in IE8 - Original [issue 58](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=58) created by orwant on 2009-09-18T04:57:11.000Z:
<b>What steps will reproduce the problem? Please provide a link to a</b>
<b>demonstration page if at all possible, or attach code.</b>
1.Make sure your IE8 is in IE8 mode, not Compatibility View or IE7
2.Load the following URL:
http://code.google.com/apis/visualization/documentation/gallery/gauge.html#Example
The gauges do not display.
<b>What component is this issue related to (PieChart, LineChart, DataTable,</b>
<b>Query, etc)?</b>
Gauges
<b>Are you using the test environment (version 1.1)?</b>
<b>(If you are not sure, answer NO)</b>
NO
<b>What operating system and browser are you using?</b>
Windows 7 64 bit RC and IE8.0.7100
For those who need a workaround add the following meta tag to your head
section (make sure it's before any CSS or other includes.
<meta http-equiv="X-UA-Compatible" content="IE=7" />
or
<meta http-equiv="X-UA-Compatible" content="IE=EmulateIE7" />
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
|
non_process
|
gauges api fails to display in original created by orwant on what steps will reproduce the problem please provide a link to a demonstration page if at all possible or attach code make sure your is in mode not compatibility view or load the following url the gauges do not display what component is this issue related to piechart linechart datatable query etc gauges are you using the test environment version if you are not sure answer no no what operating system and browser are you using windows bit rc and for those who need a workaround add the following meta tag to your head section make sure it s before any css or other includes lt meta http equiv quot x ua compatible quot content quot ie quot gt or lt meta http equiv quot x ua compatible quot content quot ie quot gt for developers viewing this issue please click the star icon to be notified of future changes and to let us know how many of you are interested in seeing it resolved
| 0
|
213,860
| 24,022,471,646
|
IssuesEvent
|
2022-09-15 08:48:10
|
sast-automation-dev/easybuggy4sb-25
|
https://api.github.com/repos/sast-automation-dev/easybuggy4sb-25
|
opened
|
bootstrap-3.3.7.min.js: 6 vulnerabilities (highest severity is: 6.1)
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /src/main/resources/templates/head.html</p>
<p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2019-8331](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1 | ❌ |
| [CVE-2018-14040](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14040) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0 | ❌ |
| [CVE-2018-20677](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20677) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0 | ❌ |
| [CVE-2018-20676](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | bootstrap - 3.4.0 | ❌ |
| [CVE-2018-14042](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0 | ❌ |
| [CVE-2016-10735](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | 3.4.0 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-8331</summary>
### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /src/main/resources/templates/head.html</p>
<p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.
<p>Publish Date: 2019-02-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-02-20</p>
<p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-14040</summary>
### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /src/main/resources/templates/head.html</p>
<p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 4.1.2, XSS is possible in the collapse data-parent attribute.
<p>Publish Date: 2018-07-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14040>CVE-2018-14040</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-07-13</p>
<p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-20677</summary>
### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /src/main/resources/templates/head.html</p>
<p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 3.4.0, XSS is possible in the affix configuration target property.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20677>CVE-2018-20677</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-20676</summary>
### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /src/main/resources/templates/head.html</p>
<p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-14042</summary>
### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /src/main/resources/templates/head.html</p>
<p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip.
<p>Publish Date: 2018-07-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042>CVE-2018-14042</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-07-13</p>
<p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2016-10735</summary>
### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /src/main/resources/templates/head.html</p>
<p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: 3.4.0</p>
</p>
<p></p>
</details>
|
True
|
bootstrap-3.3.7.min.js: 6 vulnerabilities (highest severity is: 6.1) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /src/main/resources/templates/head.html</p>
<p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2019-8331](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1 | ❌ |
| [CVE-2018-14040](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14040) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0 | ❌ |
| [CVE-2018-20677](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20677) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0 | ❌ |
| [CVE-2018-20676](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | bootstrap - 3.4.0 | ❌ |
| [CVE-2018-14042](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0 | ❌ |
| [CVE-2016-10735](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | 3.4.0 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-8331</summary>
### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /src/main/resources/templates/head.html</p>
<p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.
<p>Publish Date: 2019-02-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-02-20</p>
<p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-14040</summary>
### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /src/main/resources/templates/head.html</p>
<p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 4.1.2, XSS is possible in the collapse data-parent attribute.
<p>Publish Date: 2018-07-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14040>CVE-2018-14040</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-07-13</p>
<p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-20677</summary>
### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /src/main/resources/templates/head.html</p>
<p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 3.4.0, XSS is possible in the affix configuration target property.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20677>CVE-2018-20677</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-20676</summary>
### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /src/main/resources/templates/head.html</p>
<p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-14042</summary>
### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /src/main/resources/templates/head.html</p>
<p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip.
<p>Publish Date: 2018-07-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042>CVE-2018-14042</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-07-13</p>
<p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2016-10735</summary>
### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /src/main/resources/templates/head.html</p>
<p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: 3.4.0</p>
</p>
<p></p>
</details>
|
non_process
|
bootstrap min js vulnerabilities highest severity is vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main resources templates head html path to vulnerable library src main resources templates head html src main webapp dfi style bootstrap html found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available medium bootstrap min js direct bootstrap bootstrap sass medium bootstrap min js direct org webjars npm bootstrap org webjars bootstrap medium bootstrap min js direct bootstrap nordron angulartemplate dynamic net express projecttemplates dotnetng template znxtapp core module theme beta jmeter medium bootstrap min js direct bootstrap medium bootstrap min js direct org webjars npm bootstrap org webjars bootstrap medium bootstrap min js direct details cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main resources templates head html path to vulnerable library src main resources templates head html src main webapp dfi style bootstrap html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before and x before xss is possible in the tooltip or popover data template attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution bootstrap bootstrap sass cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main resources templates head html path to vulnerable library src main resources templates head html src main webapp dfi style bootstrap html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the collapse data parent attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution org webjars npm bootstrap org webjars bootstrap cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main resources templates head html path to vulnerable library src main resources templates head html src main webapp dfi style bootstrap html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the affix configuration target property publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap nordron angulartemplate dynamic net express projecttemplates dotnetng template znxtapp core module theme beta jmeter cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main resources templates head html path to vulnerable library src main resources templates head html src main webapp dfi style bootstrap html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the tooltip data viewport attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main resources templates head html path to vulnerable library src main resources templates head html src main webapp dfi style bootstrap html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the data container property of tooltip publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution org webjars npm bootstrap org webjars bootstrap cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main resources templates head html path to vulnerable library src main resources templates head html src main webapp dfi style bootstrap html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap x before and x beta before beta xss is possible in the data target attribute a different vulnerability than cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution
| 0
|
610,669
| 18,920,636,222
|
IssuesEvent
|
2021-11-17 00:57:35
|
googleapis/sphinx-docfx-yaml
|
https://api.github.com/repos/googleapis/sphinx-docfx-yaml
|
opened
|
Ensure handwritten libraries are well supported
|
type: feature request priority: p1
|
Googlers see b/206698969.
There are a lot of formats that the plugin does not support yet, which is a blocking issue for some of the handwritten libraries that are using much more variety of Sphinx features. The documentation does not get rendered properly, which could lead to a poor customer experience for the documentation.
|
1.0
|
Ensure handwritten libraries are well supported - Googlers see b/206698969.
There are a lot of formats that the plugin does not support yet, which is a blocking issue for some of the handwritten libraries that are using much more variety of Sphinx features. The documentation does not get rendered properly, which could lead to a poor customer experience for the documentation.
|
non_process
|
ensure handwritten libraries are well supported googlers see b there are a lot of formats that the plugin does not support yet which is a blocking issue for some of the handwritten libraries that are using much more variety of sphinx features the documentation does not get rendered properly which could lead to a poor customer experience for the documentation
| 0
|
21,843
| 6,226,906,102
|
IssuesEvent
|
2017-07-10 19:31:25
|
XceedBoucherS/TestImport5
|
https://api.github.com/repos/XceedBoucherS/TestImport5
|
closed
|
Control Request: TimePicker
|
CodePlex
|
<b>brianlagunas[CodePlex]</b> <br />Request for a new TimePicker control. This control should have similar functionality to the TimePicker in the Silverlight Toolkit
|
1.0
|
Control Request: TimePicker - <b>brianlagunas[CodePlex]</b> <br />Request for a new TimePicker control. This control should have similar functionality to the TimePicker in the Silverlight Toolkit
|
non_process
|
control request timepicker brianlagunas request for a new timepicker control this control should have similar functionality to the timepicker in the silverlight toolkit
| 0
|
10,960
| 13,766,192,902
|
IssuesEvent
|
2020-10-07 14:17:29
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
macOS in Environment
|
Pri2 devops-cicd-process/tech devops/prod product-question
|
[Enter feedback here]
Is there a way to create an environment using macOS instead of Windows or Linux?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 91d0d31f-81ee-c024-db7e-daddbf525f71
* Version Independent ID: 330f1649-386c-d0aa-5f96-b8343a1480d3
* Content: [Environment - Virtual machine resource - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments-virtual-machines?view=azure-devops)
* Content Source: [docs/pipelines/process/environments-virtual-machines.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments-virtual-machines.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
macOS in Environment -
[Enter feedback here]
Is there a way to create an environment using macOS instead of Windows or Linux?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 91d0d31f-81ee-c024-db7e-daddbf525f71
* Version Independent ID: 330f1649-386c-d0aa-5f96-b8343a1480d3
* Content: [Environment - Virtual machine resource - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments-virtual-machines?view=azure-devops)
* Content Source: [docs/pipelines/process/environments-virtual-machines.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments-virtual-machines.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
macos in environment is there a way to create an environment using macos instead of windows or linux document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
16,915
| 22,263,827,117
|
IssuesEvent
|
2022-06-10 04:54:19
|
camunda/zeebe
|
https://api.github.com/repos/camunda/zeebe
|
closed
|
Writer timer triggered event with process instance key
|
kind/feature scope/broker team/process-automation
|
**Is your feature request related to a problem? Please describe.**
I deployed a process with a timer start event. If the timer is triggered then a new process instance is created and an event of type `timer` with intent `triggered` is written.
As a monitoring application, I want to track which process instance was created by this timer. But currently, the `timer: triggered` event doesn't contain the key of the created process instance.
**Describe the solution you'd like**
The `timer: triggered` event contains the key of the created process instance.
**Describe alternatives you've considered**
Add a reference of the timer to the created process instance. But this option seems not practical because a process instance can be created by different triggers, for example, a message.
**Additional context**
In a monitoring application, I want to track that a process instance was created by a triggered timer start event.
|
1.0
|
Writer timer triggered event with process instance key - **Is your feature request related to a problem? Please describe.**
I deployed a process with a timer start event. If the timer is triggered then a new process instance is created and an event of type `timer` with intent `triggered` is written.
As a monitoring application, I want to track which process instance was created by this timer. But currently, the `timer: triggered` event doesn't contain the key of the created process instance.
**Describe the solution you'd like**
The `timer: triggered` event contains the key of the created process instance.
**Describe alternatives you've considered**
Add a reference of the timer to the created process instance. But this option seems not practical because a process instance can be created by different triggers, for example, a message.
**Additional context**
In a monitoring application, I want to track that a process instance was created by a triggered timer start event.
|
process
|
writer timer triggered event with process instance key is your feature request related to a problem please describe i deployed a process with a timer start event if the timer is triggered then a new process instance is created and an event of type timer with intent triggered is written as a monitoring application i want to track which process instance was created by this timer but currently the timer triggered event doesn t contain the key of the created process instance describe the solution you d like the timer triggered event contains the key of the created process instance describe alternatives you ve considered add a reference of the timer to the created process instance but this option seems not practical because a process instance can be created by different triggers for example a message additional context in a monitoring application i want to track that a process instance was created by a triggered timer start event
| 1
|
52,836
| 13,064,101,982
|
IssuesEvent
|
2020-07-30 17:32:50
|
GoogleCloudPlatform/golang-samples
|
https://api.github.com/repos/GoogleCloudPlatform/golang-samples
|
closed
|
asset/quickstart/batch-get-assets-history: TestMain failed
|
api: cloudasset buildcop: issue priority: p1 sample type: bug
|
This test failed!
To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop).
If I'm commenting on this issue too often, add the `buildcop: quiet` label and
I will stop commenting.
---
commit: 9d203dfe6a2bf97041383afc7c74880cdfbec364
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/f337f766-bb8e-4406-a4cb-9c2a7048957c), [Sponge](http://sponge2/f337f766-bb8e-4406-a4cb-9c2a7048957c)
status: failed
|
1.0
|
asset/quickstart/batch-get-assets-history: TestMain failed - This test failed!
To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop).
If I'm commenting on this issue too often, add the `buildcop: quiet` label and
I will stop commenting.
---
commit: 9d203dfe6a2bf97041383afc7c74880cdfbec364
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/f337f766-bb8e-4406-a4cb-9c2a7048957c), [Sponge](http://sponge2/f337f766-bb8e-4406-a4cb-9c2a7048957c)
status: failed
|
non_process
|
asset quickstart batch get assets history testmain failed this test failed to configure my behavior see if i m commenting on this issue too often add the buildcop quiet label and i will stop commenting commit buildurl status failed
| 0
|
226,832
| 18,044,208,715
|
IssuesEvent
|
2021-09-18 15:50:14
|
logicmoo/logicmoo_workspace
|
https://api.github.com/repos/logicmoo/logicmoo_workspace
|
opened
|
logicmoo.pfc.test.sanity_base.NEG_01V JUnit
|
Test_9999 logicmoo.pfc.test.sanity_base unit_test NEG_01V
|
(cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc)
GH_MASTER_ISSUE_FINFO=
ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ANEG_01V
GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3
https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/pfc/t/sanity_base/neg_01v.pfc
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/NEG_01V/logicmoo_pfc_test_sanity_base_NEG_01V_JUnit/
This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/67/testReport/logicmoo.pfc.test.sanity_base/NEG_01V/logicmoo_pfc_test_sanity_base_NEG_01V_JUnit/
GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3
https://github.com/logicmoo/logicmoo_workspace/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/pfc/t/sanity_base/neg_01v.pfc
```
%
running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/neg_01v.pfc'),
%~ /var/lib/jenkins/.local/share/swi-prolog/pack/logicmoo_utils/prolog/logicmoo_test_header.pl:92
%~ this_test_might_need( :-( use_module( library(logicmoo_plarkc))))
:- use_module(library(statistics)).
%:- mpred_notrace_exec.
% reset runtime counter
%:- mpred_notrace_exec.
% reset runtime counter
:- statistics(runtime,_Secs).
~path(X,X).
path(1,2).
path(2,3).
path(3,4).
path(1,1).
:- mpred_test(path(3, 4)).
%~ mpred_test("Test_0001_Line_0000__path_3",baseKB:path(3,4))
%~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L20
/*~
%~ mpred_test("Test_0001_Line_0000__path_3",baseKB:path(3,4))
passed=info(why_was_true(baseKB:path(3,4)))
Justifications for path(3,4):
[36m 1.1 mfl4(_,baseKB,'* https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L17 ',17) [0m
name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0001_Line_0000__path_3'.
JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'.
JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'.
% saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0001_Line_0000__path_3-junit.xml
~*/
:- mpred_test(path(2, 3)).
%~ mpred_test("Test_0002_Line_0000__path_2",baseKB:path(2,3))
%~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L21
/*~
%~ mpred_test("Test_0002_Line_0000__path_2",baseKB:path(2,3))
passed=info(why_was_true(baseKB:path(2,3)))
Justifications for path(2,3):
[36m 1.1 mfl4(_,baseKB,'* https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L16 ',16) [0m
name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0002_Line_0000__path_2'.
JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'.
JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'.
% saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0002_Line_0000__path_2-junit.xml
~*/
:- mpred_test(path(1, 2)).
%~ mpred_test("Test_0003_Line_0000__path_1",baseKB:path(1,2))
%~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L22
/*~
%~ mpred_test("Test_0003_Line_0000__path_1",baseKB:path(1,2))
passed=info(why_was_true(baseKB:path(1,2)))
Justifications for path(1,2):
[36m 1.1 mfl4(_,baseKB,'* https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L15 ',15) [0m
name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0003_Line_0000__path_1'.
JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'.
JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'.
% saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0003_Line_0000__path_1-junit.xml
~*/
:- mpred_test(~path(1,1)).
%~ mpred_test("Test_0004_Line_0000__path_1",baseKB: ~path(1,1))
/*~
%~ mpred_test("Test_0004_Line_0000__path_1",baseKB: ~path(1,1))
^ Call: (68) [baseKB] ~path(1, 1)
^ Unify: (68) [baseKB] ~ (baseKB:path(1, 1))
^ Call: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(1, 1)), info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1320, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(1, 1)))))
^ Unify: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(1, 1)), info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1320, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(1, 1)))))
Call: (76) [system] set_prolog_flag(last_call_optimisation, false)
Exit: (76) [system] set_prolog_flag(last_call_optimisation, false)
^ Call: (76) [loop_check] prolog_frame_attribute(1320, parent_goal, loop_check_term_frame(_87158, info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, _87164, _87166))
^ Fail: (76) [loop_check] prolog_frame_attribute(1320, parent_goal, loop_check_term_frame(_87158, info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, _87164, _87166))
^ Redo: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(1, 1)), info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1320, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(1, 1)))))
Call: (76) [pfc_lib] neg_in_code0(baseKB:path(1, 1))
Unify: (76) [pfc_lib] neg_in_code0(baseKB:path(1, 1))
^ Call: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(1, 1)), info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1459, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(1, 1)))))
^ Unify: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(1, 1)), info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1459, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(1, 1)))))
Call: (83) [system] set_prolog_flag(last_call_optimisation, false)
Exit: (83) [system] set_prolog_flag(last_call_optimisation, false)
^ Call: (83) [loop_check] prolog_frame_attribute(1459, parent_goal, loop_check_term_frame(_92882, info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, _92888, _92890))
^ Fail: (83) [loop_check] prolog_frame_attribute(1459, parent_goal, loop_check_term_frame(_92882, info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, _92888, _92890))
^ Redo: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(1, 1)), info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1459, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(1, 1)))))
Call: (83) [pfc_lib] neg_may_naf(baseKB:path(1, 1))
Unify: (83) [pfc_lib] neg_may_naf(baseKB:path(1, 1))
^ Call: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _96214)
^ Unify: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _96214)
^ Call: (88) [system] clause(pfc_lib:prologNegByFailure(path), true, _96214)
^ Fail: (88) [system] clause(pfc_lib:prologNegByFailure(path), true, _96214)
^ Fail: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _96214)
Unify: (83) [pfc_lib] neg_may_naf(baseKB:path(1, 1))
^ Call: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(1, 1))
^ Unify: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(1, 1))
^ Call: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(1, 1))
^ Unify: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(1, 1))
^ Fail: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(1, 1))
^ Redo: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(1, 1))
^ Exit: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(1, 1))
^ Call: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384), call(_104384)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))))
^ Unify: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384), call(_104384)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))))
^ Call: (90) [hook_database] clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384)
^ Fail: (90) [hook_database] clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384)
Call: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))
Unify: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))
^ Call: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170))
^ Unify: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170))
^ Call: (92) [baseKB] clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170)
^ Fail: (92) [baseKB] clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170)
^ Fail: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170))
Fail: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))
^ Fail: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384), call(_104384)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))))
^ Call: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _112590)
^ Unify: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, syntaxic(_113160))
^ Call: (88) [pfc_lib] mpred_database_term_syntax(path, 2, _113160)
^ Fail: (88) [pfc_lib] mpred_database_term_syntax(path, 2, _113160)
^ Redo: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _115162)
^ Unify: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _115790)
^ Call: (88) [pfc_lib] mpred_core_database_term(path, 2, _116358)
^ Fail: (88) [pfc_lib] mpred_core_database_term(path, 2, _116358)
^ Fail: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _117732)
^ Call: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336), call(_118336)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))))
^ Unify: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336), call(_118336)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))))
^ Call: (88) [hook_database] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336)
^ Fail: (88) [hook_database] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336)
Call: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))
Unify: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))
^ Call: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122))
^ Unify: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122))
^ Call: (90) [baseKB] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122)
^ Fail: (90) [baseKB] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122)
^ Fail: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122))
Fail: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))
^ Fail: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336), call(_118336)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))))
^ Call: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_126488, path, 2, prologHybrid), _126518), call(_126518)*->true;clause_b(baseKB:mpred_prop(_126488, path, 2, prologHybrid))))
^ Unify: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_126488, path, 2, prologHybrid), _126518), call(_126518)*->true;clause_b(baseKB:mpred_prop(_126488, path, 2, prologHybrid))))
^ Call: (91) [hook_database] clause(mpred_prop(_126488, path, 2, prologHybrid), _126518)
^ Fail: (91) [hook_database] clause(mpred_prop(_126488, path, 2, prologHybrid), _126518)
Call: (91) [hook_database] clause_b(baseKB:mpred_prop(_126488, path, 2, prologHybrid))
Unify: (91) [hook_database] clause_b(baseKB:mpred_prop(_126488, path, 2, prologHybrid))
^ Call: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_126488, path, 2, prologHybrid), _130304))
^ Unify: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_1434, path, 2, prologHybrid), _1490))
^ Call: (93) [baseKB] clause(mpred_prop(_1434, path, 2, prologHybrid), _1490)
^ Fail: (93) [baseKB] clause(mpred_prop(_1434, path, 2, prologHybrid), _1490)
^ Fail: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_1434, path, 2, prologHybrid), _1490))
Fail: (91) [hook_database] clause_b(baseKB:mpred_prop(_1434, path, 2, prologHybrid))
^ Fail: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_1434, path, 2, prologHybrid), _1464), call(_1464)*->true;clause_b(baseKB:mpred_prop(_1434, path, 2, prologHybrid))))
^ Call: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _5272)
^ Unify: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, syntaxic(_5842))
^ Call: (89) [pfc_lib] mpred_database_term_syntax(path, 2, _5842)
^ Fail: (89) [pfc_lib] mpred_database_term_syntax(path, 2, _5842)
^ Redo: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _7844)
^ Unify: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _8472)
^ Call: (89) [pfc_lib] mpred_core_database_term(path, 2, _9040)
^ Fail: (89) [pfc_lib] mpred_core_database_term(path, 2, _9040)
^ Fail: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _10414)
Call: (98) [$autoload] leave_sandbox(_11012)
Unify: (98) [$autoload] leave_sandbox(_11012)
Exit: (98) [$autoload] leave_sandbox(false)
Call: (97) [$autoload] restore_sandbox(false)
Unify: (97) [$autoload] restore_sandbox(false)
Exit: (97) [$autoload] restore_sandbox(false)
Fail: (83) [pfc_lib] neg_may_naf(baseKB:path(1, 1))
^ Fail: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(1, 1)), info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1459, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(1, 1)))))
Fail: (76) [pfc_lib] neg_in_code0(baseKB:path(1, 1))
^ Fail: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(1, 1)), info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1320, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(1, 1)))))
^ Fail: (68) [baseKB] ~ (baseKB:path(1, 1))
^ Call: (68) [must_sanity] must_sanity:mquietly_if(true, rtrace:tAt_normal)
^ Unify: (68) [must_sanity] must_sanity:mquietly_if(true, rtrace:tAt_normal)
failure=info((why_was_true(baseKB:(\+ ~path(1,1))),rtrace(baseKB: ~path(1,1))))
no_proof_for(\+ ~path(1,1)).
no_proof_for(\+ ~path(1,1)).
no_proof_for(\+ ~path(1,1)).
name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0004_Line_0000__path_1'.
JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'.
JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'.
% saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0004_Line_0000__path_1-junit.xml
~*/
:- mpred_test(~path(2,2)).
%~ mpred_test("Test_0005_Line_0000__path_2",baseKB: ~path(2,2))
/*~
%~ mpred_test("Test_0005_Line_0000__path_2",baseKB: ~path(2,2))
^ Call: (68) [baseKB] ~path(2, 2)
^ Unify: (68) [baseKB] ~ (baseKB:path(2, 2))
^ Call: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(2, 2)), info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1189, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(2, 2)))))
^ Unify: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(2, 2)), info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1189, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(2, 2)))))
Call: (76) [system] set_prolog_flag(last_call_optimisation, false)
Exit: (76) [system] set_prolog_flag(last_call_optimisation, false)
^ Call: (76) [loop_check] prolog_frame_attribute(1189, parent_goal, loop_check_term_frame(_121764, info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, _121770, _121772))
^ Fail: (76) [loop_check] prolog_frame_attribute(1189, parent_goal, loop_check_term_frame(_121764, info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, _121770, _121772))
^ Redo: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(2, 2)), info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1189, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(2, 2)))))
Call: (76) [pfc_lib] neg_in_code0(baseKB:path(2, 2))
Unify: (76) [pfc_lib] neg_in_code0(baseKB:path(2, 2))
^ Call: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(2, 2)), info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1328, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(2, 2)))))
^ Unify: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(2, 2)), info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1328, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(2, 2)))))
Call: (83) [system] set_prolog_flag(last_call_optimisation, false)
Exit: (83) [system] set_prolog_flag(last_call_optimisation, false)
^ Call: (83) [loop_check] prolog_frame_attribute(1328, parent_goal, loop_check_term_frame(_127488, info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, _127494, _127496))
^ Fail: (83) [loop_check] prolog_frame_attribute(1328, parent_goal, loop_check_term_frame(_127488, info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, _127494, _127496))
^ Redo: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(2, 2)), info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1328, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(2, 2)))))
Call: (83) [pfc_lib] neg_may_naf(baseKB:path(2, 2))
Unify: (83) [pfc_lib] neg_may_naf(baseKB:path(2, 2))
^ Call: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _130820)
^ Unify: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _130820)
^ Call: (88) [system] clause(pfc_lib:prologNegByFailure(path), true, _130820)
^ Fail: (88) [system] clause(pfc_lib:prologNegByFailure(path), true, _130820)
^ Fail: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _130820)
Unify: (83) [pfc_lib] neg_may_naf(baseKB:path(2, 2))
^ Call: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(2, 2))
^ Unify: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(2, 2))
^ Call: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(2, 2))
^ Unify: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(2, 2))
^ Fail: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(2, 2))
^ Redo: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(2, 2))
^ Exit: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(2, 2))
^ Call: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990), call(_138990)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))))
^ Unify: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990), call(_138990)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))))
^ Call: (90) [hook_database] clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990)
^ Fail: (90) [hook_database] clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990)
Call: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))
Unify: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))
^ Call: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776))
^ Unify: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776))
^ Call: (92) [baseKB] clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776)
^ Fail: (92) [baseKB] clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776)
^ Fail: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776))
Fail: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))
^ Fail: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990), call(_138990)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))))
^ Call: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _147196)
^ Unify: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, syntaxic(_147766))
^ Call: (88) [pfc_lib] mpred_database_term_syntax(path, 2, _147766)
^ Fail: (88) [pfc_lib] mpred_database_term_syntax(path, 2, _147766)
^ Redo: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _149768)
^ Unify: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _150396)
^ Call: (88) [pfc_lib] mpred_core_database_term(path, 2, _150964)
^ Fail: (88) [pfc_lib] mpred_core_database_term(path, 2, _150964)
^ Fail: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _152338)
^ Call: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942), call(_152942)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))))
^ Unify: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942), call(_152942)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))))
^ Call: (88) [hook_database] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942)
^ Fail: (88) [hook_database] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942)
Call: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))
Unify: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))
^ Call: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728))
^ Unify: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728))
^ Call: (90) [baseKB] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728)
^ Fail: (90) [baseKB] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728)
^ Fail: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728))
Fail: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))
^ Fail: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942), call(_152942)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))))
^ Call: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_161094, path, 2, prologHybrid), _161124), call(_161124)*->true;clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid))))
^ Unify: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_161094, path, 2, prologHybrid), _161124), call(_161124)*->true;clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid))))
^ Call: (91) [hook_database] clause(mpred_prop(_161094, path, 2, prologHybrid), _161124)
^ Fail: (91) [hook_database] clause(mpred_prop(_161094, path, 2, prologHybrid), _161124)
Call: (91) [hook_database] clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid))
Unify: (91) [hook_database] clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid))
^ Call: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_161094, path, 2, prologHybrid), _164910))
^ Unify: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_161094, path, 2, prologHybrid), _164910))
^ Call: (93) [baseKB] clause(mpred_prop(_161094, path, 2, prologHybrid), _164910)
^ Fail: (93) [baseKB] clause(mpred_prop(_161094, path, 2, prologHybrid), _164910)
^ Fail: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_161094, path, 2, prologHybrid), _164910))
Fail: (91) [hook_database] clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid))
^ Fail: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_161094, path, 2, prologHybrid), _161124), call(_161124)*->true;clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid))))
^ Call: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _169330)
^ Unify: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, syntaxic(_169900))
^ Call: (89) [pfc_lib] mpred_database_term_syntax(path, 2, _169900)
^ Fail: (89) [pfc_lib] mpred_database_term_syntax(path, 2, _169900)
^ Redo: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _171902)
^ Unify: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _172530)
^ Call: (89) [pfc_lib] mpred_core_database_term(path, 2, _173098)
^ Fail: (89) [pfc_lib] mpred_core_database_term(path, 2, _173098)
^ Fail: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _174472)
Call: (98) [$autoload] leave_sandbox(_175070)
Unify: (98) [$autoload] leave_sandbox(_175070)
Exit: (98) [$autoload] leave_sandbox(false)
Call: (97) [$autoload] restore_sandbox(false)
Unify: (97) [$autoload] restore_sandbox(false)
Exit: (97) [$autoload] restore_sandbox(false)
Fail: (83) [pfc_lib] neg_may_naf(baseKB:path(2, 2))
^ Fail: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(2, 2)), info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1328, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(2, 2)))))
Fail: (76) [pfc_lib] neg_in_code0(baseKB:path(2, 2))
^ Fail: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(2, 2)), info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1189, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(2, 2)))))
^ Fail: (68) [baseKB] ~ (baseKB:path(2, 2))
^ Call: (68) [must_sanity] must_sanity:mquietly_if(true, rtrace:tAt_normal)
^ Unify: (68) [must_sanity] must_sanity:mquietly_if(true, rtrace:tAt_normal)
failure=info((why_was_true(baseKB:(\+ ~path(2,2))),rtrace(baseKB: ~path(2,2))))
no_proof_for(\+ ~path(2,2)).
no_proof_for(\+ ~path(2,2)).
no_proof_for(\+ ~path(2,2)).
name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0005_Line_0000__path_2'.
JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'.
JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'.
% saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0005_Line_0000__path_2-junit.xml
~*/
/*
:- mpred_test(\+path(1,1)).
:- mpred_test(\+path(2,2)).
*/
:- mpred_test(\+path(3,1)).
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/neg_01v.pfc:32
%~ mpred_test("Test_0006_Line_0000__naf_path_3",baseKB:(\+path(3,1)))
/*~
%~ mpred_test("Test_0006_Line_0000__naf_path_3",baseKB:(\+path(3,1)))
passed=info(why_was_true(baseKB:(\+path(3,1))))
no_proof_for(\+path(3,1)).
no_proof_for(\+path(3,1)).
no_proof_for(\+path(3,1)).
name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0006_Line_0000__naf_path_3'.
JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'.
JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'.
% saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0006_Line_0000__naf_path_3-junit.xml
~*/
:- mpred_test(\+ ~path(1,5)).
%~ mpred_test("Test_0007_Line_0000__naf_path_1",baseKB:(\+ ~path(1,5)))
/*~
%~ mpred_test("Test_0007_Line_0000__naf_path_1",baseKB:(\+ ~path(1,5)))
passed=info(why_was_true(baseKB:(\+ ~path(1,5))))
no_proof_for(\+ ~path(1,5)).
no_proof_for(\+ ~path(1,5)).
no_proof_for(\+ ~path(1,5)).
name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0007_Line_0000__naf_path_1'.
JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'.
JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'.
% saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0007_Line_0000__naf_path_1-junit.xml
~*/
:- break.
%~ skipped(blocks_on_input,break)
%~ unused(save_junit_results)
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/neg_01v.pfc:36
%~ test_completed_exit(6)
:- dynamic junit_prop/3.
:- dynamic junit_prop/3.
:- dynamic junit_prop/3.
```
totalTime=1
ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ANEG_01V
GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3
https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/pfc/t/sanity_base/neg_01v.pfc
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/NEG_01V/logicmoo_pfc_test_sanity_base_NEG_01V_JUnit/
This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/67/testReport/logicmoo.pfc.test.sanity_base/NEG_01V/logicmoo_pfc_test_sanity_base_NEG_01V_JUnit/
GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3
https://github.com/logicmoo/logicmoo_workspace/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/pfc/t/sanity_base/neg_01v.pfc
FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k neg_01v.pfc (returned 6)
|
3.0
|
logicmoo.pfc.test.sanity_base.NEG_01V JUnit - (cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc)
GH_MASTER_ISSUE_FINFO=
ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ANEG_01V
GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3
https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/pfc/t/sanity_base/neg_01v.pfc
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/NEG_01V/logicmoo_pfc_test_sanity_base_NEG_01V_JUnit/
This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/67/testReport/logicmoo.pfc.test.sanity_base/NEG_01V/logicmoo_pfc_test_sanity_base_NEG_01V_JUnit/
GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3
https://github.com/logicmoo/logicmoo_workspace/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/pfc/t/sanity_base/neg_01v.pfc
```
%
running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/neg_01v.pfc'),
%~ /var/lib/jenkins/.local/share/swi-prolog/pack/logicmoo_utils/prolog/logicmoo_test_header.pl:92
%~ this_test_might_need( :-( use_module( library(logicmoo_plarkc))))
:- use_module(library(statistics)).
%:- mpred_notrace_exec.
% reset runtime counter
%:- mpred_notrace_exec.
% reset runtime counter
:- statistics(runtime,_Secs).
~path(X,X).
path(1,2).
path(2,3).
path(3,4).
path(1,1).
:- mpred_test(path(3, 4)).
%~ mpred_test("Test_0001_Line_0000__path_3",baseKB:path(3,4))
%~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L20
/*~
%~ mpred_test("Test_0001_Line_0000__path_3",baseKB:path(3,4))
passed=info(why_was_true(baseKB:path(3,4)))
Justifications for path(3,4):
[36m 1.1 mfl4(_,baseKB,'* https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L17 ',17) [0m
name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0001_Line_0000__path_3'.
JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'.
JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'.
% saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0001_Line_0000__path_3-junit.xml
~*/
:- mpred_test(path(2, 3)).
%~ mpred_test("Test_0002_Line_0000__path_2",baseKB:path(2,3))
%~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L21
/*~
%~ mpred_test("Test_0002_Line_0000__path_2",baseKB:path(2,3))
passed=info(why_was_true(baseKB:path(2,3)))
Justifications for path(2,3):
[36m 1.1 mfl4(_,baseKB,'* https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L16 ',16) [0m
name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0002_Line_0000__path_2'.
JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'.
JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'.
% saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0002_Line_0000__path_2-junit.xml
~*/
:- mpred_test(path(1, 2)).
%~ mpred_test("Test_0003_Line_0000__path_1",baseKB:path(1,2))
%~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L22
/*~
%~ mpred_test("Test_0003_Line_0000__path_1",baseKB:path(1,2))
passed=info(why_was_true(baseKB:path(1,2)))
Justifications for path(1,2):
[36m 1.1 mfl4(_,baseKB,'* https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L15 ',15) [0m
name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0003_Line_0000__path_1'.
JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'.
JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'.
% saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0003_Line_0000__path_1-junit.xml
~*/
:- mpred_test(~path(1,1)).
%~ mpred_test("Test_0004_Line_0000__path_1",baseKB: ~path(1,1))
/*~
%~ mpred_test("Test_0004_Line_0000__path_1",baseKB: ~path(1,1))
^ Call: (68) [baseKB] ~path(1, 1)
^ Unify: (68) [baseKB] ~ (baseKB:path(1, 1))
^ Call: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(1, 1)), info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1320, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(1, 1)))))
^ Unify: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(1, 1)), info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1320, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(1, 1)))))
Call: (76) [system] set_prolog_flag(last_call_optimisation, false)
Exit: (76) [system] set_prolog_flag(last_call_optimisation, false)
^ Call: (76) [loop_check] prolog_frame_attribute(1320, parent_goal, loop_check_term_frame(_87158, info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, _87164, _87166))
^ Fail: (76) [loop_check] prolog_frame_attribute(1320, parent_goal, loop_check_term_frame(_87158, info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, _87164, _87166))
^ Redo: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(1, 1)), info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1320, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(1, 1)))))
Call: (76) [pfc_lib] neg_in_code0(baseKB:path(1, 1))
Unify: (76) [pfc_lib] neg_in_code0(baseKB:path(1, 1))
^ Call: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(1, 1)), info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1459, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(1, 1)))))
^ Unify: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(1, 1)), info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1459, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(1, 1)))))
Call: (83) [system] set_prolog_flag(last_call_optimisation, false)
Exit: (83) [system] set_prolog_flag(last_call_optimisation, false)
^ Call: (83) [loop_check] prolog_frame_attribute(1459, parent_goal, loop_check_term_frame(_92882, info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, _92888, _92890))
^ Fail: (83) [loop_check] prolog_frame_attribute(1459, parent_goal, loop_check_term_frame(_92882, info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, _92888, _92890))
^ Redo: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(1, 1)), info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1459, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(1, 1)))))
Call: (83) [pfc_lib] neg_may_naf(baseKB:path(1, 1))
Unify: (83) [pfc_lib] neg_may_naf(baseKB:path(1, 1))
^ Call: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _96214)
^ Unify: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _96214)
^ Call: (88) [system] clause(pfc_lib:prologNegByFailure(path), true, _96214)
^ Fail: (88) [system] clause(pfc_lib:prologNegByFailure(path), true, _96214)
^ Fail: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _96214)
Unify: (83) [pfc_lib] neg_may_naf(baseKB:path(1, 1))
^ Call: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(1, 1))
^ Unify: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(1, 1))
^ Call: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(1, 1))
^ Unify: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(1, 1))
^ Fail: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(1, 1))
^ Redo: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(1, 1))
^ Exit: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(1, 1))
^ Call: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384), call(_104384)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))))
^ Unify: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384), call(_104384)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))))
^ Call: (90) [hook_database] clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384)
^ Fail: (90) [hook_database] clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384)
Call: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))
Unify: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))
^ Call: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170))
^ Unify: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170))
^ Call: (92) [baseKB] clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170)
^ Fail: (92) [baseKB] clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170)
^ Fail: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170))
Fail: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))
^ Fail: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384), call(_104384)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))))
^ Call: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _112590)
^ Unify: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, syntaxic(_113160))
^ Call: (88) [pfc_lib] mpred_database_term_syntax(path, 2, _113160)
^ Fail: (88) [pfc_lib] mpred_database_term_syntax(path, 2, _113160)
^ Redo: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _115162)
^ Unify: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _115790)
^ Call: (88) [pfc_lib] mpred_core_database_term(path, 2, _116358)
^ Fail: (88) [pfc_lib] mpred_core_database_term(path, 2, _116358)
^ Fail: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _117732)
^ Call: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336), call(_118336)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))))
^ Unify: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336), call(_118336)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))))
^ Call: (88) [hook_database] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336)
^ Fail: (88) [hook_database] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336)
Call: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))
Unify: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))
^ Call: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122))
^ Unify: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122))
^ Call: (90) [baseKB] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122)
^ Fail: (90) [baseKB] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122)
^ Fail: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122))
Fail: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))
^ Fail: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336), call(_118336)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))))
^ Call: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_126488, path, 2, prologHybrid), _126518), call(_126518)*->true;clause_b(baseKB:mpred_prop(_126488, path, 2, prologHybrid))))
^ Unify: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_126488, path, 2, prologHybrid), _126518), call(_126518)*->true;clause_b(baseKB:mpred_prop(_126488, path, 2, prologHybrid))))
^ Call: (91) [hook_database] clause(mpred_prop(_126488, path, 2, prologHybrid), _126518)
^ Fail: (91) [hook_database] clause(mpred_prop(_126488, path, 2, prologHybrid), _126518)
Call: (91) [hook_database] clause_b(baseKB:mpred_prop(_126488, path, 2, prologHybrid))
Unify: (91) [hook_database] clause_b(baseKB:mpred_prop(_126488, path, 2, prologHybrid))
^ Call: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_126488, path, 2, prologHybrid), _130304))
^ Unify: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_1434, path, 2, prologHybrid), _1490))
^ Call: (93) [baseKB] clause(mpred_prop(_1434, path, 2, prologHybrid), _1490)
^ Fail: (93) [baseKB] clause(mpred_prop(_1434, path, 2, prologHybrid), _1490)
^ Fail: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_1434, path, 2, prologHybrid), _1490))
Fail: (91) [hook_database] clause_b(baseKB:mpred_prop(_1434, path, 2, prologHybrid))
^ Fail: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_1434, path, 2, prologHybrid), _1464), call(_1464)*->true;clause_b(baseKB:mpred_prop(_1434, path, 2, prologHybrid))))
^ Call: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _5272)
^ Unify: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, syntaxic(_5842))
^ Call: (89) [pfc_lib] mpred_database_term_syntax(path, 2, _5842)
^ Fail: (89) [pfc_lib] mpred_database_term_syntax(path, 2, _5842)
^ Redo: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _7844)
^ Unify: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _8472)
^ Call: (89) [pfc_lib] mpred_core_database_term(path, 2, _9040)
^ Fail: (89) [pfc_lib] mpred_core_database_term(path, 2, _9040)
^ Fail: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _10414)
Call: (98) [$autoload] leave_sandbox(_11012)
Unify: (98) [$autoload] leave_sandbox(_11012)
Exit: (98) [$autoload] leave_sandbox(false)
Call: (97) [$autoload] restore_sandbox(false)
Unify: (97) [$autoload] restore_sandbox(false)
Exit: (97) [$autoload] restore_sandbox(false)
Fail: (83) [pfc_lib] neg_may_naf(baseKB:path(1, 1))
^ Fail: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(1, 1)), info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1459, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(1, 1)))))
Fail: (76) [pfc_lib] neg_in_code0(baseKB:path(1, 1))
^ Fail: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(1, 1)), info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1320, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(1, 1)))))
^ Fail: (68) [baseKB] ~ (baseKB:path(1, 1))
^ Call: (68) [must_sanity] must_sanity:mquietly_if(true, rtrace:tAt_normal)
^ Unify: (68) [must_sanity] must_sanity:mquietly_if(true, rtrace:tAt_normal)
failure=info((why_was_true(baseKB:(\+ ~path(1,1))),rtrace(baseKB: ~path(1,1))))
no_proof_for(\+ ~path(1,1)).
no_proof_for(\+ ~path(1,1)).
no_proof_for(\+ ~path(1,1)).
name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0004_Line_0000__path_1'.
JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'.
JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'.
% saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0004_Line_0000__path_1-junit.xml
~*/
:- mpred_test(~path(2,2)).
%~ mpred_test("Test_0005_Line_0000__path_2",baseKB: ~path(2,2))
/*~
%~ mpred_test("Test_0005_Line_0000__path_2",baseKB: ~path(2,2))
^ Call: (68) [baseKB] ~path(2, 2)
^ Unify: (68) [baseKB] ~ (baseKB:path(2, 2))
^ Call: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(2, 2)), info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1189, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(2, 2)))))
^ Unify: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(2, 2)), info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1189, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(2, 2)))))
Call: (76) [system] set_prolog_flag(last_call_optimisation, false)
Exit: (76) [system] set_prolog_flag(last_call_optimisation, false)
^ Call: (76) [loop_check] prolog_frame_attribute(1189, parent_goal, loop_check_term_frame(_121764, info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, _121770, _121772))
^ Fail: (76) [loop_check] prolog_frame_attribute(1189, parent_goal, loop_check_term_frame(_121764, info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, _121770, _121772))
^ Redo: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(2, 2)), info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1189, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(2, 2)))))
Call: (76) [pfc_lib] neg_in_code0(baseKB:path(2, 2))
Unify: (76) [pfc_lib] neg_in_code0(baseKB:path(2, 2))
^ Call: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(2, 2)), info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1328, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(2, 2)))))
^ Unify: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(2, 2)), info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1328, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(2, 2)))))
Call: (83) [system] set_prolog_flag(last_call_optimisation, false)
Exit: (83) [system] set_prolog_flag(last_call_optimisation, false)
^ Call: (83) [loop_check] prolog_frame_attribute(1328, parent_goal, loop_check_term_frame(_127488, info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, _127494, _127496))
^ Fail: (83) [loop_check] prolog_frame_attribute(1328, parent_goal, loop_check_term_frame(_127488, info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, _127494, _127496))
^ Redo: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(2, 2)), info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1328, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(2, 2)))))
Call: (83) [pfc_lib] neg_may_naf(baseKB:path(2, 2))
Unify: (83) [pfc_lib] neg_may_naf(baseKB:path(2, 2))
^ Call: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _130820)
^ Unify: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _130820)
^ Call: (88) [system] clause(pfc_lib:prologNegByFailure(path), true, _130820)
^ Fail: (88) [system] clause(pfc_lib:prologNegByFailure(path), true, _130820)
^ Fail: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _130820)
Unify: (83) [pfc_lib] neg_may_naf(baseKB:path(2, 2))
^ Call: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(2, 2))
^ Unify: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(2, 2))
^ Call: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(2, 2))
^ Unify: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(2, 2))
^ Fail: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(2, 2))
^ Redo: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(2, 2))
^ Exit: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(2, 2))
^ Call: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990), call(_138990)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))))
^ Unify: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990), call(_138990)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))))
^ Call: (90) [hook_database] clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990)
^ Fail: (90) [hook_database] clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990)
Call: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))
Unify: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))
^ Call: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776))
^ Unify: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776))
^ Call: (92) [baseKB] clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776)
^ Fail: (92) [baseKB] clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776)
^ Fail: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776))
Fail: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))
^ Fail: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990), call(_138990)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid))))
^ Call: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _147196)
^ Unify: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, syntaxic(_147766))
^ Call: (88) [pfc_lib] mpred_database_term_syntax(path, 2, _147766)
^ Fail: (88) [pfc_lib] mpred_database_term_syntax(path, 2, _147766)
^ Redo: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _149768)
^ Unify: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _150396)
^ Call: (88) [pfc_lib] mpred_core_database_term(path, 2, _150964)
^ Fail: (88) [pfc_lib] mpred_core_database_term(path, 2, _150964)
^ Fail: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _152338)
^ Call: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942), call(_152942)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))))
^ Unify: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942), call(_152942)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))))
^ Call: (88) [hook_database] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942)
^ Fail: (88) [hook_database] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942)
Call: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))
Unify: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))
^ Call: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728))
^ Unify: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728))
^ Call: (90) [baseKB] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728)
^ Fail: (90) [baseKB] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728)
^ Fail: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728))
Fail: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))
^ Fail: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942), call(_152942)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin))))
^ Call: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_161094, path, 2, prologHybrid), _161124), call(_161124)*->true;clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid))))
^ Unify: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_161094, path, 2, prologHybrid), _161124), call(_161124)*->true;clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid))))
^ Call: (91) [hook_database] clause(mpred_prop(_161094, path, 2, prologHybrid), _161124)
^ Fail: (91) [hook_database] clause(mpred_prop(_161094, path, 2, prologHybrid), _161124)
Call: (91) [hook_database] clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid))
Unify: (91) [hook_database] clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid))
^ Call: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_161094, path, 2, prologHybrid), _164910))
^ Unify: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_161094, path, 2, prologHybrid), _164910))
^ Call: (93) [baseKB] clause(mpred_prop(_161094, path, 2, prologHybrid), _164910)
^ Fail: (93) [baseKB] clause(mpred_prop(_161094, path, 2, prologHybrid), _164910)
^ Fail: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_161094, path, 2, prologHybrid), _164910))
Fail: (91) [hook_database] clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid))
^ Fail: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_161094, path, 2, prologHybrid), _161124), call(_161124)*->true;clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid))))
^ Call: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _169330)
^ Unify: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, syntaxic(_169900))
^ Call: (89) [pfc_lib] mpred_database_term_syntax(path, 2, _169900)
^ Fail: (89) [pfc_lib] mpred_database_term_syntax(path, 2, _169900)
^ Redo: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _171902)
^ Unify: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _172530)
^ Call: (89) [pfc_lib] mpred_core_database_term(path, 2, _173098)
^ Fail: (89) [pfc_lib] mpred_core_database_term(path, 2, _173098)
^ Fail: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _174472)
Call: (98) [$autoload] leave_sandbox(_175070)
Unify: (98) [$autoload] leave_sandbox(_175070)
Exit: (98) [$autoload] leave_sandbox(false)
Call: (97) [$autoload] restore_sandbox(false)
Unify: (97) [$autoload] restore_sandbox(false)
Exit: (97) [$autoload] restore_sandbox(false)
Fail: (83) [pfc_lib] neg_may_naf(baseKB:path(2, 2))
^ Fail: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(2, 2)), info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1328, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(2, 2)))))
Fail: (76) [pfc_lib] neg_in_code0(baseKB:path(2, 2))
^ Fail: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(2, 2)), info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1189, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(2, 2)))))
^ Fail: (68) [baseKB] ~ (baseKB:path(2, 2))
^ Call: (68) [must_sanity] must_sanity:mquietly_if(true, rtrace:tAt_normal)
^ Unify: (68) [must_sanity] must_sanity:mquietly_if(true, rtrace:tAt_normal)
failure=info((why_was_true(baseKB:(\+ ~path(2,2))),rtrace(baseKB: ~path(2,2))))
no_proof_for(\+ ~path(2,2)).
no_proof_for(\+ ~path(2,2)).
no_proof_for(\+ ~path(2,2)).
name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0005_Line_0000__path_2'.
JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'.
JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'.
% saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0005_Line_0000__path_2-junit.xml
~*/
/*
:- mpred_test(\+path(1,1)).
:- mpred_test(\+path(2,2)).
*/
:- mpred_test(\+path(3,1)).
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/neg_01v.pfc:32
%~ mpred_test("Test_0006_Line_0000__naf_path_3",baseKB:(\+path(3,1)))
/*~
%~ mpred_test("Test_0006_Line_0000__naf_path_3",baseKB:(\+path(3,1)))
passed=info(why_was_true(baseKB:(\+path(3,1))))
no_proof_for(\+path(3,1)).
no_proof_for(\+path(3,1)).
no_proof_for(\+path(3,1)).
name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0006_Line_0000__naf_path_3'.
JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'.
JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'.
% saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0006_Line_0000__naf_path_3-junit.xml
~*/
:- mpred_test(\+ ~path(1,5)).
%~ mpred_test("Test_0007_Line_0000__naf_path_1",baseKB:(\+ ~path(1,5)))
/*~
%~ mpred_test("Test_0007_Line_0000__naf_path_1",baseKB:(\+ ~path(1,5)))
passed=info(why_was_true(baseKB:(\+ ~path(1,5))))
no_proof_for(\+ ~path(1,5)).
no_proof_for(\+ ~path(1,5)).
no_proof_for(\+ ~path(1,5)).
name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0007_Line_0000__naf_path_1'.
JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'.
JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'.
% saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0007_Line_0000__naf_path_1-junit.xml
~*/
:- break.
%~ skipped(blocks_on_input,break)
%~ unused(save_junit_results)
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/neg_01v.pfc:36
%~ test_completed_exit(6)
:- dynamic junit_prop/3.
:- dynamic junit_prop/3.
:- dynamic junit_prop/3.
```
totalTime=1
ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ANEG_01V
GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3
https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/pfc/t/sanity_base/neg_01v.pfc
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/NEG_01V/logicmoo_pfc_test_sanity_base_NEG_01V_JUnit/
This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/67/testReport/logicmoo.pfc.test.sanity_base/NEG_01V/logicmoo_pfc_test_sanity_base_NEG_01V_JUnit/
GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3
https://github.com/logicmoo/logicmoo_workspace/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/pfc/t/sanity_base/neg_01v.pfc
FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k neg_01v.pfc (returned 6)
|
non_process
|
logicmoo pfc test sanity base neg junit cd var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base timeout foreground preserve status s sigkill k lmoo clif neg pfc gh master issue finfo issue search gitlab latest this build github running var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base neg pfc var lib jenkins local share swi prolog pack logicmoo utils prolog logicmoo test header pl this test might need use module library logicmoo plarkc use module library statistics mpred notrace exec reset runtime counter mpred notrace exec reset runtime counter statistics runtime secs path x x path path path path mpred test path mpred test test line path basekb path file mpred test test line path basekb path passed info why was true basekb path justifications for path basekb name logicmoo pfc test sanity base neg test line path junit classname logicmoo pfc test sanity base neg junit cmd timeout foreground preserve status s sigkill k lmoo clif neg pfc saving junit var lib jenkins workspace logicmoo workspace test results jenkins report logicmoo junit test sanity base units logicmoo pfc test sanity base neg test line path junit xml mpred test path mpred test test line path basekb path file mpred test test line path basekb path passed info why was true basekb path justifications for path basekb name logicmoo pfc test sanity base neg test line path junit classname logicmoo pfc test sanity base neg junit cmd timeout foreground preserve status s sigkill k lmoo clif neg pfc saving junit var lib jenkins workspace logicmoo workspace test results jenkins report logicmoo junit test sanity base units logicmoo pfc test sanity base neg test line path junit xml mpred test path mpred test test line path basekb path file mpred test test line path basekb path passed info why was true basekb path justifications for path basekb name logicmoo pfc test sanity base neg test line path junit classname logicmoo pfc test sanity base neg junit cmd timeout foreground preserve status s sigkill k lmoo clif neg pfc saving junit var lib jenkins workspace logicmoo workspace test results jenkins report logicmoo junit test sanity base units logicmoo pfc test sanity base neg test line path junit xml mpred test path mpred test test line path basekb path mpred test test line path basekb path call path unify basekb path call loop check loop check term frame pfc lib neg in basekb path info pfc lib neg in basekb path mpred core pl pfc lib trace or throw looped pfc lib neg in basekb path unify loop check loop check term frame pfc lib neg in basekb path info pfc lib neg in basekb path mpred core pl pfc lib trace or throw looped pfc lib neg in basekb path call set prolog flag last call optimisation false exit set prolog flag last call optimisation false call prolog frame attribute parent goal loop check term frame info pfc lib neg in basekb path mpred core pl fail prolog frame attribute parent goal loop check term frame info pfc lib neg in basekb path mpred core pl redo loop check loop check term frame pfc lib neg in basekb path info pfc lib neg in basekb path mpred core pl pfc lib trace or throw looped pfc lib neg in basekb path call neg in basekb path unify neg in basekb path call loop check loop check term frame pfc lib neg may naf basekb path info pfc lib neg may naf basekb path mpred core pl pfc lib trace or throw looped pfc lib neg may naf basekb path unify loop check loop check term frame pfc lib neg may naf basekb path info pfc lib neg may naf basekb path mpred core pl pfc lib trace or throw looped pfc lib neg may naf basekb path call set prolog flag last call optimisation false exit set prolog flag last call optimisation false call prolog frame attribute parent goal loop check term frame info pfc lib neg may naf basekb path mpred core pl fail prolog frame attribute parent goal loop check term frame info pfc lib neg may naf basekb path mpred core pl redo loop check loop check term frame pfc lib neg may naf basekb path info pfc lib neg may naf basekb path mpred core pl pfc lib trace or throw looped pfc lib neg may naf basekb path call neg may naf basekb path unify neg may naf basekb path call hook database clause i pfc lib prolognegbyfailure path true unify hook database clause i pfc lib prolognegbyfailure path true call clause pfc lib prolognegbyfailure path true fail clause pfc lib prolognegbyfailure path true fail hook database clause i pfc lib prolognegbyfailure path true unify neg may naf basekb path call ucatch is ftcompound basekb path unify ucatch is ftcompound basekb path call ucatch is ftvar basekb path unify ucatch is ftvar basekb path fail ucatch is ftvar basekb path redo ucatch is ftcompound basekb path exit ucatch is ftcompound basekb path call hook database pfc with quiet vars lock clause mpred prop basekb path prologhybrid call true clause b basekb mpred prop basekb path prologhybrid unify hook database pfc with quiet vars lock clause mpred prop basekb path prologhybrid call true clause b basekb mpred prop basekb path prologhybrid call clause mpred prop basekb path prologhybrid fail clause mpred prop basekb path prologhybrid call clause b basekb mpred prop basekb path prologhybrid unify clause b basekb mpred prop basekb path prologhybrid call hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologhybrid unify hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologhybrid call clause mpred prop basekb path prologhybrid fail clause mpred prop basekb path prologhybrid fail hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologhybrid fail clause b basekb mpred prop basekb path prologhybrid fail hook database pfc with quiet vars lock clause mpred prop basekb path prologhybrid call true clause b basekb mpred prop basekb path prologhybrid call basekb mpred database term path unify basekb mpred database term path syntaxic call mpred database term syntax path fail mpred database term syntax path redo basekb mpred database term path unify basekb mpred database term path call mpred core database term path fail mpred core database term path fail basekb mpred database term path call hook database pfc with quiet vars lock clause mpred prop basekb path prologbuiltin call true clause b basekb mpred prop basekb path prologbuiltin unify hook database pfc with quiet vars lock clause mpred prop basekb path prologbuiltin call true clause b basekb mpred prop basekb path prologbuiltin call clause mpred prop basekb path prologbuiltin fail clause mpred prop basekb path prologbuiltin call clause b basekb mpred prop basekb path prologbuiltin unify clause b basekb mpred prop basekb path prologbuiltin call hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologbuiltin unify hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologbuiltin call clause mpred prop basekb path prologbuiltin fail clause mpred prop basekb path prologbuiltin fail hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologbuiltin fail clause b basekb mpred prop basekb path prologbuiltin fail hook database pfc with quiet vars lock clause mpred prop basekb path prologbuiltin call true clause b basekb mpred prop basekb path prologbuiltin call hook database pfc with quiet vars lock clause mpred prop path prologhybrid call true clause b basekb mpred prop path prologhybrid unify hook database pfc with quiet vars lock clause mpred prop path prologhybrid call true clause b basekb mpred prop path prologhybrid call clause mpred prop path prologhybrid fail clause mpred prop path prologhybrid call clause b basekb mpred prop path prologhybrid unify clause b basekb mpred prop path prologhybrid call hook database pfc with quiet vars lock basekb clause mpred prop path prologhybrid unify hook database pfc with quiet vars lock basekb clause mpred prop path prologhybrid call clause mpred prop path prologhybrid fail clause mpred prop path prologhybrid fail hook database pfc with quiet vars lock basekb clause mpred prop path prologhybrid fail clause b basekb mpred prop path prologhybrid fail hook database pfc with quiet vars lock clause mpred prop path prologhybrid call true clause b basekb mpred prop path prologhybrid call basekb mpred database term path unify basekb mpred database term path syntaxic call mpred database term syntax path fail mpred database term syntax path redo basekb mpred database term path unify basekb mpred database term path call mpred core database term path fail mpred core database term path fail basekb mpred database term path call leave sandbox unify leave sandbox exit leave sandbox false call restore sandbox false unify restore sandbox false exit restore sandbox false fail neg may naf basekb path fail loop check loop check term frame pfc lib neg may naf basekb path info pfc lib neg may naf basekb path mpred core pl pfc lib trace or throw looped pfc lib neg may naf basekb path fail neg in basekb path fail loop check loop check term frame pfc lib neg in basekb path info pfc lib neg in basekb path mpred core pl pfc lib trace or throw looped pfc lib neg in basekb path fail basekb path call must sanity mquietly if true rtrace tat normal unify must sanity mquietly if true rtrace tat normal failure info why was true basekb path rtrace basekb path no proof for path no proof for path no proof for path name logicmoo pfc test sanity base neg test line path junit classname logicmoo pfc test sanity base neg junit cmd timeout foreground preserve status s sigkill k lmoo clif neg pfc saving junit var lib jenkins workspace logicmoo workspace test results jenkins report logicmoo junit test sanity base units logicmoo pfc test sanity base neg test line path junit xml mpred test path mpred test test line path basekb path mpred test test line path basekb path call path unify basekb path call loop check loop check term frame pfc lib neg in basekb path info pfc lib neg in basekb path mpred core pl pfc lib trace or throw looped pfc lib neg in basekb path unify loop check loop check term frame pfc lib neg in basekb path info pfc lib neg in basekb path mpred core pl pfc lib trace or throw looped pfc lib neg in basekb path call set prolog flag last call optimisation false exit set prolog flag last call optimisation false call prolog frame attribute parent goal loop check term frame info pfc lib neg in basekb path mpred core pl fail prolog frame attribute parent goal loop check term frame info pfc lib neg in basekb path mpred core pl redo loop check loop check term frame pfc lib neg in basekb path info pfc lib neg in basekb path mpred core pl pfc lib trace or throw looped pfc lib neg in basekb path call neg in basekb path unify neg in basekb path call loop check loop check term frame pfc lib neg may naf basekb path info pfc lib neg may naf basekb path mpred core pl pfc lib trace or throw looped pfc lib neg may naf basekb path unify loop check loop check term frame pfc lib neg may naf basekb path info pfc lib neg may naf basekb path mpred core pl pfc lib trace or throw looped pfc lib neg may naf basekb path call set prolog flag last call optimisation false exit set prolog flag last call optimisation false call prolog frame attribute parent goal loop check term frame info pfc lib neg may naf basekb path mpred core pl fail prolog frame attribute parent goal loop check term frame info pfc lib neg may naf basekb path mpred core pl redo loop check loop check term frame pfc lib neg may naf basekb path info pfc lib neg may naf basekb path mpred core pl pfc lib trace or throw looped pfc lib neg may naf basekb path call neg may naf basekb path unify neg may naf basekb path call hook database clause i pfc lib prolognegbyfailure path true unify hook database clause i pfc lib prolognegbyfailure path true call clause pfc lib prolognegbyfailure path true fail clause pfc lib prolognegbyfailure path true fail hook database clause i pfc lib prolognegbyfailure path true unify neg may naf basekb path call ucatch is ftcompound basekb path unify ucatch is ftcompound basekb path call ucatch is ftvar basekb path unify ucatch is ftvar basekb path fail ucatch is ftvar basekb path redo ucatch is ftcompound basekb path exit ucatch is ftcompound basekb path call hook database pfc with quiet vars lock clause mpred prop basekb path prologhybrid call true clause b basekb mpred prop basekb path prologhybrid unify hook database pfc with quiet vars lock clause mpred prop basekb path prologhybrid call true clause b basekb mpred prop basekb path prologhybrid call clause mpred prop basekb path prologhybrid fail clause mpred prop basekb path prologhybrid call clause b basekb mpred prop basekb path prologhybrid unify clause b basekb mpred prop basekb path prologhybrid call hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologhybrid unify hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologhybrid call clause mpred prop basekb path prologhybrid fail clause mpred prop basekb path prologhybrid fail hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologhybrid fail clause b basekb mpred prop basekb path prologhybrid fail hook database pfc with quiet vars lock clause mpred prop basekb path prologhybrid call true clause b basekb mpred prop basekb path prologhybrid call basekb mpred database term path unify basekb mpred database term path syntaxic call mpred database term syntax path fail mpred database term syntax path redo basekb mpred database term path unify basekb mpred database term path call mpred core database term path fail mpred core database term path fail basekb mpred database term path call hook database pfc with quiet vars lock clause mpred prop basekb path prologbuiltin call true clause b basekb mpred prop basekb path prologbuiltin unify hook database pfc with quiet vars lock clause mpred prop basekb path prologbuiltin call true clause b basekb mpred prop basekb path prologbuiltin call clause mpred prop basekb path prologbuiltin fail clause mpred prop basekb path prologbuiltin call clause b basekb mpred prop basekb path prologbuiltin unify clause b basekb mpred prop basekb path prologbuiltin call hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologbuiltin unify hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologbuiltin call clause mpred prop basekb path prologbuiltin fail clause mpred prop basekb path prologbuiltin fail hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologbuiltin fail clause b basekb mpred prop basekb path prologbuiltin fail hook database pfc with quiet vars lock clause mpred prop basekb path prologbuiltin call true clause b basekb mpred prop basekb path prologbuiltin call hook database pfc with quiet vars lock clause mpred prop path prologhybrid call true clause b basekb mpred prop path prologhybrid unify hook database pfc with quiet vars lock clause mpred prop path prologhybrid call true clause b basekb mpred prop path prologhybrid call clause mpred prop path prologhybrid fail clause mpred prop path prologhybrid call clause b basekb mpred prop path prologhybrid unify clause b basekb mpred prop path prologhybrid call hook database pfc with quiet vars lock basekb clause mpred prop path prologhybrid unify hook database pfc with quiet vars lock basekb clause mpred prop path prologhybrid call clause mpred prop path prologhybrid fail clause mpred prop path prologhybrid fail hook database pfc with quiet vars lock basekb clause mpred prop path prologhybrid fail clause b basekb mpred prop path prologhybrid fail hook database pfc with quiet vars lock clause mpred prop path prologhybrid call true clause b basekb mpred prop path prologhybrid call basekb mpred database term path unify basekb mpred database term path syntaxic call mpred database term syntax path fail mpred database term syntax path redo basekb mpred database term path unify basekb mpred database term path call mpred core database term path fail mpred core database term path fail basekb mpred database term path call leave sandbox unify leave sandbox exit leave sandbox false call restore sandbox false unify restore sandbox false exit restore sandbox false fail neg may naf basekb path fail loop check loop check term frame pfc lib neg may naf basekb path info pfc lib neg may naf basekb path mpred core pl pfc lib trace or throw looped pfc lib neg may naf basekb path fail neg in basekb path fail loop check loop check term frame pfc lib neg in basekb path info pfc lib neg in basekb path mpred core pl pfc lib trace or throw looped pfc lib neg in basekb path fail basekb path call must sanity mquietly if true rtrace tat normal unify must sanity mquietly if true rtrace tat normal failure info why was true basekb path rtrace basekb path no proof for path no proof for path no proof for path name logicmoo pfc test sanity base neg test line path junit classname logicmoo pfc test sanity base neg junit cmd timeout foreground preserve status s sigkill k lmoo clif neg pfc saving junit var lib jenkins workspace logicmoo workspace test results jenkins report logicmoo junit test sanity base units logicmoo pfc test sanity base neg test line path junit xml mpred test path mpred test path mpred test path var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base neg pfc mpred test test line naf path basekb path mpred test test line naf path basekb path passed info why was true basekb path no proof for path no proof for path no proof for path name logicmoo pfc test sanity base neg test line naf path junit classname logicmoo pfc test sanity base neg junit cmd timeout foreground preserve status s sigkill k lmoo clif neg pfc saving junit var lib jenkins workspace logicmoo workspace test results jenkins report logicmoo junit test sanity base units logicmoo pfc test sanity base neg test line naf path junit xml mpred test path mpred test test line naf path basekb path mpred test test line naf path basekb path passed info why was true basekb path no proof for path no proof for path no proof for path name logicmoo pfc test sanity base neg test line naf path junit classname logicmoo pfc test sanity base neg junit cmd timeout foreground preserve status s sigkill k lmoo clif neg pfc saving junit var lib jenkins workspace logicmoo workspace test results jenkins report logicmoo junit test sanity base units logicmoo pfc test sanity base neg test line naf path junit xml break skipped blocks on input break unused save junit results var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base neg pfc test completed exit dynamic junit prop dynamic junit prop dynamic junit prop totaltime issue search gitlab latest this build github failed var lib jenkins workspace logicmoo workspace bin lmoo junit minor k neg pfc returned
| 0
|
317
| 2,762,978,072
|
IssuesEvent
|
2015-04-29 04:42:05
|
sysown/proxysql-0.2
|
https://api.github.com/repos/sysown/proxysql-0.2
|
opened
|
Enabling fast_forward at runtime can cause crashes
|
ADMIN MYSQL PROTOCOL QUERY PROCESSOR
|
Enabling fast_forward at runtime can cause crashes if there are already established connections
|
1.0
|
Enabling fast_forward at runtime can cause crashes - Enabling fast_forward at runtime can cause crashes if there are already established connections
|
process
|
enabling fast forward at runtime can cause crashes enabling fast forward at runtime can cause crashes if there are already established connections
| 1
|
726
| 3,213,368,795
|
IssuesEvent
|
2015-10-06 19:34:09
|
nationalparkservice/places-data
|
https://api.github.com/repos/nationalparkservice/places-data
|
closed
|
Add highway=steps
|
pending-other-process schema
|
http://wiki.openstreetmap.org/wiki/Tag:highway%3Dsteps
There are a lot of staircases in parks, and this would be useful to have in our map
|
1.0
|
Add highway=steps - http://wiki.openstreetmap.org/wiki/Tag:highway%3Dsteps
There are a lot of staircases in parks, and this would be useful to have in our map
|
process
|
add highway steps there are a lot of staircases in parks and this would be useful to have in our map
| 1
|
692,133
| 23,723,871,713
|
IssuesEvent
|
2022-08-30 17:41:03
|
project-pareto/project-pareto
|
https://api.github.com/repos/project-pareto/project-pareto
|
closed
|
Network errors in toy_case_study cause occasional CI failures
|
bug Priority:Normal
|
- The error message seems to occur for requests to `router.project-osrm.org
- However, it's not immediately clear if this is due to rate limiting enforced by that server (which is triggered by multiple CI jobs running in parallel), or other network issues
## Traceback
From https://github.com/project-pareto/project-pareto/runs/6984942337?check_suite_focus=true
```txt
Traceback (most recent call last):
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connection.py", line 174, in _new_conn
conn = connection.create_connection(
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\util\connection.py", line 72, in create_connection
for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM):
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\socket.py", line 954, in getaddrinfo
for res in _socket.getaddrinfo(host, port, family, type, proto, flags):
socket.gaierror: [Errno 11001] getaddrinfo failed
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connectionpool.py", line 703, in urlopen
httplib_response = self._make_request(
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connectionpool.py", line 386, in _make_request
self._validate_conn(conn)
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connectionpool.py", line 1040, in _validate_conn
conn.connect()
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connection.py", line 358, in connect
self.sock = conn = self._new_conn()
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connection.py", line 186, in _new_conn
raise NewConnectionError(
urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x0000013C6C5A44C0>: Failed to establish a new connection: [Errno 11001] getaddrinfo failed
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\adapters.py", line 489, in send
resp = conn.urlopen(
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connectionpool.py", line 785, in urlopen
retries = retries.increment(
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\util\retry.py", line 592, in increment
raise MaxRetryError(_pool, url, error or ResponseError(cause))
urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='router.project-osrm.org', port=443): Max retries exceeded with url: /table/v1/driving/-122.36685609717435,47.62390651082141;-122.39902392056128,47.544942555956496;-122.2875801822827,47.54628135125307;-122.2901468966058,47.59922966761389;-122.3345,47.6044;-122.38514810411333,47.6785514001581;-122.31941699831951,47.671605016754775;-122.37256636119979,47.51373173275276;-122.30833430642372,47.5982109494788;-122.27578180785954,47.54501307370556;-122.25392487570376,47.53198073978911;-122.23615203748196,47.56473303789118;-122.38490267916536,47.52979263294095;-122.43178036972577,47.63168538306254;-122.3748467760494,47.66220329642665;-122.33832847843796,47.650576946946856;-122.24953147273567,47.655834380297506;-122.27548462416144,47.68078709171784;-122.41679135415569,47.53282219470006;-122.29739693081503,47.63214139655833;-122.39872536852423,47.606697845554585;-122.26887716808142,47.61641535939208;-122.27286360239516,47.60037158057717;-122.27032458981431,47.53461603738988;-122.36685609717435,47.62390651082141;-122.39902392056128,47.544942555956496;-122.2875801822827,47.54628135125307;-122.2901468966058,47.59922966761389;-122.3345,47.6044;-122.38514810411333,47.6785514001581;-122.31941699831951,47.671605016754775;-122.37256636119979,47.51373173275276;-122.30833430642372,47.5982109494788;-122.27578180785954,47.54501307370556;-122.25392487570376,47.53198073978911;-122.23615203748196,47.56473303789118;-122.38490267916536,47.52979263294095;-122.43178036972577,47.63168538306254;-122.3748467760494,47.66220329642665;-122.33832847843796,47.650576946946856;-122.24953147273567,47.655834380297506;-122.27548462416144,47.68078709171784;-122.41679135415569,47.53282219470006;-122.29739693081503,47.63214139655833;-122.39872536852423,47.606697845554585;-122.26887716808142,47.61641535939208;-122.27286360239516,47.60037158057717;-122.27032458981431,47.53461603738988?sources=0;1;2;3;4;5;6;7;8;9;10;11;12;13;14;15;16;17;18;19;20;21;22;23&destinations=24;25;26;27;28;29;30;31;32;33;34;35;36;37;38;39;40;41;42;43;44;45;46;47&annotations=duration,distance (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x0000013C6C5A44C0>: Failed to establish a new connection: [Errno 11001] getaddrinfo failed'))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\runpy.py", line 197, in _run_module_as_main
return _run_code(code, main_globals, None,
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\runpy.py", line 87, in _run_code
exec(code, run_globals)
File "D:\a\project-pareto\project-pareto\pareto\case_studies\toy_case_study.py", line 203, in <module>
df_parameters["DriveTimes"] = od_matrix(od_matrix_input)
File "D:\a\project-pareto\project-pareto\pareto\utilities\get_data.py", line 462, in od_matrix
response = requests.get(
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\api.py", line 73, in get
return request("get", url, params=params, **kwargs)
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\api.py", line 59, in request
return session.request(method=method, url=url, **kwargs)
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\sessions.py", line 587, in request
resp = self.send(prep, **send_kwargs)
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\sessions.py", line 701, in send
r = adapter.send(request, **kwargs)
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\adapters.py", line 565, in send
raise ConnectionError(e, request=request)
requests.exceptions.ConnectionError: HTTPSConnectionPool(host='router.project-osrm.org', port=443): Max retries exceeded with url: /table/v1/driving/-122.36685609717435,47.62390651082141;-122.39902392056128,47.544942555956496;-122.2875801822827,47.54628135125307;-122.2901468966058,47.59922966761389;-122.3345,47.6044;-122.38514810411333,47.6785514001581;-122.31941699831951,47.671605016754775;-122.37256636119979,47.51373173275276;-122.30833430642372,47.5982109494788;-122.27578180785954,47.54501307370556;-122.25392487570376,47.53198073978911;-122.23615203748196,47.56473303789118;-122.38490267916536,47.52979263294095;-122.43178036972577,47.63168538306254;-122.3748467760494,47.66220329642665;-122.33832847843796,47.650576946946856;-122.24953147273567,47.655834380297506;-122.27548462416144,47.68078709171784;-122.41679135415569,47.53282219470006;-122.29739693081503,47.63214139655833;-122.39872536852423,47.606697845554585;-122.26887716808142,47.61641535939208;-122.27286360239516,47.60037158057717;-122.27032458981431,47.53461603738988;-122.36685609717435,47.62390651082141;-122.39902392056128,47.544942555956496;-122.2875801822827,47.54628135125307;-122.2901468966058,47.59922966761389;-122.3345,47.6044;-122.38514810411333,47.6785514001581;-122.31941699831951,47.671605016754775;-122.37256636119979,47.51373173275276;-122.30833430642372,47.5982109494788;-122.27578180785954,47.54501307370556;-122.25392487570376,47.53198073978911;-122.23615203748196,47.56473303789118;-122.38490267916536,47.52979263294095;-122.43178036972577,47.63168538306254;-122.3748467760494,47.66220329642665;-122.33832847843796,47.650576946946856;-122.24953147273567,47.655834380297506;-122.27548462416144,47.68078709171784;-122.41679135415569,47.53282219470006;-122.29739693081503,47.63214139655833;-122.39872536852423,47.606697845554585;-122.26887716808142,47.61641535939208;-122.27286360239516,47.60037158057717;-122.27032458981431,47.53461603738988?sources=0;1;2;3;4;5;6;7;8;9;10;11;12;13;14;15;16;17;18;19;20;21;22;23&destinations=24;25;26;27;28;29;30;31;32;33;34;35;36;37;38;39;40;41;42;43;44;45;46;47&annotations=duration,distance (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x0000013C6C5A44C0>: Failed to establish a new connection: [Errno 11001] getaddrinfo failed'))
```
|
1.0
|
Network errors in toy_case_study cause occasional CI failures - - The error message seems to occur for requests to `router.project-osrm.org
- However, it's not immediately clear if this is due to rate limiting enforced by that server (which is triggered by multiple CI jobs running in parallel), or other network issues
## Traceback
From https://github.com/project-pareto/project-pareto/runs/6984942337?check_suite_focus=true
```txt
Traceback (most recent call last):
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connection.py", line 174, in _new_conn
conn = connection.create_connection(
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\util\connection.py", line 72, in create_connection
for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM):
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\socket.py", line 954, in getaddrinfo
for res in _socket.getaddrinfo(host, port, family, type, proto, flags):
socket.gaierror: [Errno 11001] getaddrinfo failed
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connectionpool.py", line 703, in urlopen
httplib_response = self._make_request(
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connectionpool.py", line 386, in _make_request
self._validate_conn(conn)
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connectionpool.py", line 1040, in _validate_conn
conn.connect()
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connection.py", line 358, in connect
self.sock = conn = self._new_conn()
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connection.py", line 186, in _new_conn
raise NewConnectionError(
urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x0000013C6C5A44C0>: Failed to establish a new connection: [Errno 11001] getaddrinfo failed
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\adapters.py", line 489, in send
resp = conn.urlopen(
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connectionpool.py", line 785, in urlopen
retries = retries.increment(
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\util\retry.py", line 592, in increment
raise MaxRetryError(_pool, url, error or ResponseError(cause))
urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='router.project-osrm.org', port=443): Max retries exceeded with url: /table/v1/driving/-122.36685609717435,47.62390651082141;-122.39902392056128,47.544942555956496;-122.2875801822827,47.54628135125307;-122.2901468966058,47.59922966761389;-122.3345,47.6044;-122.38514810411333,47.6785514001581;-122.31941699831951,47.671605016754775;-122.37256636119979,47.51373173275276;-122.30833430642372,47.5982109494788;-122.27578180785954,47.54501307370556;-122.25392487570376,47.53198073978911;-122.23615203748196,47.56473303789118;-122.38490267916536,47.52979263294095;-122.43178036972577,47.63168538306254;-122.3748467760494,47.66220329642665;-122.33832847843796,47.650576946946856;-122.24953147273567,47.655834380297506;-122.27548462416144,47.68078709171784;-122.41679135415569,47.53282219470006;-122.29739693081503,47.63214139655833;-122.39872536852423,47.606697845554585;-122.26887716808142,47.61641535939208;-122.27286360239516,47.60037158057717;-122.27032458981431,47.53461603738988;-122.36685609717435,47.62390651082141;-122.39902392056128,47.544942555956496;-122.2875801822827,47.54628135125307;-122.2901468966058,47.59922966761389;-122.3345,47.6044;-122.38514810411333,47.6785514001581;-122.31941699831951,47.671605016754775;-122.37256636119979,47.51373173275276;-122.30833430642372,47.5982109494788;-122.27578180785954,47.54501307370556;-122.25392487570376,47.53198073978911;-122.23615203748196,47.56473303789118;-122.38490267916536,47.52979263294095;-122.43178036972577,47.63168538306254;-122.3748467760494,47.66220329642665;-122.33832847843796,47.650576946946856;-122.24953147273567,47.655834380297506;-122.27548462416144,47.68078709171784;-122.41679135415569,47.53282219470006;-122.29739693081503,47.63214139655833;-122.39872536852423,47.606697845554585;-122.26887716808142,47.61641535939208;-122.27286360239516,47.60037158057717;-122.27032458981431,47.53461603738988?sources=0;1;2;3;4;5;6;7;8;9;10;11;12;13;14;15;16;17;18;19;20;21;22;23&destinations=24;25;26;27;28;29;30;31;32;33;34;35;36;37;38;39;40;41;42;43;44;45;46;47&annotations=duration,distance (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x0000013C6C5A44C0>: Failed to establish a new connection: [Errno 11001] getaddrinfo failed'))
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\runpy.py", line 197, in _run_module_as_main
return _run_code(code, main_globals, None,
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\runpy.py", line 87, in _run_code
exec(code, run_globals)
File "D:\a\project-pareto\project-pareto\pareto\case_studies\toy_case_study.py", line 203, in <module>
df_parameters["DriveTimes"] = od_matrix(od_matrix_input)
File "D:\a\project-pareto\project-pareto\pareto\utilities\get_data.py", line 462, in od_matrix
response = requests.get(
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\api.py", line 73, in get
return request("get", url, params=params, **kwargs)
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\api.py", line 59, in request
return session.request(method=method, url=url, **kwargs)
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\sessions.py", line 587, in request
resp = self.send(prep, **send_kwargs)
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\sessions.py", line 701, in send
r = adapter.send(request, **kwargs)
File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\adapters.py", line 565, in send
raise ConnectionError(e, request=request)
requests.exceptions.ConnectionError: HTTPSConnectionPool(host='router.project-osrm.org', port=443): Max retries exceeded with url: /table/v1/driving/-122.36685609717435,47.62390651082141;-122.39902392056128,47.544942555956496;-122.2875801822827,47.54628135125307;-122.2901468966058,47.59922966761389;-122.3345,47.6044;-122.38514810411333,47.6785514001581;-122.31941699831951,47.671605016754775;-122.37256636119979,47.51373173275276;-122.30833430642372,47.5982109494788;-122.27578180785954,47.54501307370556;-122.25392487570376,47.53198073978911;-122.23615203748196,47.56473303789118;-122.38490267916536,47.52979263294095;-122.43178036972577,47.63168538306254;-122.3748467760494,47.66220329642665;-122.33832847843796,47.650576946946856;-122.24953147273567,47.655834380297506;-122.27548462416144,47.68078709171784;-122.41679135415569,47.53282219470006;-122.29739693081503,47.63214139655833;-122.39872536852423,47.606697845554585;-122.26887716808142,47.61641535939208;-122.27286360239516,47.60037158057717;-122.27032458981431,47.53461603738988;-122.36685609717435,47.62390651082141;-122.39902392056128,47.544942555956496;-122.2875801822827,47.54628135125307;-122.2901468966058,47.59922966761389;-122.3345,47.6044;-122.38514810411333,47.6785514001581;-122.31941699831951,47.671605016754775;-122.37256636119979,47.51373173275276;-122.30833430642372,47.5982109494788;-122.27578180785954,47.54501307370556;-122.25392487570376,47.53198073978911;-122.23615203748196,47.56473303789118;-122.38490267916536,47.52979263294095;-122.43178036972577,47.63168538306254;-122.3748467760494,47.66220329642665;-122.33832847843796,47.650576946946856;-122.24953147273567,47.655834380297506;-122.27548462416144,47.68078709171784;-122.41679135415569,47.53282219470006;-122.29739693081503,47.63214139655833;-122.39872536852423,47.606697845554585;-122.26887716808142,47.61641535939208;-122.27286360239516,47.60037158057717;-122.27032458981431,47.53461603738988?sources=0;1;2;3;4;5;6;7;8;9;10;11;12;13;14;15;16;17;18;19;20;21;22;23&destinations=24;25;26;27;28;29;30;31;32;33;34;35;36;37;38;39;40;41;42;43;44;45;46;47&annotations=duration,distance (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x0000013C6C5A44C0>: Failed to establish a new connection: [Errno 11001] getaddrinfo failed'))
```
|
non_process
|
network errors in toy case study cause occasional ci failures the error message seems to occur for requests to router project osrm org however it s not immediately clear if this is due to rate limiting enforced by that server which is triggered by multiple ci jobs running in parallel or other network issues traceback from txt traceback most recent call last file c hostedtoolcache windows python lib site packages connection py line in new conn conn connection create connection file c hostedtoolcache windows python lib site packages util connection py line in create connection for res in socket getaddrinfo host port family socket sock stream file c hostedtoolcache windows python lib socket py line in getaddrinfo for res in socket getaddrinfo host port family type proto flags socket gaierror getaddrinfo failed during handling of the above exception another exception occurred traceback most recent call last file c hostedtoolcache windows python lib site packages connectionpool py line in urlopen httplib response self make request file c hostedtoolcache windows python lib site packages connectionpool py line in make request self validate conn conn file c hostedtoolcache windows python lib site packages connectionpool py line in validate conn conn connect file c hostedtoolcache windows python lib site packages connection py line in connect self sock conn self new conn file c hostedtoolcache windows python lib site packages connection py line in new conn raise newconnectionerror exceptions newconnectionerror failed to establish a new connection getaddrinfo failed during handling of the above exception another exception occurred traceback most recent call last file c hostedtoolcache windows python lib site packages requests adapters py line in send resp conn urlopen file c hostedtoolcache windows python lib site packages connectionpool py line in urlopen retries retries increment file c hostedtoolcache windows python lib site packages util retry py line in increment raise maxretryerror pool url error or responseerror cause exceptions maxretryerror httpsconnectionpool host router project osrm org port max retries exceeded with url table driving sources destinations annotations duration distance caused by newconnectionerror failed to establish a new connection getaddrinfo failed during handling of the above exception another exception occurred traceback most recent call last file c hostedtoolcache windows python lib runpy py line in run module as main return run code code main globals none file c hostedtoolcache windows python lib runpy py line in run code exec code run globals file d a project pareto project pareto pareto case studies toy case study py line in df parameters od matrix od matrix input file d a project pareto project pareto pareto utilities get data py line in od matrix response requests get file c hostedtoolcache windows python lib site packages requests api py line in get return request get url params params kwargs file c hostedtoolcache windows python lib site packages requests api py line in request return session request method method url url kwargs file c hostedtoolcache windows python lib site packages requests sessions py line in request resp self send prep send kwargs file c hostedtoolcache windows python lib site packages requests sessions py line in send r adapter send request kwargs file c hostedtoolcache windows python lib site packages requests adapters py line in send raise connectionerror e request request requests exceptions connectionerror httpsconnectionpool host router project osrm org port max retries exceeded with url table driving sources destinations annotations duration distance caused by newconnectionerror failed to establish a new connection getaddrinfo failed
| 0
|
93,947
| 15,946,447,026
|
IssuesEvent
|
2021-04-15 01:05:29
|
jgeraigery/core
|
https://api.github.com/repos/jgeraigery/core
|
opened
|
CVE-2021-25122 (High) detected in tomcat-embed-core-8.5.34.jar
|
security vulnerability
|
## CVE-2021-25122 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.34.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p>
<p>Path to dependency file: core/nimbus-entity-dsl/pom.xml</p>
<p>Path to vulnerable library: core/8.5.34/tomcat-embed-core-8.5.34.jar,core/8.5.34/tomcat-embed-core-8.5.34.jar,core/8.5.34/tomcat-embed-core-8.5.34.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.0.5.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-2.0.5.RELEASE.jar
- :x: **tomcat-embed-core-8.5.34.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When responding to new h2c connection requests, Apache Tomcat versions 10.0.0-M1 to 10.0.0, 9.0.0.M1 to 9.0.41 and 8.5.0 to 8.5.61 could duplicate request headers and a limited amount of request body from one request to another meaning user A and user B could both see the results of user A's request.
<p>Publish Date: 2021-03-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25122>CVE-2021-25122</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://lists.apache.org/thread.html/r7b95bc248603360501f18c8eb03bb6001ec0ee3296205b34b07105b7%40%3Cannounce.tomcat.apache.org%3E">https://lists.apache.org/thread.html/r7b95bc248603360501f18c8eb03bb6001ec0ee3296205b34b07105b7%40%3Cannounce.tomcat.apache.org%3E</a></p>
<p>Release Date: 2021-03-01</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:8.5.62,9.0.42,10.0.2;org.apache.tomcat:tomcat-coyote:8.5.62,9.0.42,10.0.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"8.5.34","packageFilePaths":["/nimbus-entity-dsl/pom.xml","/nimbus-test/pom.xml","/nimbus-core/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.0.5.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:2.0.5.RELEASE;org.apache.tomcat.embed:tomcat-embed-core:8.5.34","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:8.5.62,9.0.42,10.0.2;org.apache.tomcat:tomcat-coyote:8.5.62,9.0.42,10.0.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-25122","vulnerabilityDetails":"When responding to new h2c connection requests, Apache Tomcat versions 10.0.0-M1 to 10.0.0, 9.0.0.M1 to 9.0.41 and 8.5.0 to 8.5.61 could duplicate request headers and a limited amount of request body from one request to another meaning user A and user B could both see the results of user A\u0027s request.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25122","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-25122 (High) detected in tomcat-embed-core-8.5.34.jar - ## CVE-2021-25122 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.34.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p>
<p>Path to dependency file: core/nimbus-entity-dsl/pom.xml</p>
<p>Path to vulnerable library: core/8.5.34/tomcat-embed-core-8.5.34.jar,core/8.5.34/tomcat-embed-core-8.5.34.jar,core/8.5.34/tomcat-embed-core-8.5.34.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.0.5.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-2.0.5.RELEASE.jar
- :x: **tomcat-embed-core-8.5.34.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When responding to new h2c connection requests, Apache Tomcat versions 10.0.0-M1 to 10.0.0, 9.0.0.M1 to 9.0.41 and 8.5.0 to 8.5.61 could duplicate request headers and a limited amount of request body from one request to another meaning user A and user B could both see the results of user A's request.
<p>Publish Date: 2021-03-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25122>CVE-2021-25122</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://lists.apache.org/thread.html/r7b95bc248603360501f18c8eb03bb6001ec0ee3296205b34b07105b7%40%3Cannounce.tomcat.apache.org%3E">https://lists.apache.org/thread.html/r7b95bc248603360501f18c8eb03bb6001ec0ee3296205b34b07105b7%40%3Cannounce.tomcat.apache.org%3E</a></p>
<p>Release Date: 2021-03-01</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:8.5.62,9.0.42,10.0.2;org.apache.tomcat:tomcat-coyote:8.5.62,9.0.42,10.0.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"8.5.34","packageFilePaths":["/nimbus-entity-dsl/pom.xml","/nimbus-test/pom.xml","/nimbus-core/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.0.5.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:2.0.5.RELEASE;org.apache.tomcat.embed:tomcat-embed-core:8.5.34","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:8.5.62,9.0.42,10.0.2;org.apache.tomcat:tomcat-coyote:8.5.62,9.0.42,10.0.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-25122","vulnerabilityDetails":"When responding to new h2c connection requests, Apache Tomcat versions 10.0.0-M1 to 10.0.0, 9.0.0.M1 to 9.0.41 and 8.5.0 to 8.5.61 could duplicate request headers and a limited amount of request body from one request to another meaning user A and user B could both see the results of user A\u0027s request.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25122","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in tomcat embed core jar cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation library home page a href path to dependency file core nimbus entity dsl pom xml path to vulnerable library core tomcat embed core jar core tomcat embed core jar core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library found in base branch master vulnerability details when responding to new connection requests apache tomcat versions to to and to could duplicate request headers and a limited amount of request body from one request to another meaning user a and user b could both see the results of user a s request publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat coyote isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org springframework boot spring boot starter web release org springframework boot spring boot starter tomcat release org apache tomcat embed tomcat embed core isminimumfixversionavailable true minimumfixversion org apache tomcat embed tomcat embed core org apache tomcat tomcat coyote basebranches vulnerabilityidentifier cve vulnerabilitydetails when responding to new connection requests apache tomcat versions to to and to could duplicate request headers and a limited amount of request body from one request to another meaning user a and user b could both see the results of user a request vulnerabilityurl
| 0
|
6,452
| 9,546,480,214
|
IssuesEvent
|
2019-05-01 20:04:17
|
openopps/openopps-platform
|
https://api.github.com/repos/openopps/openopps-platform
|
closed
|
Department of State: Experience Questions- Add character limit
|
Apply Process Approved Requirements Ready State Dept.
|
Who: Student
What:Character Account
Why: As an applicant I would like to know the character limit to a text box
A/C
- On the Experience & References Page Under the "Other" selection
- the "If you chose "other", please specify - provide a character limit (250 characters remaining)
- Please indicate the total length of your overseas experience(s) provide a character limit (250 characters remaining)
- The character limit will change as the user types in the box and reflect the number of remaining characters.
|
1.0
|
Department of State: Experience Questions- Add character limit - Who: Student
What:Character Account
Why: As an applicant I would like to know the character limit to a text box
A/C
- On the Experience & References Page Under the "Other" selection
- the "If you chose "other", please specify - provide a character limit (250 characters remaining)
- Please indicate the total length of your overseas experience(s) provide a character limit (250 characters remaining)
- The character limit will change as the user types in the box and reflect the number of remaining characters.
|
process
|
department of state experience questions add character limit who student what character account why as an applicant i would like to know the character limit to a text box a c on the experience references page under the other selection the if you chose other please specify provide a character limit characters remaining please indicate the total length of your overseas experience s provide a character limit characters remaining the character limit will change as the user types in the box and reflect the number of remaining characters
| 1
|
14,310
| 17,316,261,846
|
IssuesEvent
|
2021-07-27 06:36:44
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
System.Diagnostics.Process.Tests fails on iOS/tvOS
|
area-System.Diagnostics.Process in pr os-ios os-tvos
|
System.Diagnostics.Process.Tests
System.Diagnostics.Tests.ProcessStartInfoTests.TestEnvironmentVariablesPropertyUnix
System.Diagnostics.Tests.ProcessStartInfoTests.TestEnvironmentProperty
System.Diagnostics.Tests.ProcessTests.MaxWorkingSet_GetNotStarted_ThrowsInvalidOperationException
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes
System.Diagnostics.Tests.ProcessTests.Kill_NotStarted_ThrowsInvalidOperationException
System.Diagnostics.Tests.ProcessTests.ProcessStart_TryOpenFolder_UseShellExecuteIsFalse_ThrowsWin32Exception
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_WorkingDirectory
System.Diagnostics.Tests.ProcessTests.TestStartOnUnixWithBadPermissions
System.Diagnostics.Tests.ProcessTests.ProcessNameMatchesScriptName
System.Diagnostics.Tests.ProcessTests.TestStartWithMissingFile
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecuteFalse_FilenameIsUrl_ThrowsWin32Exception
System.Diagnostics.Tests.ProcessTests.TestStartOnUnixWithBadFormat
System.Diagnostics.Tests.ProcessTests.MinWorkingSet_GetNotStarted_ThrowsInvalidOperationException
System.Diagnostics.Tests.ProcessTests.ProcessStart_TryExitCommandAsFileName_ThrowsWin32Exception
```
System.Diagnostics.Process.Tests.dll Failed: 16
Test collection for System.Diagnostics.Process.Tests.dll
System.Diagnostics.Tests.ProcessStartInfoTests.TestEnvironmentVariablesPropertyUnix
Assert.NotEqual() Failure\nExpected: Not 0\nActual: 0
System.Diagnostics.Tests.ProcessStartInfoTests.TestEnvironmentProperty
Assert.NotEqual() Failure\nExpected: Not 0\nActual: 0
System.Diagnostics.Tests.ProcessTests.MaxWorkingSet_GetNotStarted_ThrowsInvalidOperationException
Assert.Throws() Failure\nExpected: typeof(System.InvalidOperationException)\nActual: typeof(System.PlatformNotSupportedException): Getting or setting the working set limits on other processes is not supported on this platform.\n---- System.PlatformNotSupportedException : Getting or setting the working set limits on other processes is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes(filenameAsUrl: False)
System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes(filenameAsUrl: True)
System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.Kill_NotStarted_ThrowsInvalidOperationException
Assert.Throws() Failure\nExpected: typeof(System.InvalidOperationException)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.ProcessStart_TryOpenFolder_UseShellExecuteIsFalse_ThrowsWin32Exception
Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_WorkingDirectory
System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.TestStartOnUnixWithBadPermissions
Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.ProcessNameMatchesScriptName
System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.TestStartWithMissingFile(fullPath: True)
Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.TestStartWithMissingFile(fullPath: False)
Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecuteFalse_FilenameIsUrl_ThrowsWin32Exception
Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.TestStartOnUnixWithBadFormat
Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.MinWorkingSet_GetNotStarted_ThrowsInvalidOperationException
Assert.Throws() Failure\nExpected: typeof(System.InvalidOperationException)\nActual: typeof(System.PlatformNotSupportedException): Getting or setting the working set limits on other processes is not supported on this platform.\n---- System.PlatformNotSupportedException : Getting or setting the working set limits on other processes is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.ProcessStart_TryExitCommandAsFileName_ThrowsWin32Exception
Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
|
1.0
|
System.Diagnostics.Process.Tests fails on iOS/tvOS - System.Diagnostics.Process.Tests
System.Diagnostics.Tests.ProcessStartInfoTests.TestEnvironmentVariablesPropertyUnix
System.Diagnostics.Tests.ProcessStartInfoTests.TestEnvironmentProperty
System.Diagnostics.Tests.ProcessTests.MaxWorkingSet_GetNotStarted_ThrowsInvalidOperationException
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes
System.Diagnostics.Tests.ProcessTests.Kill_NotStarted_ThrowsInvalidOperationException
System.Diagnostics.Tests.ProcessTests.ProcessStart_TryOpenFolder_UseShellExecuteIsFalse_ThrowsWin32Exception
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_WorkingDirectory
System.Diagnostics.Tests.ProcessTests.TestStartOnUnixWithBadPermissions
System.Diagnostics.Tests.ProcessTests.ProcessNameMatchesScriptName
System.Diagnostics.Tests.ProcessTests.TestStartWithMissingFile
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecuteFalse_FilenameIsUrl_ThrowsWin32Exception
System.Diagnostics.Tests.ProcessTests.TestStartOnUnixWithBadFormat
System.Diagnostics.Tests.ProcessTests.MinWorkingSet_GetNotStarted_ThrowsInvalidOperationException
System.Diagnostics.Tests.ProcessTests.ProcessStart_TryExitCommandAsFileName_ThrowsWin32Exception
```
System.Diagnostics.Process.Tests.dll Failed: 16
Test collection for System.Diagnostics.Process.Tests.dll
System.Diagnostics.Tests.ProcessStartInfoTests.TestEnvironmentVariablesPropertyUnix
Assert.NotEqual() Failure\nExpected: Not 0\nActual: 0
System.Diagnostics.Tests.ProcessStartInfoTests.TestEnvironmentProperty
Assert.NotEqual() Failure\nExpected: Not 0\nActual: 0
System.Diagnostics.Tests.ProcessTests.MaxWorkingSet_GetNotStarted_ThrowsInvalidOperationException
Assert.Throws() Failure\nExpected: typeof(System.InvalidOperationException)\nActual: typeof(System.PlatformNotSupportedException): Getting or setting the working set limits on other processes is not supported on this platform.\n---- System.PlatformNotSupportedException : Getting or setting the working set limits on other processes is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes(filenameAsUrl: False)
System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes(filenameAsUrl: True)
System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.Kill_NotStarted_ThrowsInvalidOperationException
Assert.Throws() Failure\nExpected: typeof(System.InvalidOperationException)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.ProcessStart_TryOpenFolder_UseShellExecuteIsFalse_ThrowsWin32Exception
Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_WorkingDirectory
System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.TestStartOnUnixWithBadPermissions
Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.ProcessNameMatchesScriptName
System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.TestStartWithMissingFile(fullPath: True)
Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.TestStartWithMissingFile(fullPath: False)
Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecuteFalse_FilenameIsUrl_ThrowsWin32Exception
Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.TestStartOnUnixWithBadFormat
Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.MinWorkingSet_GetNotStarted_ThrowsInvalidOperationException
Assert.Throws() Failure\nExpected: typeof(System.InvalidOperationException)\nActual: typeof(System.PlatformNotSupportedException): Getting or setting the working set limits on other processes is not supported on this platform.\n---- System.PlatformNotSupportedException : Getting or setting the working set limits on other processes is not supported on this platform.
System.Diagnostics.Tests.ProcessTests.ProcessStart_TryExitCommandAsFileName_ThrowsWin32Exception
Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
|
process
|
system diagnostics process tests fails on ios tvos system diagnostics process tests system diagnostics tests processstartinfotests testenvironmentvariablespropertyunix system diagnostics tests processstartinfotests testenvironmentproperty system diagnostics tests processtests maxworkingset getnotstarted throwsinvalidoperationexception system diagnostics tests processtests processstart useshellexecute executes system diagnostics tests processtests kill notstarted throwsinvalidoperationexception system diagnostics tests processtests processstart tryopenfolder useshellexecuteisfalse system diagnostics tests processtests processstart useshellexecute workingdirectory system diagnostics tests processtests teststartonunixwithbadpermissions system diagnostics tests processtests processnamematchesscriptname system diagnostics tests processtests teststartwithmissingfile system diagnostics tests processtests processstart useshellexecutefalse filenameisurl system diagnostics tests processtests teststartonunixwithbadformat system diagnostics tests processtests minworkingset getnotstarted throwsinvalidoperationexception system diagnostics tests processtests processstart tryexitcommandasfilename system diagnostics process tests dll failed test collection for system diagnostics process tests dll system diagnostics tests processstartinfotests testenvironmentvariablespropertyunix assert notequal failure nexpected not nactual system diagnostics tests processstartinfotests testenvironmentproperty assert notequal failure nexpected not nactual system diagnostics tests processtests maxworkingset getnotstarted throwsinvalidoperationexception assert throws failure nexpected typeof system invalidoperationexception nactual typeof system platformnotsupportedexception getting or setting the working set limits on other processes is not supported on this platform n system platformnotsupportedexception getting or setting the working set limits on other processes is not supported on this platform system diagnostics tests processtests processstart useshellexecute executes filenameasurl false system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests processstart useshellexecute executes filenameasurl true system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests kill notstarted throwsinvalidoperationexception assert throws failure nexpected typeof system invalidoperationexception nactual typeof system platformnotsupportedexception operation is not supported on this platform n system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests processstart tryopenfolder useshellexecuteisfalse assert throws failure nexpected typeof system componentmodel nactual typeof system platformnotsupportedexception operation is not supported on this platform n system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests processstart useshellexecute workingdirectory system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests teststartonunixwithbadpermissions assert throws failure nexpected typeof system componentmodel nactual typeof system platformnotsupportedexception operation is not supported on this platform n system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests processnamematchesscriptname system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests teststartwithmissingfile fullpath true assert throws failure nexpected typeof system componentmodel nactual typeof system platformnotsupportedexception operation is not supported on this platform n system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests teststartwithmissingfile fullpath false assert throws failure nexpected typeof system componentmodel nactual typeof system platformnotsupportedexception operation is not supported on this platform n system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests processstart useshellexecutefalse filenameisurl assert throws failure nexpected typeof system componentmodel nactual typeof system platformnotsupportedexception operation is not supported on this platform n system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests teststartonunixwithbadformat assert throws failure nexpected typeof system componentmodel nactual typeof system platformnotsupportedexception operation is not supported on this platform n system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests minworkingset getnotstarted throwsinvalidoperationexception assert throws failure nexpected typeof system invalidoperationexception nactual typeof system platformnotsupportedexception getting or setting the working set limits on other processes is not supported on this platform n system platformnotsupportedexception getting or setting the working set limits on other processes is not supported on this platform system diagnostics tests processtests processstart tryexitcommandasfilename assert throws failure nexpected typeof system componentmodel nactual typeof system platformnotsupportedexception operation is not supported on this platform n system platformnotsupportedexception operation is not supported on this platform
| 1
|
56,611
| 32,076,419,892
|
IssuesEvent
|
2023-09-25 11:19:19
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
opt: slower plans chosen due to poor stats for TPC-H queries
|
C-performance A-sql-optimizer X-stale no-issue-activity
|
With join reordering, slower plans are chose for TPC-H queries 16 and 20. Q16 goes from ~600ms to ~900ms, and Q20 goes from ~13s to ~18s.
Q16 fast-plan cost: `1090503.24`
Q16 slow-plan cost: `342084.224`
Q20 false-plan cost: `5790840.88`
Q20 slow-plan cost: `5782416.86`
Both queries have a `LIKE` filter for which selectivity is severely underestimated; this is because the default conjunct selectivity estimate of 1/3 is used for `LIKE` filters. This leads to row-count overestimation.
Q20 has an additional problem with stats estimation; selectivity for the `l_partkey:31 = ps_partkey:14 AND l_suppkey:32 = ps_suppkey:15` join filter is overestimated (as opposed to the `LIKE` filter, the selectivity of which is underestimated). This leads to row-count underestimation.
Jira issue: CRDB-3976
|
True
|
opt: slower plans chosen due to poor stats for TPC-H queries - With join reordering, slower plans are chose for TPC-H queries 16 and 20. Q16 goes from ~600ms to ~900ms, and Q20 goes from ~13s to ~18s.
Q16 fast-plan cost: `1090503.24`
Q16 slow-plan cost: `342084.224`
Q20 false-plan cost: `5790840.88`
Q20 slow-plan cost: `5782416.86`
Both queries have a `LIKE` filter for which selectivity is severely underestimated; this is because the default conjunct selectivity estimate of 1/3 is used for `LIKE` filters. This leads to row-count overestimation.
Q20 has an additional problem with stats estimation; selectivity for the `l_partkey:31 = ps_partkey:14 AND l_suppkey:32 = ps_suppkey:15` join filter is overestimated (as opposed to the `LIKE` filter, the selectivity of which is underestimated). This leads to row-count underestimation.
Jira issue: CRDB-3976
|
non_process
|
opt slower plans chosen due to poor stats for tpc h queries with join reordering slower plans are chose for tpc h queries and goes from to and goes from to fast plan cost slow plan cost false plan cost slow plan cost both queries have a like filter for which selectivity is severely underestimated this is because the default conjunct selectivity estimate of is used for like filters this leads to row count overestimation has an additional problem with stats estimation selectivity for the l partkey ps partkey and l suppkey ps suppkey join filter is overestimated as opposed to the like filter the selectivity of which is underestimated this leads to row count underestimation jira issue crdb
| 0
|
75,919
| 7,496,217,559
|
IssuesEvent
|
2018-04-08 07:00:31
|
MajkiIT/polish-ads-filter
|
https://api.github.com/repos/MajkiIT/polish-ads-filter
|
closed
|
hrubie.pl
|
reguły gotowe/testowanie reklama
|
Ad empty container
### Zrzut ekranu

### Link bezpośredni
http://www.hrubie.pl/
### Co trzeba zrobić, aby pojawił się element, reklama albo błąd?
1. Visit the page
### Moja konfiguracja
Win 7 SP1 [64bit]
Firefox 59.0.2 [64bit]
uBlock Origin 1.15.24
### Filtry:
<details>
<summary>Click to expand</summary>
Nano Defender 13.48
Nano Defender filter
Nano Base filter
Nano Whitelist filter
uBlock filters
uBlock filters – Annoyances
uBlock filters – Badware risks
uBlock filters – Privacy
uBlock filters – Resource abuse
uBlock filters – Unbreak
Adguard’s Annoyance List
Adguard Base Filters
Adguard Spyware Filters
Adblock & uBlock polish filter - AdGuard supplement
Adblock Warning Removal List
EasyList
EasyList Polish
EasyPrivacy
Fanboy’s Annoyance List
Fanboy’s Anti-Thirdparty Social (see warning inside list)
Fanboy’s Cookiemonster List
Fanboy’s Enhanced Tracking List
Fanboy's Polish
Fanboy’s Social Blocking List
I'm OK with cookies
I don't care about cookies
hpHosts’ Ad and tracking servers
Peter Lowe’s Ad and tracking server list
POL: polskie filtry do Adblocka i uBlocka
Polish Privacy Filters
Polskie Filtry Anty-Donacyjne
Polskie Filtry Ciasteczkowe
Polskie Filtry Elementów Irytujących
Polskie Filtry RSS
Polskie Filtry Społecznościowe
Polskie Filtry Wewnętrzne
AlleBlock
KAD - Przekręty
Web Annoyances Ultralist
</details>
|
1.0
|
hrubie.pl - Ad empty container
### Zrzut ekranu

### Link bezpośredni
http://www.hrubie.pl/
### Co trzeba zrobić, aby pojawił się element, reklama albo błąd?
1. Visit the page
### Moja konfiguracja
Win 7 SP1 [64bit]
Firefox 59.0.2 [64bit]
uBlock Origin 1.15.24
### Filtry:
<details>
<summary>Click to expand</summary>
Nano Defender 13.48
Nano Defender filter
Nano Base filter
Nano Whitelist filter
uBlock filters
uBlock filters – Annoyances
uBlock filters – Badware risks
uBlock filters – Privacy
uBlock filters – Resource abuse
uBlock filters – Unbreak
Adguard’s Annoyance List
Adguard Base Filters
Adguard Spyware Filters
Adblock & uBlock polish filter - AdGuard supplement
Adblock Warning Removal List
EasyList
EasyList Polish
EasyPrivacy
Fanboy’s Annoyance List
Fanboy’s Anti-Thirdparty Social (see warning inside list)
Fanboy’s Cookiemonster List
Fanboy’s Enhanced Tracking List
Fanboy's Polish
Fanboy’s Social Blocking List
I'm OK with cookies
I don't care about cookies
hpHosts’ Ad and tracking servers
Peter Lowe’s Ad and tracking server list
POL: polskie filtry do Adblocka i uBlocka
Polish Privacy Filters
Polskie Filtry Anty-Donacyjne
Polskie Filtry Ciasteczkowe
Polskie Filtry Elementów Irytujących
Polskie Filtry RSS
Polskie Filtry Społecznościowe
Polskie Filtry Wewnętrzne
AlleBlock
KAD - Przekręty
Web Annoyances Ultralist
</details>
|
non_process
|
hrubie pl ad empty container zrzut ekranu link bezpośredni co trzeba zrobić aby pojawił się element reklama albo błąd visit the page moja konfiguracja win firefox ublock origin filtry click to expand nano defender nano defender filter nano base filter nano whitelist filter ublock filters ublock filters – annoyances ublock filters – badware risks ublock filters – privacy ublock filters – resource abuse ublock filters – unbreak adguard’s annoyance list adguard base filters adguard spyware filters adblock ublock polish filter adguard supplement adblock warning removal list easylist easylist polish easyprivacy fanboy’s annoyance list fanboy’s anti thirdparty social see warning inside list fanboy’s cookiemonster list fanboy’s enhanced tracking list fanboy s polish fanboy’s social blocking list i m ok with cookies i don t care about cookies hphosts’ ad and tracking servers peter lowe’s ad and tracking server list pol polskie filtry do adblocka i ublocka polish privacy filters polskie filtry anty donacyjne polskie filtry ciasteczkowe polskie filtry elementów irytujących polskie filtry rss polskie filtry społecznościowe polskie filtry wewnętrzne alleblock kad przekręty web annoyances ultralist
| 0
|
186,125
| 15,047,976,016
|
IssuesEvent
|
2021-02-03 09:37:01
|
NotOnlyStudents/Documentazione
|
https://api.github.com/repos/NotOnlyStudents/Documentazione
|
closed
|
UC3 Come funzionalità e non implementazione
|
adr bug documentation
|
Tullio: "UC3 non rappresenta una funzionalità, ma è già calato in una possibile implementazione che prevede l’uso di “menù”. I sotto-casi individuati soffrono del medesimo problema."
|
1.0
|
UC3 Come funzionalità e non implementazione - Tullio: "UC3 non rappresenta una funzionalità, ma è già calato in una possibile implementazione che prevede l’uso di “menù”. I sotto-casi individuati soffrono del medesimo problema."
|
non_process
|
come funzionalità e non implementazione tullio non rappresenta una funzionalità ma è già calato in una possibile implementazione che prevede l’uso di “menù” i sotto casi individuati soffrono del medesimo problema
| 0
|
330,609
| 10,053,250,311
|
IssuesEvent
|
2019-07-21 15:14:02
|
ticket721/web-app
|
https://api.github.com/repos/ticket721/web-app
|
closed
|
d.feat: generic price / balance displayer
|
[priority] [➖➖➖➖] critical [status] to do [type] feat
|
Find a way to display value in $ -> hover shows ether and DAI value, highlights fixed currency
|
1.0
|
d.feat: generic price / balance displayer - Find a way to display value in $ -> hover shows ether and DAI value, highlights fixed currency
|
non_process
|
d feat generic price balance displayer find a way to display value in hover shows ether and dai value highlights fixed currency
| 0
|
10,748
| 7,300,697,978
|
IssuesEvent
|
2018-02-27 01:00:30
|
dotnet/coreclr
|
https://api.github.com/repos/dotnet/coreclr
|
opened
|
[Perf] Investigate CscBench Regression between release/2.0.0 and release/2.1
|
area-Benchmarks tenet-performance
|
We saw a regression of 5% between 2.0 and 2.1
|
True
|
[Perf] Investigate CscBench Regression between release/2.0.0 and release/2.1 - We saw a regression of 5% between 2.0 and 2.1
|
non_process
|
investigate cscbench regression between release and release we saw a regression of between and
| 0
|
1,068
| 3,536,075,422
|
IssuesEvent
|
2016-01-17 00:28:31
|
MaretEngineering/MROV
|
https://api.github.com/repos/MaretEngineering/MROV
|
closed
|
Add constants to replace numbers
|
Arduino enhancement Processing
|
There should be almost no numbers left on their own in the main body of the program. They should all be stored in a constant with a good name to add clarity.
|
1.0
|
Add constants to replace numbers - There should be almost no numbers left on their own in the main body of the program. They should all be stored in a constant with a good name to add clarity.
|
process
|
add constants to replace numbers there should be almost no numbers left on their own in the main body of the program they should all be stored in a constant with a good name to add clarity
| 1
|
826,352
| 31,592,092,172
|
IssuesEvent
|
2023-09-05 00:03:33
|
Selody-project/Backend
|
https://api.github.com/repos/Selody-project/Backend
|
closed
|
[BACK-TASK] 개인 일정 공유 그룹 or 비공유 그룹
|
Priority 2
|
### 설명
- 공유 그룹인 경우에는 기존과 같은 방식으로 개인 일정을 사용
- 일정 비공유 그룹인 경우에는 해당 유저의 개인 일정을 그룹에서 사용하지 않도록 수정
|
1.0
|
[BACK-TASK] 개인 일정 공유 그룹 or 비공유 그룹 - ### 설명
- 공유 그룹인 경우에는 기존과 같은 방식으로 개인 일정을 사용
- 일정 비공유 그룹인 경우에는 해당 유저의 개인 일정을 그룹에서 사용하지 않도록 수정
|
non_process
|
개인 일정 공유 그룹 or 비공유 그룹 설명 공유 그룹인 경우에는 기존과 같은 방식으로 개인 일정을 사용 일정 비공유 그룹인 경우에는 해당 유저의 개인 일정을 그룹에서 사용하지 않도록 수정
| 0
|
17,030
| 22,406,824,612
|
IssuesEvent
|
2022-06-18 04:46:58
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
[metricstransformprocessor] Summary datapoint merging is not handled.
|
bug proc: metricstransformprocessor
|
**Describe the bug**
In `datapoint_aggregation.go`, the `MetricDescriptor_SUMMARY` type is not handled.
**Steps to reproduce**
1. Get a Summary metric into the pipeline. For example, the `http.server.requests` metric emitted from Spring Boot + `micrometer-registry-statsd` is a "timer" type, and becomes a "summary" when received by the `statsdreceiver`, with this config:
```
statsd:
timer_histogram_mapping:
- statsd_type: histogram
observer_type: summary
- statsd_type: timer
observer_type: summary
```
2. Apply a `metricstransform` processor with a "sum" aggregation type to the metric, such as with this config:
```
metricstransform:
transforms:
- include: http.server.requests
action: update
operations:
- action: aggregate_labels
label_set: [ uri ]
aggregation_type: sum
```
**What did you expect to see?**
I expect the labels of the `http.server.requests` metric besides `uri` to be removed, and the summaries for those datapoints to be aggregated.
**What did you see instead?**
The `http.server.requests` metric is dropped entirely.
**What version did you use?**
Version: aws-opentelemetry-collector v0.15
**Additional context**
The `datapoint_aggregation.go` file already has logic for aggregating histogram datapoints, but is missing logic for summary datapoints (specifically, it has logic for `MetricDescriptor_GAUGE_DISTRIBUTION`, but not `MetricDescriptor_SUMMARY`)
|
1.0
|
[metricstransformprocessor] Summary datapoint merging is not handled. - **Describe the bug**
In `datapoint_aggregation.go`, the `MetricDescriptor_SUMMARY` type is not handled.
**Steps to reproduce**
1. Get a Summary metric into the pipeline. For example, the `http.server.requests` metric emitted from Spring Boot + `micrometer-registry-statsd` is a "timer" type, and becomes a "summary" when received by the `statsdreceiver`, with this config:
```
statsd:
timer_histogram_mapping:
- statsd_type: histogram
observer_type: summary
- statsd_type: timer
observer_type: summary
```
2. Apply a `metricstransform` processor with a "sum" aggregation type to the metric, such as with this config:
```
metricstransform:
transforms:
- include: http.server.requests
action: update
operations:
- action: aggregate_labels
label_set: [ uri ]
aggregation_type: sum
```
**What did you expect to see?**
I expect the labels of the `http.server.requests` metric besides `uri` to be removed, and the summaries for those datapoints to be aggregated.
**What did you see instead?**
The `http.server.requests` metric is dropped entirely.
**What version did you use?**
Version: aws-opentelemetry-collector v0.15
**Additional context**
The `datapoint_aggregation.go` file already has logic for aggregating histogram datapoints, but is missing logic for summary datapoints (specifically, it has logic for `MetricDescriptor_GAUGE_DISTRIBUTION`, but not `MetricDescriptor_SUMMARY`)
|
process
|
summary datapoint merging is not handled describe the bug in datapoint aggregation go the metricdescriptor summary type is not handled steps to reproduce get a summary metric into the pipeline for example the http server requests metric emitted from spring boot micrometer registry statsd is a timer type and becomes a summary when received by the statsdreceiver with this config statsd timer histogram mapping statsd type histogram observer type summary statsd type timer observer type summary apply a metricstransform processor with a sum aggregation type to the metric such as with this config metricstransform transforms include http server requests action update operations action aggregate labels label set aggregation type sum what did you expect to see i expect the labels of the http server requests metric besides uri to be removed and the summaries for those datapoints to be aggregated what did you see instead the http server requests metric is dropped entirely what version did you use version aws opentelemetry collector additional context the datapoint aggregation go file already has logic for aggregating histogram datapoints but is missing logic for summary datapoints specifically it has logic for metricdescriptor gauge distribution but not metricdescriptor summary
| 1
|
18,546
| 24,555,302,205
|
IssuesEvent
|
2022-10-12 15:25:41
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[iOS] Formstep > text choice > user is able to select multiple answer options even though 'Single select' option is selected in the SB
|
Bug P1 iOS Process: Fixed Process: Tested dev
|
Steps:
1. In SB, create a study with a 'Text choice' response type by selecting single select option
2. Launch the study
3. Sign up or sign in to the mobile app
4. Enroll to the study
5. Click on text choice activity
6. Try to select multiple answer options
AR: Participant is able to select multiple answer options
ER: Participant should not able to select multiple answer options when the 'single select' option is selected in the SB
SB:

ER:

|
2.0
|
[iOS] Formstep > text choice > user is able to select multiple answer options even though 'Single select' option is selected in the SB - Steps:
1. In SB, create a study with a 'Text choice' response type by selecting single select option
2. Launch the study
3. Sign up or sign in to the mobile app
4. Enroll to the study
5. Click on text choice activity
6. Try to select multiple answer options
AR: Participant is able to select multiple answer options
ER: Participant should not able to select multiple answer options when the 'single select' option is selected in the SB
SB:

ER:

|
process
|
formstep text choice user is able to select multiple answer options even though single select option is selected in the sb steps in sb create a study with a text choice response type by selecting single select option launch the study sign up or sign in to the mobile app enroll to the study click on text choice activity try to select multiple answer options ar participant is able to select multiple answer options er participant should not able to select multiple answer options when the single select option is selected in the sb sb er
| 1
|
597,295
| 18,160,590,165
|
IssuesEvent
|
2021-09-27 09:09:38
|
CaptureCoop/SnipSniper
|
https://api.github.com/repos/CaptureCoop/SnipSniper
|
opened
|
ConfigWindow -> Better tabs
|
enhancement Medium Priority
|
Make each tab its own class for cleaner code.
Also implement that when you close the window it prompts you if window is dirty.
Also ask the same when switching from a dirty tab to another one.
|
1.0
|
ConfigWindow -> Better tabs - Make each tab its own class for cleaner code.
Also implement that when you close the window it prompts you if window is dirty.
Also ask the same when switching from a dirty tab to another one.
|
non_process
|
configwindow better tabs make each tab its own class for cleaner code also implement that when you close the window it prompts you if window is dirty also ask the same when switching from a dirty tab to another one
| 0
|
334,063
| 10,135,605,353
|
IssuesEvent
|
2019-08-02 10:35:31
|
aiidateam/aiida-quantumespresso
|
https://api.github.com/repos/aiidateam/aiida-quantumespresso
|
closed
|
Import error in `BasePwCpInputGenerator.prepare_for_submission`
|
priority/important topic/calculations type/bug
|
```
also_bands = settings.pop('ALSO_BANDS', None)
if also_bands:
import logging
from aiida.common.log import get_dblogger_extra
```
|
1.0
|
Import error in `BasePwCpInputGenerator.prepare_for_submission` - ```
also_bands = settings.pop('ALSO_BANDS', None)
if also_bands:
import logging
from aiida.common.log import get_dblogger_extra
```
|
non_process
|
import error in basepwcpinputgenerator prepare for submission also bands settings pop also bands none if also bands import logging from aiida common log import get dblogger extra
| 0
|
707,642
| 24,312,757,552
|
IssuesEvent
|
2022-09-30 01:16:13
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.amazon.co.jp - site is not usable
|
priority-important status-needsinfo-karlcow browser-fenix engine-gecko
|
<!-- @browser: Firefox Mobile 107.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 12; Mobile; rv:107.0) Gecko/107.0 Firefox/107.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/111454 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.amazon.co.jp/dp/B09P83BLHW?binding=kindle_edition&ref=dbs_dp_rwt_sb_pc_tukn
**Browser / Version**: Firefox Mobile 107.0
**Operating System**: Android 12
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Buttons or links not working
**Steps to Reproduce**:
I can not purchase by "1-Click to buy now"(1-clickで今すぐ買う)button.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/9/d362e4c9-aabb-4b21-86e7-44fc0a572f94.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220926093803</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/9/225da95d-fd1f-4939-91ff-03ff4fc2cd01)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.amazon.co.jp - site is not usable - <!-- @browser: Firefox Mobile 107.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 12; Mobile; rv:107.0) Gecko/107.0 Firefox/107.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/111454 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.amazon.co.jp/dp/B09P83BLHW?binding=kindle_edition&ref=dbs_dp_rwt_sb_pc_tukn
**Browser / Version**: Firefox Mobile 107.0
**Operating System**: Android 12
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Buttons or links not working
**Steps to Reproduce**:
I can not purchase by "1-Click to buy now"(1-clickで今すぐ買う)button.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/9/d362e4c9-aabb-4b21-86e7-44fc0a572f94.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220926093803</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/9/225da95d-fd1f-4939-91ff-03ff4fc2cd01)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
site is not usable url browser version firefox mobile operating system android tested another browser yes chrome problem type site is not usable description buttons or links not working steps to reproduce i can not purchase by click to buy now ( clickで今すぐ買う)button view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
159,791
| 25,046,457,495
|
IssuesEvent
|
2022-11-05 10:07:34
|
hypha-dao/dho-web-client
|
https://api.github.com/repos/hypha-dao/dho-web-client
|
closed
|
Proposal Cards should always be the same height
|
Bug Design
|
Reproduce:
1. Go to https://dao.hypha.earth/hypha/proposals/
2. Check height of cards:
<img width="666" alt="image" src="https://user-images.githubusercontent.com/75991832/192541781-8d579d85-c6aa-4251-83b7-a2c70550f31b.png">
Note: probably because 1 card has a comment?
|
1.0
|
Proposal Cards should always be the same height - Reproduce:
1. Go to https://dao.hypha.earth/hypha/proposals/
2. Check height of cards:
<img width="666" alt="image" src="https://user-images.githubusercontent.com/75991832/192541781-8d579d85-c6aa-4251-83b7-a2c70550f31b.png">
Note: probably because 1 card has a comment?
|
non_process
|
proposal cards should always be the same height reproduce go to check height of cards img width alt image src note probably because card has a comment
| 0
|
71,645
| 18,819,987,248
|
IssuesEvent
|
2021-11-10 06:55:04
|
appsmithorg/appsmith
|
https://api.github.com/repos/appsmithorg/appsmith
|
closed
|
[Bug]: Toast should be having a max width
|
Bug High Release UI Building Pod regression UI Builders Pod hotfix
|
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
Toast messages are too lengthy

### Steps To Reproduce
Use `showAlert` to show a large message
### Environment
Release
### Version
Self-Hosted
|
2.0
|
[Bug]: Toast should be having a max width - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
Toast messages are too lengthy

### Steps To Reproduce
Use `showAlert` to show a large message
### Environment
Release
### Version
Self-Hosted
|
non_process
|
toast should be having a max width is there an existing issue for this i have searched the existing issues current behavior toast messages are too lengthy steps to reproduce use showalert to show a large message environment release version self hosted
| 0
|
53,295
| 13,795,765,869
|
IssuesEvent
|
2020-10-09 18:36:13
|
fsjs-academy/aroa-ui
|
https://api.github.com/repos/fsjs-academy/aroa-ui
|
closed
|
CVE-2012-6708 (Medium) detected in jquery-1.8.1.min.js
|
security vulnerability
|
## CVE-2012-6708 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p>
<p>Path to dependency file: aroa-ui/node_modules/redeyed/examples/browser/index.html</p>
<p>Path to vulnerable library: aroa-ui/node_modules/redeyed/examples/browser/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.8.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fsjs-academy/aroa-ui/commit/86ab73fa8c8dcf1da44f79b5fa99aebd57adcd1d">86ab73fa8c8dcf1da44f79b5fa99aebd57adcd1d</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v1.9.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2012-6708 (Medium) detected in jquery-1.8.1.min.js - ## CVE-2012-6708 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p>
<p>Path to dependency file: aroa-ui/node_modules/redeyed/examples/browser/index.html</p>
<p>Path to vulnerable library: aroa-ui/node_modules/redeyed/examples/browser/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.8.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fsjs-academy/aroa-ui/commit/86ab73fa8c8dcf1da44f79b5fa99aebd57adcd1d">86ab73fa8c8dcf1da44f79b5fa99aebd57adcd1d</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v1.9.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file aroa ui node modules redeyed examples browser index html path to vulnerable library aroa ui node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch main vulnerability details jquery before is vulnerable to cross site scripting xss attacks the jquery strinput function does not differentiate selectors from html in a reliable fashion in vulnerable versions jquery determined whether the input was html by looking for the character anywhere in the string giving attackers more flexibility when attempting to construct a malicious payload in fixed versions jquery only deems the input to be html if it explicitly starts with the character limiting exploitability only to attackers who can control the beginning of a string which is far less common publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
| 0
|
9,319
| 12,338,228,672
|
IssuesEvent
|
2020-05-14 16:07:51
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Runtime parameters - out of date banner
|
Pri1 devops-cicd-process/tech devops/prod doc-bug
|
On the page with runtime parameters there is this banner:

I think this banner is out of date right now. This feature is probably available to almost all organizations.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 790318bb-8220-3241-4ca7-73351074492f
* Version Independent ID: db1da9db-3694-779b-17aa-1ed67fcecf86
* Content: [Use runtime and type-safe parameters - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/runtime-parameters?view=azure-devops&tabs=script)
* Content Source: [docs/pipelines/process/runtime-parameters.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/runtime-parameters.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
Runtime parameters - out of date banner - On the page with runtime parameters there is this banner:

I think this banner is out of date right now. This feature is probably available to almost all organizations.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 790318bb-8220-3241-4ca7-73351074492f
* Version Independent ID: db1da9db-3694-779b-17aa-1ed67fcecf86
* Content: [Use runtime and type-safe parameters - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/runtime-parameters?view=azure-devops&tabs=script)
* Content Source: [docs/pipelines/process/runtime-parameters.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/runtime-parameters.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
runtime parameters out of date banner on the page with runtime parameters there is this banner i think this banner is out of date right now this feature is probably available to almost all organizations document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.