Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
22,254
30,802,636,546
IssuesEvent
2023-08-01 03:34:55
emily-writes-poems/emily-writes-poems-processing
https://api.github.com/repos/emily-writes-poems/emily-writes-poems-processing
closed
display linked poems
processing refinement
in poems table. maybe as a modal? can also redesign the link poem option to be in this modal...?
1.0
display linked poems - in poems table. maybe as a modal? can also redesign the link poem option to be in this modal...?
process
display linked poems in poems table maybe as a modal can also redesign the link poem option to be in this modal
1
4,324
7,229,376,901
IssuesEvent
2018-02-11 19:14:49
pwittchen/ReactiveNetwork
https://api.github.com/repos/pwittchen/ReactiveNetwork
closed
Release 0.12.3 (RxJava1.x)
RxJava1.x release process
**Initial release notes**: - bumped RxJava: 1.3.3 -> 1.3.5 - updated Gradle Build Tools: 3.0.0 -> 3.0.1 - added script for publishing JavaDoc on gh-pages - added script for publishing documentation on gh-pages **Things to do**: - [x] update JavaDoc on `gh-pages` - [x] update documentation on `gh-pages` - [x] bump library version - [x] upload archives to Maven Central - [x] close and release artifact on Maven Central - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` - [x] update docs on gh-pages after updating `README.md` - [x] create new GitHub release
1.0
Release 0.12.3 (RxJava1.x) - **Initial release notes**: - bumped RxJava: 1.3.3 -> 1.3.5 - updated Gradle Build Tools: 3.0.0 -> 3.0.1 - added script for publishing JavaDoc on gh-pages - added script for publishing documentation on gh-pages **Things to do**: - [x] update JavaDoc on `gh-pages` - [x] update documentation on `gh-pages` - [x] bump library version - [x] upload archives to Maven Central - [x] close and release artifact on Maven Central - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` - [x] update docs on gh-pages after updating `README.md` - [x] create new GitHub release
process
release x initial release notes bumped rxjava updated gradle build tools added script for publishing javadoc on gh pages added script for publishing documentation on gh pages things to do update javadoc on gh pages update documentation on gh pages bump library version upload archives to maven central close and release artifact on maven central update changelog md after maven sync bump library version in readme md update docs on gh pages after updating readme md create new github release
1
182,036
30,782,689,347
IssuesEvent
2023-07-31 11:08:42
DeveloperAcademy-POSTECH/MC3-G5T15-DreamEgg
https://api.github.com/repos/DeveloperAcademy-POSTECH/MC3-G5T15-DreamEgg
opened
[Feat] UI Design 다듬기
🎨 Design
### 📝 작업 목적 / Miro / Figma <!-- 기능 혹은 버그에 대한 설명 --> 디벨롭하거나 디자인 변경이 안된 부분을 반영하기 --- ### 🛠️ Tasks * [ ] line 1 * [ ] line 2
1.0
[Feat] UI Design 다듬기 - ### 📝 작업 목적 / Miro / Figma <!-- 기능 혹은 버그에 대한 설명 --> 디벨롭하거나 디자인 변경이 안된 부분을 반영하기 --- ### 🛠️ Tasks * [ ] line 1 * [ ] line 2
non_process
ui design 다듬기 📝 작업 목적 miro figma 디벨롭하거나 디자인 변경이 안된 부분을 반영하기 🛠️ tasks line line
0
17,411
23,228,628,390
IssuesEvent
2022-08-03 04:44:37
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
Allow copying equivalent Python command directly from Processing algorithm dialog, without running algorithm first (Request in QGIS)
Processing 3.24
### Request for documentation From pull request QGIS/qgis#46499 Author: @nyalldawson QGIS version: 3.24 **Allow copying equivalent Python command directly from Processing algorithm dialog, without running algorithm first** ### PR Description: Adds an advanced button to the toolbox algorithm dialog, which contains an option to copy the equivalent Python command as the parameters define in the dialog. While this command is also available from the history dialog, the advanced button provides a way for users to generate these commands WITHOUT actually having to run the algorithm in advance. This menu will be extended with additional items (such as copy parameter as json, paste parameters, copy as qgis_process command) in follow up PRs. Sponsored by the Research Institute for Nature and Forest, Flemish Govt ### Commits tagged with [need-docs] or [FEATURE] "[feature][processing] Add an advanced button to the toolbox algorithm\ndialog, which contains an option to copy the equivalent Python\ncommand as the parameters define in the dialog\n\nWhile this command is also available from the history dialog, the\nadvanced button provides a way for users to generate these\ncommands WITHOUT actually having to run the algorithm in advance.\n\nSponsored by the Research Institute for Nature and Forest, Flemish Govt"
1.0
Allow copying equivalent Python command directly from Processing algorithm dialog, without running algorithm first (Request in QGIS) - ### Request for documentation From pull request QGIS/qgis#46499 Author: @nyalldawson QGIS version: 3.24 **Allow copying equivalent Python command directly from Processing algorithm dialog, without running algorithm first** ### PR Description: Adds an advanced button to the toolbox algorithm dialog, which contains an option to copy the equivalent Python command as the parameters define in the dialog. While this command is also available from the history dialog, the advanced button provides a way for users to generate these commands WITHOUT actually having to run the algorithm in advance. This menu will be extended with additional items (such as copy parameter as json, paste parameters, copy as qgis_process command) in follow up PRs. Sponsored by the Research Institute for Nature and Forest, Flemish Govt ### Commits tagged with [need-docs] or [FEATURE] "[feature][processing] Add an advanced button to the toolbox algorithm\ndialog, which contains an option to copy the equivalent Python\ncommand as the parameters define in the dialog\n\nWhile this command is also available from the history dialog, the\nadvanced button provides a way for users to generate these\ncommands WITHOUT actually having to run the algorithm in advance.\n\nSponsored by the Research Institute for Nature and Forest, Flemish Govt"
process
allow copying equivalent python command directly from processing algorithm dialog without running algorithm first request in qgis request for documentation from pull request qgis qgis author nyalldawson qgis version allow copying equivalent python command directly from processing algorithm dialog without running algorithm first pr description adds an advanced button to the toolbox algorithm dialog which contains an option to copy the equivalent python command as the parameters define in the dialog while this command is also available from the history dialog the advanced button provides a way for users to generate these commands without actually having to run the algorithm in advance this menu will be extended with additional items such as copy parameter as json paste parameters copy as qgis process command in follow up prs sponsored by the research institute for nature and forest flemish govt commits tagged with or add an advanced button to the toolbox algorithm ndialog which contains an option to copy the equivalent python ncommand as the parameters define in the dialog n nwhile this command is also available from the history dialog the nadvanced button provides a way for users to generate these ncommands without actually having to run the algorithm in advance n nsponsored by the research institute for nature and forest flemish govt
1
221,914
17,377,057,227
IssuesEvent
2021-07-31 00:10:23
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
closed
The blob which has virtual directory doesn't display under the Versions/Snapshots view when managing versions/snapshots from the root under flat list mode
:gear: blobs :heavy_check_mark: merged 🧪 testing
**Storage Explorer Version**: 1.21.0-dev **Build Number**: 20210727.2 **Branch**: main **Platform/OS**: Windows 10 **Architecture**: ia32 **How Found**: Exploratory testing **Regression From**: Not a regression ## Steps to Reproduce ## 1. Expand one Non-ADLS Gen2 storage account -> Blob Containers. 2. Create a blob container -> Create a folder then upload one blob to it -> Click 'Back'. 3. Click 'Show View Options' -> Select 'Flat'. 4. Hide view options panel -> Right click the blob. 5. Click 'Manage History -> Manage Versions'. 6. Check whether the blob displays under the Versions view. ## Expected Experience ## The blob displays under the Versions view. ## Actual Experience ## The blob does not display under the Versions view.
1.0
The blob which has virtual directory doesn't display under the Versions/Snapshots view when managing versions/snapshots from the root under flat list mode - **Storage Explorer Version**: 1.21.0-dev **Build Number**: 20210727.2 **Branch**: main **Platform/OS**: Windows 10 **Architecture**: ia32 **How Found**: Exploratory testing **Regression From**: Not a regression ## Steps to Reproduce ## 1. Expand one Non-ADLS Gen2 storage account -> Blob Containers. 2. Create a blob container -> Create a folder then upload one blob to it -> Click 'Back'. 3. Click 'Show View Options' -> Select 'Flat'. 4. Hide view options panel -> Right click the blob. 5. Click 'Manage History -> Manage Versions'. 6. Check whether the blob displays under the Versions view. ## Expected Experience ## The blob displays under the Versions view. ## Actual Experience ## The blob does not display under the Versions view.
non_process
the blob which has virtual directory doesn t display under the versions snapshots view when managing versions snapshots from the root under flat list mode storage explorer version dev build number branch main platform os windows architecture how found exploratory testing regression from not a regression steps to reproduce expand one non adls storage account blob containers create a blob container create a folder then upload one blob to it click back click show view options select flat hide view options panel right click the blob click manage history manage versions check whether the blob displays under the versions view expected experience the blob displays under the versions view actual experience the blob does not display under the versions view
0
18,108
24,134,131,721
IssuesEvent
2022-09-21 09:50:19
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Taxon constraint: GO:0032922
taxon constraints cellular processes
Hello, GO:0032922 (circadian regulation of gene expression) should not be excluded from cyanobacteria, and maybe not from archaea either. The circadian clock is highly studied in cyanobacteria, especially Synechococcus elongatus PCC 7942 and Thermosynechococcus elongatus BP-1 (now Thermosynechococcus vestitus), see review PMID:33419320. This paper (https://bmcecolevol.biomedcentral.com/articles/10.1186/s12862-017-0999-7) discusses a possible clock in archaea, although it is not yet clear if they really exist. There may be a clock in a few other (non-oxygenic) photosynthetic bacteria as well based simply on gene conservation. I think you should remove this taxonomic restraint. If you want it could be limited to cyanobacteria, but I wonder if we would have to revisit this issue one day if you chose to limit it to that taxonomic level. Thanks, Andrea
1.0
Taxon constraint: GO:0032922 - Hello, GO:0032922 (circadian regulation of gene expression) should not be excluded from cyanobacteria, and maybe not from archaea either. The circadian clock is highly studied in cyanobacteria, especially Synechococcus elongatus PCC 7942 and Thermosynechococcus elongatus BP-1 (now Thermosynechococcus vestitus), see review PMID:33419320. This paper (https://bmcecolevol.biomedcentral.com/articles/10.1186/s12862-017-0999-7) discusses a possible clock in archaea, although it is not yet clear if they really exist. There may be a clock in a few other (non-oxygenic) photosynthetic bacteria as well based simply on gene conservation. I think you should remove this taxonomic restraint. If you want it could be limited to cyanobacteria, but I wonder if we would have to revisit this issue one day if you chose to limit it to that taxonomic level. Thanks, Andrea
process
taxon constraint go hello go circadian regulation of gene expression should not be excluded from cyanobacteria and maybe not from archaea either the circadian clock is highly studied in cyanobacteria especially synechococcus elongatus pcc and thermosynechococcus elongatus bp now thermosynechococcus vestitus see review pmid this paper discusses a possible clock in archaea although it is not yet clear if they really exist there may be a clock in a few other non oxygenic photosynthetic bacteria as well based simply on gene conservation i think you should remove this taxonomic restraint if you want it could be limited to cyanobacteria but i wonder if we would have to revisit this issue one day if you chose to limit it to that taxonomic level thanks andrea
1
14,796
18,072,553,905
IssuesEvent
2021-09-21 05:41:39
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Obsoletion notice: GO:1902186 regulation of viral release from host cell & children
obsoletion multi-species process
Dear all, The proposal has been made to obsolete : GO:1902186 regulation of viral release from host cell GO:0044789 modulation by host of viral release from host cell GO:1902188 positive regulation of viral release from host cell GO:0044791 positive regulation by host of viral release from host cell The reason for obsoletion is that there is no evidence that this process is regulated. Annotations are to mutants, and should be moved to 'GO:0019076 viral release from host cell' or a child. There are 23 EXP annotations by UniProt (18) and AgBase (5). No action is needed, since annotations will be replaced by GO:0019076 viral release from host cell' . There are no mappings to these terms; these terms are not present in any subsets. You may comment on the ticket: Thanks, Pascale
1.0
Obsoletion notice: GO:1902186 regulation of viral release from host cell & children - Dear all, The proposal has been made to obsolete : GO:1902186 regulation of viral release from host cell GO:0044789 modulation by host of viral release from host cell GO:1902188 positive regulation of viral release from host cell GO:0044791 positive regulation by host of viral release from host cell The reason for obsoletion is that there is no evidence that this process is regulated. Annotations are to mutants, and should be moved to 'GO:0019076 viral release from host cell' or a child. There are 23 EXP annotations by UniProt (18) and AgBase (5). No action is needed, since annotations will be replaced by GO:0019076 viral release from host cell' . There are no mappings to these terms; these terms are not present in any subsets. You may comment on the ticket: Thanks, Pascale
process
obsoletion notice go regulation of viral release from host cell children dear all the proposal has been made to obsolete go regulation of viral release from host cell go modulation by host of viral release from host cell go positive regulation of viral release from host cell go positive regulation by host of viral release from host cell the reason for obsoletion is that there is no evidence that this process is regulated annotations are to mutants and should be moved to go viral release from host cell or a child there are exp annotations by uniprot and agbase no action is needed since annotations will be replaced by go viral release from host cell there are no mappings to these terms these terms are not present in any subsets you may comment on the ticket thanks pascale
1
6,278
9,255,975,562
IssuesEvent
2019-03-16 15:14:47
chuminh712/BookStorage---Group-2
https://api.github.com/repos/chuminh712/BookStorage---Group-2
closed
Detail Design
In Process
Design class diagram for Use Case Manage Customer Design class diagram for Use Case Manage Supplier
1.0
Detail Design - Design class diagram for Use Case Manage Customer Design class diagram for Use Case Manage Supplier
process
detail design design class diagram for use case manage customer design class diagram for use case manage supplier
1
306,508
9,395,829,872
IssuesEvent
2019-04-08 04:42:32
wso2/product-ei
https://api.github.com/repos/wso2/product-ei
opened
Proxy service results in a errorneous state if multiple transports are defined with incorrect configurations
Priority/Low Severity/Minor
**Description:** Assume a proxy is defined with mutiple transports vfs and http. If the proxy service is not configured for the vfs transport properly, the service will be marked as faulty with the following warn message ``` WARN - VFSTransportListener Unable to configure the service SplitAggregateProxy for the VFS transport: Service doesn't have configuration information for transport vfs. This service is being marked as faulty and will not be available over the VFS transport. ``` and will not be listed by the admin service (or in the management console). However, it will be acessible via http. This is an inconsistent state. sample proxy: ``` <proxy name="StockQuoteProxy"> <target> <inSequence> <filter source="$trp:priority" regex="1"> <then> <enqueue priority="1" sequence="priority_sequence" executor="exec"/> </then> <else> <enqueue priority="10" sequence="priority_sequence" executor="exec"/> </else> </filter> </inSequence> <outSequence> <send/> </outSequence> </target> <publishWSDL uri="file:samples/service-bus/resources/proxy/sample_proxy_1.wsdl"/> </proxy> ``` **Affected Product Version:** All EI versions **Related Issues:** https://github.com/wso2/product-ei/issues/3631
1.0
Proxy service results in a errorneous state if multiple transports are defined with incorrect configurations - **Description:** Assume a proxy is defined with mutiple transports vfs and http. If the proxy service is not configured for the vfs transport properly, the service will be marked as faulty with the following warn message ``` WARN - VFSTransportListener Unable to configure the service SplitAggregateProxy for the VFS transport: Service doesn't have configuration information for transport vfs. This service is being marked as faulty and will not be available over the VFS transport. ``` and will not be listed by the admin service (or in the management console). However, it will be acessible via http. This is an inconsistent state. sample proxy: ``` <proxy name="StockQuoteProxy"> <target> <inSequence> <filter source="$trp:priority" regex="1"> <then> <enqueue priority="1" sequence="priority_sequence" executor="exec"/> </then> <else> <enqueue priority="10" sequence="priority_sequence" executor="exec"/> </else> </filter> </inSequence> <outSequence> <send/> </outSequence> </target> <publishWSDL uri="file:samples/service-bus/resources/proxy/sample_proxy_1.wsdl"/> </proxy> ``` **Affected Product Version:** All EI versions **Related Issues:** https://github.com/wso2/product-ei/issues/3631
non_process
proxy service results in a errorneous state if multiple transports are defined with incorrect configurations description assume a proxy is defined with mutiple transports vfs and http if the proxy service is not configured for the vfs transport properly the service will be marked as faulty with the following warn message warn vfstransportlistener unable to configure the service splitaggregateproxy for the vfs transport service doesn t have configuration information for transport vfs this service is being marked as faulty and will not be available over the vfs transport and will not be listed by the admin service or in the management console however it will be acessible via http this is an inconsistent state sample proxy affected product version all ei versions related issues
0
5,792
8,638,867,707
IssuesEvent
2018-11-23 16:13:13
Dweepa/DataAnalytics-
https://api.github.com/repos/Dweepa/DataAnalytics-
closed
Filling in missing values using clusters
Cleaning and pre-processing
When dealing with missing values, for continuous variables, chose to group by fields that are correlated with those variables, and then take the mean of the grouped values to fill the missing variable
1.0
Filling in missing values using clusters - When dealing with missing values, for continuous variables, chose to group by fields that are correlated with those variables, and then take the mean of the grouped values to fill the missing variable
process
filling in missing values using clusters when dealing with missing values for continuous variables chose to group by fields that are correlated with those variables and then take the mean of the grouped values to fill the missing variable
1
414,175
27,980,163,256
IssuesEvent
2023-03-26 03:12:12
ahg3/mywebclass-simulation
https://api.github.com/repos/ahg3/mywebclass-simulation
opened
Imaging and File Optimization
documentation good first issue
Test that the website's images and files are optimized for faster loading times.
1.0
Imaging and File Optimization - Test that the website's images and files are optimized for faster loading times.
non_process
imaging and file optimization test that the website s images and files are optimized for faster loading times
0
16,009
10,423,688,746
IssuesEvent
2019-09-16 12:04:05
kyma-project/kyma
https://api.github.com/repos/kyma-project/kyma
closed
Release process should work both for the master branch and dedicated api-server branch
area/service-catalog
**Description** The release process should work both for the master and dedicated api-server branch. The release is done by the Travis job. The current release process is documented here: https://github.com/kubernetes-sigs/service-catalog/wiki/Release-Process **AC** - release process for CRD only from master is working - release process for the dedicated api-server branch is working
1.0
Release process should work both for the master branch and dedicated api-server branch - **Description** The release process should work both for the master and dedicated api-server branch. The release is done by the Travis job. The current release process is documented here: https://github.com/kubernetes-sigs/service-catalog/wiki/Release-Process **AC** - release process for CRD only from master is working - release process for the dedicated api-server branch is working
non_process
release process should work both for the master branch and dedicated api server branch description the release process should work both for the master and dedicated api server branch the release is done by the travis job the current release process is documented here ac release process for crd only from master is working release process for the dedicated api server branch is working
0
339,163
10,243,460,719
IssuesEvent
2019-08-20 08:13:45
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.estv.admin.ch - design is broken
browser-firefox engine-gecko priority-normal
<!-- @browser: Firefox 69.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:69.0) Gecko/20100101 Firefox/69.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html **Browser / Version**: Firefox 69.0 **Operating System**: Windows 7 **Tested Another Browser**: No **Problem type**: Design is broken **Description**: no access to login **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/8/3f045f63-c65e-4b6b-9708-ee580cf21921-thumb.jpeg)](https://webcompat.com/uploads/2019/8/3f045f63-c65e-4b6b-9708-ee580cf21921.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190815163925</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "Request to access cookie or storage on https://googleads.g.doubleclick.net/pagead/id was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://static.doubleclick.net/instream/ad_status.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://googleads.g.doubleclick.net/pagead/id was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://googleads.g.doubleclick.net/pagead/id?slf_rd=1 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://googleads.g.doubleclick.net/pagead/id?slf_rd=1 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "MouseEvent.mozPressure is deprecated. Use PointerEvent.pressure instead." {file: "https://www.youtube.com/yts/jsbin/player_ias-vflubst9M/en_US/base.js" line: 1135}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.estv.admin.ch - design is broken - <!-- @browser: Firefox 69.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:69.0) Gecko/20100101 Firefox/69.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html **Browser / Version**: Firefox 69.0 **Operating System**: Windows 7 **Tested Another Browser**: No **Problem type**: Design is broken **Description**: no access to login **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/8/3f045f63-c65e-4b6b-9708-ee580cf21921-thumb.jpeg)](https://webcompat.com/uploads/2019/8/3f045f63-c65e-4b6b-9708-ee580cf21921.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190815163925</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "Request to access cookie or storage on https://googleads.g.doubleclick.net/pagead/id was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://static.doubleclick.net/instream/ad_status.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://googleads.g.doubleclick.net/pagead/id was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://googleads.g.doubleclick.net/pagead/id?slf_rd=1 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://googleads.g.doubleclick.net/pagead/id?slf_rd=1 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.estv.admin.ch/estv/it/home/mehrwertsteuer/dienstleistungen/mwst-abrechnung-online.html" line: 0}]', u'[JavaScript Warning: "MouseEvent.mozPressure is deprecated. Use PointerEvent.pressure instead." {file: "https://www.youtube.com/yts/jsbin/player_ias-vflubst9M/en_US/base.js" line: 1135}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
design is broken url browser version firefox operating system windows tested another browser no problem type design is broken description no access to login steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta console messages u u u u u from with ❤️
0
8,319
11,486,378,628
IssuesEvent
2020-02-11 09:50:14
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Multi-organism processes tidying up: feeding behavior
multi-species process
Hello, Following up on the discussion we had on the ontology call regarding #14956 I will merge: GO:0044366 feeding on or from other organism GO:0044367 feeding from tissue of other organism (no annotations) into GO:0042755 eating behavior and move GO:0002120 predatory behavior and GO:0060756 foraging behavior as children of GO:0007631 feeding behavior Thanks, Pascale
1.0
Multi-organism processes tidying up: feeding behavior - Hello, Following up on the discussion we had on the ontology call regarding #14956 I will merge: GO:0044366 feeding on or from other organism GO:0044367 feeding from tissue of other organism (no annotations) into GO:0042755 eating behavior and move GO:0002120 predatory behavior and GO:0060756 foraging behavior as children of GO:0007631 feeding behavior Thanks, Pascale
process
multi organism processes tidying up feeding behavior hello following up on the discussion we had on the ontology call regarding i will merge go feeding on or from other organism go feeding from tissue of other organism no annotations into go eating behavior and move go predatory behavior and go foraging behavior as children of go feeding behavior thanks pascale
1
68,746
14,948,852,750
IssuesEvent
2021-01-26 10:39:58
symfony/symfony
https://api.github.com/repos/symfony/symfony
closed
[Security] TypeError: Argument 1 passed to Symfony\Component\Security\Core\Exception\AuthenticationException::__unserialize() must be of the type array, null given
Bug Security Status: Needs Review
**Symfony version(s) affected**: 4.4.17 **Description** In some specific circumstances, PHP 7.4 session could not be unserialized in PHP 7.3. I asusme it's related to https://bugs.php.net/bug.php?id=79759. This is a common scenario in multi-server environment, when PHP is upgraded server by server. It's strange that there are no previous bug reports on this subject. Errors I collected: ``` TypeError: Argument 1 passed to Symfony\Component\Security\Core\Exception\AuthenticationException::__unserialize() must be of the type array, null given, called in vendor/symfony/symfony/src/Symfony/Component/Security/Core/Exception/AuthenticationException.php on line 137 #24 /vendor/symfony/symfony/src/Symfony/Component/Security/Core/Exception/AuthenticationException.php(96): Symfony\Component\Security\Core\Exception\AuthenticationException::__unserialize #23 /vendor/symfony/symfony/src/Symfony/Component/Security/Core/Exception/AuthenticationException.php(137): Symfony\Component\Security\Core\Exception\AuthenticationException::__wakeup #22 [internal](0): session_start #21 /vendor/symfony/symfony/src/Symfony/Component/HttpFoundation/Session/Storage/NativeSessionStorage.php(156): Symfony\Component\HttpFoundation\Session\Storage\NativeSessionStorage::start #20 /vendor/symfony/symfony/src/Symfony/Component/HttpFoundation/Session/Storage/NativeSessionStorage.php(323): Symfony\Component\HttpFoundation\Session\Storage\NativeSessionStorage::getBag ``` How does `session_start()` knows about `AuthenticationException`? ``` "PHP message: PHP Warning: session_start(): Session object destruction failed in vendor/symfony/symfony/src/Symfony/Component/HttpFoundation/Session/Storage/NativeSessionStorage.php on line 156" ``` **How to reproduce** Save session in PHP 7.3, and load it in PHP 7.4. but bug only affects small portion of users, so it's not easy to reproduce. **Possible Improvement** One minor improvement for better DX and error handling is to check if `$this->serialized` is `null` before trying to unserialize and fail with a better error.
True
[Security] TypeError: Argument 1 passed to Symfony\Component\Security\Core\Exception\AuthenticationException::__unserialize() must be of the type array, null given - **Symfony version(s) affected**: 4.4.17 **Description** In some specific circumstances, PHP 7.4 session could not be unserialized in PHP 7.3. I asusme it's related to https://bugs.php.net/bug.php?id=79759. This is a common scenario in multi-server environment, when PHP is upgraded server by server. It's strange that there are no previous bug reports on this subject. Errors I collected: ``` TypeError: Argument 1 passed to Symfony\Component\Security\Core\Exception\AuthenticationException::__unserialize() must be of the type array, null given, called in vendor/symfony/symfony/src/Symfony/Component/Security/Core/Exception/AuthenticationException.php on line 137 #24 /vendor/symfony/symfony/src/Symfony/Component/Security/Core/Exception/AuthenticationException.php(96): Symfony\Component\Security\Core\Exception\AuthenticationException::__unserialize #23 /vendor/symfony/symfony/src/Symfony/Component/Security/Core/Exception/AuthenticationException.php(137): Symfony\Component\Security\Core\Exception\AuthenticationException::__wakeup #22 [internal](0): session_start #21 /vendor/symfony/symfony/src/Symfony/Component/HttpFoundation/Session/Storage/NativeSessionStorage.php(156): Symfony\Component\HttpFoundation\Session\Storage\NativeSessionStorage::start #20 /vendor/symfony/symfony/src/Symfony/Component/HttpFoundation/Session/Storage/NativeSessionStorage.php(323): Symfony\Component\HttpFoundation\Session\Storage\NativeSessionStorage::getBag ``` How does `session_start()` knows about `AuthenticationException`? ``` "PHP message: PHP Warning: session_start(): Session object destruction failed in vendor/symfony/symfony/src/Symfony/Component/HttpFoundation/Session/Storage/NativeSessionStorage.php on line 156" ``` **How to reproduce** Save session in PHP 7.3, and load it in PHP 7.4. but bug only affects small portion of users, so it's not easy to reproduce. **Possible Improvement** One minor improvement for better DX and error handling is to check if `$this->serialized` is `null` before trying to unserialize and fail with a better error.
non_process
typeerror argument passed to symfony component security core exception authenticationexception unserialize must be of the type array null given symfony version s affected description in some specific circumstances php session could not be unserialized in php i asusme it s related to this is a common scenario in multi server environment when php is upgraded server by server it s strange that there are no previous bug reports on this subject errors i collected typeerror argument passed to symfony component security core exception authenticationexception unserialize must be of the type array null given called in vendor symfony symfony src symfony component security core exception authenticationexception php on line vendor symfony symfony src symfony component security core exception authenticationexception php symfony component security core exception authenticationexception unserialize vendor symfony symfony src symfony component security core exception authenticationexception php symfony component security core exception authenticationexception wakeup session start vendor symfony symfony src symfony component httpfoundation session storage nativesessionstorage php symfony component httpfoundation session storage nativesessionstorage start vendor symfony symfony src symfony component httpfoundation session storage nativesessionstorage php symfony component httpfoundation session storage nativesessionstorage getbag how does session start knows about authenticationexception php message php warning session start session object destruction failed in vendor symfony symfony src symfony component httpfoundation session storage nativesessionstorage php on line how to reproduce save session in php and load it in php but bug only affects small portion of users so it s not easy to reproduce possible improvement one minor improvement for better dx and error handling is to check if this serialized is null before trying to unserialize and fail with a better error
0
735,616
25,406,705,155
IssuesEvent
2022-11-22 15:47:18
grpc/grpc
https://api.github.com/repos/grpc/grpc
closed
Unable to install grpcio 1.51.0 for Python on Mac OS Apple M1
kind/bug lang/Python priority/P2 disposition/requires reporter action
<!-- PLEASE DO NOT POST A QUESTION HERE. This form is for bug reports and feature requests ONLY! For general questions and troubleshooting, please ask/look for answers at StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc For questions that specifically need to be answered by gRPC team members, please ask/look for answers at grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new) --> ### What version of gRPC and what language are you using? grpcio==1.51.0 grpcio-tools=1.51.0 ### What operating system (Linux, Windows,...) and version? Mac OS Ventura (Apple M1) 13.0.1 ### What runtime / compiler are you using (e.g. python version or version of gcc) Python 3.11.0 / 3.10.8 / 3.9.15 ### What did you do? Please provide either 1) A unit test for reproducing the bug or 2) Specific steps for us to follow to reproduce the bug. If there’s not enough information to debug the problem, gRPC team may close the issue at their discretion. You’re welcome to re-open the issue once you have a reproduction. I wanted to upgrade a Poetry project depending on `grpcio`, and I'm facing a `setuptools` errors while trying to install the new wheel. I've tried to install the project manually through `pip` with this command line: ```sh $ pip install --upgrade grpcio grpcio-tools ``` And I was still unable to update my project dependency. ### What did you expect to see? A successful installation. ### What did you see instead? ```sh $ pip install --upgrade grpcio grpcio-tools Collecting grpcio Using cached grpcio-1.51.0.tar.gz (22.1 MB) Preparing metadata (setup.py): started Preparing metadata (setup.py): finished with status 'done' Collecting grpcio-tools Downloading grpcio-tools-1.51.0.tar.gz (2.3 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.3/2.3 MB 24.5 MB/s eta 0:00:00 Preparing metadata (setup.py): started Preparing metadata (setup.py): finished with status 'done' Requirement already satisfied: protobuf<5.0dev,>=4.21.6 in ./.venv/lib/python3.9/site-packages (from grpcio-tools) (4.21.9) Requirement already satisfied: setuptools in ./.venv/lib/python3.9/site-packages (from grpcio-tools) (65.6.0) Building wheels for collected packages: grpcio, grpcio-tools Building wheel for grpcio (setup.py): started Building wheel for grpcio (setup.py): finished with status 'error' error: subprocess-exited-with-error × python setup.py bdist_wheel did not run successfully. │ exit code: 1 ╰─> [2106 lines of output] Boringssl currently does not support macOS arm64, so we'll try to use the system installation of 'openssl' to build by default, make sure you have 'openssl' installed in this case Found cython-generated files... running bdist_wheel running build running build_py running build_project_metadata ... Truncated mkdir and file copy ... clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX12.sdk -D_WIN32_WINNT=1536 -DGRPC_XDS_USER_AGENT_NAME_SUFFIX=\"Python\" -DGRPC_XDS_USER_AGENT_VERSION_SUFFIX=\"1.51.0\" -DOPENSSL_NO_ASM=1 -DGPR_BACKWARDS_COMPATIBILITY_MODE=1 -DHAVE_CONFIG_H=1 -DGRPC_ENABLE_FORK_SUPPORT=1 "-DPyMODINIT_FUNC=extern \"C\" __attribute__((visibility (\"default\"))) PyObject*" -DGRPC_POSIX_FORK_ALLOW_PTHREAD_ATFORK=1 -Isrc/python/grpcio -Iinclude -I. -Ithird_party/abseil-cpp -Ithird_party/address_sorting/include -Ithird_party/cares/cares/include -Ithird_party/cares -Ithird_party/cares/cares -Ithird_party/cares/config_darwin -Ithird_party/re2 -I/usr/include/openssl -Ithird_party/upb -Isrc/core/ext/upb-generated -Isrc/core/ext/upbdefs-generated -Ithird_party/xxhash -Ithird_party/zlib -I/path/to/.venv/include -I/opt/homebrew/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/include/python3.9 -c src/core/ext/filters/census/grpc_context.cc -o python_build/temp.macosx-12-arm64-cpython-39/src/core/ext/filters/census/grpc_context.o -std=c++14 -stdlib=libc++ -fvisibility=hidden -fno-wrapv -fno-exceptions -DHAVE_UNISTD_H -pthread clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX12.sdk -D_WIN32_WINNT=1536 -DGRPC_XDS_USER_AGENT_NAME_SUFFIX=\"Python\" -DGRPC_XDS_USER_AGENT_VERSION_SUFFIX=\"1.51.0\" -DOPENSSL_NO_ASM=1 -DGPR_BACKWARDS_COMPATIBILITY_MODE=1 -DHAVE_CONFIG_H=1 -DGRPC_ENABLE_FORK_SUPPORT=1 "-DPyMODINIT_FUNC=extern \"C\" __attribute__((visibility (\"default\"))) PyObject*" -DGRPC_POSIX_FORK_ALLOW_PTHREAD_ATFORK=1 -Isrc/python/grpcio -Iinclude -I. -Ithird_party/abseil-cpp -Ithird_party/address_sorting/include -Ithird_party/cares/cares/include -Ithird_party/cares -Ithird_party/cares/cares -Ithird_party/cares/config_darwin -Ithird_party/re2 -I/usr/include/openssl -Ithird_party/upb -Isrc/core/ext/upb-generated -Isrc/core/ext/upbdefs-generated -Ithird_party/xxhash -Ithird_party/zlib -I/path/to/.venv/include -I/opt/homebrew/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/include/python3.9 -c src/core/ext/filters/client_channel/lb_policy/weighted_target/weighted_target.cc -o python_build/temp.macosx-12-arm64-cpython-39/src/core/ext/filters/client_channel/lb_policy/weighted_target/weighted_target.o -std=c++14 -stdlib=libc++ -fvisibility=hidden -fno-wrapv -fno-exceptions -DHAVE_UNISTD_H -pthread ... Truncated clang invokations ... clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX12.sdk -D_WIN32_WINNT=1536 -DGRPC_XDS_USER_AGENT_NAME_SUFFIX=\"Python\" -DGRPC_XDS_USER_AGENT_VERSION_SUFFIX=\"1.51.0\" -DOPENSSL_NO_ASM=1 -DGPR_BACKWARDS_COMPATIBILITY_MODE=1 -DHAVE_CONFIG_H=1 -DGRPC_ENABLE_FORK_SUPPORT=1 "-DPyMODINIT_FUNC=extern \"C\" __attribute__((visibility (\"default\"))) PyObject*" -DGRPC_POSIX_FORK_ALLOW_PTHREAD_ATFORK=1 -Isrc/python/grpcio -Iinclude -I. -Ithird_party/abseil-cpp -Ithird_party/address_sorting/include -Ithird_party/cares/cares/include -Ithird_party/cares -Ithird_party/cares/cares -Ithird_party/cares/config_darwin -Ithird_party/re2 -I/usr/include/openssl -Ithird_party/upb -Isrc/core/ext/upb-generated -Isrc/core/ext/upbdefs-generated -Ithird_party/xxhash -Ithird_party/zlib -I/path/to/.venv/include -I/opt/homebrew/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/include/python3.9 -c third_party/upb/upb/array.c -o python_build/temp.macosx-12-arm64-cpython-39/third_party/upb/upb/array.o -stdlib=libc++ -fvisibility=hidden -fno-wrapv -fno-exceptions -DHAVE_UNISTD_H -pthread creating None creating None/var creating None/var/folders creating None/var/folders/9g creating None/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn creating None/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T creating None/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/tmp6oq83wvk clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX12.sdk -I/path/to/.venv/include -I/opt/homebrew/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/include/python3.9 -c /var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/tmp6oq83wvk/a.c -o None/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/tmp6oq83wvk/a.o Traceback (most recent call last): File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/unixccompiler.py", line 186, in _compile self.spawn(compiler_so + cc_args + [src, '-o', obj] + extra_postargs) File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_spawn_patch.py", line 54, in _commandfile_spawn _classic_spawn(self, command) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/ccompiler.py", line 1007, in spawn spawn(cmd, dry_run=self.dry_run, **kwargs) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/spawn.py", line 70, in spawn raise DistutilsExecError( distutils.errors.DistutilsExecError: command '/usr/bin/clang' failed with exit code 1 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 280, in build_extensions build_ext.build_ext.build_extensions(self) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 468, in build_extensions self._build_extensions_serial() File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 494, in _build_extensions_serial self.build_extension(ext) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/command/build_ext.py", line 246, in build_extension _build_ext.build_extension(self, ext) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 549, in build_extension objects = self.compiler.compile( File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_parallel_compile_patch.py", line 58, in _parallel_compile multiprocessing.pool.ThreadPool(BUILD_EXT_COMPILER_JOBS).map( File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 364, in map return self._map_async(func, iterable, mapstar, chunksize).get() File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 771, in get raise self._value File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 125, in worker result = (True, func(*args, **kwds)) File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 48, in mapstar return list(map(*args)) File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_parallel_compile_patch.py", line 54, in _compile_single_file self._compile(obj, src, ext, cc_args, extra_postargs, pp_opts) File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 263, in new_compile return old_compile(obj, src, ext, cc_args, extra_postargs, File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/unixccompiler.py", line 188, in _compile raise CompileError(msg) distutils.errors.CompileError: command '/usr/bin/clang' failed with exit code 1 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "<string>", line 2, in <module> File "<pip-setuptools-caller>", line 34, in <module> File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/setup.py", line 546, in <module> setuptools.setup( File "/path/to/.venv/lib/python3.9/site-packages/setuptools/__init__.py", line 87, in setup return distutils.core.setup(**attrs) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/core.py", line 185, in setup return run_commands(dist) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/core.py", line 201, in run_commands dist.run_commands() File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 969, in run_commands self.run_command(cmd) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/dist.py", line 1217, in run_command super().run_command(command) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 988, in run_command cmd_obj.run() File "/path/to/.venv/lib/python3.9/site-packages/wheel/bdist_wheel.py", line 299, in run self.run_command('build') File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command self.distribution.run_command(command) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/dist.py", line 1217, in run_command super().run_command(command) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 988, in run_command cmd_obj.run() File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build.py", line 132, in run self.run_command(cmd_name) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command self.distribution.run_command(command) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/dist.py", line 1217, in run_command super().run_command(command) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 988, in run_command cmd_obj.run() File "/path/to/.venv/lib/python3.9/site-packages/setuptools/command/build_ext.py", line 84, in run _build_ext.run(self) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 346, in run self.build_extensions() File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 284, in build_extensions raise CommandError( commands.CommandError: Failed `build_ext` step: Traceback (most recent call last): File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/unixccompiler.py", line 186, in _compile self.spawn(compiler_so + cc_args + [src, '-o', obj] + extra_postargs) File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_spawn_patch.py", line 54, in _commandfile_spawn _classic_spawn(self, command) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/ccompiler.py", line 1007, in spawn spawn(cmd, dry_run=self.dry_run, **kwargs) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/spawn.py", line 70, in spawn raise DistutilsExecError( distutils.errors.DistutilsExecError: command '/usr/bin/clang' failed with exit code 1 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 280, in build_extensions build_ext.build_ext.build_extensions(self) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 468, in build_extensions self._build_extensions_serial() File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 494, in _build_extensions_serial self.build_extension(ext) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/command/build_ext.py", line 246, in build_extension _build_ext.build_extension(self, ext) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 549, in build_extension objects = self.compiler.compile( File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_parallel_compile_patch.py", line 58, in _parallel_compile multiprocessing.pool.ThreadPool(BUILD_EXT_COMPILER_JOBS).map( File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 364, in map return self._map_async(func, iterable, mapstar, chunksize).get() File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 771, in get raise self._value File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 125, in worker result = (True, func(*args, **kwds)) File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 48, in mapstar return list(map(*args)) File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_parallel_compile_patch.py", line 54, in _compile_single_file self._compile(obj, src, ext, cc_args, extra_postargs, pp_opts) File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 263, in new_compile return old_compile(obj, src, ext, cc_args, extra_postargs, File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/unixccompiler.py", line 188, in _compile raise CompileError(msg) distutils.errors.CompileError: command '/usr/bin/clang' failed with exit code 1 [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. ERROR: Failed building wheel for grpcio Running setup.py clean for grpcio Building wheel for grpcio-tools (setup.py): started Building wheel for grpcio-tools (setup.py): finished with status 'done' Created wheel for grpcio-tools: filename=grpcio_tools-1.51.0-cp39-cp39-macosx_12_0_arm64.whl size=1889028 sha256=01a33f1fdde7f11a87ad82211aa73c9f79cc9c85ba735aec144d0a857b81d351 Stored in directory: /Users/rigma/Library/Caches/pip/wheels/17/1d/7a/1d007fdff1149d47af524dbb20479c5ee32a81636b302a6fbd Successfully built grpcio-tools Failed to build grpcio ... Same errors with legacy pip installation ... × Encountered error while trying to install package. ╰─> grpcio note: This is an issue with the package mentioned above, not pip. hint: See above for output from the failure. ``` I've got the same output for Python 3.10 and 3.11. ### Anything else we should know about your project / environment? My project is setup thanks to Poetry with the following `pyproject.toml` subset: ```toml [build-system] requires = ["poetry-core"] build-backend = "poetry.core.masonry.api" [tool.poetry] name = "example-project" version = "0.0.0" readme = "README.md" packages = [ { include = "app" }, ] [tool.poetry.dependencies] python = "^3.9" dependency-injector = { version = "4.40.0", extras = ["pydantic"] } fastapi = "0.87.0" google-cloud-storage = "2.6.0" orjson = "3.8.2" pydantic = { version = "1.10.2", extras = ["email"] } typing-extensions = "4.4.0" [tool.poetry.group.dev.dependencies] alembic = "1.8.1" black = "*" pyproject-parser = "0.7.0" pytest = "7.2.0" pytest-asyncio = "0.20.2" pytest-cov = "4.0.0" uvicorn = "0.19.0" [tool.pytest.ini_options] addopts = "--cov app --cov-report term --cov-report xml:cov.xml --disable-warnings --strict-markers" asyncio_mode = "strict" testpaths = ["tests"] markers = [] filterwarnings = [] [tool.coverage.run] branch = true source = ["app"] omit = ["app/__main__.py"] ```
1.0
Unable to install grpcio 1.51.0 for Python on Mac OS Apple M1 - <!-- PLEASE DO NOT POST A QUESTION HERE. This form is for bug reports and feature requests ONLY! For general questions and troubleshooting, please ask/look for answers at StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc For questions that specifically need to be answered by gRPC team members, please ask/look for answers at grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io Issues specific to *grpc-java*, *grpc-go*, *grpc-node*, *grpc-dart*, *grpc-web* should be created in the repository they belong to (e.g. https://github.com/grpc/grpc-LANGUAGE/issues/new) --> ### What version of gRPC and what language are you using? grpcio==1.51.0 grpcio-tools=1.51.0 ### What operating system (Linux, Windows,...) and version? Mac OS Ventura (Apple M1) 13.0.1 ### What runtime / compiler are you using (e.g. python version or version of gcc) Python 3.11.0 / 3.10.8 / 3.9.15 ### What did you do? Please provide either 1) A unit test for reproducing the bug or 2) Specific steps for us to follow to reproduce the bug. If there’s not enough information to debug the problem, gRPC team may close the issue at their discretion. You’re welcome to re-open the issue once you have a reproduction. I wanted to upgrade a Poetry project depending on `grpcio`, and I'm facing a `setuptools` errors while trying to install the new wheel. I've tried to install the project manually through `pip` with this command line: ```sh $ pip install --upgrade grpcio grpcio-tools ``` And I was still unable to update my project dependency. ### What did you expect to see? A successful installation. ### What did you see instead? ```sh $ pip install --upgrade grpcio grpcio-tools Collecting grpcio Using cached grpcio-1.51.0.tar.gz (22.1 MB) Preparing metadata (setup.py): started Preparing metadata (setup.py): finished with status 'done' Collecting grpcio-tools Downloading grpcio-tools-1.51.0.tar.gz (2.3 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.3/2.3 MB 24.5 MB/s eta 0:00:00 Preparing metadata (setup.py): started Preparing metadata (setup.py): finished with status 'done' Requirement already satisfied: protobuf<5.0dev,>=4.21.6 in ./.venv/lib/python3.9/site-packages (from grpcio-tools) (4.21.9) Requirement already satisfied: setuptools in ./.venv/lib/python3.9/site-packages (from grpcio-tools) (65.6.0) Building wheels for collected packages: grpcio, grpcio-tools Building wheel for grpcio (setup.py): started Building wheel for grpcio (setup.py): finished with status 'error' error: subprocess-exited-with-error × python setup.py bdist_wheel did not run successfully. │ exit code: 1 ╰─> [2106 lines of output] Boringssl currently does not support macOS arm64, so we'll try to use the system installation of 'openssl' to build by default, make sure you have 'openssl' installed in this case Found cython-generated files... running bdist_wheel running build running build_py running build_project_metadata ... Truncated mkdir and file copy ... clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX12.sdk -D_WIN32_WINNT=1536 -DGRPC_XDS_USER_AGENT_NAME_SUFFIX=\"Python\" -DGRPC_XDS_USER_AGENT_VERSION_SUFFIX=\"1.51.0\" -DOPENSSL_NO_ASM=1 -DGPR_BACKWARDS_COMPATIBILITY_MODE=1 -DHAVE_CONFIG_H=1 -DGRPC_ENABLE_FORK_SUPPORT=1 "-DPyMODINIT_FUNC=extern \"C\" __attribute__((visibility (\"default\"))) PyObject*" -DGRPC_POSIX_FORK_ALLOW_PTHREAD_ATFORK=1 -Isrc/python/grpcio -Iinclude -I. -Ithird_party/abseil-cpp -Ithird_party/address_sorting/include -Ithird_party/cares/cares/include -Ithird_party/cares -Ithird_party/cares/cares -Ithird_party/cares/config_darwin -Ithird_party/re2 -I/usr/include/openssl -Ithird_party/upb -Isrc/core/ext/upb-generated -Isrc/core/ext/upbdefs-generated -Ithird_party/xxhash -Ithird_party/zlib -I/path/to/.venv/include -I/opt/homebrew/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/include/python3.9 -c src/core/ext/filters/census/grpc_context.cc -o python_build/temp.macosx-12-arm64-cpython-39/src/core/ext/filters/census/grpc_context.o -std=c++14 -stdlib=libc++ -fvisibility=hidden -fno-wrapv -fno-exceptions -DHAVE_UNISTD_H -pthread clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX12.sdk -D_WIN32_WINNT=1536 -DGRPC_XDS_USER_AGENT_NAME_SUFFIX=\"Python\" -DGRPC_XDS_USER_AGENT_VERSION_SUFFIX=\"1.51.0\" -DOPENSSL_NO_ASM=1 -DGPR_BACKWARDS_COMPATIBILITY_MODE=1 -DHAVE_CONFIG_H=1 -DGRPC_ENABLE_FORK_SUPPORT=1 "-DPyMODINIT_FUNC=extern \"C\" __attribute__((visibility (\"default\"))) PyObject*" -DGRPC_POSIX_FORK_ALLOW_PTHREAD_ATFORK=1 -Isrc/python/grpcio -Iinclude -I. -Ithird_party/abseil-cpp -Ithird_party/address_sorting/include -Ithird_party/cares/cares/include -Ithird_party/cares -Ithird_party/cares/cares -Ithird_party/cares/config_darwin -Ithird_party/re2 -I/usr/include/openssl -Ithird_party/upb -Isrc/core/ext/upb-generated -Isrc/core/ext/upbdefs-generated -Ithird_party/xxhash -Ithird_party/zlib -I/path/to/.venv/include -I/opt/homebrew/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/include/python3.9 -c src/core/ext/filters/client_channel/lb_policy/weighted_target/weighted_target.cc -o python_build/temp.macosx-12-arm64-cpython-39/src/core/ext/filters/client_channel/lb_policy/weighted_target/weighted_target.o -std=c++14 -stdlib=libc++ -fvisibility=hidden -fno-wrapv -fno-exceptions -DHAVE_UNISTD_H -pthread ... Truncated clang invokations ... clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX12.sdk -D_WIN32_WINNT=1536 -DGRPC_XDS_USER_AGENT_NAME_SUFFIX=\"Python\" -DGRPC_XDS_USER_AGENT_VERSION_SUFFIX=\"1.51.0\" -DOPENSSL_NO_ASM=1 -DGPR_BACKWARDS_COMPATIBILITY_MODE=1 -DHAVE_CONFIG_H=1 -DGRPC_ENABLE_FORK_SUPPORT=1 "-DPyMODINIT_FUNC=extern \"C\" __attribute__((visibility (\"default\"))) PyObject*" -DGRPC_POSIX_FORK_ALLOW_PTHREAD_ATFORK=1 -Isrc/python/grpcio -Iinclude -I. -Ithird_party/abseil-cpp -Ithird_party/address_sorting/include -Ithird_party/cares/cares/include -Ithird_party/cares -Ithird_party/cares/cares -Ithird_party/cares/config_darwin -Ithird_party/re2 -I/usr/include/openssl -Ithird_party/upb -Isrc/core/ext/upb-generated -Isrc/core/ext/upbdefs-generated -Ithird_party/xxhash -Ithird_party/zlib -I/path/to/.venv/include -I/opt/homebrew/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/include/python3.9 -c third_party/upb/upb/array.c -o python_build/temp.macosx-12-arm64-cpython-39/third_party/upb/upb/array.o -stdlib=libc++ -fvisibility=hidden -fno-wrapv -fno-exceptions -DHAVE_UNISTD_H -pthread creating None creating None/var creating None/var/folders creating None/var/folders/9g creating None/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn creating None/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T creating None/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/tmp6oq83wvk clang -Wno-unused-result -Wsign-compare -Wunreachable-code -fno-common -dynamic -DNDEBUG -g -fwrapv -O3 -Wall -isysroot /Library/Developer/CommandLineTools/SDKs/MacOSX12.sdk -I/path/to/.venv/include -I/opt/homebrew/opt/python@3.9/Frameworks/Python.framework/Versions/3.9/include/python3.9 -c /var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/tmp6oq83wvk/a.c -o None/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/tmp6oq83wvk/a.o Traceback (most recent call last): File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/unixccompiler.py", line 186, in _compile self.spawn(compiler_so + cc_args + [src, '-o', obj] + extra_postargs) File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_spawn_patch.py", line 54, in _commandfile_spawn _classic_spawn(self, command) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/ccompiler.py", line 1007, in spawn spawn(cmd, dry_run=self.dry_run, **kwargs) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/spawn.py", line 70, in spawn raise DistutilsExecError( distutils.errors.DistutilsExecError: command '/usr/bin/clang' failed with exit code 1 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 280, in build_extensions build_ext.build_ext.build_extensions(self) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 468, in build_extensions self._build_extensions_serial() File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 494, in _build_extensions_serial self.build_extension(ext) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/command/build_ext.py", line 246, in build_extension _build_ext.build_extension(self, ext) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 549, in build_extension objects = self.compiler.compile( File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_parallel_compile_patch.py", line 58, in _parallel_compile multiprocessing.pool.ThreadPool(BUILD_EXT_COMPILER_JOBS).map( File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 364, in map return self._map_async(func, iterable, mapstar, chunksize).get() File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 771, in get raise self._value File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 125, in worker result = (True, func(*args, **kwds)) File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 48, in mapstar return list(map(*args)) File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_parallel_compile_patch.py", line 54, in _compile_single_file self._compile(obj, src, ext, cc_args, extra_postargs, pp_opts) File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 263, in new_compile return old_compile(obj, src, ext, cc_args, extra_postargs, File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/unixccompiler.py", line 188, in _compile raise CompileError(msg) distutils.errors.CompileError: command '/usr/bin/clang' failed with exit code 1 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "<string>", line 2, in <module> File "<pip-setuptools-caller>", line 34, in <module> File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/setup.py", line 546, in <module> setuptools.setup( File "/path/to/.venv/lib/python3.9/site-packages/setuptools/__init__.py", line 87, in setup return distutils.core.setup(**attrs) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/core.py", line 185, in setup return run_commands(dist) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/core.py", line 201, in run_commands dist.run_commands() File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 969, in run_commands self.run_command(cmd) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/dist.py", line 1217, in run_command super().run_command(command) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 988, in run_command cmd_obj.run() File "/path/to/.venv/lib/python3.9/site-packages/wheel/bdist_wheel.py", line 299, in run self.run_command('build') File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command self.distribution.run_command(command) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/dist.py", line 1217, in run_command super().run_command(command) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 988, in run_command cmd_obj.run() File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build.py", line 132, in run self.run_command(cmd_name) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command self.distribution.run_command(command) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/dist.py", line 1217, in run_command super().run_command(command) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/dist.py", line 988, in run_command cmd_obj.run() File "/path/to/.venv/lib/python3.9/site-packages/setuptools/command/build_ext.py", line 84, in run _build_ext.run(self) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 346, in run self.build_extensions() File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 284, in build_extensions raise CommandError( commands.CommandError: Failed `build_ext` step: Traceback (most recent call last): File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/unixccompiler.py", line 186, in _compile self.spawn(compiler_so + cc_args + [src, '-o', obj] + extra_postargs) File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_spawn_patch.py", line 54, in _commandfile_spawn _classic_spawn(self, command) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/ccompiler.py", line 1007, in spawn spawn(cmd, dry_run=self.dry_run, **kwargs) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/spawn.py", line 70, in spawn raise DistutilsExecError( distutils.errors.DistutilsExecError: command '/usr/bin/clang' failed with exit code 1 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 280, in build_extensions build_ext.build_ext.build_extensions(self) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 468, in build_extensions self._build_extensions_serial() File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 494, in _build_extensions_serial self.build_extension(ext) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/command/build_ext.py", line 246, in build_extension _build_ext.build_extension(self, ext) File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.py", line 549, in build_extension objects = self.compiler.compile( File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_parallel_compile_patch.py", line 58, in _parallel_compile multiprocessing.pool.ThreadPool(BUILD_EXT_COMPILER_JOBS).map( File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 364, in map return self._map_async(func, iterable, mapstar, chunksize).get() File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 771, in get raise self._value File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 125, in worker result = (True, func(*args, **kwds)) File "/opt/homebrew/Cellar/python@3.9/3.9.15/Frameworks/Python.framework/Versions/3.9/lib/python3.9/multiprocessing/pool.py", line 48, in mapstar return list(map(*args)) File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/_parallel_compile_patch.py", line 54, in _compile_single_file self._compile(obj, src, ext, cc_args, extra_postargs, pp_opts) File "/private/var/folders/9g/r8cmr5nd44s4z_c_2379ntym0000gn/T/pip-install-og0qtd4c/grpcio_c8eb1bc5f558437e82b7ca1b2ea4139b/src/python/grpcio/commands.py", line 263, in new_compile return old_compile(obj, src, ext, cc_args, extra_postargs, File "/path/to/.venv/lib/python3.9/site-packages/setuptools/_distutils/unixccompiler.py", line 188, in _compile raise CompileError(msg) distutils.errors.CompileError: command '/usr/bin/clang' failed with exit code 1 [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. ERROR: Failed building wheel for grpcio Running setup.py clean for grpcio Building wheel for grpcio-tools (setup.py): started Building wheel for grpcio-tools (setup.py): finished with status 'done' Created wheel for grpcio-tools: filename=grpcio_tools-1.51.0-cp39-cp39-macosx_12_0_arm64.whl size=1889028 sha256=01a33f1fdde7f11a87ad82211aa73c9f79cc9c85ba735aec144d0a857b81d351 Stored in directory: /Users/rigma/Library/Caches/pip/wheels/17/1d/7a/1d007fdff1149d47af524dbb20479c5ee32a81636b302a6fbd Successfully built grpcio-tools Failed to build grpcio ... Same errors with legacy pip installation ... × Encountered error while trying to install package. ╰─> grpcio note: This is an issue with the package mentioned above, not pip. hint: See above for output from the failure. ``` I've got the same output for Python 3.10 and 3.11. ### Anything else we should know about your project / environment? My project is setup thanks to Poetry with the following `pyproject.toml` subset: ```toml [build-system] requires = ["poetry-core"] build-backend = "poetry.core.masonry.api" [tool.poetry] name = "example-project" version = "0.0.0" readme = "README.md" packages = [ { include = "app" }, ] [tool.poetry.dependencies] python = "^3.9" dependency-injector = { version = "4.40.0", extras = ["pydantic"] } fastapi = "0.87.0" google-cloud-storage = "2.6.0" orjson = "3.8.2" pydantic = { version = "1.10.2", extras = ["email"] } typing-extensions = "4.4.0" [tool.poetry.group.dev.dependencies] alembic = "1.8.1" black = "*" pyproject-parser = "0.7.0" pytest = "7.2.0" pytest-asyncio = "0.20.2" pytest-cov = "4.0.0" uvicorn = "0.19.0" [tool.pytest.ini_options] addopts = "--cov app --cov-report term --cov-report xml:cov.xml --disable-warnings --strict-markers" asyncio_mode = "strict" testpaths = ["tests"] markers = [] filterwarnings = [] [tool.coverage.run] branch = true source = ["app"] omit = ["app/__main__.py"] ```
non_process
unable to install grpcio for python on mac os apple please do not post a question here this form is for bug reports and feature requests only for general questions and troubleshooting please ask look for answers at stackoverflow with grpc tag for questions that specifically need to be answered by grpc team members please ask look for answers at grpc io mailing list issues specific to grpc java grpc go grpc node grpc dart grpc web should be created in the repository they belong to e g what version of grpc and what language are you using grpcio grpcio tools what operating system linux windows and version mac os ventura apple what runtime compiler are you using e g python version or version of gcc python what did you do please provide either a unit test for reproducing the bug or specific steps for us to follow to reproduce the bug if there’s not enough information to debug the problem grpc team may close the issue at their discretion you’re welcome to re open the issue once you have a reproduction i wanted to upgrade a poetry project depending on grpcio and i m facing a setuptools errors while trying to install the new wheel i ve tried to install the project manually through pip with this command line sh pip install upgrade grpcio grpcio tools and i was still unable to update my project dependency what did you expect to see a successful installation what did you see instead sh pip install upgrade grpcio grpcio tools collecting grpcio using cached grpcio tar gz mb preparing metadata setup py started preparing metadata setup py finished with status done collecting grpcio tools downloading grpcio tools tar gz mb ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ mb mb s eta preparing metadata setup py started preparing metadata setup py finished with status done requirement already satisfied protobuf in venv lib site packages from grpcio tools requirement already satisfied setuptools in venv lib site packages from grpcio tools building wheels for collected packages grpcio grpcio tools building wheel for grpcio setup py started building wheel for grpcio setup py finished with status error error subprocess exited with error × python setup py bdist wheel did not run successfully │ exit code ╰─ boringssl currently does not support macos so we ll try to use the system installation of openssl to build by default make sure you have openssl installed in this case found cython generated files running bdist wheel running build running build py running build project metadata truncated mkdir and file copy clang wno unused result wsign compare wunreachable code fno common dynamic dndebug g fwrapv wall isysroot library developer commandlinetools sdks sdk d winnt dgrpc xds user agent name suffix python dgrpc xds user agent version suffix dopenssl no asm dgpr backwards compatibility mode dhave config h dgrpc enable fork support dpymodinit func extern c attribute visibility default pyobject dgrpc posix fork allow pthread atfork isrc python grpcio iinclude i ithird party abseil cpp ithird party address sorting include ithird party cares cares include ithird party cares ithird party cares cares ithird party cares config darwin ithird party i usr include openssl ithird party upb isrc core ext upb generated isrc core ext upbdefs generated ithird party xxhash ithird party zlib i path to venv include i opt homebrew opt python frameworks python framework versions include c src core ext filters census grpc context cc o python build temp macosx cpython src core ext filters census grpc context o std c stdlib libc fvisibility hidden fno wrapv fno exceptions dhave unistd h pthread clang wno unused result wsign compare wunreachable code fno common dynamic dndebug g fwrapv wall isysroot library developer commandlinetools sdks sdk d winnt dgrpc xds user agent name suffix python dgrpc xds user agent version suffix dopenssl no asm dgpr backwards compatibility mode dhave config h dgrpc enable fork support dpymodinit func extern c attribute visibility default pyobject dgrpc posix fork allow pthread atfork isrc python grpcio iinclude i ithird party abseil cpp ithird party address sorting include ithird party cares cares include ithird party cares ithird party cares cares ithird party cares config darwin ithird party i usr include openssl ithird party upb isrc core ext upb generated isrc core ext upbdefs generated ithird party xxhash ithird party zlib i path to venv include i opt homebrew opt python frameworks python framework versions include c src core ext filters client channel lb policy weighted target weighted target cc o python build temp macosx cpython src core ext filters client channel lb policy weighted target weighted target o std c stdlib libc fvisibility hidden fno wrapv fno exceptions dhave unistd h pthread truncated clang invokations clang wno unused result wsign compare wunreachable code fno common dynamic dndebug g fwrapv wall isysroot library developer commandlinetools sdks sdk d winnt dgrpc xds user agent name suffix python dgrpc xds user agent version suffix dopenssl no asm dgpr backwards compatibility mode dhave config h dgrpc enable fork support dpymodinit func extern c attribute visibility default pyobject dgrpc posix fork allow pthread atfork isrc python grpcio iinclude i ithird party abseil cpp ithird party address sorting include ithird party cares cares include ithird party cares ithird party cares cares ithird party cares config darwin ithird party i usr include openssl ithird party upb isrc core ext upb generated isrc core ext upbdefs generated ithird party xxhash ithird party zlib i path to venv include i opt homebrew opt python frameworks python framework versions include c third party upb upb array c o python build temp macosx cpython third party upb upb array o stdlib libc fvisibility hidden fno wrapv fno exceptions dhave unistd h pthread creating none creating none var creating none var folders creating none var folders creating none var folders c creating none var folders c t creating none var folders c t clang wno unused result wsign compare wunreachable code fno common dynamic dndebug g fwrapv wall isysroot library developer commandlinetools sdks sdk i path to venv include i opt homebrew opt python frameworks python framework versions include c var folders c t a c o none var folders c t a o traceback most recent call last file path to venv lib site packages setuptools distutils unixccompiler py line in compile self spawn compiler so cc args extra postargs file private var folders c t pip install grpcio src python grpcio spawn patch py line in commandfile spawn classic spawn self command file path to venv lib site packages setuptools distutils ccompiler py line in spawn spawn cmd dry run self dry run kwargs file path to venv lib site packages setuptools distutils spawn py line in spawn raise distutilsexecerror distutils errors distutilsexecerror command usr bin clang failed with exit code during handling of the above exception another exception occurred traceback most recent call last file private var folders c t pip install grpcio src python grpcio commands py line in build extensions build ext build ext build extensions self file path to venv lib site packages setuptools distutils command build ext py line in build extensions self build extensions serial file path to venv lib site packages setuptools distutils command build ext py line in build extensions serial self build extension ext file path to venv lib site packages setuptools command build ext py line in build extension build ext build extension self ext file path to venv lib site packages setuptools distutils command build ext py line in build extension objects self compiler compile file private var folders c t pip install grpcio src python grpcio parallel compile patch py line in parallel compile multiprocessing pool threadpool build ext compiler jobs map file opt homebrew cellar python frameworks python framework versions lib multiprocessing pool py line in map return self map async func iterable mapstar chunksize get file opt homebrew cellar python frameworks python framework versions lib multiprocessing pool py line in get raise self value file opt homebrew cellar python frameworks python framework versions lib multiprocessing pool py line in worker result true func args kwds file opt homebrew cellar python frameworks python framework versions lib multiprocessing pool py line in mapstar return list map args file private var folders c t pip install grpcio src python grpcio parallel compile patch py line in compile single file self compile obj src ext cc args extra postargs pp opts file private var folders c t pip install grpcio src python grpcio commands py line in new compile return old compile obj src ext cc args extra postargs file path to venv lib site packages setuptools distutils unixccompiler py line in compile raise compileerror msg distutils errors compileerror command usr bin clang failed with exit code during handling of the above exception another exception occurred traceback most recent call last file line in file line in file private var folders c t pip install grpcio setup py line in setuptools setup file path to venv lib site packages setuptools init py line in setup return distutils core setup attrs file path to venv lib site packages setuptools distutils core py line in setup return run commands dist file path to venv lib site packages setuptools distutils core py line in run commands dist run commands file path to venv lib site packages setuptools distutils dist py line in run commands self run command cmd file path to venv lib site packages setuptools dist py line in run command super run command command file path to venv lib site packages setuptools distutils dist py line in run command cmd obj run file path to venv lib site packages wheel bdist wheel py line in run self run command build file path to venv lib site packages setuptools distutils cmd py line in run command self distribution run command command file path to venv lib site packages setuptools dist py line in run command super run command command file path to venv lib site packages setuptools distutils dist py line in run command cmd obj run file path to venv lib site packages setuptools distutils command build py line in run self run command cmd name file path to venv lib site packages setuptools distutils cmd py line in run command self distribution run command command file path to venv lib site packages setuptools dist py line in run command super run command command file path to venv lib site packages setuptools distutils dist py line in run command cmd obj run file path to venv lib site packages setuptools command build ext py line in run build ext run self file path to venv lib site packages setuptools distutils command build ext py line in run self build extensions file private var folders c t pip install grpcio src python grpcio commands py line in build extensions raise commanderror commands commanderror failed build ext step traceback most recent call last file path to venv lib site packages setuptools distutils unixccompiler py line in compile self spawn compiler so cc args extra postargs file private var folders c t pip install grpcio src python grpcio spawn patch py line in commandfile spawn classic spawn self command file path to venv lib site packages setuptools distutils ccompiler py line in spawn spawn cmd dry run self dry run kwargs file path to venv lib site packages setuptools distutils spawn py line in spawn raise distutilsexecerror distutils errors distutilsexecerror command usr bin clang failed with exit code during handling of the above exception another exception occurred traceback most recent call last file private var folders c t pip install grpcio src python grpcio commands py line in build extensions build ext build ext build extensions self file path to venv lib site packages setuptools distutils command build ext py line in build extensions self build extensions serial file path to venv lib site packages setuptools distutils command build ext py line in build extensions serial self build extension ext file path to venv lib site packages setuptools command build ext py line in build extension build ext build extension self ext file path to venv lib site packages setuptools distutils command build ext py line in build extension objects self compiler compile file private var folders c t pip install grpcio src python grpcio parallel compile patch py line in parallel compile multiprocessing pool threadpool build ext compiler jobs map file opt homebrew cellar python frameworks python framework versions lib multiprocessing pool py line in map return self map async func iterable mapstar chunksize get file opt homebrew cellar python frameworks python framework versions lib multiprocessing pool py line in get raise self value file opt homebrew cellar python frameworks python framework versions lib multiprocessing pool py line in worker result true func args kwds file opt homebrew cellar python frameworks python framework versions lib multiprocessing pool py line in mapstar return list map args file private var folders c t pip install grpcio src python grpcio parallel compile patch py line in compile single file self compile obj src ext cc args extra postargs pp opts file private var folders c t pip install grpcio src python grpcio commands py line in new compile return old compile obj src ext cc args extra postargs file path to venv lib site packages setuptools distutils unixccompiler py line in compile raise compileerror msg distutils errors compileerror command usr bin clang failed with exit code note this error originates from a subprocess and is likely not a problem with pip error failed building wheel for grpcio running setup py clean for grpcio building wheel for grpcio tools setup py started building wheel for grpcio tools setup py finished with status done created wheel for grpcio tools filename grpcio tools macosx whl size stored in directory users rigma library caches pip wheels successfully built grpcio tools failed to build grpcio same errors with legacy pip installation × encountered error while trying to install package ╰─ grpcio note this is an issue with the package mentioned above not pip hint see above for output from the failure i ve got the same output for python and anything else we should know about your project environment my project is setup thanks to poetry with the following pyproject toml subset toml requires build backend poetry core masonry api name example project version readme readme md packages include app python dependency injector version extras fastapi google cloud storage orjson pydantic version extras typing extensions alembic black pyproject parser pytest pytest asyncio pytest cov uvicorn addopts cov app cov report term cov report xml cov xml disable warnings strict markers asyncio mode strict testpaths markers filterwarnings branch true source omit
0
336
2,792,238,767
IssuesEvent
2015-05-10 20:42:43
t3kt/vjzual
https://api.github.com/repos/t3kt/vjzual
opened
add more blend modes
enhancement video processing
possible modes: * tima-based time distortion * masking using an external mask source
1.0
add more blend modes - possible modes: * tima-based time distortion * masking using an external mask source
process
add more blend modes possible modes tima based time distortion masking using an external mask source
1
163,779
12,744,477,720
IssuesEvent
2020-06-26 12:35:21
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
certified-conformance : ./sonobuoy run --mode certified-conformance --wait -v 5 failed with timeout error
kind/failing-test needs-sig
certified-conformance : ./sonobuoy run --mode certified-conformance --wait -v 5 failed with below error level=error msg="Result processing error (409): result systemd-logs/cluster already received" level=error msg="Result processing error (409): result systemd-logs/cluster already received" level=error msg="Result processing error (409): result systemd-logs/cluster already received" level=error msg="Result processing error (409): result systemd-logs/cluster already received" level=error msg="Timeout waiting for plugin e2e" level=error msg="error running plugins: timeout occurred when waiting for plugin results" level=error msg="Timeout occurred when running plugins. Inspect logs further for details."
1.0
certified-conformance : ./sonobuoy run --mode certified-conformance --wait -v 5 failed with timeout error - certified-conformance : ./sonobuoy run --mode certified-conformance --wait -v 5 failed with below error level=error msg="Result processing error (409): result systemd-logs/cluster already received" level=error msg="Result processing error (409): result systemd-logs/cluster already received" level=error msg="Result processing error (409): result systemd-logs/cluster already received" level=error msg="Result processing error (409): result systemd-logs/cluster already received" level=error msg="Timeout waiting for plugin e2e" level=error msg="error running plugins: timeout occurred when waiting for plugin results" level=error msg="Timeout occurred when running plugins. Inspect logs further for details."
non_process
certified conformance sonobuoy run mode certified conformance wait v failed with timeout error certified conformance sonobuoy run mode certified conformance wait v failed with below error level error msg result processing error result systemd logs cluster already received level error msg result processing error result systemd logs cluster already received level error msg result processing error result systemd logs cluster already received level error msg result processing error result systemd logs cluster already received level error msg timeout waiting for plugin level error msg error running plugins timeout occurred when waiting for plugin results level error msg timeout occurred when running plugins inspect logs further for details
0
151,416
19,650,296,376
IssuesEvent
2022-01-10 05:50:32
panasalap/linux-4.19.72
https://api.github.com/repos/panasalap/linux-4.19.72
opened
CVE-2021-26932 (Medium) detected in multiple libraries
security vulnerability
## CVE-2021-26932 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-yoctov5.4.51</b>, <b>linuxlinux-4.19.224</b>, <b>linux-yoctov5.4.51</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel 3.2 through 5.10.16, as used by Xen. Grant mapping operations often occur in batch hypercalls, where a number of operations are done in a single hypercall, the success or failure of each one is reported to the backend driver, and the backend driver then loops over the results, performing follow-up actions based on the success or failure of each operation. Unfortunately, when running in PV mode, the Linux backend drivers mishandle this: Some errors are ignored, effectively implying their success from the success of related batch elements. In other cases, errors resulting from one batch element lead to further batch elements not being inspected, and hence successful ones to not be possible to properly unmap upon error recovery. Only systems with Linux backends running in PV mode are vulnerable. Linux backends run in HVM / PVH modes are not vulnerable. This affects arch/*/xen/p2m.c and drivers/xen/gntdev.c. <p>Publish Date: 2021-02-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-26932>CVE-2021-26932</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-26932 (Medium) detected in multiple libraries - ## CVE-2021-26932 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-yoctov5.4.51</b>, <b>linuxlinux-4.19.224</b>, <b>linux-yoctov5.4.51</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel 3.2 through 5.10.16, as used by Xen. Grant mapping operations often occur in batch hypercalls, where a number of operations are done in a single hypercall, the success or failure of each one is reported to the backend driver, and the backend driver then loops over the results, performing follow-up actions based on the success or failure of each operation. Unfortunately, when running in PV mode, the Linux backend drivers mishandle this: Some errors are ignored, effectively implying their success from the success of related batch elements. In other cases, errors resulting from one batch element lead to further batch elements not being inspected, and hence successful ones to not be possible to properly unmap upon error recovery. Only systems with Linux backends running in PV mode are vulnerable. Linux backends run in HVM / PVH modes are not vulnerable. This affects arch/*/xen/p2m.c and drivers/xen/gntdev.c. <p>Publish Date: 2021-02-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-26932>CVE-2021-26932</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries linux linuxlinux linux vulnerability details an issue was discovered in the linux kernel through as used by xen grant mapping operations often occur in batch hypercalls where a number of operations are done in a single hypercall the success or failure of each one is reported to the backend driver and the backend driver then loops over the results performing follow up actions based on the success or failure of each operation unfortunately when running in pv mode the linux backend drivers mishandle this some errors are ignored effectively implying their success from the success of related batch elements in other cases errors resulting from one batch element lead to further batch elements not being inspected and hence successful ones to not be possible to properly unmap upon error recovery only systems with linux backends running in pv mode are vulnerable linux backends run in hvm pvh modes are not vulnerable this affects arch xen c and drivers xen gntdev c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource
0
57,646
6,552,434,475
IssuesEvent
2017-09-05 18:20:34
Aubron/scoreshots-templates
https://api.github.com/repos/Aubron/scoreshots-templates
closed
Modify Football Statshot, XML Integration
Priority: High Status: Testing Needed
![381](https://user-images.githubusercontent.com/26412274/29973855-f5c55384-8efe-11e7-850a-16f61d796e5e.jpg) ### Requested by: Internal ## Dynamic Considerations: Adding XML integration to template 381, pictured above. No other edits should be needed.
1.0
Modify Football Statshot, XML Integration - ![381](https://user-images.githubusercontent.com/26412274/29973855-f5c55384-8efe-11e7-850a-16f61d796e5e.jpg) ### Requested by: Internal ## Dynamic Considerations: Adding XML integration to template 381, pictured above. No other edits should be needed.
non_process
modify football statshot xml integration requested by internal dynamic considerations adding xml integration to template pictured above no other edits should be needed
0
22,341
31,018,532,597
IssuesEvent
2023-08-10 02:00:07
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Thu, 10 Aug 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### Constructing Holistic Spatio-Temporal Scene Graph for Video Semantic Role Labeling - **Authors:** Yu Zhao, Hao Fei, Yixin Cao, Bobo Li, Meishan Zhang, Jianguo Wei, Min Zhang, Tat-Seng Chua - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL) - **Arxiv link:** https://arxiv.org/abs/2308.05081 - **Pdf link:** https://arxiv.org/pdf/2308.05081 - **Abstract** Video Semantic Role Labeling (VidSRL) aims to detect the salient events from given videos, by recognizing the predict-argument event structures and the interrelationships between events. While recent endeavors have put forth methods for VidSRL, they can be mostly subject to two key drawbacks, including the lack of fine-grained spatial scene perception and the insufficiently modeling of video temporality. Towards this end, this work explores a novel holistic spatio-temporal scene graph (namely HostSG) representation based on the existing dynamic scene graph structures, which well model both the fine-grained spatial semantics and temporal dynamics of videos for VidSRL. Built upon the HostSG, we present a nichetargeting VidSRL framework. A scene-event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high-level event semantic structure, resulting in an overall hierarchical scene-event (termed ICE) graph structure. We further perform iterative structure refinement to optimize the ICE graph, such that the overall structure representation can best coincide with end task demand. Finally, three subtask predictions of VidSRL are jointly decoded, where the end-to-end paradigm effectively avoids error propagation. On the benchmark dataset, our framework boosts significantly over the current best-performing model. Further analyses are shown for a better understanding of the advances of our methods. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB ### Rendering Humans from Object-Occluded Monocular Videos - **Authors:** Tiange Xiang, Adam Sun, Jiajun Wu, Ehsan Adeli, Li Fei-Fei - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.04622 - **Pdf link:** https://arxiv.org/pdf/2308.04622 - **Abstract** 3D understanding and rendering of moving humans from monocular videos is a challenging task. Despite recent progress, the task remains difficult in real-world scenarios, where obstacles may block the camera view and cause partial occlusions in the captured videos. Existing methods cannot handle such defects due to two reasons. First, the standard rendering strategy relies on point-point mapping, which could lead to dramatic disparities between the visible and occluded areas of the body. Second, the naive direct regression approach does not consider any feasibility criteria (ie, prior information) for rendering under occlusions. To tackle the above drawbacks, we present OccNeRF, a neural rendering method that achieves better rendering of humans in severely occluded scenes. As direct solutions to the two drawbacks, we propose surface-based rendering by integrating geometry and visibility priors. We validate our method on both simulated and real-world occlusions and demonstrate our method's superiority. ### Constructing Holistic Spatio-Temporal Scene Graph for Video Semantic Role Labeling - **Authors:** Yu Zhao, Hao Fei, Yixin Cao, Bobo Li, Meishan Zhang, Jianguo Wei, Min Zhang, Tat-Seng Chua - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL) - **Arxiv link:** https://arxiv.org/abs/2308.05081 - **Pdf link:** https://arxiv.org/pdf/2308.05081 - **Abstract** Video Semantic Role Labeling (VidSRL) aims to detect the salient events from given videos, by recognizing the predict-argument event structures and the interrelationships between events. While recent endeavors have put forth methods for VidSRL, they can be mostly subject to two key drawbacks, including the lack of fine-grained spatial scene perception and the insufficiently modeling of video temporality. Towards this end, this work explores a novel holistic spatio-temporal scene graph (namely HostSG) representation based on the existing dynamic scene graph structures, which well model both the fine-grained spatial semantics and temporal dynamics of videos for VidSRL. Built upon the HostSG, we present a nichetargeting VidSRL framework. A scene-event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high-level event semantic structure, resulting in an overall hierarchical scene-event (termed ICE) graph structure. We further perform iterative structure refinement to optimize the ICE graph, such that the overall structure representation can best coincide with end task demand. Finally, three subtask predictions of VidSRL are jointly decoded, where the end-to-end paradigm effectively avoids error propagation. On the benchmark dataset, our framework boosts significantly over the current best-performing model. Further analyses are shown for a better understanding of the advances of our methods. ## Keyword: ISP ### Rendering Humans from Object-Occluded Monocular Videos - **Authors:** Tiange Xiang, Adam Sun, Jiajun Wu, Ehsan Adeli, Li Fei-Fei - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.04622 - **Pdf link:** https://arxiv.org/pdf/2308.04622 - **Abstract** 3D understanding and rendering of moving humans from monocular videos is a challenging task. Despite recent progress, the task remains difficult in real-world scenarios, where obstacles may block the camera view and cause partial occlusions in the captured videos. Existing methods cannot handle such defects due to two reasons. First, the standard rendering strategy relies on point-point mapping, which could lead to dramatic disparities between the visible and occluded areas of the body. Second, the naive direct regression approach does not consider any feasibility criteria (ie, prior information) for rendering under occlusions. To tackle the above drawbacks, we present OccNeRF, a neural rendering method that achieves better rendering of humans in severely occluded scenes. As direct solutions to the two drawbacks, we propose surface-based rendering by integrating geometry and visibility priors. We validate our method on both simulated and real-world occlusions and demonstrate our method's superiority. ### Addressing Racial Bias in Facial Emotion Recognition - **Authors:** Alex Fan, Xingshuo Xiao, Peter Washington - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Computers and Society (cs.CY) - **Arxiv link:** https://arxiv.org/abs/2308.04674 - **Pdf link:** https://arxiv.org/pdf/2308.04674 - **Abstract** Fairness in deep learning models trained with high-dimensional inputs and subjective labels remains a complex and understudied area. Facial emotion recognition, a domain where datasets are often racially imbalanced, can lead to models that yield disparate outcomes across racial groups. This study focuses on analyzing racial bias by sub-sampling training sets with varied racial distributions and assessing test performance across these simulations. Our findings indicate that smaller datasets with posed faces improve on both fairness and performance metrics as the simulations approach racial balance. Notably, the F1-score increases by $27.2\%$ points, and demographic parity increases by $15.7\%$ points on average across the simulations. However, in larger datasets with greater facial variation, fairness metrics generally remain constant, suggesting that racial balance by itself is insufficient to achieve parity in test performance across different racial groups. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### Resource Constrained Model Compression via Minimax Optimization for Spiking Neural Networks - **Authors:** Jue Chen, Huan Yuan, Jianchao Tan, Bin Chen, Chengru Song, Di Zhang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Multimedia (cs.MM) - **Arxiv link:** https://arxiv.org/abs/2308.04672 - **Pdf link:** https://arxiv.org/pdf/2308.04672 - **Abstract** Brain-inspired Spiking Neural Networks (SNNs) have the characteristics of event-driven and high energy-efficient, which are different from traditional Artificial Neural Networks (ANNs) when deployed on edge devices such as neuromorphic chips. Most previous work focuses on SNNs training strategies to improve model performance and brings larger and deeper network architectures. It is difficult to deploy these complex networks on resource-limited edge devices directly. To meet such demand, people compress SNNs very cautiously to balance the performance and the computation efficiency. Existing compression methods either iteratively pruned SNNs using weights norm magnitude or formulated the problem as a sparse learning optimization. We propose an improved end-to-end Minimax optimization method for this sparse learning problem to better balance the model performance and the computation efficiency. We also demonstrate that jointly applying compression and finetuning on SNNs is better than sequentially, especially for extreme compression ratios. The compressed SNN models achieved state-of-the-art (SOTA) performance on various benchmark datasets and architectures. Our code is available at https://github.com/chenjallen/Resource-Constrained-Compression-on-SNN. ## Keyword: RAW ### Rendering Humans from Object-Occluded Monocular Videos - **Authors:** Tiange Xiang, Adam Sun, Jiajun Wu, Ehsan Adeli, Li Fei-Fei - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.04622 - **Pdf link:** https://arxiv.org/pdf/2308.04622 - **Abstract** 3D understanding and rendering of moving humans from monocular videos is a challenging task. Despite recent progress, the task remains difficult in real-world scenarios, where obstacles may block the camera view and cause partial occlusions in the captured videos. Existing methods cannot handle such defects due to two reasons. First, the standard rendering strategy relies on point-point mapping, which could lead to dramatic disparities between the visible and occluded areas of the body. Second, the naive direct regression approach does not consider any feasibility criteria (ie, prior information) for rendering under occlusions. To tackle the above drawbacks, we present OccNeRF, a neural rendering method that achieves better rendering of humans in severely occluded scenes. As direct solutions to the two drawbacks, we propose surface-based rendering by integrating geometry and visibility priors. We validate our method on both simulated and real-world occlusions and demonstrate our method's superiority. ### SAfER: Layer-Level Sensitivity Assessment for Efficient and Robust Neural Network Inference - **Authors:** Edouard Yvinec, Arnaud Dapogny, Kevin Bailly - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.04753 - **Pdf link:** https://arxiv.org/pdf/2308.04753 - **Abstract** Deep neural networks (DNNs) demonstrate outstanding performance across most computer vision tasks. Some critical applications, such as autonomous driving or medical imaging, also require investigation into their behavior and the reasons behind the decisions they make. In this vein, DNN attribution consists in studying the relationship between the predictions of a DNN and its inputs. Attribution methods have been adapted to highlight the most relevant weights or neurons in a DNN, allowing to more efficiently select which weights or neurons can be pruned. However, a limitation of these approaches is that weights are typically compared within each layer separately, while some layers might appear as more critical than others. In this work, we propose to investigate DNN layer importance, i.e. to estimate the sensitivity of the accuracy w.r.t. perturbations applied at the layer level. To do so, we propose a novel dataset to evaluate our method as well as future works. We benchmark a number of criteria and draw conclusions regarding how to assess DNN layer importance and, consequently, how to budgetize layers for increased DNN efficiency (with applications for DNN pruning and quantization), as well as robustness to hardware failure (e.g. bit swaps). ### View while Moving: Efficient Video Recognition in Long-untrimmed Videos - **Authors:** Ye Tian, Mengyu Yang, Lanshan Zhang, Zhizhen Zhang, Yang Liu, Xiaohui Xie, Xirong Que, Wendong Wang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.04834 - **Pdf link:** https://arxiv.org/pdf/2308.04834 - **Abstract** Recent adaptive methods for efficient video recognition mostly follow the two-stage paradigm of "preview-then-recognition" and have achieved great success on multiple video benchmarks. However, this two-stage paradigm involves two visits of raw frames from coarse-grained to fine-grained during inference (cannot be parallelized), and the captured spatiotemporal features cannot be reused in the second stage (due to varying granularity), being not friendly to efficiency and computation optimization. To this end, inspired by human cognition, we propose a novel recognition paradigm of "View while Moving" for efficient long-untrimmed video recognition. In contrast to the two-stage paradigm, our paradigm only needs to access the raw frame once. The two phases of coarse-grained sampling and fine-grained recognition are combined into unified spatiotemporal modeling, showing great performance. Moreover, we investigate the properties of semantic units in video and propose a hierarchical mechanism to efficiently capture and reason about the unit-level and video-level temporal semantics in long-untrimmed videos respectively. Extensive experiments on both long-untrimmed and short-trimmed videos demonstrate that our approach outperforms state-of-the-art methods in terms of accuracy as well as efficiency, yielding new efficiency and accuracy trade-offs for video spatiotemporal modeling. ### Constructing Holistic Spatio-Temporal Scene Graph for Video Semantic Role Labeling - **Authors:** Yu Zhao, Hao Fei, Yixin Cao, Bobo Li, Meishan Zhang, Jianguo Wei, Min Zhang, Tat-Seng Chua - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL) - **Arxiv link:** https://arxiv.org/abs/2308.05081 - **Pdf link:** https://arxiv.org/pdf/2308.05081 - **Abstract** Video Semantic Role Labeling (VidSRL) aims to detect the salient events from given videos, by recognizing the predict-argument event structures and the interrelationships between events. While recent endeavors have put forth methods for VidSRL, they can be mostly subject to two key drawbacks, including the lack of fine-grained spatial scene perception and the insufficiently modeling of video temporality. Towards this end, this work explores a novel holistic spatio-temporal scene graph (namely HostSG) representation based on the existing dynamic scene graph structures, which well model both the fine-grained spatial semantics and temporal dynamics of videos for VidSRL. Built upon the HostSG, we present a nichetargeting VidSRL framework. A scene-event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high-level event semantic structure, resulting in an overall hierarchical scene-event (termed ICE) graph structure. We further perform iterative structure refinement to optimize the ICE graph, such that the overall structure representation can best coincide with end task demand. Finally, three subtask predictions of VidSRL are jointly decoded, where the end-to-end paradigm effectively avoids error propagation. On the benchmark dataset, our framework boosts significantly over the current best-performing model. Further analyses are shown for a better understanding of the advances of our methods. ## Keyword: raw image There is no result
2.0
New submissions for Thu, 10 Aug 23 - ## Keyword: events ### Constructing Holistic Spatio-Temporal Scene Graph for Video Semantic Role Labeling - **Authors:** Yu Zhao, Hao Fei, Yixin Cao, Bobo Li, Meishan Zhang, Jianguo Wei, Min Zhang, Tat-Seng Chua - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL) - **Arxiv link:** https://arxiv.org/abs/2308.05081 - **Pdf link:** https://arxiv.org/pdf/2308.05081 - **Abstract** Video Semantic Role Labeling (VidSRL) aims to detect the salient events from given videos, by recognizing the predict-argument event structures and the interrelationships between events. While recent endeavors have put forth methods for VidSRL, they can be mostly subject to two key drawbacks, including the lack of fine-grained spatial scene perception and the insufficiently modeling of video temporality. Towards this end, this work explores a novel holistic spatio-temporal scene graph (namely HostSG) representation based on the existing dynamic scene graph structures, which well model both the fine-grained spatial semantics and temporal dynamics of videos for VidSRL. Built upon the HostSG, we present a nichetargeting VidSRL framework. A scene-event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high-level event semantic structure, resulting in an overall hierarchical scene-event (termed ICE) graph structure. We further perform iterative structure refinement to optimize the ICE graph, such that the overall structure representation can best coincide with end task demand. Finally, three subtask predictions of VidSRL are jointly decoded, where the end-to-end paradigm effectively avoids error propagation. On the benchmark dataset, our framework boosts significantly over the current best-performing model. Further analyses are shown for a better understanding of the advances of our methods. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB ### Rendering Humans from Object-Occluded Monocular Videos - **Authors:** Tiange Xiang, Adam Sun, Jiajun Wu, Ehsan Adeli, Li Fei-Fei - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.04622 - **Pdf link:** https://arxiv.org/pdf/2308.04622 - **Abstract** 3D understanding and rendering of moving humans from monocular videos is a challenging task. Despite recent progress, the task remains difficult in real-world scenarios, where obstacles may block the camera view and cause partial occlusions in the captured videos. Existing methods cannot handle such defects due to two reasons. First, the standard rendering strategy relies on point-point mapping, which could lead to dramatic disparities between the visible and occluded areas of the body. Second, the naive direct regression approach does not consider any feasibility criteria (ie, prior information) for rendering under occlusions. To tackle the above drawbacks, we present OccNeRF, a neural rendering method that achieves better rendering of humans in severely occluded scenes. As direct solutions to the two drawbacks, we propose surface-based rendering by integrating geometry and visibility priors. We validate our method on both simulated and real-world occlusions and demonstrate our method's superiority. ### Constructing Holistic Spatio-Temporal Scene Graph for Video Semantic Role Labeling - **Authors:** Yu Zhao, Hao Fei, Yixin Cao, Bobo Li, Meishan Zhang, Jianguo Wei, Min Zhang, Tat-Seng Chua - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL) - **Arxiv link:** https://arxiv.org/abs/2308.05081 - **Pdf link:** https://arxiv.org/pdf/2308.05081 - **Abstract** Video Semantic Role Labeling (VidSRL) aims to detect the salient events from given videos, by recognizing the predict-argument event structures and the interrelationships between events. While recent endeavors have put forth methods for VidSRL, they can be mostly subject to two key drawbacks, including the lack of fine-grained spatial scene perception and the insufficiently modeling of video temporality. Towards this end, this work explores a novel holistic spatio-temporal scene graph (namely HostSG) representation based on the existing dynamic scene graph structures, which well model both the fine-grained spatial semantics and temporal dynamics of videos for VidSRL. Built upon the HostSG, we present a nichetargeting VidSRL framework. A scene-event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high-level event semantic structure, resulting in an overall hierarchical scene-event (termed ICE) graph structure. We further perform iterative structure refinement to optimize the ICE graph, such that the overall structure representation can best coincide with end task demand. Finally, three subtask predictions of VidSRL are jointly decoded, where the end-to-end paradigm effectively avoids error propagation. On the benchmark dataset, our framework boosts significantly over the current best-performing model. Further analyses are shown for a better understanding of the advances of our methods. ## Keyword: ISP ### Rendering Humans from Object-Occluded Monocular Videos - **Authors:** Tiange Xiang, Adam Sun, Jiajun Wu, Ehsan Adeli, Li Fei-Fei - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.04622 - **Pdf link:** https://arxiv.org/pdf/2308.04622 - **Abstract** 3D understanding and rendering of moving humans from monocular videos is a challenging task. Despite recent progress, the task remains difficult in real-world scenarios, where obstacles may block the camera view and cause partial occlusions in the captured videos. Existing methods cannot handle such defects due to two reasons. First, the standard rendering strategy relies on point-point mapping, which could lead to dramatic disparities between the visible and occluded areas of the body. Second, the naive direct regression approach does not consider any feasibility criteria (ie, prior information) for rendering under occlusions. To tackle the above drawbacks, we present OccNeRF, a neural rendering method that achieves better rendering of humans in severely occluded scenes. As direct solutions to the two drawbacks, we propose surface-based rendering by integrating geometry and visibility priors. We validate our method on both simulated and real-world occlusions and demonstrate our method's superiority. ### Addressing Racial Bias in Facial Emotion Recognition - **Authors:** Alex Fan, Xingshuo Xiao, Peter Washington - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Computers and Society (cs.CY) - **Arxiv link:** https://arxiv.org/abs/2308.04674 - **Pdf link:** https://arxiv.org/pdf/2308.04674 - **Abstract** Fairness in deep learning models trained with high-dimensional inputs and subjective labels remains a complex and understudied area. Facial emotion recognition, a domain where datasets are often racially imbalanced, can lead to models that yield disparate outcomes across racial groups. This study focuses on analyzing racial bias by sub-sampling training sets with varied racial distributions and assessing test performance across these simulations. Our findings indicate that smaller datasets with posed faces improve on both fairness and performance metrics as the simulations approach racial balance. Notably, the F1-score increases by $27.2\%$ points, and demographic parity increases by $15.7\%$ points on average across the simulations. However, in larger datasets with greater facial variation, fairness metrics generally remain constant, suggesting that racial balance by itself is insufficient to achieve parity in test performance across different racial groups. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### Resource Constrained Model Compression via Minimax Optimization for Spiking Neural Networks - **Authors:** Jue Chen, Huan Yuan, Jianchao Tan, Bin Chen, Chengru Song, Di Zhang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Multimedia (cs.MM) - **Arxiv link:** https://arxiv.org/abs/2308.04672 - **Pdf link:** https://arxiv.org/pdf/2308.04672 - **Abstract** Brain-inspired Spiking Neural Networks (SNNs) have the characteristics of event-driven and high energy-efficient, which are different from traditional Artificial Neural Networks (ANNs) when deployed on edge devices such as neuromorphic chips. Most previous work focuses on SNNs training strategies to improve model performance and brings larger and deeper network architectures. It is difficult to deploy these complex networks on resource-limited edge devices directly. To meet such demand, people compress SNNs very cautiously to balance the performance and the computation efficiency. Existing compression methods either iteratively pruned SNNs using weights norm magnitude or formulated the problem as a sparse learning optimization. We propose an improved end-to-end Minimax optimization method for this sparse learning problem to better balance the model performance and the computation efficiency. We also demonstrate that jointly applying compression and finetuning on SNNs is better than sequentially, especially for extreme compression ratios. The compressed SNN models achieved state-of-the-art (SOTA) performance on various benchmark datasets and architectures. Our code is available at https://github.com/chenjallen/Resource-Constrained-Compression-on-SNN. ## Keyword: RAW ### Rendering Humans from Object-Occluded Monocular Videos - **Authors:** Tiange Xiang, Adam Sun, Jiajun Wu, Ehsan Adeli, Li Fei-Fei - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.04622 - **Pdf link:** https://arxiv.org/pdf/2308.04622 - **Abstract** 3D understanding and rendering of moving humans from monocular videos is a challenging task. Despite recent progress, the task remains difficult in real-world scenarios, where obstacles may block the camera view and cause partial occlusions in the captured videos. Existing methods cannot handle such defects due to two reasons. First, the standard rendering strategy relies on point-point mapping, which could lead to dramatic disparities between the visible and occluded areas of the body. Second, the naive direct regression approach does not consider any feasibility criteria (ie, prior information) for rendering under occlusions. To tackle the above drawbacks, we present OccNeRF, a neural rendering method that achieves better rendering of humans in severely occluded scenes. As direct solutions to the two drawbacks, we propose surface-based rendering by integrating geometry and visibility priors. We validate our method on both simulated and real-world occlusions and demonstrate our method's superiority. ### SAfER: Layer-Level Sensitivity Assessment for Efficient and Robust Neural Network Inference - **Authors:** Edouard Yvinec, Arnaud Dapogny, Kevin Bailly - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.04753 - **Pdf link:** https://arxiv.org/pdf/2308.04753 - **Abstract** Deep neural networks (DNNs) demonstrate outstanding performance across most computer vision tasks. Some critical applications, such as autonomous driving or medical imaging, also require investigation into their behavior and the reasons behind the decisions they make. In this vein, DNN attribution consists in studying the relationship between the predictions of a DNN and its inputs. Attribution methods have been adapted to highlight the most relevant weights or neurons in a DNN, allowing to more efficiently select which weights or neurons can be pruned. However, a limitation of these approaches is that weights are typically compared within each layer separately, while some layers might appear as more critical than others. In this work, we propose to investigate DNN layer importance, i.e. to estimate the sensitivity of the accuracy w.r.t. perturbations applied at the layer level. To do so, we propose a novel dataset to evaluate our method as well as future works. We benchmark a number of criteria and draw conclusions regarding how to assess DNN layer importance and, consequently, how to budgetize layers for increased DNN efficiency (with applications for DNN pruning and quantization), as well as robustness to hardware failure (e.g. bit swaps). ### View while Moving: Efficient Video Recognition in Long-untrimmed Videos - **Authors:** Ye Tian, Mengyu Yang, Lanshan Zhang, Zhizhen Zhang, Yang Liu, Xiaohui Xie, Xirong Que, Wendong Wang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.04834 - **Pdf link:** https://arxiv.org/pdf/2308.04834 - **Abstract** Recent adaptive methods for efficient video recognition mostly follow the two-stage paradigm of "preview-then-recognition" and have achieved great success on multiple video benchmarks. However, this two-stage paradigm involves two visits of raw frames from coarse-grained to fine-grained during inference (cannot be parallelized), and the captured spatiotemporal features cannot be reused in the second stage (due to varying granularity), being not friendly to efficiency and computation optimization. To this end, inspired by human cognition, we propose a novel recognition paradigm of "View while Moving" for efficient long-untrimmed video recognition. In contrast to the two-stage paradigm, our paradigm only needs to access the raw frame once. The two phases of coarse-grained sampling and fine-grained recognition are combined into unified spatiotemporal modeling, showing great performance. Moreover, we investigate the properties of semantic units in video and propose a hierarchical mechanism to efficiently capture and reason about the unit-level and video-level temporal semantics in long-untrimmed videos respectively. Extensive experiments on both long-untrimmed and short-trimmed videos demonstrate that our approach outperforms state-of-the-art methods in terms of accuracy as well as efficiency, yielding new efficiency and accuracy trade-offs for video spatiotemporal modeling. ### Constructing Holistic Spatio-Temporal Scene Graph for Video Semantic Role Labeling - **Authors:** Yu Zhao, Hao Fei, Yixin Cao, Bobo Li, Meishan Zhang, Jianguo Wei, Min Zhang, Tat-Seng Chua - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL) - **Arxiv link:** https://arxiv.org/abs/2308.05081 - **Pdf link:** https://arxiv.org/pdf/2308.05081 - **Abstract** Video Semantic Role Labeling (VidSRL) aims to detect the salient events from given videos, by recognizing the predict-argument event structures and the interrelationships between events. While recent endeavors have put forth methods for VidSRL, they can be mostly subject to two key drawbacks, including the lack of fine-grained spatial scene perception and the insufficiently modeling of video temporality. Towards this end, this work explores a novel holistic spatio-temporal scene graph (namely HostSG) representation based on the existing dynamic scene graph structures, which well model both the fine-grained spatial semantics and temporal dynamics of videos for VidSRL. Built upon the HostSG, we present a nichetargeting VidSRL framework. A scene-event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high-level event semantic structure, resulting in an overall hierarchical scene-event (termed ICE) graph structure. We further perform iterative structure refinement to optimize the ICE graph, such that the overall structure representation can best coincide with end task demand. Finally, three subtask predictions of VidSRL are jointly decoded, where the end-to-end paradigm effectively avoids error propagation. On the benchmark dataset, our framework boosts significantly over the current best-performing model. Further analyses are shown for a better understanding of the advances of our methods. ## Keyword: raw image There is no result
process
new submissions for thu aug keyword events constructing holistic spatio temporal scene graph for video semantic role labeling authors yu zhao hao fei yixin cao bobo li meishan zhang jianguo wei min zhang tat seng chua subjects computer vision and pattern recognition cs cv computation and language cs cl arxiv link pdf link abstract video semantic role labeling vidsrl aims to detect the salient events from given videos by recognizing the predict argument event structures and the interrelationships between events while recent endeavors have put forth methods for vidsrl they can be mostly subject to two key drawbacks including the lack of fine grained spatial scene perception and the insufficiently modeling of video temporality towards this end this work explores a novel holistic spatio temporal scene graph namely hostsg representation based on the existing dynamic scene graph structures which well model both the fine grained spatial semantics and temporal dynamics of videos for vidsrl built upon the hostsg we present a nichetargeting vidsrl framework a scene event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high level event semantic structure resulting in an overall hierarchical scene event termed ice graph structure we further perform iterative structure refinement to optimize the ice graph such that the overall structure representation can best coincide with end task demand finally three subtask predictions of vidsrl are jointly decoded where the end to end paradigm effectively avoids error propagation on the benchmark dataset our framework boosts significantly over the current best performing model further analyses are shown for a better understanding of the advances of our methods keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb rendering humans from object occluded monocular videos authors tiange xiang adam sun jiajun wu ehsan adeli li fei fei subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract understanding and rendering of moving humans from monocular videos is a challenging task despite recent progress the task remains difficult in real world scenarios where obstacles may block the camera view and cause partial occlusions in the captured videos existing methods cannot handle such defects due to two reasons first the standard rendering strategy relies on point point mapping which could lead to dramatic disparities between the visible and occluded areas of the body second the naive direct regression approach does not consider any feasibility criteria ie prior information for rendering under occlusions to tackle the above drawbacks we present occnerf a neural rendering method that achieves better rendering of humans in severely occluded scenes as direct solutions to the two drawbacks we propose surface based rendering by integrating geometry and visibility priors we validate our method on both simulated and real world occlusions and demonstrate our method s superiority constructing holistic spatio temporal scene graph for video semantic role labeling authors yu zhao hao fei yixin cao bobo li meishan zhang jianguo wei min zhang tat seng chua subjects computer vision and pattern recognition cs cv computation and language cs cl arxiv link pdf link abstract video semantic role labeling vidsrl aims to detect the salient events from given videos by recognizing the predict argument event structures and the interrelationships between events while recent endeavors have put forth methods for vidsrl they can be mostly subject to two key drawbacks including the lack of fine grained spatial scene perception and the insufficiently modeling of video temporality towards this end this work explores a novel holistic spatio temporal scene graph namely hostsg representation based on the existing dynamic scene graph structures which well model both the fine grained spatial semantics and temporal dynamics of videos for vidsrl built upon the hostsg we present a nichetargeting vidsrl framework a scene event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high level event semantic structure resulting in an overall hierarchical scene event termed ice graph structure we further perform iterative structure refinement to optimize the ice graph such that the overall structure representation can best coincide with end task demand finally three subtask predictions of vidsrl are jointly decoded where the end to end paradigm effectively avoids error propagation on the benchmark dataset our framework boosts significantly over the current best performing model further analyses are shown for a better understanding of the advances of our methods keyword isp rendering humans from object occluded monocular videos authors tiange xiang adam sun jiajun wu ehsan adeli li fei fei subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract understanding and rendering of moving humans from monocular videos is a challenging task despite recent progress the task remains difficult in real world scenarios where obstacles may block the camera view and cause partial occlusions in the captured videos existing methods cannot handle such defects due to two reasons first the standard rendering strategy relies on point point mapping which could lead to dramatic disparities between the visible and occluded areas of the body second the naive direct regression approach does not consider any feasibility criteria ie prior information for rendering under occlusions to tackle the above drawbacks we present occnerf a neural rendering method that achieves better rendering of humans in severely occluded scenes as direct solutions to the two drawbacks we propose surface based rendering by integrating geometry and visibility priors we validate our method on both simulated and real world occlusions and demonstrate our method s superiority addressing racial bias in facial emotion recognition authors alex fan xingshuo xiao peter washington subjects computer vision and pattern recognition cs cv artificial intelligence cs ai computers and society cs cy arxiv link pdf link abstract fairness in deep learning models trained with high dimensional inputs and subjective labels remains a complex and understudied area facial emotion recognition a domain where datasets are often racially imbalanced can lead to models that yield disparate outcomes across racial groups this study focuses on analyzing racial bias by sub sampling training sets with varied racial distributions and assessing test performance across these simulations our findings indicate that smaller datasets with posed faces improve on both fairness and performance metrics as the simulations approach racial balance notably the score increases by points and demographic parity increases by points on average across the simulations however in larger datasets with greater facial variation fairness metrics generally remain constant suggesting that racial balance by itself is insufficient to achieve parity in test performance across different racial groups keyword image signal processing there is no result keyword image signal process there is no result keyword compression resource constrained model compression via minimax optimization for spiking neural networks authors jue chen huan yuan jianchao tan bin chen chengru song di zhang subjects computer vision and pattern recognition cs cv artificial intelligence cs ai multimedia cs mm arxiv link pdf link abstract brain inspired spiking neural networks snns have the characteristics of event driven and high energy efficient which are different from traditional artificial neural networks anns when deployed on edge devices such as neuromorphic chips most previous work focuses on snns training strategies to improve model performance and brings larger and deeper network architectures it is difficult to deploy these complex networks on resource limited edge devices directly to meet such demand people compress snns very cautiously to balance the performance and the computation efficiency existing compression methods either iteratively pruned snns using weights norm magnitude or formulated the problem as a sparse learning optimization we propose an improved end to end minimax optimization method for this sparse learning problem to better balance the model performance and the computation efficiency we also demonstrate that jointly applying compression and finetuning on snns is better than sequentially especially for extreme compression ratios the compressed snn models achieved state of the art sota performance on various benchmark datasets and architectures our code is available at keyword raw rendering humans from object occluded monocular videos authors tiange xiang adam sun jiajun wu ehsan adeli li fei fei subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract understanding and rendering of moving humans from monocular videos is a challenging task despite recent progress the task remains difficult in real world scenarios where obstacles may block the camera view and cause partial occlusions in the captured videos existing methods cannot handle such defects due to two reasons first the standard rendering strategy relies on point point mapping which could lead to dramatic disparities between the visible and occluded areas of the body second the naive direct regression approach does not consider any feasibility criteria ie prior information for rendering under occlusions to tackle the above drawbacks we present occnerf a neural rendering method that achieves better rendering of humans in severely occluded scenes as direct solutions to the two drawbacks we propose surface based rendering by integrating geometry and visibility priors we validate our method on both simulated and real world occlusions and demonstrate our method s superiority safer layer level sensitivity assessment for efficient and robust neural network inference authors edouard yvinec arnaud dapogny kevin bailly subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract deep neural networks dnns demonstrate outstanding performance across most computer vision tasks some critical applications such as autonomous driving or medical imaging also require investigation into their behavior and the reasons behind the decisions they make in this vein dnn attribution consists in studying the relationship between the predictions of a dnn and its inputs attribution methods have been adapted to highlight the most relevant weights or neurons in a dnn allowing to more efficiently select which weights or neurons can be pruned however a limitation of these approaches is that weights are typically compared within each layer separately while some layers might appear as more critical than others in this work we propose to investigate dnn layer importance i e to estimate the sensitivity of the accuracy w r t perturbations applied at the layer level to do so we propose a novel dataset to evaluate our method as well as future works we benchmark a number of criteria and draw conclusions regarding how to assess dnn layer importance and consequently how to budgetize layers for increased dnn efficiency with applications for dnn pruning and quantization as well as robustness to hardware failure e g bit swaps view while moving efficient video recognition in long untrimmed videos authors ye tian mengyu yang lanshan zhang zhizhen zhang yang liu xiaohui xie xirong que wendong wang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract recent adaptive methods for efficient video recognition mostly follow the two stage paradigm of preview then recognition and have achieved great success on multiple video benchmarks however this two stage paradigm involves two visits of raw frames from coarse grained to fine grained during inference cannot be parallelized and the captured spatiotemporal features cannot be reused in the second stage due to varying granularity being not friendly to efficiency and computation optimization to this end inspired by human cognition we propose a novel recognition paradigm of view while moving for efficient long untrimmed video recognition in contrast to the two stage paradigm our paradigm only needs to access the raw frame once the two phases of coarse grained sampling and fine grained recognition are combined into unified spatiotemporal modeling showing great performance moreover we investigate the properties of semantic units in video and propose a hierarchical mechanism to efficiently capture and reason about the unit level and video level temporal semantics in long untrimmed videos respectively extensive experiments on both long untrimmed and short trimmed videos demonstrate that our approach outperforms state of the art methods in terms of accuracy as well as efficiency yielding new efficiency and accuracy trade offs for video spatiotemporal modeling constructing holistic spatio temporal scene graph for video semantic role labeling authors yu zhao hao fei yixin cao bobo li meishan zhang jianguo wei min zhang tat seng chua subjects computer vision and pattern recognition cs cv computation and language cs cl arxiv link pdf link abstract video semantic role labeling vidsrl aims to detect the salient events from given videos by recognizing the predict argument event structures and the interrelationships between events while recent endeavors have put forth methods for vidsrl they can be mostly subject to two key drawbacks including the lack of fine grained spatial scene perception and the insufficiently modeling of video temporality towards this end this work explores a novel holistic spatio temporal scene graph namely hostsg representation based on the existing dynamic scene graph structures which well model both the fine grained spatial semantics and temporal dynamics of videos for vidsrl built upon the hostsg we present a nichetargeting vidsrl framework a scene event mapping mechanism is first designed to bridge the gap between the underlying scene structure and the high level event semantic structure resulting in an overall hierarchical scene event termed ice graph structure we further perform iterative structure refinement to optimize the ice graph such that the overall structure representation can best coincide with end task demand finally three subtask predictions of vidsrl are jointly decoded where the end to end paradigm effectively avoids error propagation on the benchmark dataset our framework boosts significantly over the current best performing model further analyses are shown for a better understanding of the advances of our methods keyword raw image there is no result
1
521
2,994,323,890
IssuesEvent
2015-07-22 11:00:03
genomizer/genomizer-server
https://api.github.com/repos/genomizer/genomizer-server
closed
genomizer-server-tester fails nondeterministically on Travis
bug Medium priority Processing
Sometimes I see nondeterministic single-test failures in `genomizer-server-tester` tests on Travis ([example](https://travis-ci.org/genomizer/genomizer-server/builds/65365650)) that go away after restarting the build. May be a race condition somewhere - if the issue persists, it can be temporarily papered over by ignoring `genomizer-server-tester`'s exit code. Won't have time to investigate, so it's just FYI.
1.0
genomizer-server-tester fails nondeterministically on Travis - Sometimes I see nondeterministic single-test failures in `genomizer-server-tester` tests on Travis ([example](https://travis-ci.org/genomizer/genomizer-server/builds/65365650)) that go away after restarting the build. May be a race condition somewhere - if the issue persists, it can be temporarily papered over by ignoring `genomizer-server-tester`'s exit code. Won't have time to investigate, so it's just FYI.
process
genomizer server tester fails nondeterministically on travis sometimes i see nondeterministic single test failures in genomizer server tester tests on travis that go away after restarting the build may be a race condition somewhere if the issue persists it can be temporarily papered over by ignoring genomizer server tester s exit code won t have time to investigate so it s just fyi
1
7,753
10,866,673,790
IssuesEvent
2019-11-14 21:48:14
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[FEATURE] Add feature filter algorithm
3.2 Automatic new feature Processing Alg
Original commit: https://github.com/qgis/QGIS/commit/147f6e71c46a5f1fd3d5b0da97f0a7028568552f by m-kuhn Adds a new "feature filter" algorithm. This algorithm takes a list of expressions and creates an output for each of them. Matching features from the input layer are sent to the outputs. This allows for an algorithm to only handle a subset of incoming features or to send different features to different outputs. The algorithm is only available in the modeler.
1.0
[FEATURE] Add feature filter algorithm - Original commit: https://github.com/qgis/QGIS/commit/147f6e71c46a5f1fd3d5b0da97f0a7028568552f by m-kuhn Adds a new "feature filter" algorithm. This algorithm takes a list of expressions and creates an output for each of them. Matching features from the input layer are sent to the outputs. This allows for an algorithm to only handle a subset of incoming features or to send different features to different outputs. The algorithm is only available in the modeler.
process
add feature filter algorithm original commit by m kuhn adds a new feature filter algorithm this algorithm takes a list of expressions and creates an output for each of them matching features from the input layer are sent to the outputs this allows for an algorithm to only handle a subset of incoming features or to send different features to different outputs the algorithm is only available in the modeler
1
13,814
16,576,213,168
IssuesEvent
2021-05-31 05:24:56
arcus-azure/arcus.messaging
https://api.github.com/repos/arcus-azure/arcus.messaging
closed
Add ContentType to AzureServiceBusSystemProperties (Arcus.Messaging.Abstractions.ServiceBus)
area:message-processing enhancement
**Is your feature request related to a problem? Please describe.** Currently we are trying to make a distinction between messages based on the contentType but unfortunately, arcus doesn't support this property. **Describe the solution you'd like** In AzureServiceBusSystemProperties add the following property : ContentType. This is already a property from ServiceBusReceivedMessage but is currently not getting used ![image](https://user-images.githubusercontent.com/79199130/119816699-4d662b00-beed-11eb-8303-356fc6e9488b.png)
1.0
Add ContentType to AzureServiceBusSystemProperties (Arcus.Messaging.Abstractions.ServiceBus) - **Is your feature request related to a problem? Please describe.** Currently we are trying to make a distinction between messages based on the contentType but unfortunately, arcus doesn't support this property. **Describe the solution you'd like** In AzureServiceBusSystemProperties add the following property : ContentType. This is already a property from ServiceBusReceivedMessage but is currently not getting used ![image](https://user-images.githubusercontent.com/79199130/119816699-4d662b00-beed-11eb-8303-356fc6e9488b.png)
process
add contenttype to azureservicebussystemproperties arcus messaging abstractions servicebus is your feature request related to a problem please describe currently we are trying to make a distinction between messages based on the contenttype but unfortunately arcus doesn t support this property describe the solution you d like in azureservicebussystemproperties add the following property contenttype this is already a property from servicebusreceivedmessage but is currently not getting used
1
371
2,499,901,565
IssuesEvent
2015-01-08 07:42:41
fossology/fossology
https://api.github.com/repos/fossology/fossology
opened
Binary analysis for open source discovery
Component: Rank Component: Tester Priority: Low Status: New Tracker: Feature
--- Author Name: **Paul Holland** Original Redmine Issue: 2457, http://www.fossology.org/issues/2457 Original Date: 2012/06/15 --- None
1.0
Binary analysis for open source discovery - --- Author Name: **Paul Holland** Original Redmine Issue: 2457, http://www.fossology.org/issues/2457 Original Date: 2012/06/15 --- None
non_process
binary analysis for open source discovery author name paul holland original redmine issue original date none
0
102,105
12,745,233,297
IssuesEvent
2020-06-26 13:55:21
magento/magento2
https://api.github.com/repos/magento/magento2
closed
Shipping Title translation is not being translated when you switch the langauge in Checkout Cart Page
Area: Design/Frontend Component: Checkout Component: Shipping Issue: Confirmed Issue: Format is valid Issue: Ready for Work Progress: dev in progress Reproduced on 2.4.x
<!--- Please review our guidelines before adding a new issue: https://github.com/magento/magento2/wiki/Issue-reporting-guidelines Fields marked with (*) are required. Please don't remove the template. --> ### Preconditions (*) <!--- Provide the exact Magento version (example: 2.3.2) and any important information on the environment where bug is reproducible. --> 1. Magneot 2.3.3 ### Steps to reproduce (*) <!--- Important: Provide a set of clear steps to reproduce this bug. We can not provide support without clear instructions on how to reproduce. --> 1. Add Two Store views 2. Change the Label for Free Shipping method in the second store 3. add a new product to be able to order it 4. disable multi shipping address 5. add the product to cart 6. go to the Checkout cart page 7. switch the languages 8. look at the shipping label in the right sidebar ### Expected result (*) <!--- Tell us what do you expect to happen. --> 1. when we switch the language we have to see the correct label per store view ### Actual result (*) <!--- Tell us what happened instead. Include error messages and issues. --> 1. we see don't see the correct language and we see the other language , when we change any information in the estimation shipping cost we got the correct label
1.0
Shipping Title translation is not being translated when you switch the langauge in Checkout Cart Page - <!--- Please review our guidelines before adding a new issue: https://github.com/magento/magento2/wiki/Issue-reporting-guidelines Fields marked with (*) are required. Please don't remove the template. --> ### Preconditions (*) <!--- Provide the exact Magento version (example: 2.3.2) and any important information on the environment where bug is reproducible. --> 1. Magneot 2.3.3 ### Steps to reproduce (*) <!--- Important: Provide a set of clear steps to reproduce this bug. We can not provide support without clear instructions on how to reproduce. --> 1. Add Two Store views 2. Change the Label for Free Shipping method in the second store 3. add a new product to be able to order it 4. disable multi shipping address 5. add the product to cart 6. go to the Checkout cart page 7. switch the languages 8. look at the shipping label in the right sidebar ### Expected result (*) <!--- Tell us what do you expect to happen. --> 1. when we switch the language we have to see the correct label per store view ### Actual result (*) <!--- Tell us what happened instead. Include error messages and issues. --> 1. we see don't see the correct language and we see the other language , when we change any information in the estimation shipping cost we got the correct label
non_process
shipping title translation is not being translated when you switch the langauge in checkout cart page please review our guidelines before adding a new issue fields marked with are required please don t remove the template preconditions provide the exact magento version example and any important information on the environment where bug is reproducible magneot steps to reproduce important provide a set of clear steps to reproduce this bug we can not provide support without clear instructions on how to reproduce add two store views change the label for free shipping method in the second store add a new product to be able to order it disable multi shipping address add the product to cart go to the checkout cart page switch the languages look at the shipping label in the right sidebar expected result when we switch the language we have to see the correct label per store view actual result we see don t see the correct language and we see the other language when we change any information in the estimation shipping cost we got the correct label
0
83,633
24,107,857,969
IssuesEvent
2022-09-20 08:56:47
openvinotoolkit/openvino
https://api.github.com/repos/openvinotoolkit/openvino
closed
Build error when build python api from resource
bug category: build support_request
##### System information (version) <!-- Example - OpenVINO => 2022.1 - Operating System / Platform => Windows 64 Bit - Compiler => Visual Studio 2017 - Problem classification: Model Conversion - Framework: TensorFlow (if applicable) - Model name: ResNet50 (if applicable) --> - OpenVINO=> 2022.1 - Operating System / Platform => windows11 - Compiler => vs2022 - Problem classification =>build error ##### Detailed description Hi, when I want to build python api from resource follow https://github.com/openvinotoolkit/openvino/wiki/BuildingForWindows. I used `` cmake -DENABLE_TESTS=ON -DENABLE_FUNCTIONAL_TESTS=ON -DENABLE_PYTHON=ON -DPYTHON_EXECUTABLE="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\python.exe" -DPYTHON_LIBRARY="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\libs\python38.lib" -DPYTHON_INCLUDE_DIR="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\include" -DENABLE_OV_ONNX_FRONTEND=ON .. `` then `` cmake --build . --config Release -j 8 `` but there were some wrong : ``` C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(10,33): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend. vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(613,19): message : see previous definition of 'private: static FeStat FrontEndMockPy::m_stat' [C:\openvino\build\src\bindings\pyth on\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(10,33): error C2491: 'FrontEndMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mock \mock_py_frontend\openvino_mock_py_frontend.vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(11,38): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend. vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(429,22): message : see previous definition of 'private: static ModelStat InputModelMockPy::m_stat' [C:\openvino\build\src\bindings \python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(11,38): error C2491: 'InputModelMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mo ck\mock_py_frontend\openvino_mock_py_frontend.vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(12,33): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend. vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(91,22): message : see previous definition of 'private: static PlaceStat PlaceMockPy::m_stat' [C:\openvino\build\src\bindings\pytho n\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(12,33): error C2491: 'PlaceMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mock\mo ck_py_frontend\openvino_mock_py_frontend.vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(17,53): error C2491: 'GetAPIVersion': definition of dllimport function not allowed [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\o penvino_mock_py_frontend.vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(21,45): error C2491: 'GetFrontEndData': definition of dllimport function not allowed [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend \openvino_mock_py_frontend.vcxproj] ``` Any help? Thanks! ##### Steps to reproduce <!-- Describe your problem and steps you've done before you got to this point. to add code example fence it with triple backticks and optional file extension ```.cpp // C++ code example ``` or attach as .txt or .zip file --> ##### Issue submission checklist - [ ] I report the issue, it's not a question <!-- OpenVINO team works with support forum, Stack Overflow and other communities to discuss problems. Tickets with question without real issue statement will be closed. --> - [ ] I checked the problem with documentation, FAQ, open issues, Stack Overflow, etc and have not found solution <!-- Places to check: * OpenVINO documentation: https://docs.openvino.ai/ * OpenVINO forum: https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/bd-p/distribution-openvino-toolkit * OpenVINO issue tracker: https://github.com/openvinotoolkit/openvino/issues?q=is%3Aissue * Stack Overflow branch: https://stackoverflow.com/questions/tagged/openvino --> - [ ] There is reproducer code and related data files: images, videos, models, etc. <!-- The best reproducer -- test case for OpenVINO that we can add to the library. -->
1.0
Build error when build python api from resource - ##### System information (version) <!-- Example - OpenVINO => 2022.1 - Operating System / Platform => Windows 64 Bit - Compiler => Visual Studio 2017 - Problem classification: Model Conversion - Framework: TensorFlow (if applicable) - Model name: ResNet50 (if applicable) --> - OpenVINO=> 2022.1 - Operating System / Platform => windows11 - Compiler => vs2022 - Problem classification =>build error ##### Detailed description Hi, when I want to build python api from resource follow https://github.com/openvinotoolkit/openvino/wiki/BuildingForWindows. I used `` cmake -DENABLE_TESTS=ON -DENABLE_FUNCTIONAL_TESTS=ON -DENABLE_PYTHON=ON -DPYTHON_EXECUTABLE="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\python.exe" -DPYTHON_LIBRARY="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\libs\python38.lib" -DPYTHON_INCLUDE_DIR="C:\Users\Local_Admin\AppData\Local\Programs\Python\Python38\include" -DENABLE_OV_ONNX_FRONTEND=ON .. `` then `` cmake --build . --config Release -j 8 `` but there were some wrong : ``` C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(10,33): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend. vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(613,19): message : see previous definition of 'private: static FeStat FrontEndMockPy::m_stat' [C:\openvino\build\src\bindings\pyth on\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(10,33): error C2491: 'FrontEndMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mock \mock_py_frontend\openvino_mock_py_frontend.vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(11,38): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend. vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(429,22): message : see previous definition of 'private: static ModelStat InputModelMockPy::m_stat' [C:\openvino\build\src\bindings \python\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(11,38): error C2491: 'InputModelMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mo ck\mock_py_frontend\openvino_mock_py_frontend.vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(12,33): warning C4273: 'm_stat': inconsistent dll linkage [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\openvino_mock_py_frontend. vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\include\mock_py_frontend/mock_py_frontend.hpp(91,22): message : see previous definition of 'private: static PlaceStat PlaceMockPy::m_stat' [C:\openvino\build\src\bindings\pytho n\tests\mock\mock_py_frontend\openvino_mock_py_frontend.vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(12,33): error C2491: 'PlaceMockPy::m_stat': definition of dllimport static data member not allowed [C:\openvino\build\src\bindings\python\tests\mock\mo ck_py_frontend\openvino_mock_py_frontend.vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(17,53): error C2491: 'GetAPIVersion': definition of dllimport function not allowed [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend\o penvino_mock_py_frontend.vcxproj] C:\openvino\src\bindings\python\tests\mock\mock_py_frontend\src\mock_py_frontend.cpp(21,45): error C2491: 'GetFrontEndData': definition of dllimport function not allowed [C:\openvino\build\src\bindings\python\tests\mock\mock_py_frontend \openvino_mock_py_frontend.vcxproj] ``` Any help? Thanks! ##### Steps to reproduce <!-- Describe your problem and steps you've done before you got to this point. to add code example fence it with triple backticks and optional file extension ```.cpp // C++ code example ``` or attach as .txt or .zip file --> ##### Issue submission checklist - [ ] I report the issue, it's not a question <!-- OpenVINO team works with support forum, Stack Overflow and other communities to discuss problems. Tickets with question without real issue statement will be closed. --> - [ ] I checked the problem with documentation, FAQ, open issues, Stack Overflow, etc and have not found solution <!-- Places to check: * OpenVINO documentation: https://docs.openvino.ai/ * OpenVINO forum: https://community.intel.com/t5/Intel-Distribution-of-OpenVINO/bd-p/distribution-openvino-toolkit * OpenVINO issue tracker: https://github.com/openvinotoolkit/openvino/issues?q=is%3Aissue * Stack Overflow branch: https://stackoverflow.com/questions/tagged/openvino --> - [ ] There is reproducer code and related data files: images, videos, models, etc. <!-- The best reproducer -- test case for OpenVINO that we can add to the library. -->
non_process
build error when build python api from resource system information version example openvino operating system platform windows bit compiler visual studio problem classification model conversion framework tensorflow if applicable model name if applicable openvino operating system platform compiler problem classification build error detailed description hi when i want to build python api from resource follow i used cmake denable tests on denable functional tests on denable python on dpython executable c users local admin appdata local programs python python exe dpython library c users local admin appdata local programs python libs lib dpython include dir c users local admin appdata local programs python include denable ov onnx frontend on then cmake build config release j but there were some wrong c openvino src bindings python tests mock mock py frontend src mock py frontend cpp warning m stat inconsistent dll linkage c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend include mock py frontend mock py frontend hpp message see previous definition of private static festat frontendmockpy m stat c openvino build src bindings pyth on tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp error frontendmockpy m stat definition of dllimport static data member not allowed c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp warning m stat inconsistent dll linkage c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend include mock py frontend mock py frontend hpp message see previous definition of private static modelstat inputmodelmockpy m stat c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp error inputmodelmockpy m stat definition of dllimport static data member not allowed c openvino build src bindings python tests mo ck mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp warning m stat inconsistent dll linkage c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend include mock py frontend mock py frontend hpp message see previous definition of private static placestat placemockpy m stat c openvino build src bindings pytho n tests mock mock py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp error placemockpy m stat definition of dllimport static data member not allowed c openvino build src bindings python tests mock mo ck py frontend openvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp error getapiversion definition of dllimport function not allowed c openvino build src bindings python tests mock mock py frontend o penvino mock py frontend vcxproj c openvino src bindings python tests mock mock py frontend src mock py frontend cpp error getfrontenddata definition of dllimport function not allowed c openvino build src bindings python tests mock mock py frontend openvino mock py frontend vcxproj any help thanks steps to reproduce describe your problem and steps you ve done before you got to this point to add code example fence it with triple backticks and optional file extension cpp c code example or attach as txt or zip file issue submission checklist i report the issue it s not a question openvino team works with support forum stack overflow and other communities to discuss problems tickets with question without real issue statement will be closed i checked the problem with documentation faq open issues stack overflow etc and have not found solution places to check openvino documentation openvino forum openvino issue tracker stack overflow branch there is reproducer code and related data files images videos models etc the best reproducer test case for openvino that we can add to the library
0
31,512
7,375,323,920
IssuesEvent
2018-03-13 23:49:40
MicrosoftDocs/live-share
https://api.github.com/repos/MicrosoftDocs/live-share
closed
[VS Code] Sign-in failed.
area: identity and sign-in duplicate needs more info vscode
## Error: Sign-in failed. ## Steps to Reproduce: 1. Connect to NetMotion (VPN) 2. Open Visual Studio or Visual Studio Code. 3. Attempt to start a new sharing session (Visual Studio) or sign into Live Share (VS Code) 4. Visual Studio will immediately fail with the error: > [2018-03-09 22:02:37.396 VSIX I] Acquired an access token for the VSO account '<14:eb047f03>' > [2018-03-09 22:02:37.420 VSIX E] Failed to create a collaboration session. An internal error occurred. Please contact Microsoft Product Support Services. VS Code will get to 'Ready to Collaborate' page on browser, but will display 'Sign-in Failed' on actual instance of VS Code. Note: If I disable VPNing, which probably also disables use of my corporate proxy, everything works as expected. Note: My system currently works (on the VPN + Proxy) with Spotify, MS Teams, NuGet in Visual Studio, and Docker, which either natively support IE proxy settings or have their own setup for proxy details. Note: When I use bash shell on windows and CURL https://download.microsoft.com, it fails with `(56) Proxy CONNECT aborted`. Note: I have the HTTP_PROXY and HTTPS_PROXY environment variables set up with my corporate proxy url (url only, no credentials). ||Version Data| |-:|:-| |**extensionName**|VSLS| |**extensionVersion**|0.2.206| |**protocolVersion**|2.1| |**applicationName**|VSCode| |**applicationVersion**|1.21.0| |**platformName**|Windows| |**platformVersion**|10.0.14393|
1.0
[VS Code] Sign-in failed. - ## Error: Sign-in failed. ## Steps to Reproduce: 1. Connect to NetMotion (VPN) 2. Open Visual Studio or Visual Studio Code. 3. Attempt to start a new sharing session (Visual Studio) or sign into Live Share (VS Code) 4. Visual Studio will immediately fail with the error: > [2018-03-09 22:02:37.396 VSIX I] Acquired an access token for the VSO account '<14:eb047f03>' > [2018-03-09 22:02:37.420 VSIX E] Failed to create a collaboration session. An internal error occurred. Please contact Microsoft Product Support Services. VS Code will get to 'Ready to Collaborate' page on browser, but will display 'Sign-in Failed' on actual instance of VS Code. Note: If I disable VPNing, which probably also disables use of my corporate proxy, everything works as expected. Note: My system currently works (on the VPN + Proxy) with Spotify, MS Teams, NuGet in Visual Studio, and Docker, which either natively support IE proxy settings or have their own setup for proxy details. Note: When I use bash shell on windows and CURL https://download.microsoft.com, it fails with `(56) Proxy CONNECT aborted`. Note: I have the HTTP_PROXY and HTTPS_PROXY environment variables set up with my corporate proxy url (url only, no credentials). ||Version Data| |-:|:-| |**extensionName**|VSLS| |**extensionVersion**|0.2.206| |**protocolVersion**|2.1| |**applicationName**|VSCode| |**applicationVersion**|1.21.0| |**platformName**|Windows| |**platformVersion**|10.0.14393|
non_process
sign in failed error sign in failed steps to reproduce connect to netmotion vpn open visual studio or visual studio code attempt to start a new sharing session visual studio or sign into live share vs code visual studio will immediately fail with the error acquired an access token for the vso account failed to create a collaboration session an internal error occurred please contact microsoft product support services vs code will get to ready to collaborate page on browser but will display sign in failed on actual instance of vs code note if i disable vpning which probably also disables use of my corporate proxy everything works as expected note my system currently works on the vpn proxy with spotify ms teams nuget in visual studio and docker which either natively support ie proxy settings or have their own setup for proxy details note when i use bash shell on windows and curl it fails with proxy connect aborted note i have the http proxy and https proxy environment variables set up with my corporate proxy url url only no credentials version data extensionname vsls extensionversion protocolversion applicationname vscode applicationversion platformname windows platformversion
0
20,542
27,192,655,179
IssuesEvent
2023-02-20 00:05:19
cse442-at-ub/project_s23-team-infinity
https://api.github.com/repos/cse442-at-ub/project_s23-team-infinity
opened
Syntax Taskes
Processing Task
Task 1: Print Hello World in php Test 1: Check if Hello World was printed: Task 2: Loop and print number 1 to 10 in php Test 2: Check if all numbers have been printed Task 3: Created a function that when called with parameters of 2 numbers will add then compare if it is greater than 10 and print out yes or no Test 3.1: Input: 3 and 4 Output: no Test 3.2: Input: 5 and 6 Output: yes
1.0
Syntax Taskes - Task 1: Print Hello World in php Test 1: Check if Hello World was printed: Task 2: Loop and print number 1 to 10 in php Test 2: Check if all numbers have been printed Task 3: Created a function that when called with parameters of 2 numbers will add then compare if it is greater than 10 and print out yes or no Test 3.1: Input: 3 and 4 Output: no Test 3.2: Input: 5 and 6 Output: yes
process
syntax taskes task print hello world in php test check if hello world was printed task loop and print number to in php test check if all numbers have been printed task created a function that when called with parameters of numbers will add then compare if it is greater than and print out yes or no test input and output no test input and output yes
1
158,293
12,412,181,869
IssuesEvent
2020-05-22 10:02:37
aliasrobotics/RVD
https://api.github.com/repos/aliasrobotics/RVD
opened
(warning) %d in format string (no. 7) requires 'int' but the argument type is 'unsigned int'.
bug cppcheck static analysis testing triage
```yaml { "id": 1, "title": "(warning) %d in format string (no. 7) requires 'int' but the argument type is 'unsigned int'.", "type": "bug", "description": "[src/opencv3/3rdparty/libjasper/jpc_cs.c:568]: (warning) %d in format string (no. 7) requires 'int' but the argument type is 'unsigned int'.", "cwe": "None", "cve": "None", "keywords": [ "cppcheck", "static analysis", "testing", "triage", "bug" ], "system": "src/opencv3/3rdparty/libjasper/jpc_cs.c", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": "", "flaw": { "phase": "testing", "specificity": "N/A", "architectural-location": "N/A", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-05-22 (10:02)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-05-22 (10:02)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "", "reproducibility": "always", "trace": "", "reproduction": "See artifacts below (if available)", "reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_kinetic/-/jobs/563367426/artifacts/download" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
1.0
(warning) %d in format string (no. 7) requires 'int' but the argument type is 'unsigned int'. - ```yaml { "id": 1, "title": "(warning) %d in format string (no. 7) requires 'int' but the argument type is 'unsigned int'.", "type": "bug", "description": "[src/opencv3/3rdparty/libjasper/jpc_cs.c:568]: (warning) %d in format string (no. 7) requires 'int' but the argument type is 'unsigned int'.", "cwe": "None", "cve": "None", "keywords": [ "cppcheck", "static analysis", "testing", "triage", "bug" ], "system": "src/opencv3/3rdparty/libjasper/jpc_cs.c", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": "", "flaw": { "phase": "testing", "specificity": "N/A", "architectural-location": "N/A", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-05-22 (10:02)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-05-22 (10:02)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "", "reproducibility": "always", "trace": "", "reproduction": "See artifacts below (if available)", "reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_kinetic/-/jobs/563367426/artifacts/download" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
non_process
warning d in format string no requires int but the argument type is unsigned int yaml id title warning d in format string no requires int but the argument type is unsigned int type bug description warning d in format string no requires int but the argument type is unsigned int cwe none cve none keywords cppcheck static analysis testing triage bug system src libjasper jpc cs c vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity n a architectural location n a application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace reproduction see artifacts below if available reproduction image gitlab com aliasrobotics offensive alurity pipelines active pipeline ros kinetic jobs artifacts download exploitation description exploitation image exploitation vector mitigation description pull request date mitigation
0
6,778
9,915,064,612
IssuesEvent
2019-06-28 15:50:50
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Unable to edit the variables after deploying the Solution
assigned-to-author automation/svc process-automation/subsvc product-question triaged
Hi, I need to exclude the VM's for this solution and trying to edit External_ExcludeVMNames variable , im not getting any option to edit . --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 225c9d05-83dd-b006-0025-3753f5ab25bf * Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096 * Content: [Start/Stop VMs during off-hours solution](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management#solution-components) * Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @georgewallace * Microsoft Alias: **gwallace**
1.0
Unable to edit the variables after deploying the Solution - Hi, I need to exclude the VM's for this solution and trying to edit External_ExcludeVMNames variable , im not getting any option to edit . --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 225c9d05-83dd-b006-0025-3753f5ab25bf * Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096 * Content: [Start/Stop VMs during off-hours solution](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management#solution-components) * Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @georgewallace * Microsoft Alias: **gwallace**
process
unable to edit the variables after deploying the solution hi i need to exclude the vm s for this solution and trying to edit external excludevmnames variable im not getting any option to edit document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login georgewallace microsoft alias gwallace
1
773,025
27,143,097,000
IssuesEvent
2023-02-16 17:44:16
containrrr/watchtower
https://api.github.com/repos/containrrr/watchtower
opened
Non-label method for excluding containers
Type: Enhancement Priority: Low Status: Available
### Is your feature request related to a problem? Please describe. On Synology (and I assume other environments - UniFi comes to mind), there are often "private" containers that we don't control the creation of. As a consequence, watchtower tries to update those but fails. Since there's no way to update labels for those containers after they were created, and we don't create them, there's no good way to exclude them, short of ### Describe the solution you'd like I'd be happy with a `--exclude` or similar flag (possibly taking a regex?) to tell watchtower what to exclude - that's easy to update. ### Describe alternatives you've considered * Using `--label-enable` is harder to manage (requires managing labels for all the other containers) * Creating derived containers just to add the label is not really an option for a lot of these private containers, especially as some of them have their own update mechanism. * On Synology specifically, updating environment variables is easier (doesn't require recreating the container), so that's definitely a good option there * There could simply be a flag to silently ignore any containers that aren't on the registry (would solve the Synology docviewer case, might not solve the UniFi case where they create common containers too) * A file with the list of exclusions might also work ### Additional context Related to https://github.com/containrrr/watchtower/issues/923 .
1.0
Non-label method for excluding containers - ### Is your feature request related to a problem? Please describe. On Synology (and I assume other environments - UniFi comes to mind), there are often "private" containers that we don't control the creation of. As a consequence, watchtower tries to update those but fails. Since there's no way to update labels for those containers after they were created, and we don't create them, there's no good way to exclude them, short of ### Describe the solution you'd like I'd be happy with a `--exclude` or similar flag (possibly taking a regex?) to tell watchtower what to exclude - that's easy to update. ### Describe alternatives you've considered * Using `--label-enable` is harder to manage (requires managing labels for all the other containers) * Creating derived containers just to add the label is not really an option for a lot of these private containers, especially as some of them have their own update mechanism. * On Synology specifically, updating environment variables is easier (doesn't require recreating the container), so that's definitely a good option there * There could simply be a flag to silently ignore any containers that aren't on the registry (would solve the Synology docviewer case, might not solve the UniFi case where they create common containers too) * A file with the list of exclusions might also work ### Additional context Related to https://github.com/containrrr/watchtower/issues/923 .
non_process
non label method for excluding containers is your feature request related to a problem please describe on synology and i assume other environments unifi comes to mind there are often private containers that we don t control the creation of as a consequence watchtower tries to update those but fails since there s no way to update labels for those containers after they were created and we don t create them there s no good way to exclude them short of describe the solution you d like i d be happy with a exclude or similar flag possibly taking a regex to tell watchtower what to exclude that s easy to update describe alternatives you ve considered using label enable is harder to manage requires managing labels for all the other containers creating derived containers just to add the label is not really an option for a lot of these private containers especially as some of them have their own update mechanism on synology specifically updating environment variables is easier doesn t require recreating the container so that s definitely a good option there there could simply be a flag to silently ignore any containers that aren t on the registry would solve the synology docviewer case might not solve the unifi case where they create common containers too a file with the list of exclusions might also work additional context related to
0
78,861
3,518,417,793
IssuesEvent
2016-01-12 12:49:04
blue-systems/pm-img
https://api.github.com/repos/blue-systems/pm-img
closed
[XWayland]: Provide XWayland 1.18
high priority ready for QA
Please provide the latest XWayland server (1.18) in the image. It got released last week and brings important improvements for X11 applications like touch support.
1.0
[XWayland]: Provide XWayland 1.18 - Please provide the latest XWayland server (1.18) in the image. It got released last week and brings important improvements for X11 applications like touch support.
non_process
provide xwayland please provide the latest xwayland server in the image it got released last week and brings important improvements for applications like touch support
0
4,275
7,189,966,494
IssuesEvent
2018-02-02 15:44:40
parcel-bundler/parcel
https://api.github.com/repos/parcel-bundler/parcel
closed
🐛 Using type="text" on input in HTML gets stripped
#Question HTML Preprocessing
<!--- Thanks for filing an issue 😄 ! Before you submit, please read the following: Search open/closed issues before submitting since someone might have asked the same thing before! --> **Choose one:** is this a 🐛 bug report or 🙋 feature request? 🐛 <!--- Provide a general summary of the issue in the title above --> ### 🎛 Configuration (.babelrc, package.json, cli command) <!--- If describing a bug, tell us what your babel configuration looks like --> No config. Just a HTML5 Boilerplate with Parcel. ### 🤔 Expected Behavior <!--- If you're describing a bug, tell us what should happen --> <!--- If you're suggesting a change/improvement, tell us how it should work --> When running ```parcel build index.html --public-url ./```, any HTML <input type="text"> should not be altered. ### 😯 Current Behavior <!--- If describing a bug, tell us what happens instead of the expected behavior --> <!--- If you are seeing an error, please include the full error message and stack trace --> <!--- If suggesting a change/improvement, explain the difference from current behavior --> When running ```parcel build index.html --public-url ./```: ```<input type="text" name="FirstName" class="form-control" id="FirstName" placeholder="*Firstname" required>``` changes to ```<input name="FirstName" class="form-control" id="FirstName" placeholder="*Firstname" required="">``` ### 💁 Possible Solution Not sure. ### 🔦 Context Need type="text" to stay, especially when using CSS [type="text"]. ### 🌍 Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> | Software | Version(s) | | ---------------- | ---------- | | Parcel | 1.4.1 | Node | 8.5.0 | npm/Yarn | 5.3.0 | Operating System | macOS 10.13.2 <!-- Love parcel? Please consider supporting our collective: 👉 https://opencollective.com/parcel/donate -->
1.0
🐛 Using type="text" on input in HTML gets stripped - <!--- Thanks for filing an issue 😄 ! Before you submit, please read the following: Search open/closed issues before submitting since someone might have asked the same thing before! --> **Choose one:** is this a 🐛 bug report or 🙋 feature request? 🐛 <!--- Provide a general summary of the issue in the title above --> ### 🎛 Configuration (.babelrc, package.json, cli command) <!--- If describing a bug, tell us what your babel configuration looks like --> No config. Just a HTML5 Boilerplate with Parcel. ### 🤔 Expected Behavior <!--- If you're describing a bug, tell us what should happen --> <!--- If you're suggesting a change/improvement, tell us how it should work --> When running ```parcel build index.html --public-url ./```, any HTML <input type="text"> should not be altered. ### 😯 Current Behavior <!--- If describing a bug, tell us what happens instead of the expected behavior --> <!--- If you are seeing an error, please include the full error message and stack trace --> <!--- If suggesting a change/improvement, explain the difference from current behavior --> When running ```parcel build index.html --public-url ./```: ```<input type="text" name="FirstName" class="form-control" id="FirstName" placeholder="*Firstname" required>``` changes to ```<input name="FirstName" class="form-control" id="FirstName" placeholder="*Firstname" required="">``` ### 💁 Possible Solution Not sure. ### 🔦 Context Need type="text" to stay, especially when using CSS [type="text"]. ### 🌍 Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> | Software | Version(s) | | ---------------- | ---------- | | Parcel | 1.4.1 | Node | 8.5.0 | npm/Yarn | 5.3.0 | Operating System | macOS 10.13.2 <!-- Love parcel? Please consider supporting our collective: 👉 https://opencollective.com/parcel/donate -->
process
🐛 using type text on input in html gets stripped thanks for filing an issue 😄 before you submit please read the following search open closed issues before submitting since someone might have asked the same thing before choose one is this a 🐛 bug report or 🙋 feature request 🐛 🎛 configuration babelrc package json cli command no config just a boilerplate with parcel 🤔 expected behavior when running parcel build index html public url any html should not be altered 😯 current behavior when running parcel build index html public url changes to 💁 possible solution not sure 🔦 context need type text to stay especially when using css 🌍 your environment software version s parcel node npm yarn operating system macos love parcel please consider supporting our collective 👉
1
20,632
27,314,544,652
IssuesEvent
2023-02-24 14:45:35
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
closed
Possibility to provide a '*' to the definition of a dataset in a recipe
enhancement preprocessor
Hello everyone, this issue follows our discussion during the Is-ENES3 GA with @valeriupredoi With CliMAF we found very interesting to be able to specify a wildcard '*' to our dataset definitions, like for instance model='*', realization='*' to work on all the models or realizations available. In the same way, we implemented the possibility to specify period='last_XXY', 'first_XXY' or '*', with XX being a number of years, to retrieve the last XX, first XX years available, or the full period. Do you guys think you could consider adding this functionality? Cheers, J.
1.0
Possibility to provide a '*' to the definition of a dataset in a recipe - Hello everyone, this issue follows our discussion during the Is-ENES3 GA with @valeriupredoi With CliMAF we found very interesting to be able to specify a wildcard '*' to our dataset definitions, like for instance model='*', realization='*' to work on all the models or realizations available. In the same way, we implemented the possibility to specify period='last_XXY', 'first_XXY' or '*', with XX being a number of years, to retrieve the last XX, first XX years available, or the full period. Do you guys think you could consider adding this functionality? Cheers, J.
process
possibility to provide a to the definition of a dataset in a recipe hello everyone this issue follows our discussion during the is ga with valeriupredoi with climaf we found very interesting to be able to specify a wildcard to our dataset definitions like for instance model realization to work on all the models or realizations available in the same way we implemented the possibility to specify period last xxy first xxy or with xx being a number of years to retrieve the last xx first xx years available or the full period do you guys think you could consider adding this functionality cheers j
1
22,025
30,541,186,737
IssuesEvent
2023-07-19 21:34:11
tradingstrategy-ai/frontend
https://api.github.com/repos/tradingstrategy-ai/frontend
closed
Aave: Add lending reserve pages to sitemap.xml
enhancement process size: XS
# Goal As a Trading Strategy marketer, I want Aave lending reserve pages to be discoverable / indexable by search engines so consumers can find them via web search. # Details Add `/trading-view/lending-reserves` and all `/trading-view/polygon/lending/[protocol]/[reserve]` pages to `sitemap.xml`
1.0
Aave: Add lending reserve pages to sitemap.xml - # Goal As a Trading Strategy marketer, I want Aave lending reserve pages to be discoverable / indexable by search engines so consumers can find them via web search. # Details Add `/trading-view/lending-reserves` and all `/trading-view/polygon/lending/[protocol]/[reserve]` pages to `sitemap.xml`
process
aave add lending reserve pages to sitemap xml goal as a trading strategy marketer i want aave lending reserve pages to be discoverable indexable by search engines so consumers can find them via web search details add trading view lending reserves and all trading view polygon lending pages to sitemap xml
1
15,058
18,763,401,040
IssuesEvent
2021-11-05 19:27:01
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
closed
Available Heat Algorithm
Process Heating
Will need to do something things on the Desktop side to account for Suite Issue #623
1.0
Available Heat Algorithm - Will need to do something things on the Desktop side to account for Suite Issue #623
process
available heat algorithm will need to do something things on the desktop side to account for suite issue
1
137,610
18,755,107,429
IssuesEvent
2021-11-05 09:43:49
Dima2022/node-kms
https://api.github.com/repos/Dima2022/node-kms
opened
CVE-2020-7656 (Medium) detected in jquery-1.7.1.min.js
security vulnerability
## CVE-2020-7656 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to dependency file: node-kms/node_modules/vm-browserify/example/run/index.html</p> <p>Path to vulnerable library: /node_modules/vm-browserify/example/run/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Dima2022/node-kms/commit/0174d52c9bbe1d27d5a8e0359134e9ffdbe6db6f">0174d52c9bbe1d27d5a8e0359134e9ffdbe6db6f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed. <p>Publish Date: 2020-05-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-q4m3-2j7h-f7xw">https://github.com/advisories/GHSA-q4m3-2j7h-f7xw</a></p> <p>Release Date: 2020-05-28</p> <p>Fix Resolution: jquery - 1.9.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.7.1","packageFilePaths":["/node_modules/vm-browserify/example/run/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jquery - 1.9.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-7656","vulnerabilityDetails":"jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove \"\u003cscript\u003e\" HTML tags that contain a whitespace character, i.e: \"\u003c/script \u003e\", which results in the enclosed script logic to be executed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-7656 (Medium) detected in jquery-1.7.1.min.js - ## CVE-2020-7656 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to dependency file: node-kms/node_modules/vm-browserify/example/run/index.html</p> <p>Path to vulnerable library: /node_modules/vm-browserify/example/run/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Dima2022/node-kms/commit/0174d52c9bbe1d27d5a8e0359134e9ffdbe6db6f">0174d52c9bbe1d27d5a8e0359134e9ffdbe6db6f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed. <p>Publish Date: 2020-05-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-q4m3-2j7h-f7xw">https://github.com/advisories/GHSA-q4m3-2j7h-f7xw</a></p> <p>Release Date: 2020-05-28</p> <p>Fix Resolution: jquery - 1.9.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.7.1","packageFilePaths":["/node_modules/vm-browserify/example/run/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jquery - 1.9.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-7656","vulnerabilityDetails":"jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove \"\u003cscript\u003e\" HTML tags that contain a whitespace character, i.e: \"\u003c/script \u003e\", which results in the enclosed script logic to be executed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file node kms node modules vm browserify example run index html path to vulnerable library node modules vm browserify example run index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery prior to allows cross site scripting attacks via the load method the load method fails to recognize and remove html tags that contain a whitespace character i e which results in the enclosed script logic to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion jquery basebranches vulnerabilityidentifier cve vulnerabilitydetails jquery prior to allows cross site scripting attacks via the load method the load method fails to recognize and remove html tags that contain a whitespace character i e script which results in the enclosed script logic to be executed vulnerabilityurl
0
12,414
14,920,392,045
IssuesEvent
2021-01-23 04:29:03
e4exp/paper_manager_abstract
https://api.github.com/repos/e4exp/paper_manager_abstract
opened
The Pile: An 800GB Dataset of Diverse Text for Language Modeling
2020 Dataset Natural Language Processing
* https://arxiv.org/abs/2101.00027 * 2020 最近の研究では、学習データセットの多様性を高めることで、大規模言語モデルの一般的なクロスドメイン知識と下流の一般化能力が向上することが実証されている。 このことを念頭に置いて,我々は大規模言語モデルの学習を目的とした825GiBの英文コーパスであるtextit{the Pile}を提示する. このコーパスは,既存のものと新たに構築されたものの両方を含む22の多様で質の高いサブセットから構成されており,その多くは学術的・専門的な情報源に由来するものである. GPT-2とGPT-3のPile上での未調整の性能を評価したところ、これらのモデルはアカデミックライティングなどの多くの構成要素で苦戦していることがわかりました。 逆に、Pile上で訓練されたモデルは、Pileのすべての構成要素においてRaw CCとCC-100を大幅に上回り、下流の評価でのパフォーマンスを向上させました。 徹底的な探索的分析により、将来のユーザーのためにデータの潜在的な側面を文書化しています。その構築に使用されたコードを公開しています。
1.0
The Pile: An 800GB Dataset of Diverse Text for Language Modeling - * https://arxiv.org/abs/2101.00027 * 2020 最近の研究では、学習データセットの多様性を高めることで、大規模言語モデルの一般的なクロスドメイン知識と下流の一般化能力が向上することが実証されている。 このことを念頭に置いて,我々は大規模言語モデルの学習を目的とした825GiBの英文コーパスであるtextit{the Pile}を提示する. このコーパスは,既存のものと新たに構築されたものの両方を含む22の多様で質の高いサブセットから構成されており,その多くは学術的・専門的な情報源に由来するものである. GPT-2とGPT-3のPile上での未調整の性能を評価したところ、これらのモデルはアカデミックライティングなどの多くの構成要素で苦戦していることがわかりました。 逆に、Pile上で訓練されたモデルは、Pileのすべての構成要素においてRaw CCとCC-100を大幅に上回り、下流の評価でのパフォーマンスを向上させました。 徹底的な探索的分析により、将来のユーザーのためにデータの潜在的な側面を文書化しています。その構築に使用されたコードを公開しています。
process
the pile an dataset of diverse text for language modeling 最近の研究では、学習データセットの多様性を高めることで、大規模言語モデルの一般的なクロスドメイン知識と下流の一般化能力が向上することが実証されている。 このことを念頭に置いて, the pile を提示する. このコーパスは, ,その多くは学術的・専門的な情報源に由来するものである. gpt 、これらのモデルはアカデミックライティングなどの多くの構成要素で苦戦していることがわかりました。 逆に、pile上で訓練されたモデルは、pileのすべての構成要素においてraw ccとcc 、下流の評価でのパフォーマンスを向上させました。 徹底的な探索的分析により、将来のユーザーのためにデータの潜在的な側面を文書化しています。その構築に使用されたコードを公開しています。
1
273,833
29,831,099,222
IssuesEvent
2023-06-18 09:31:46
RG4421/ampere-centos-kernel
https://api.github.com/repos/RG4421/ampere-centos-kernel
closed
CVE-2020-15437 (Medium) detected in linuxv5.2 - autoclosed
Mend: dependency security vulnerability
## CVE-2020-15437 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/serial/8250/8250_core.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/serial/8250/8250_core.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/serial/8250/8250_core.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The Linux kernel before version 5.8 is vulnerable to a NULL pointer dereference in drivers/tty/serial/8250/8250_core.c:serial8250_isa_init_ports() that allows local users to cause a denial of service by using the p->serial_in pointer which uninitialized. Mend Note: After conducting further research, Mend has determined that versions v3.9-rc5 through v4.4.231, v4.5-rc1 through v4.9.231, v4.10-rc1 through v4.14.189, v4.15-rc1 through v4.19.134, v5.0-rc1 through v5.4.53, v5.5-rc1 through v5.7.10 and v5.8-rc1 through v5.8-rc6 of Linux Kernel are vulnerable to CVE-2020-15437. <p>Publish Date: 2020-11-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-15437>CVE-2020-15437</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-15437">https://www.linuxkernelcves.com/cves/CVE-2020-15437</a></p> <p>Release Date: 2020-11-23</p> <p>Fix Resolution: v4.4.232, v4.9.232, v4.14.190, v4.19.135, v5.4.54, v5.7.11, v5.8-rc7</p> </p> </details> <p></p>
True
CVE-2020-15437 (Medium) detected in linuxv5.2 - autoclosed - ## CVE-2020-15437 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/serial/8250/8250_core.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/serial/8250/8250_core.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/serial/8250/8250_core.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The Linux kernel before version 5.8 is vulnerable to a NULL pointer dereference in drivers/tty/serial/8250/8250_core.c:serial8250_isa_init_ports() that allows local users to cause a denial of service by using the p->serial_in pointer which uninitialized. Mend Note: After conducting further research, Mend has determined that versions v3.9-rc5 through v4.4.231, v4.5-rc1 through v4.9.231, v4.10-rc1 through v4.14.189, v4.15-rc1 through v4.19.134, v5.0-rc1 through v5.4.53, v5.5-rc1 through v5.7.10 and v5.8-rc1 through v5.8-rc6 of Linux Kernel are vulnerable to CVE-2020-15437. <p>Publish Date: 2020-11-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-15437>CVE-2020-15437</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-15437">https://www.linuxkernelcves.com/cves/CVE-2020-15437</a></p> <p>Release Date: 2020-11-23</p> <p>Fix Resolution: v4.4.232, v4.9.232, v4.14.190, v4.19.135, v5.4.54, v5.7.11, v5.8-rc7</p> </p> </details> <p></p>
non_process
cve medium detected in autoclosed cve medium severity vulnerability vulnerable library linux kernel source tree library home page a href found in base branch amp centos kernel vulnerable source files drivers tty serial core c drivers tty serial core c drivers tty serial core c vulnerability details the linux kernel before version is vulnerable to a null pointer dereference in drivers tty serial core c isa init ports that allows local users to cause a denial of service by using the p serial in pointer which uninitialized mend note after conducting further research mend has determined that versions through through through through through through and through of linux kernel are vulnerable to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
316,378
23,628,475,513
IssuesEvent
2022-08-25 07:14:15
iotaledger/inx-chronicle
https://api.github.com/repos/iotaledger/inx-chronicle
closed
Document MongoDB `transactionLifetimeLimitSeconds` configuration
documentation
When we call `ReadUnspentOutputs` the resulting transaction can take a really long time, depending on the amount of outputs. In some cases, the default timeout from MongoDB kills the transaction before completion. To work around this, users should configure their MongoDB instance with a larger timeout using the `transactionLifetimeLimitSeconds` option. We should document that behavior, provide an example on how to set the flag, and adapt our `docker-compose.yml` file accordingly.
1.0
Document MongoDB `transactionLifetimeLimitSeconds` configuration - When we call `ReadUnspentOutputs` the resulting transaction can take a really long time, depending on the amount of outputs. In some cases, the default timeout from MongoDB kills the transaction before completion. To work around this, users should configure their MongoDB instance with a larger timeout using the `transactionLifetimeLimitSeconds` option. We should document that behavior, provide an example on how to set the flag, and adapt our `docker-compose.yml` file accordingly.
non_process
document mongodb transactionlifetimelimitseconds configuration when we call readunspentoutputs the resulting transaction can take a really long time depending on the amount of outputs in some cases the default timeout from mongodb kills the transaction before completion to work around this users should configure their mongodb instance with a larger timeout using the transactionlifetimelimitseconds option we should document that behavior provide an example on how to set the flag and adapt our docker compose yml file accordingly
0
20,157
26,710,314,262
IssuesEvent
2023-01-27 22:47:33
openxla/stablehlo
https://api.github.com/repos/openxla/stablehlo
opened
Consider standardizing test files
Process
Currently, several of our tests do not follow a consistent style in the way the tests are written. Though not an exhaustive list some issues are: - No standard test naming convention - Inconsistent placement of CHECK labels - Inconsistent string we are comparing CHECK-NEXT directives on - Ordering of the tests - Non-pretty printed ops - Using default SSA values (e.g. %0, %1, etc.) - etc. Although #951 tackles this for future tests, this is not the case for already existing tests that is not part of this workflow. This ticket should track the progress of how consistent the tests are written.
1.0
Consider standardizing test files - Currently, several of our tests do not follow a consistent style in the way the tests are written. Though not an exhaustive list some issues are: - No standard test naming convention - Inconsistent placement of CHECK labels - Inconsistent string we are comparing CHECK-NEXT directives on - Ordering of the tests - Non-pretty printed ops - Using default SSA values (e.g. %0, %1, etc.) - etc. Although #951 tackles this for future tests, this is not the case for already existing tests that is not part of this workflow. This ticket should track the progress of how consistent the tests are written.
process
consider standardizing test files currently several of our tests do not follow a consistent style in the way the tests are written though not an exhaustive list some issues are no standard test naming convention inconsistent placement of check labels inconsistent string we are comparing check next directives on ordering of the tests non pretty printed ops using default ssa values e g etc etc although tackles this for future tests this is not the case for already existing tests that is not part of this workflow this ticket should track the progress of how consistent the tests are written
1
7,144
10,288,573,369
IssuesEvent
2019-08-27 10:43:30
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
The Runbook will Add the AZ modules but cannot remove the old AzureRM
Pri2 automation/svc cxp process-automation/subsvc product-question triaged
When a new Automation account is created it comes by pre-installed with AzureRM modules. This runbook will add the AZ modules no worries. But it does not remove the AzureRM. Neither you can delete the old AzureRm modules --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 55d92628-d657-759d-d301-06029cd61138 * Version Independent ID: ca381b85-2be5-24af-be18-00c13a7e53df * Content: [Update Azure modules in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/automation-update-azure-modules#feedback) * Content Source: [articles/automation/automation-update-azure-modules.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-update-azure-modules.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @bobbytreed * Microsoft Alias: **robreed**
1.0
The Runbook will Add the AZ modules but cannot remove the old AzureRM - When a new Automation account is created it comes by pre-installed with AzureRM modules. This runbook will add the AZ modules no worries. But it does not remove the AzureRM. Neither you can delete the old AzureRm modules --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 55d92628-d657-759d-d301-06029cd61138 * Version Independent ID: ca381b85-2be5-24af-be18-00c13a7e53df * Content: [Update Azure modules in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/automation-update-azure-modules#feedback) * Content Source: [articles/automation/automation-update-azure-modules.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-update-azure-modules.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @bobbytreed * Microsoft Alias: **robreed**
process
the runbook will add the az modules but cannot remove the old azurerm when a new automation account is created it comes by pre installed with azurerm modules this runbook will add the az modules no worries but it does not remove the azurerm neither you can delete the old azurerm modules document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login bobbytreed microsoft alias robreed
1
205,149
15,964,633,018
IssuesEvent
2021-04-16 06:34:25
Icelenaugust/pe
https://api.github.com/repos/Icelenaugust/pe
opened
Mismatch of diagram in UG findp
severity.VeryLow type.DocumentationBug
in UG the displayed message is 2 persons listed! ![Screenshot 2021-04-16 at 2.34.21 PM.png](https://raw.githubusercontent.com/Icelenaugust/pe/main/files/6223bd8a-79b3-4f24-9e7a-bfe2fae051a0.png) while in actual product, the displayed message is 2 person(s) listed! It might be a good habit to update UG! ![Screenshot 2021-04-16 at 2.34.10 PM.png](https://raw.githubusercontent.com/Icelenaugust/pe/main/files/cd0f85ce-45cb-43ab-bf91-cb90e6321618.png) <!--session: 1618551311074-1a994e8c-6641-4e86-94f8-3737348b1ccc-->
1.0
Mismatch of diagram in UG findp - in UG the displayed message is 2 persons listed! ![Screenshot 2021-04-16 at 2.34.21 PM.png](https://raw.githubusercontent.com/Icelenaugust/pe/main/files/6223bd8a-79b3-4f24-9e7a-bfe2fae051a0.png) while in actual product, the displayed message is 2 person(s) listed! It might be a good habit to update UG! ![Screenshot 2021-04-16 at 2.34.10 PM.png](https://raw.githubusercontent.com/Icelenaugust/pe/main/files/cd0f85ce-45cb-43ab-bf91-cb90e6321618.png) <!--session: 1618551311074-1a994e8c-6641-4e86-94f8-3737348b1ccc-->
non_process
mismatch of diagram in ug findp in ug the displayed message is persons listed while in actual product the displayed message is person s listed it might be a good habit to update ug
0
20,669
27,335,048,724
IssuesEvent
2023-02-26 04:35:00
cse442-at-ub/project_s23-cinco
https://api.github.com/repos/cse442-at-ub/project_s23-cinco
closed
Create a sign up frame as well as interactivity to go to and from the new frame
Processing Task Sprint 1
Test 1: go to figma page: https://www.figma.com/file/5qJUyXFUAdbtiobIQYqH20/Project-Prototype?node-id=151%3A3&t=QH31e7QFg894LkNj-0 click sign up in the top right corner confirm that it takes you to a sign up page confirm that clicking sign up brings you back to the login screen where you can log in with your new credentials
1.0
Create a sign up frame as well as interactivity to go to and from the new frame - Test 1: go to figma page: https://www.figma.com/file/5qJUyXFUAdbtiobIQYqH20/Project-Prototype?node-id=151%3A3&t=QH31e7QFg894LkNj-0 click sign up in the top right corner confirm that it takes you to a sign up page confirm that clicking sign up brings you back to the login screen where you can log in with your new credentials
process
create a sign up frame as well as interactivity to go to and from the new frame test go to figma page click sign up in the top right corner confirm that it takes you to a sign up page confirm that clicking sign up brings you back to the login screen where you can log in with your new credentials
1
2,362
5,166,321,914
IssuesEvent
2017-01-17 15:57:46
inasafe/inasafe
https://api.github.com/repos/inasafe/inasafe
closed
Communicate the limitations of running with big datasets to users
Feature request Needs review Preprocessing User Interface
Either by docuemntation or the GUI, when users select detailed datasets with a big viewport we should set their expectations instead of them having to wonder wether or not it is going to finish successfully. Perhaps warnings when the data is too big? Like OSM does. (Ask Liana for more info).
1.0
Communicate the limitations of running with big datasets to users - Either by docuemntation or the GUI, when users select detailed datasets with a big viewport we should set their expectations instead of them having to wonder wether or not it is going to finish successfully. Perhaps warnings when the data is too big? Like OSM does. (Ask Liana for more info).
process
communicate the limitations of running with big datasets to users either by docuemntation or the gui when users select detailed datasets with a big viewport we should set their expectations instead of them having to wonder wether or not it is going to finish successfully perhaps warnings when the data is too big like osm does ask liana for more info
1
444,624
31,081,006,601
IssuesEvent
2023-08-13 04:13:11
aws/aws-sdk-ruby
https://api.github.com/repos/aws/aws-sdk-ruby
opened
Stub response_target
documentation needs-triage
### Describe the issue I am trying to stub the `response_target` in the `get_object` function call. I currently have this setup in my test suite ``` Aws.config[:s3] = { stub_responses: { list_buckets: { buckets: [name: 'bucket_name' }, list_objects_v2: { contents: [{key: "some_key"}] }, get_object: { body: "lorem ipsum", response_target: "trying_to_stub/this" } } } ``` I get a `ArgumentError: unexpected value at params[:response_target]` I don't see any documentation on this. If I remove the `response_target` key, it will try to create the file(which i don't want) ### Links https://docs.aws.amazon.com/sdk-for-ruby/v3/api/Aws/ClientStubs.html
1.0
Stub response_target - ### Describe the issue I am trying to stub the `response_target` in the `get_object` function call. I currently have this setup in my test suite ``` Aws.config[:s3] = { stub_responses: { list_buckets: { buckets: [name: 'bucket_name' }, list_objects_v2: { contents: [{key: "some_key"}] }, get_object: { body: "lorem ipsum", response_target: "trying_to_stub/this" } } } ``` I get a `ArgumentError: unexpected value at params[:response_target]` I don't see any documentation on this. If I remove the `response_target` key, it will try to create the file(which i don't want) ### Links https://docs.aws.amazon.com/sdk-for-ruby/v3/api/Aws/ClientStubs.html
non_process
stub response target describe the issue i am trying to stub the response target in the get object function call i currently have this setup in my test suite aws config stub responses list buckets buckets name bucket name list objects contents get object body lorem ipsum response target trying to stub this i get a argumenterror unexpected value at params i don t see any documentation on this if i remove the response target key it will try to create the file which i don t want links
0
540
3,001,553,233
IssuesEvent
2015-07-24 12:14:25
genomizer/genomizer-server
https://api.github.com/repos/genomizer/genomizer-server
closed
Smooth file truncated by IGB, should be investigated
bug High priority Processing
Check result in Integrated Genome Browser. Smoothing should use a moving window and compute the mean or median of the window and assign to the rounded (any direction) center coordinate. This can be rewritten in Java with relative ease.
1.0
Smooth file truncated by IGB, should be investigated - Check result in Integrated Genome Browser. Smoothing should use a moving window and compute the mean or median of the window and assign to the rounded (any direction) center coordinate. This can be rewritten in Java with relative ease.
process
smooth file truncated by igb should be investigated check result in integrated genome browser smoothing should use a moving window and compute the mean or median of the window and assign to the rounded any direction center coordinate this can be rewritten in java with relative ease
1
98,444
11,083,446,007
IssuesEvent
2019-12-13 14:29:37
esdc-devcop/esdc-devcop.github.io
https://api.github.com/repos/esdc-devcop/esdc-devcop.github.io
closed
Update Strategy
documentation
The strategy is changing to include streams and hasn't had the time to focus on management support. So the documentation on this should be updated.
1.0
Update Strategy - The strategy is changing to include streams and hasn't had the time to focus on management support. So the documentation on this should be updated.
non_process
update strategy the strategy is changing to include streams and hasn t had the time to focus on management support so the documentation on this should be updated
0
327,818
9,981,678,467
IssuesEvent
2019-07-10 08:06:02
Sinapse-Energia/IoT-Hub-Home
https://api.github.com/repos/Sinapse-Energia/IoT-Hub-Home
opened
[BUG] External Metering is not working with channel 2 & 3
Client Priority: high Size: 2 Status: new Type: bug
We connect a load to each 3 phases of the EASTRON but is only working the first one. The phase 2 return negative values and the phase 3 return 0. `L3_READ_EXT_METERING_R;855022;1;2;0.0;0.0;0.0;0.0;0.0;-21.81;-1;-186.1;0.0;780.2;-21.8;1562742958;` `L3_READ_EXT_METERING_R;855022;1;3;0.0;0.0;0.0;0.0;0.0;0.00;1;0.0;0.0;0.0;0.0;1562742985;` Also, when there is not load on the channel 2, we get also strange values. The values displayed in the EASTRON screen are OK
1.0
[BUG] External Metering is not working with channel 2 & 3 - We connect a load to each 3 phases of the EASTRON but is only working the first one. The phase 2 return negative values and the phase 3 return 0. `L3_READ_EXT_METERING_R;855022;1;2;0.0;0.0;0.0;0.0;0.0;-21.81;-1;-186.1;0.0;780.2;-21.8;1562742958;` `L3_READ_EXT_METERING_R;855022;1;3;0.0;0.0;0.0;0.0;0.0;0.00;1;0.0;0.0;0.0;0.0;1562742985;` Also, when there is not load on the channel 2, we get also strange values. The values displayed in the EASTRON screen are OK
non_process
external metering is not working with channel we connect a load to each phases of the eastron but is only working the first one the phase return negative values and the phase return read ext metering r read ext metering r also when there is not load on the channel we get also strange values the values displayed in the eastron screen are ok
0
20,720
27,419,639,087
IssuesEvent
2023-03-01 15:54:35
xataio/xata-py
https://api.github.com/repos/xataio/xata-py
closed
`throw_exception` option in BulkProcessor
bulk-processor
Currently, every exception in the bulk processor throws an exception and terminates the thread. Add an option `throw_exception` that allows a bool flag to trigger throw or not to throw. Default: `False`.
1.0
`throw_exception` option in BulkProcessor - Currently, every exception in the bulk processor throws an exception and terminates the thread. Add an option `throw_exception` that allows a bool flag to trigger throw or not to throw. Default: `False`.
process
throw exception option in bulkprocessor currently every exception in the bulk processor throws an exception and terminates the thread add an option throw exception that allows a bool flag to trigger throw or not to throw default false
1
83,289
10,324,745,366
IssuesEvent
2019-09-01 11:56:47
coredns/coredns
https://api.github.com/repos/coredns/coredns
closed
Need document on how to set up gRPC server
documentation plugin/grpc
Hi, I'm trying to setup CoreDNS client and server to communicate via gRPC, but totally cannot figure out how to. I tried this config for server: `.grpc://test.com:4343 { forward . 8.8.8.8 }` and this for client on another machine: `.:4343 { grpc . 192.168.1.19:4343 { tls_servername test.com } }` But looks it's not working and no any error printed. For gRPC server, I think we need http2 and https enabled? If so I'd like to use self-signed cert, but don't know where to specify the certs, and how to let client don't check the cert. I could neither find related article on Google, so could you add some basic instruction on this? Thanks.
1.0
Need document on how to set up gRPC server - Hi, I'm trying to setup CoreDNS client and server to communicate via gRPC, but totally cannot figure out how to. I tried this config for server: `.grpc://test.com:4343 { forward . 8.8.8.8 }` and this for client on another machine: `.:4343 { grpc . 192.168.1.19:4343 { tls_servername test.com } }` But looks it's not working and no any error printed. For gRPC server, I think we need http2 and https enabled? If so I'd like to use self-signed cert, but don't know where to specify the certs, and how to let client don't check the cert. I could neither find related article on Google, so could you add some basic instruction on this? Thanks.
non_process
need document on how to set up grpc server hi i m trying to setup coredns client and server to communicate via grpc but totally cannot figure out how to i tried this config for server grpc test com forward and this for client on another machine grpc tls servername test com but looks it s not working and no any error printed for grpc server i think we need and https enabled if so i d like to use self signed cert but don t know where to specify the certs and how to let client don t check the cert i could neither find related article on google so could you add some basic instruction on this thanks
0
417,531
12,167,037,057
IssuesEvent
2020-04-27 10:13:12
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.chase.com - site is not usable
browser-firefox engine-gecko priority-important
<!-- @browser: Firefox 75.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:75.0) Gecko/20100101 Firefox/75.0 --> <!-- @reported_with: --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/52130 --> **URL**: http://www.chase.com **Browser / Version**: Firefox 75.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Browser unsupported **Steps to Reproduce**: site does not load images or stylesheets, only test including, "Please update your browser. We don't support this browser anymore. Using another one will help protect your accounts and provide a better experience. " despite claiming to support firefox. Console: Loading failed for the <script> with source “blob:https://www.chase.com/51046d69-62b6-4c5b-bb2e-e4f6407d4948”. www.chase.com:1:1 Loading failed for the <script> with source “https://www.chase.com/c/041820/etc/designs/chase-ux/clientlibs/jquery/js/jquery.min.js”. www.chase.com:32:1 Loading failed for the <script> with source “https://www.chase.com/c/041820/etc/designs/chase-ux/clientlibs/require/js/require.min.js”. www.chase.com:33:1 Loading failed for the <script> with source “https://www.chase.com/c/041820/apps/chase/clientlibs/foundation/scripts/Reporting.js”. www.chase.com:41:1 ReferenceError: requirejs is not defined www.chase.com:49:15 Loading failed for the <script> with source “https://www.chase.com/c/041820/etc/designs/chase-ux/clientlibs/chase-ux/js/dist/home.min.js”. www.chase.com:64:1 ReferenceError: require is not defined www.chase.com:66:7 ReferenceError: require is not defined www.chase.com:76:11 ReferenceError: require is not defined www.chase.com:166:5 ReferenceError: require is not defined www.chase.com:546:7 ReferenceError: require is not defined www.chase.com:627:31 Content Security Policy: Ignoring ‘x-frame-options’ because of ‘frame-ancestors’ directive. ReferenceError: CHASE is not defined <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/4/2ec9b877-4e62-4abf-b916-764181138c3f.jpg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.chase.com - site is not usable - <!-- @browser: Firefox 75.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:75.0) Gecko/20100101 Firefox/75.0 --> <!-- @reported_with: --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/52130 --> **URL**: http://www.chase.com **Browser / Version**: Firefox 75.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Browser unsupported **Steps to Reproduce**: site does not load images or stylesheets, only test including, "Please update your browser. We don't support this browser anymore. Using another one will help protect your accounts and provide a better experience. " despite claiming to support firefox. Console: Loading failed for the <script> with source “blob:https://www.chase.com/51046d69-62b6-4c5b-bb2e-e4f6407d4948”. www.chase.com:1:1 Loading failed for the <script> with source “https://www.chase.com/c/041820/etc/designs/chase-ux/clientlibs/jquery/js/jquery.min.js”. www.chase.com:32:1 Loading failed for the <script> with source “https://www.chase.com/c/041820/etc/designs/chase-ux/clientlibs/require/js/require.min.js”. www.chase.com:33:1 Loading failed for the <script> with source “https://www.chase.com/c/041820/apps/chase/clientlibs/foundation/scripts/Reporting.js”. www.chase.com:41:1 ReferenceError: requirejs is not defined www.chase.com:49:15 Loading failed for the <script> with source “https://www.chase.com/c/041820/etc/designs/chase-ux/clientlibs/chase-ux/js/dist/home.min.js”. www.chase.com:64:1 ReferenceError: require is not defined www.chase.com:66:7 ReferenceError: require is not defined www.chase.com:76:11 ReferenceError: require is not defined www.chase.com:166:5 ReferenceError: require is not defined www.chase.com:546:7 ReferenceError: require is not defined www.chase.com:627:31 Content Security Policy: Ignoring ‘x-frame-options’ because of ‘frame-ancestors’ directive. ReferenceError: CHASE is not defined <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/4/2ec9b877-4e62-4abf-b916-764181138c3f.jpg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
site is not usable url browser version firefox operating system windows tested another browser yes chrome problem type site is not usable description browser unsupported steps to reproduce site does not load images or stylesheets only test including please update your browser we don t support this browser anymore using another one will help protect your accounts and provide a better experience despite claiming to support firefox console loading failed for the with source “blob loading failed for the with source “ loading failed for the with source “ loading failed for the with source “ referenceerror requirejs is not defined loading failed for the with source “ referenceerror require is not defined referenceerror require is not defined referenceerror require is not defined referenceerror require is not defined referenceerror require is not defined content security policy ignoring ‘x frame options’ because of ‘frame ancestors’ directive referenceerror chase is not defined view the screenshot img alt screenshot src browser configuration none from with ❤️
0
54,832
7,926,704,136
IssuesEvent
2018-07-06 03:55:02
pkulchenko/ZeroBraneStudio
https://api.github.com/repos/pkulchenko/ZeroBraneStudio
closed
Suggetion - Start debug then run directly until break point
Documentation
Currently, start debug will begin step at the first line. I think this is not necessary, because user could put break point at first line if they need. As the title said, maybe 'Start debug then run directly until break point' is more handy.
1.0
Suggetion - Start debug then run directly until break point - Currently, start debug will begin step at the first line. I think this is not necessary, because user could put break point at first line if they need. As the title said, maybe 'Start debug then run directly until break point' is more handy.
non_process
suggetion start debug then run directly until break point currently start debug will begin step at the first line i think this is not necessary because user could put break point at first line if they need as the title said maybe start debug then run directly until break point is more handy
0
288,504
31,861,420,698
IssuesEvent
2023-09-15 11:12:31
nidhi7598/linux-v4.19.72_CVE-2022-3564
https://api.github.com/repos/nidhi7598/linux-v4.19.72_CVE-2022-3564
opened
CVE-2021-3347 (High) detected in linuxlinux-4.19.294
Mend: dependency security vulnerability
## CVE-2021-3347 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-v4.19.72_CVE-2022-3564/commit/9ffee08efa44c7887e2babb8f304df0fa1094efb">9ffee08efa44c7887e2babb8f304df0fa1094efb</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/futex.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel through 5.10.11. PI futexes have a kernel stack use-after-free during fault handling, allowing local users to execute code in the kernel, aka CID-34b1a1ce1458. <p>Publish Date: 2021-01-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3347>CVE-2021-3347</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-3347">https://www.linuxkernelcves.com/cves/CVE-2021-3347</a></p> <p>Release Date: 2021-01-29</p> <p>Fix Resolution: v4.14.218,v4.19.172,v5.10.12,v5.4.94</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-3347 (High) detected in linuxlinux-4.19.294 - ## CVE-2021-3347 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-v4.19.72_CVE-2022-3564/commit/9ffee08efa44c7887e2babb8f304df0fa1094efb">9ffee08efa44c7887e2babb8f304df0fa1094efb</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/futex.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel through 5.10.11. PI futexes have a kernel stack use-after-free during fault handling, allowing local users to execute code in the kernel, aka CID-34b1a1ce1458. <p>Publish Date: 2021-01-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3347>CVE-2021-3347</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-3347">https://www.linuxkernelcves.com/cves/CVE-2021-3347</a></p> <p>Release Date: 2021-01-29</p> <p>Fix Resolution: v4.14.218,v4.19.172,v5.10.12,v5.4.94</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files kernel futex c vulnerability details an issue was discovered in the linux kernel through pi futexes have a kernel stack use after free during fault handling allowing local users to execute code in the kernel aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
16,334
20,990,704,947
IssuesEvent
2022-03-29 09:02:26
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Move GO:1990109 rejection of pollen from other species
multi-species process
Hi GO:1990109 rejection of pollen from other species is a child of 'GO:0044419 biological process involved in interspecies interaction between organisms' In the multiorg call we concluded it should be under 'GO:0048544 recognition of pollen' @tberardini is that OK with you? Thanks, Pascale
1.0
Move GO:1990109 rejection of pollen from other species - Hi GO:1990109 rejection of pollen from other species is a child of 'GO:0044419 biological process involved in interspecies interaction between organisms' In the multiorg call we concluded it should be under 'GO:0048544 recognition of pollen' @tberardini is that OK with you? Thanks, Pascale
process
move go rejection of pollen from other species hi go rejection of pollen from other species is a child of go biological process involved in interspecies interaction between organisms in the multiorg call we concluded it should be under go recognition of pollen tberardini is that ok with you thanks pascale
1
418,075
12,192,906,405
IssuesEvent
2020-04-29 13:39:40
semanticarts/gist
https://api.github.com/repos/semanticarts/gist
closed
Clean up documentation and tools
effort: small impact: patch priority: should have status: triaged
- bundle.bat can be deleted - now replaced by ontology-toolkit. Instructions in README.md should be updated to reflect the use of ontology-toolkit. - versionize scripts can be deleted - serialize scripts can be deleted - version.txt - delete - curl_requests.txt - do we need this? - Only rdf-toolkit is needed in tools directory
1.0
Clean up documentation and tools - - bundle.bat can be deleted - now replaced by ontology-toolkit. Instructions in README.md should be updated to reflect the use of ontology-toolkit. - versionize scripts can be deleted - serialize scripts can be deleted - version.txt - delete - curl_requests.txt - do we need this? - Only rdf-toolkit is needed in tools directory
non_process
clean up documentation and tools bundle bat can be deleted now replaced by ontology toolkit instructions in readme md should be updated to reflect the use of ontology toolkit versionize scripts can be deleted serialize scripts can be deleted version txt delete curl requests txt do we need this only rdf toolkit is needed in tools directory
0
168,904
13,106,875,378
IssuesEvent
2020-08-04 14:28:22
foodcoopshop/foodcoopshop
https://api.github.com/repos/foodcoopshop/foodcoopshop
closed
Implement integration tests with IntegrationTestTrait
cleanup 🧹 tests
Since v3.7 CakePHP provides IntegrationTestTrait which is really great! This refactoring * should make the tests quicker * reduces code - as methods (eg. assert methods) from the cake core are used **Todos** **done** * [x] netzwerk modul testen * [x] remove html output in SendShells (unit tests break on travis) * [x] rechnngs-download geht nicht: https://demo-de.foodcoopshop.com/admin/action-logs?types=&types%5B%5D=cronjob_send_invoices&customerId=&productId=&dateFrom=06.05.2020&dateTo=03.08.2020 * [x] open order lists inline in browser and do not force download * [x] PaymentsController: file upload not yet working * [x] Changelog: speed up travis integration testing by 50%! * [x] ListsController: link with path in order_lists.php (wrong), tests not yet working * [x] as the new LoginTrait just fakes a login (to be quick), the "real" login needs to get an own test * [x] Wrap up 'headers' => [ 'X_REQUESTED_WITH' => 'XMLHttpRequest', 'ACCEPT' => application/json' ], * [x] do not show stack trace: ProductsControllerTest.php:257 * [x] InstantOrderTests are not working (no AppSession in AppAuth) * [x] AppCakeTestCase should contain much less assertMethods. Use the built-in assert-methods in IntegrationTestCase. * [x] remove usage of httpClient in all tests * [x] HttpClient removed: BlogPostControllerTest (#529), ManufacturersFrontendController, ProductsFrontendController (#531) * [x] CartControllerTest::testAddLoggedOut fails due to missing exit on autorization * [x] CartController authorize does not work any more * [x] ProductsControllerTest::testDeleteProduct fails * [x] Refactor CartsControllerTest * [x] Refactor OrderDetailsControllerAddFeedbackTest * [x] Refactor OrderDetailsControllerCancellationTest.php * [x] Refactor OrderDetailsControllerEditAmountTest.php * [x] Refactor OrderDetailsControllerEdittCustomerTest.php * [x] Refactor OrderDetailsControllerEditPickupDayTest * [x] Refactor OrderDetailsControllerEditPriceTest * [x] Refactor OrderDetailsControllerEditQuantityTest * [x] addProductToCart() and finishCart without httpClient * [x] enable testCancellationWithTimebasedCurrency * [x] if (!$_SERVER['REQUEST_METHOD'] == 'POST') { is not set with IntegrationTest * [x] testNewPasswordRequestWithValidEmail test login with new password
1.0
Implement integration tests with IntegrationTestTrait - Since v3.7 CakePHP provides IntegrationTestTrait which is really great! This refactoring * should make the tests quicker * reduces code - as methods (eg. assert methods) from the cake core are used **Todos** **done** * [x] netzwerk modul testen * [x] remove html output in SendShells (unit tests break on travis) * [x] rechnngs-download geht nicht: https://demo-de.foodcoopshop.com/admin/action-logs?types=&types%5B%5D=cronjob_send_invoices&customerId=&productId=&dateFrom=06.05.2020&dateTo=03.08.2020 * [x] open order lists inline in browser and do not force download * [x] PaymentsController: file upload not yet working * [x] Changelog: speed up travis integration testing by 50%! * [x] ListsController: link with path in order_lists.php (wrong), tests not yet working * [x] as the new LoginTrait just fakes a login (to be quick), the "real" login needs to get an own test * [x] Wrap up 'headers' => [ 'X_REQUESTED_WITH' => 'XMLHttpRequest', 'ACCEPT' => application/json' ], * [x] do not show stack trace: ProductsControllerTest.php:257 * [x] InstantOrderTests are not working (no AppSession in AppAuth) * [x] AppCakeTestCase should contain much less assertMethods. Use the built-in assert-methods in IntegrationTestCase. * [x] remove usage of httpClient in all tests * [x] HttpClient removed: BlogPostControllerTest (#529), ManufacturersFrontendController, ProductsFrontendController (#531) * [x] CartControllerTest::testAddLoggedOut fails due to missing exit on autorization * [x] CartController authorize does not work any more * [x] ProductsControllerTest::testDeleteProduct fails * [x] Refactor CartsControllerTest * [x] Refactor OrderDetailsControllerAddFeedbackTest * [x] Refactor OrderDetailsControllerCancellationTest.php * [x] Refactor OrderDetailsControllerEditAmountTest.php * [x] Refactor OrderDetailsControllerEdittCustomerTest.php * [x] Refactor OrderDetailsControllerEditPickupDayTest * [x] Refactor OrderDetailsControllerEditPriceTest * [x] Refactor OrderDetailsControllerEditQuantityTest * [x] addProductToCart() and finishCart without httpClient * [x] enable testCancellationWithTimebasedCurrency * [x] if (!$_SERVER['REQUEST_METHOD'] == 'POST') { is not set with IntegrationTest * [x] testNewPasswordRequestWithValidEmail test login with new password
non_process
implement integration tests with integrationtesttrait since cakephp provides integrationtesttrait which is really great this refactoring should make the tests quicker reduces code as methods eg assert methods from the cake core are used todos done netzwerk modul testen remove html output in sendshells unit tests break on travis rechnngs download geht nicht open order lists inline in browser and do not force download paymentscontroller file upload not yet working changelog speed up travis integration testing by listscontroller link with path in order lists php wrong tests not yet working as the new logintrait just fakes a login to be quick the real login needs to get an own test wrap up headers do not show stack trace productscontrollertest php instantordertests are not working no appsession in appauth appcaketestcase should contain much less assertmethods use the built in assert methods in integrationtestcase remove usage of httpclient in all tests httpclient removed blogpostcontrollertest manufacturersfrontendcontroller productsfrontendcontroller cartcontrollertest testaddloggedout fails due to missing exit on autorization cartcontroller authorize does not work any more productscontrollertest testdeleteproduct fails refactor cartscontrollertest refactor orderdetailscontrolleraddfeedbacktest refactor orderdetailscontrollercancellationtest php refactor orderdetailscontrollereditamounttest php refactor orderdetailscontrolleredittcustomertest php refactor orderdetailscontrollereditpickupdaytest refactor orderdetailscontrollereditpricetest refactor orderdetailscontrollereditquantitytest addproducttocart and finishcart without httpclient enable testcancellationwithtimebasedcurrency if server post is not set with integrationtest testnewpasswordrequestwithvalidemail test login with new password
0
223,600
17,117,010,832
IssuesEvent
2021-07-11 15:10:03
WebDevsOrg/calculator
https://api.github.com/repos/WebDevsOrg/calculator
opened
Capture standard mode requirements
documentation
This should have been done earlier, but we were following MS calculator and try to replicate some part of it. But before we start user interface its good to capture these requirements and pick good features, UI layout etc. from other calculator projects too. Its also important to understand the use case 1. Is it targeted for mobile first customer 2. Is it targeted for desktop users 3. Or both Its also important to consider who is going to use this calculator, so we should think about accessibility issues, like themes, colors, button sizes, labels, font size, contrast ![image](https://user-images.githubusercontent.com/6313263/125200288-0b6c2b00-e288-11eb-9265-d5a716d2440b.png) We should think mobile first design but initially we may target for desktop and later on mobile as it may require different setup.
1.0
Capture standard mode requirements - This should have been done earlier, but we were following MS calculator and try to replicate some part of it. But before we start user interface its good to capture these requirements and pick good features, UI layout etc. from other calculator projects too. Its also important to understand the use case 1. Is it targeted for mobile first customer 2. Is it targeted for desktop users 3. Or both Its also important to consider who is going to use this calculator, so we should think about accessibility issues, like themes, colors, button sizes, labels, font size, contrast ![image](https://user-images.githubusercontent.com/6313263/125200288-0b6c2b00-e288-11eb-9265-d5a716d2440b.png) We should think mobile first design but initially we may target for desktop and later on mobile as it may require different setup.
non_process
capture standard mode requirements this should have been done earlier but we were following ms calculator and try to replicate some part of it but before we start user interface its good to capture these requirements and pick good features ui layout etc from other calculator projects too its also important to understand the use case is it targeted for mobile first customer is it targeted for desktop users or both its also important to consider who is going to use this calculator so we should think about accessibility issues like themes colors button sizes labels font size contrast we should think mobile first design but initially we may target for desktop and later on mobile as it may require different setup
0
18,920
24,867,914,700
IssuesEvent
2022-10-27 13:20:02
OpenDataScotland/the_od_bods
https://api.github.com/repos/OpenDataScotland/the_od_bods
opened
Scottish OpenStreetMap repository
data processing back end new source
**Is your feature request related to a problem? Please describe.** OpenStreetMap holds a lot of great GIS data for Scotland that we could scrape and store in a GitHub repository to then be surfaced on OpenData.Scot **Describe the solution you'd like** A script for making queries to the OSM Overpass API to then save in various formats (CSV, JSON, GEOJSON) in a GitHub repo. The script could run on a variable scheduled basis depending on how often we reckon the data would be updated (e.g. weekly, monthly, annually) **Describe alternatives you've considered** We could just link directly to overpass queries but that's fairly basic, can have slow API response times and requires more work for the user vs. serving up ready-made and curated datasets **Additional context** N/A
1.0
Scottish OpenStreetMap repository - **Is your feature request related to a problem? Please describe.** OpenStreetMap holds a lot of great GIS data for Scotland that we could scrape and store in a GitHub repository to then be surfaced on OpenData.Scot **Describe the solution you'd like** A script for making queries to the OSM Overpass API to then save in various formats (CSV, JSON, GEOJSON) in a GitHub repo. The script could run on a variable scheduled basis depending on how often we reckon the data would be updated (e.g. weekly, monthly, annually) **Describe alternatives you've considered** We could just link directly to overpass queries but that's fairly basic, can have slow API response times and requires more work for the user vs. serving up ready-made and curated datasets **Additional context** N/A
process
scottish openstreetmap repository is your feature request related to a problem please describe openstreetmap holds a lot of great gis data for scotland that we could scrape and store in a github repository to then be surfaced on opendata scot describe the solution you d like a script for making queries to the osm overpass api to then save in various formats csv json geojson in a github repo the script could run on a variable scheduled basis depending on how often we reckon the data would be updated e g weekly monthly annually describe alternatives you ve considered we could just link directly to overpass queries but that s fairly basic can have slow api response times and requires more work for the user vs serving up ready made and curated datasets additional context n a
1
334,004
24,399,523,080
IssuesEvent
2022-10-04 23:08:54
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
better description of solveh_banded limitations
scipy.linalg Documentation
Related to the discussion in the original pull request for linalg.solveh_banded, #4373, the solver requires additional constraints on the matrix compared to linalg.solve and linalg.solve_banded. The routine solveh_banded requires leading minors of the matrix A in the equation A * x = b be positive definite. I do not understand this requirement at all, so I can not begin to delve into the linear algebra discussion. This requirement is very confusing as it stands, however, because it is not raised in very similar solvers that it is meant to supersede. In an ideal world, I would like to help relax this requirement to be more in line with the other solvers. I do not know how to go about this as it stands. At a minimum, could we work on a more informative error message that mentions the other solvers may still work in situations where solveh_banded fails? Here is an example of an error that is raised in an application finding eigenvectors from eigenvalues by inverse iteration ``` import numpy as np from scipy.linalg import solve, solveh_banded, solve_banded numz = 105 d = np.full(numz, -1.9964864348874727) C = np.diag(d) + np.diag(np.ones(numz - 1), k=1) + np.diag(np.ones(numz - 1), k=-1) # first iteration to find eigen-vector phi_test = np.ones(numz) phi1 = solve(C, phi_test) C_b = np.ones((3, numz)) C_b[1, :] = d phi2 = solve_banded((1, 1), C_b, phi_test) C_h = np.ones((2, numz)) C_h[1, :] = d phi3 = solveh_banded(C_h, phi_test) ```
1.0
better description of solveh_banded limitations - Related to the discussion in the original pull request for linalg.solveh_banded, #4373, the solver requires additional constraints on the matrix compared to linalg.solve and linalg.solve_banded. The routine solveh_banded requires leading minors of the matrix A in the equation A * x = b be positive definite. I do not understand this requirement at all, so I can not begin to delve into the linear algebra discussion. This requirement is very confusing as it stands, however, because it is not raised in very similar solvers that it is meant to supersede. In an ideal world, I would like to help relax this requirement to be more in line with the other solvers. I do not know how to go about this as it stands. At a minimum, could we work on a more informative error message that mentions the other solvers may still work in situations where solveh_banded fails? Here is an example of an error that is raised in an application finding eigenvectors from eigenvalues by inverse iteration ``` import numpy as np from scipy.linalg import solve, solveh_banded, solve_banded numz = 105 d = np.full(numz, -1.9964864348874727) C = np.diag(d) + np.diag(np.ones(numz - 1), k=1) + np.diag(np.ones(numz - 1), k=-1) # first iteration to find eigen-vector phi_test = np.ones(numz) phi1 = solve(C, phi_test) C_b = np.ones((3, numz)) C_b[1, :] = d phi2 = solve_banded((1, 1), C_b, phi_test) C_h = np.ones((2, numz)) C_h[1, :] = d phi3 = solveh_banded(C_h, phi_test) ```
non_process
better description of solveh banded limitations related to the discussion in the original pull request for linalg solveh banded the solver requires additional constraints on the matrix compared to linalg solve and linalg solve banded the routine solveh banded requires leading minors of the matrix a in the equation a x b be positive definite i do not understand this requirement at all so i can not begin to delve into the linear algebra discussion this requirement is very confusing as it stands however because it is not raised in very similar solvers that it is meant to supersede in an ideal world i would like to help relax this requirement to be more in line with the other solvers i do not know how to go about this as it stands at a minimum could we work on a more informative error message that mentions the other solvers may still work in situations where solveh banded fails here is an example of an error that is raised in an application finding eigenvectors from eigenvalues by inverse iteration import numpy as np from scipy linalg import solve solveh banded solve banded numz d np full numz c np diag d np diag np ones numz k np diag np ones numz k first iteration to find eigen vector phi test np ones numz solve c phi test c b np ones numz c b d solve banded c b phi test c h np ones numz c h d solveh banded c h phi test
0
61,954
7,533,917,929
IssuesEvent
2018-04-16 05:42:59
ParabolInc/action
https://api.github.com/repos/ParabolInc/action
closed
Design: duplicate/clone cards
design enhancement user request
## Issue - Enhancement User `google-oauth2|103150684875644068831` requested the following: > [I] need to duplicate cards This is an alternative to [#1071 Assign to All](https://github.com/ParabolInc/action/issues/1071), and probably easier to implement. We'd need to design where we'd put this control. ### Acceptance Criteria (optional) A design in which users can: - Quickly access a control that allows them to clone a card - When the control is pressed, it duplicates the card in its current context... - ...cloning the card within its column if pressed on the dashboard or _Project Updates_ - ...cloning the card when processing an agenda item - **Estimated effort:** 13 points ([see CONTRIBUTING.md](https://github.com/ParabolInc/action/blob/master/CONTRIBUTING.md#points-and-sizes))
1.0
Design: duplicate/clone cards - ## Issue - Enhancement User `google-oauth2|103150684875644068831` requested the following: > [I] need to duplicate cards This is an alternative to [#1071 Assign to All](https://github.com/ParabolInc/action/issues/1071), and probably easier to implement. We'd need to design where we'd put this control. ### Acceptance Criteria (optional) A design in which users can: - Quickly access a control that allows them to clone a card - When the control is pressed, it duplicates the card in its current context... - ...cloning the card within its column if pressed on the dashboard or _Project Updates_ - ...cloning the card when processing an agenda item - **Estimated effort:** 13 points ([see CONTRIBUTING.md](https://github.com/ParabolInc/action/blob/master/CONTRIBUTING.md#points-and-sizes))
non_process
design duplicate clone cards issue enhancement user google requested the following need to duplicate cards this is an alternative to and probably easier to implement we d need to design where we d put this control acceptance criteria optional a design in which users can quickly access a control that allows them to clone a card when the control is pressed it duplicates the card in its current context cloning the card within its column if pressed on the dashboard or project updates cloning the card when processing an agenda item estimated effort points
0
10,271
13,125,342,255
IssuesEvent
2020-08-06 06:27:35
linuxdeepin/developer-center
https://api.github.com/repos/linuxdeepin/developer-center
closed
Change keyboard layout shortcut doesn't work on lock screen
Solved | resolved bug | functional behavior other | delay processing place | WH
I'm running Deepin Desktop 15.11. In the lock screen, changing keyboard layout using shortcuts `Alt + Shift` or `Win + Space` to enter password doesn't work.
1.0
Change keyboard layout shortcut doesn't work on lock screen - I'm running Deepin Desktop 15.11. In the lock screen, changing keyboard layout using shortcuts `Alt + Shift` or `Win + Space` to enter password doesn't work.
process
change keyboard layout shortcut doesn t work on lock screen i m running deepin desktop in the lock screen changing keyboard layout using shortcuts alt shift or win space to enter password doesn t work
1
15,279
19,269,704,153
IssuesEvent
2021-12-10 02:41:14
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Merge Vector Layer missing in Saga
Feedback stale Processing Bug
### What is the bug or the crash? The merge vector layer tool is not available in the SAGA processing toolbox. I am running ubuntu 20.04 and the latest version of qgis. ### Steps to reproduce the issue Go to QGIS Processing Toolbox Open the Saga drop down select vector general merge vector layer missing does not show up in search either. ### Versions <!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd"> <html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /><style type="text/css"> p, li { white-space: pre-wrap; } </style></head><body> QGIS version | 3.20.3-Odense | QGIS code revision | 495fbaecaf -- | -- | -- | -- Qt version | 5.12.8 Python version | 3.8.10 GDAL/OGR version | 3.0.4 PROJ version | 6.3.1 EPSG Registry database version | v9.8.6 (2020-01-22) Compiled against GEOS | 3.8.0-CAPI-1.13.1 | Running against GEOS | 3.8.0-CAPI-1.13.1 SQLite version | 3.31.1 PDAL version | 2.0.1 PostgreSQL client version | 12.8 (Ubuntu 12.8-0ubuntu0.20.04.1) SpatiaLite version | 4.3.0a QWT version | 6.1.4 QScintilla2 version | 2.11.2 OS version | Ubuntu 20.04.3 LTS   |   |   |   Active Python plugins | quick_map_servicescartography_toolsQuickOSMAutomaticBackup-masterMetaSearchdb_managerprocessing </body></html> ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context _No response_
1.0
Merge Vector Layer missing in Saga - ### What is the bug or the crash? The merge vector layer tool is not available in the SAGA processing toolbox. I am running ubuntu 20.04 and the latest version of qgis. ### Steps to reproduce the issue Go to QGIS Processing Toolbox Open the Saga drop down select vector general merge vector layer missing does not show up in search either. ### Versions <!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd"> <html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /><style type="text/css"> p, li { white-space: pre-wrap; } </style></head><body> QGIS version | 3.20.3-Odense | QGIS code revision | 495fbaecaf -- | -- | -- | -- Qt version | 5.12.8 Python version | 3.8.10 GDAL/OGR version | 3.0.4 PROJ version | 6.3.1 EPSG Registry database version | v9.8.6 (2020-01-22) Compiled against GEOS | 3.8.0-CAPI-1.13.1 | Running against GEOS | 3.8.0-CAPI-1.13.1 SQLite version | 3.31.1 PDAL version | 2.0.1 PostgreSQL client version | 12.8 (Ubuntu 12.8-0ubuntu0.20.04.1) SpatiaLite version | 4.3.0a QWT version | 6.1.4 QScintilla2 version | 2.11.2 OS version | Ubuntu 20.04.3 LTS   |   |   |   Active Python plugins | quick_map_servicescartography_toolsQuickOSMAutomaticBackup-masterMetaSearchdb_managerprocessing </body></html> ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context _No response_
process
merge vector layer missing in saga what is the bug or the crash the merge vector layer tool is not available in the saga processing toolbox i am running ubuntu and the latest version of qgis steps to reproduce the issue go to qgis processing toolbox open the saga drop down select vector general merge vector layer missing does not show up in search either versions doctype html public dtd html en p li white space pre wrap qgis version odense qgis code revision qt version python version gdal ogr version proj version epsg registry database version compiled against geos capi running against geos capi sqlite version pdal version postgresql client version ubuntu spatialite version qwt version version os version ubuntu lts         active python plugins quick map servicescartography toolsquickosmautomaticbackup mastermetasearchdb managerprocessing supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
1
9,126
8,528,049,705
IssuesEvent
2018-11-02 21:46:07
goharbor/harbor
https://api.github.com/repos/goharbor/harbor
closed
Dead loop redis event after fired demo job
area/job-services kind/bug
**Expected behavior and actual behavior:** expect job service could receive cancel job event, and do CancelJob at once. **Steps to reproduce the problem:** ``` POST /api/v1/jobs/a5fe27325fbfb9a5df7f9e2a { "action": "stop" } ``` **Versions:** - job service: [v.1.6.1] **Additional context:** I'm running the raw job service, and want to use it as back service just sth like celery, and I found that there is dead loop events while do canceljob or stopjob. steps are: 1. we fired job operations, and redis pool received at [here](https://github.com/goharbor/harbor/blob/v1.6.1/src/jobservice/pool/redis_pool.go#L135); 2. gcwp do [`SendCommand`](https://github.com/goharbor/harbor/blob/v1.6.1/src/jobservice/pool/redis_pool.go#L606); 3. at redis stats mgr, we `fired` msg at [here](https://github.com/goharbor/harbor/blob/v1.6.1/src/jobservice/opm/redis_job_stats_mgr.go#L221), and it [published to redis channel](https://github.com/goharbor/harbor/blob/v1.6.1/src/jobservice/opm/op_commands.go#L90) 4. and redis pool received msg again... so why we need to fire this again and again? or there is something which i misunderstand? Thanks.
1.0
Dead loop redis event after fired demo job - **Expected behavior and actual behavior:** expect job service could receive cancel job event, and do CancelJob at once. **Steps to reproduce the problem:** ``` POST /api/v1/jobs/a5fe27325fbfb9a5df7f9e2a { "action": "stop" } ``` **Versions:** - job service: [v.1.6.1] **Additional context:** I'm running the raw job service, and want to use it as back service just sth like celery, and I found that there is dead loop events while do canceljob or stopjob. steps are: 1. we fired job operations, and redis pool received at [here](https://github.com/goharbor/harbor/blob/v1.6.1/src/jobservice/pool/redis_pool.go#L135); 2. gcwp do [`SendCommand`](https://github.com/goharbor/harbor/blob/v1.6.1/src/jobservice/pool/redis_pool.go#L606); 3. at redis stats mgr, we `fired` msg at [here](https://github.com/goharbor/harbor/blob/v1.6.1/src/jobservice/opm/redis_job_stats_mgr.go#L221), and it [published to redis channel](https://github.com/goharbor/harbor/blob/v1.6.1/src/jobservice/opm/op_commands.go#L90) 4. and redis pool received msg again... so why we need to fire this again and again? or there is something which i misunderstand? Thanks.
non_process
dead loop redis event after fired demo job expected behavior and actual behavior expect job service could receive cancel job event and do canceljob at once steps to reproduce the problem post api jobs action stop versions job service additional context i m running the raw job service and want to use it as back service just sth like celery and i found that there is dead loop events while do canceljob or stopjob steps are we fired job operations and redis pool received at gcwp do at redis stats mgr we fired msg at and it and redis pool received msg again so why we need to fire this again and again or there is something which i misunderstand thanks
0
596
3,071,598,424
IssuesEvent
2015-08-19 13:06:22
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Cannot publish remote HTTP DITA Map to XHTML [DOT 2.x develop branch]
bug P2 preprocess
I'm publishing a DITA Map from a webdav HTTP server without authentication. I have images referenced in my topics, at some point the publishing breaks with this error: ```shell BUILD FAILED D:\projects\eXml\frameworks\dita\DITA-OT2.x\build.xml:41: The following error occurred while executing this line: D:\projects\eXml\frameworks\dita\DITA-OT2.x\plugins\org.dita.xhtml\build_general.xml:112: java.lang.IllegalArgumentException: URI scheme is not "file" at java.io.File.<init>(Unknown Source) at org.dita.dost.writer.ImageMetadataFilter.getImageFile(ImageMetadataFilter.java:172) at org.dita.dost.writer.ImageMetadataFilter.startElement(ImageMetadataFilter.java:84) at org.apache.xerces.parsers.AbstractSAXParser.startElement(Unknown Source) at org.apache.xerces.parsers.AbstractXMLDocumentParser.emptyElement(Unknown Source) at org.ditang.relaxng.defaults.RelaxNGDefaultsComponent.emptyElement(RelaxNGDefaultsComponent.java:635) at org.apache.xerces.impl.XMLNSDocumentScannerImpl.scanStartElement(XMLNSDocumentScannerImpl.java:260) at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl$FragmentContentDispatcher.dispatch(XMLDocumentFragmentScannerImpl.java:1655) at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl.scanDocument(XMLDocumentFragmentScannerImpl.java:325) at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source) at org.ditang.relaxng.defaults.RelaxDefaultsParserConfiguration.parse(RelaxDefaultsParserConfiguration.java:150) at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source) at org.apache.xerces.parsers.XMLParser.parse(Unknown Source) at org.apache.xerces.parsers.AbstractSAXParser.parse(Unknown Source) at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source) at net.sf.saxon.event.Sender.sendSAXSource(Sender.java:441) at net.sf.saxon.event.Sender.send(Sender.java:146) at net.sf.saxon.jaxp.IdentityTransformer.transform(IdentityTransformer.java:363) at org.dita.dost.util.XMLUtils.transform(XMLUtils.java:244) at org.dita.dost.util.XMLUtils.transform(XMLUtils.java:204) at org.dita.dost.writer.AbstractXMLFilter.write(AbstractXMLFilter.java:32) at org.dita.dost.writer.ImageMetadataFilter.write(ImageMetadataFilter.java:73) at org.dita.dost.module.ImageMetadataModule.execute(ImageMetadataModule.java:46) at org.dita.dost.pipeline.PipelineFacade.execute(PipelineFacade.java:68) at org.dita.dost.invoker.ExtensibleAntInvoker.execute(ExtensibleAntInvoker.java:193) at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292) at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) ```
1.0
Cannot publish remote HTTP DITA Map to XHTML [DOT 2.x develop branch] - I'm publishing a DITA Map from a webdav HTTP server without authentication. I have images referenced in my topics, at some point the publishing breaks with this error: ```shell BUILD FAILED D:\projects\eXml\frameworks\dita\DITA-OT2.x\build.xml:41: The following error occurred while executing this line: D:\projects\eXml\frameworks\dita\DITA-OT2.x\plugins\org.dita.xhtml\build_general.xml:112: java.lang.IllegalArgumentException: URI scheme is not "file" at java.io.File.<init>(Unknown Source) at org.dita.dost.writer.ImageMetadataFilter.getImageFile(ImageMetadataFilter.java:172) at org.dita.dost.writer.ImageMetadataFilter.startElement(ImageMetadataFilter.java:84) at org.apache.xerces.parsers.AbstractSAXParser.startElement(Unknown Source) at org.apache.xerces.parsers.AbstractXMLDocumentParser.emptyElement(Unknown Source) at org.ditang.relaxng.defaults.RelaxNGDefaultsComponent.emptyElement(RelaxNGDefaultsComponent.java:635) at org.apache.xerces.impl.XMLNSDocumentScannerImpl.scanStartElement(XMLNSDocumentScannerImpl.java:260) at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl$FragmentContentDispatcher.dispatch(XMLDocumentFragmentScannerImpl.java:1655) at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl.scanDocument(XMLDocumentFragmentScannerImpl.java:325) at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source) at org.ditang.relaxng.defaults.RelaxDefaultsParserConfiguration.parse(RelaxDefaultsParserConfiguration.java:150) at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source) at org.apache.xerces.parsers.XMLParser.parse(Unknown Source) at org.apache.xerces.parsers.AbstractSAXParser.parse(Unknown Source) at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source) at net.sf.saxon.event.Sender.sendSAXSource(Sender.java:441) at net.sf.saxon.event.Sender.send(Sender.java:146) at net.sf.saxon.jaxp.IdentityTransformer.transform(IdentityTransformer.java:363) at org.dita.dost.util.XMLUtils.transform(XMLUtils.java:244) at org.dita.dost.util.XMLUtils.transform(XMLUtils.java:204) at org.dita.dost.writer.AbstractXMLFilter.write(AbstractXMLFilter.java:32) at org.dita.dost.writer.ImageMetadataFilter.write(ImageMetadataFilter.java:73) at org.dita.dost.module.ImageMetadataModule.execute(ImageMetadataModule.java:46) at org.dita.dost.pipeline.PipelineFacade.execute(PipelineFacade.java:68) at org.dita.dost.invoker.ExtensibleAntInvoker.execute(ExtensibleAntInvoker.java:193) at org.apache.tools.ant.UnknownElement.execute(UnknownElement.java:292) at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) ```
process
cannot publish remote http dita map to xhtml i m publishing a dita map from a webdav http server without authentication i have images referenced in my topics at some point the publishing breaks with this error shell build failed d projects exml frameworks dita dita x build xml the following error occurred while executing this line d projects exml frameworks dita dita x plugins org dita xhtml build general xml java lang illegalargumentexception uri scheme is not file at java io file unknown source at org dita dost writer imagemetadatafilter getimagefile imagemetadatafilter java at org dita dost writer imagemetadatafilter startelement imagemetadatafilter java at org apache xerces parsers abstractsaxparser startelement unknown source at org apache xerces parsers abstractxmldocumentparser emptyelement unknown source at org ditang relaxng defaults relaxngdefaultscomponent emptyelement relaxngdefaultscomponent java at org apache xerces impl xmlnsdocumentscannerimpl scanstartelement xmlnsdocumentscannerimpl java at org apache xerces impl xmldocumentfragmentscannerimpl fragmentcontentdispatcher dispatch xmldocumentfragmentscannerimpl java at org apache xerces impl xmldocumentfragmentscannerimpl scandocument xmldocumentfragmentscannerimpl java at org apache xerces parsers parse unknown source at org ditang relaxng defaults relaxdefaultsparserconfiguration parse relaxdefaultsparserconfiguration java at org apache xerces parsers parse unknown source at org apache xerces parsers xmlparser parse unknown source at org apache xerces parsers abstractsaxparser parse unknown source at org xml sax helpers xmlfilterimpl parse unknown source at net sf saxon event sender sendsaxsource sender java at net sf saxon event sender send sender java at net sf saxon jaxp identitytransformer transform identitytransformer java at org dita dost util xmlutils transform xmlutils java at org dita dost util xmlutils transform xmlutils java at org dita dost writer abstractxmlfilter write abstractxmlfilter java at org dita dost writer imagemetadatafilter write imagemetadatafilter java at org dita dost module imagemetadatamodule execute imagemetadatamodule java at org dita dost pipeline pipelinefacade execute pipelinefacade java at org dita dost invoker extensibleantinvoker execute extensibleantinvoker java at org apache tools ant unknownelement execute unknownelement java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke unknown source
1
21,214
28,292,127,141
IssuesEvent
2023-04-09 10:44:57
firebase/firebase-cpp-sdk
https://api.github.com/repos/firebase/firebase-cpp-sdk
closed
[C++] Nightly Integration Testing Report
type: process nightly-testing
Note: This report excludes firestore. Please also check **[the report for firestore](https://github.com/firebase/firebase-cpp-sdk/issues/1178)** *** <hidden value="integration-test-status-comment"></hidden> ### [build against repo] Integration test with FLAKINESS (succeeded after retry) Requested by @DellaBitta on commit 8b4226383959ef9558c4bf13e3b40012b4f75d08 Last updated: Sat Apr 8 03:50 PDT 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4644379829)** | Failures | Configs | |----------|---------| | messaging | [TEST] [FLAKINESS] [Android] [1/3 os: windows] [1/4 android_device: android_target]<details><summary>(1 failed tests)</summary>&nbsp;&nbsp;CRASH/TIMEOUT</details> | | storage | [TEST] [FLAKINESS] [iOS] [macos] [1/6 ios_device: ios_target]<details><summary>(1 failed tests)</summary>&nbsp;&nbsp;FirebaseStorageTest.TestWriteAndReadFileWithCustomMetadata</details> | Add flaky tests to **[go/fpl-cpp-flake-tracker](http://go/fpl-cpp-flake-tracker)** <hidden value="integration-test-status-comment"></hidden> *** ### ✅&nbsp; [build against SDK] Integration test succeeded! Requested by @firebase-workflow-trigger[bot] on commit 8b4226383959ef9558c4bf13e3b40012b4f75d08 Last updated: Sat Apr 8 11:44 PDT 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4645148803)** <hidden value="integration-test-status-comment"></hidden>
1.0
[C++] Nightly Integration Testing Report - Note: This report excludes firestore. Please also check **[the report for firestore](https://github.com/firebase/firebase-cpp-sdk/issues/1178)** *** <hidden value="integration-test-status-comment"></hidden> ### [build against repo] Integration test with FLAKINESS (succeeded after retry) Requested by @DellaBitta on commit 8b4226383959ef9558c4bf13e3b40012b4f75d08 Last updated: Sat Apr 8 03:50 PDT 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4644379829)** | Failures | Configs | |----------|---------| | messaging | [TEST] [FLAKINESS] [Android] [1/3 os: windows] [1/4 android_device: android_target]<details><summary>(1 failed tests)</summary>&nbsp;&nbsp;CRASH/TIMEOUT</details> | | storage | [TEST] [FLAKINESS] [iOS] [macos] [1/6 ios_device: ios_target]<details><summary>(1 failed tests)</summary>&nbsp;&nbsp;FirebaseStorageTest.TestWriteAndReadFileWithCustomMetadata</details> | Add flaky tests to **[go/fpl-cpp-flake-tracker](http://go/fpl-cpp-flake-tracker)** <hidden value="integration-test-status-comment"></hidden> *** ### ✅&nbsp; [build against SDK] Integration test succeeded! Requested by @firebase-workflow-trigger[bot] on commit 8b4226383959ef9558c4bf13e3b40012b4f75d08 Last updated: Sat Apr 8 11:44 PDT 2023 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4645148803)** <hidden value="integration-test-status-comment"></hidden>
process
nightly integration testing report note this report excludes firestore please also check integration test with flakiness succeeded after retry requested by dellabitta on commit last updated sat apr pdt failures configs messaging failed tests nbsp nbsp crash timeout storage failed tests nbsp nbsp firebasestoragetest testwriteandreadfilewithcustommetadata add flaky tests to ✅ nbsp integration test succeeded requested by firebase workflow trigger on commit last updated sat apr pdt
1
1,264
3,797,110,332
IssuesEvent
2016-03-23 05:26:15
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
No link present in TOC (DITA OT 2.2.2)
bug P1 preprocess/keyref
I'm attaching a sample DITA Map: http://www.oxygenxml.com/forum/files/noLinkInTOC.zip When published to XHTML, the TOC has no link in it. It should have one because inside the subordinate DITA Map there is a topicref with keyref: <topicref keyref="topicref_intro"/> This works with DITA OT 1.8. I tested on the 2.2.3 branch and does not seem to work. This also does not seem to work on the "develop" branch.
1.0
No link present in TOC (DITA OT 2.2.2) - I'm attaching a sample DITA Map: http://www.oxygenxml.com/forum/files/noLinkInTOC.zip When published to XHTML, the TOC has no link in it. It should have one because inside the subordinate DITA Map there is a topicref with keyref: <topicref keyref="topicref_intro"/> This works with DITA OT 1.8. I tested on the 2.2.3 branch and does not seem to work. This also does not seem to work on the "develop" branch.
process
no link present in toc dita ot i m attaching a sample dita map when published to xhtml the toc has no link in it it should have one because inside the subordinate dita map there is a topicref with keyref this works with dita ot i tested on the branch and does not seem to work this also does not seem to work on the develop branch
1
327,617
28,073,959,400
IssuesEvent
2023-03-29 21:24:13
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Go to Test Alt+Enter doesn't work on Linux
bug testing
Testing microsoft/vscode-python#17039 testing on Linux Debian build Select a test in the Test Explorer and press Alt+Enter (called out in the context menu). Focus goes to the File menu and the cursor does not go to the selected test in the editor.
1.0
Go to Test Alt+Enter doesn't work on Linux - Testing microsoft/vscode-python#17039 testing on Linux Debian build Select a test in the Test Explorer and press Alt+Enter (called out in the context menu). Focus goes to the File menu and the cursor does not go to the selected test in the editor.
non_process
go to test alt enter doesn t work on linux testing microsoft vscode python testing on linux debian build select a test in the test explorer and press alt enter called out in the context menu focus goes to the file menu and the cursor does not go to the selected test in the editor
0
85,635
3,697,061,759
IssuesEvent
2016-02-27 11:57:17
TwidereProject/Twidere-Android
https://api.github.com/repos/TwidereProject/Twidere-Android
closed
"Like"/fav star is not saved between timeline and tweet view
category:ui priority:high type:bug
When I "like" a tweet in my timeline and then go to the tweet detail view the "like" is not shown? Seems like a recent regression.
1.0
"Like"/fav star is not saved between timeline and tweet view - When I "like" a tweet in my timeline and then go to the tweet detail view the "like" is not shown? Seems like a recent regression.
non_process
like fav star is not saved between timeline and tweet view when i like a tweet in my timeline and then go to the tweet detail view the like is not shown seems like a recent regression
0
13,918
2,789,755,504
IssuesEvent
2015-05-08 21:17:26
google/google-visualization-api-issues
https://api.github.com/repos/google/google-visualization-api-issues
closed
Gauges API fails to display in IE8
Priority-Medium Type-Defect
Original [issue 58](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=58) created by orwant on 2009-09-18T04:57:11.000Z: <b>What steps will reproduce the problem? Please provide a link to a</b> <b>demonstration page if at all possible, or attach code.</b> 1.Make sure your IE8 is in IE8 mode, not Compatibility View or IE7 2.Load the following URL: http://code.google.com/apis/visualization/documentation/gallery/gauge.html#Example The gauges do not display. <b>What component is this issue related to (PieChart, LineChart, DataTable,</b> <b>Query, etc)?</b> Gauges <b>Are you using the test environment (version 1.1)?</b> <b>(If you are not sure, answer NO)</b> NO <b>What operating system and browser are you using?</b> Windows 7 64 bit RC and IE8.0.7100 For those who need a workaround add the following meta tag to your head section (make sure it's before any CSS or other includes. &lt;meta http-equiv=&quot;X-UA-Compatible&quot; content=&quot;IE=7&quot; /&gt; or &lt;meta http-equiv=&quot;X-UA-Compatible&quot; content=&quot;IE=EmulateIE7&quot; /&gt; <b>*********************************************************</b> <b>For developers viewing this issue: please click the 'star' icon to be</b> <b>notified of future changes, and to let us know how many of you are</b> <b>interested in seeing it resolved.</b> <b>*********************************************************</b>
1.0
Gauges API fails to display in IE8 - Original [issue 58](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=58) created by orwant on 2009-09-18T04:57:11.000Z: <b>What steps will reproduce the problem? Please provide a link to a</b> <b>demonstration page if at all possible, or attach code.</b> 1.Make sure your IE8 is in IE8 mode, not Compatibility View or IE7 2.Load the following URL: http://code.google.com/apis/visualization/documentation/gallery/gauge.html#Example The gauges do not display. <b>What component is this issue related to (PieChart, LineChart, DataTable,</b> <b>Query, etc)?</b> Gauges <b>Are you using the test environment (version 1.1)?</b> <b>(If you are not sure, answer NO)</b> NO <b>What operating system and browser are you using?</b> Windows 7 64 bit RC and IE8.0.7100 For those who need a workaround add the following meta tag to your head section (make sure it's before any CSS or other includes. &lt;meta http-equiv=&quot;X-UA-Compatible&quot; content=&quot;IE=7&quot; /&gt; or &lt;meta http-equiv=&quot;X-UA-Compatible&quot; content=&quot;IE=EmulateIE7&quot; /&gt; <b>*********************************************************</b> <b>For developers viewing this issue: please click the 'star' icon to be</b> <b>notified of future changes, and to let us know how many of you are</b> <b>interested in seeing it resolved.</b> <b>*********************************************************</b>
non_process
gauges api fails to display in original created by orwant on what steps will reproduce the problem please provide a link to a demonstration page if at all possible or attach code make sure your is in mode not compatibility view or load the following url the gauges do not display what component is this issue related to piechart linechart datatable query etc gauges are you using the test environment version if you are not sure answer no no what operating system and browser are you using windows bit rc and for those who need a workaround add the following meta tag to your head section make sure it s before any css or other includes lt meta http equiv quot x ua compatible quot content quot ie quot gt or lt meta http equiv quot x ua compatible quot content quot ie quot gt for developers viewing this issue please click the star icon to be notified of future changes and to let us know how many of you are interested in seeing it resolved
0
213,860
24,022,471,646
IssuesEvent
2022-09-15 08:48:10
sast-automation-dev/easybuggy4sb-25
https://api.github.com/repos/sast-automation-dev/easybuggy4sb-25
opened
bootstrap-3.3.7.min.js: 6 vulnerabilities (highest severity is: 6.1)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/resources/templates/head.html</p> <p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p> <p> <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2019-8331](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1 | &#10060; | | [CVE-2018-14040](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14040) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0 | &#10060; | | [CVE-2018-20677](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20677) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0 | &#10060; | | [CVE-2018-20676](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | bootstrap - 3.4.0 | &#10060; | | [CVE-2018-14042](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0 | &#10060; | | [CVE-2016-10735](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | 3.4.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-8331</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/resources/templates/head.html</p> <p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute. <p>Publish Date: 2019-02-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-02-20</p> <p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-14040</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/resources/templates/head.html</p> <p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 4.1.2, XSS is possible in the collapse data-parent attribute. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14040>CVE-2018-14040</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-20677</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/resources/templates/head.html</p> <p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.0, XSS is possible in the affix configuration target property. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20677>CVE-2018-20677</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-20676</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/resources/templates/head.html</p> <p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: bootstrap - 3.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-14042</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/resources/templates/head.html</p> <p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042>CVE-2018-14042</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2016-10735</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/resources/templates/head.html</p> <p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: 3.4.0</p> </p> <p></p> </details>
True
bootstrap-3.3.7.min.js: 6 vulnerabilities (highest severity is: 6.1) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/resources/templates/head.html</p> <p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p> <p> <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2019-8331](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1 | &#10060; | | [CVE-2018-14040](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14040) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0 | &#10060; | | [CVE-2018-20677](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20677) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0 | &#10060; | | [CVE-2018-20676](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | bootstrap - 3.4.0 | &#10060; | | [CVE-2018-14042](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0 | &#10060; | | [CVE-2016-10735](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | 3.4.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-8331</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/resources/templates/head.html</p> <p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute. <p>Publish Date: 2019-02-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-02-20</p> <p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-14040</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/resources/templates/head.html</p> <p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 4.1.2, XSS is possible in the collapse data-parent attribute. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14040>CVE-2018-14040</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-20677</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/resources/templates/head.html</p> <p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.0, XSS is possible in the affix configuration target property. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20677>CVE-2018-20677</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-20676</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/resources/templates/head.html</p> <p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: bootstrap - 3.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-14042</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/resources/templates/head.html</p> <p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042>CVE-2018-14042</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2016-10735</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/resources/templates/head.html</p> <p>Path to vulnerable library: /src/main/resources/templates/head.html,/src/main/webapp/dfi/style_bootstrap.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/easybuggy4sb-25/commit/5db13b8ddf701e4c247d78eea18c6b360b1687e1">5db13b8ddf701e4c247d78eea18c6b360b1687e1</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: 3.4.0</p> </p> <p></p> </details>
non_process
bootstrap min js vulnerabilities highest severity is vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main resources templates head html path to vulnerable library src main resources templates head html src main webapp dfi style bootstrap html found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available medium bootstrap min js direct bootstrap bootstrap sass medium bootstrap min js direct org webjars npm bootstrap org webjars bootstrap medium bootstrap min js direct bootstrap nordron angulartemplate dynamic net express projecttemplates dotnetng template znxtapp core module theme beta jmeter medium bootstrap min js direct bootstrap medium bootstrap min js direct org webjars npm bootstrap org webjars bootstrap medium bootstrap min js direct details cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main resources templates head html path to vulnerable library src main resources templates head html src main webapp dfi style bootstrap html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before and x before xss is possible in the tooltip or popover data template attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution bootstrap bootstrap sass cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main resources templates head html path to vulnerable library src main resources templates head html src main webapp dfi style bootstrap html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the collapse data parent attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution org webjars npm bootstrap org webjars bootstrap cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main resources templates head html path to vulnerable library src main resources templates head html src main webapp dfi style bootstrap html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the affix configuration target property publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap nordron angulartemplate dynamic net express projecttemplates dotnetng template znxtapp core module theme beta jmeter cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main resources templates head html path to vulnerable library src main resources templates head html src main webapp dfi style bootstrap html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the tooltip data viewport attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main resources templates head html path to vulnerable library src main resources templates head html src main webapp dfi style bootstrap html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the data container property of tooltip publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution org webjars npm bootstrap org webjars bootstrap cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main resources templates head html path to vulnerable library src main resources templates head html src main webapp dfi style bootstrap html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap x before and x beta before beta xss is possible in the data target attribute a different vulnerability than cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution
0
610,669
18,920,636,222
IssuesEvent
2021-11-17 00:57:35
googleapis/sphinx-docfx-yaml
https://api.github.com/repos/googleapis/sphinx-docfx-yaml
opened
Ensure handwritten libraries are well supported
type: feature request priority: p1
Googlers see b/206698969. There are a lot of formats that the plugin does not support yet, which is a blocking issue for some of the handwritten libraries that are using much more variety of Sphinx features. The documentation does not get rendered properly, which could lead to a poor customer experience for the documentation.
1.0
Ensure handwritten libraries are well supported - Googlers see b/206698969. There are a lot of formats that the plugin does not support yet, which is a blocking issue for some of the handwritten libraries that are using much more variety of Sphinx features. The documentation does not get rendered properly, which could lead to a poor customer experience for the documentation.
non_process
ensure handwritten libraries are well supported googlers see b there are a lot of formats that the plugin does not support yet which is a blocking issue for some of the handwritten libraries that are using much more variety of sphinx features the documentation does not get rendered properly which could lead to a poor customer experience for the documentation
0
21,843
6,226,906,102
IssuesEvent
2017-07-10 19:31:25
XceedBoucherS/TestImport5
https://api.github.com/repos/XceedBoucherS/TestImport5
closed
Control Request: TimePicker
CodePlex
<b>brianlagunas[CodePlex]</b> <br />Request for a new TimePicker control. This control should have similar functionality to the TimePicker in the Silverlight Toolkit
1.0
Control Request: TimePicker - <b>brianlagunas[CodePlex]</b> <br />Request for a new TimePicker control. This control should have similar functionality to the TimePicker in the Silverlight Toolkit
non_process
control request timepicker brianlagunas request for a new timepicker control this control should have similar functionality to the timepicker in the silverlight toolkit
0
10,960
13,766,192,902
IssuesEvent
2020-10-07 14:17:29
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
macOS in Environment
Pri2 devops-cicd-process/tech devops/prod product-question
[Enter feedback here] Is there a way to create an environment using macOS instead of Windows or Linux? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 91d0d31f-81ee-c024-db7e-daddbf525f71 * Version Independent ID: 330f1649-386c-d0aa-5f96-b8343a1480d3 * Content: [Environment - Virtual machine resource - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments-virtual-machines?view=azure-devops) * Content Source: [docs/pipelines/process/environments-virtual-machines.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments-virtual-machines.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
macOS in Environment - [Enter feedback here] Is there a way to create an environment using macOS instead of Windows or Linux? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 91d0d31f-81ee-c024-db7e-daddbf525f71 * Version Independent ID: 330f1649-386c-d0aa-5f96-b8343a1480d3 * Content: [Environment - Virtual machine resource - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments-virtual-machines?view=azure-devops) * Content Source: [docs/pipelines/process/environments-virtual-machines.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments-virtual-machines.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
macos in environment is there a way to create an environment using macos instead of windows or linux document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
16,915
22,263,827,117
IssuesEvent
2022-06-10 04:54:19
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
Writer timer triggered event with process instance key
kind/feature scope/broker team/process-automation
**Is your feature request related to a problem? Please describe.** I deployed a process with a timer start event. If the timer is triggered then a new process instance is created and an event of type `timer` with intent `triggered` is written. As a monitoring application, I want to track which process instance was created by this timer. But currently, the `timer: triggered` event doesn't contain the key of the created process instance. **Describe the solution you'd like** The `timer: triggered` event contains the key of the created process instance. **Describe alternatives you've considered** Add a reference of the timer to the created process instance. But this option seems not practical because a process instance can be created by different triggers, for example, a message. **Additional context** In a monitoring application, I want to track that a process instance was created by a triggered timer start event.
1.0
Writer timer triggered event with process instance key - **Is your feature request related to a problem? Please describe.** I deployed a process with a timer start event. If the timer is triggered then a new process instance is created and an event of type `timer` with intent `triggered` is written. As a monitoring application, I want to track which process instance was created by this timer. But currently, the `timer: triggered` event doesn't contain the key of the created process instance. **Describe the solution you'd like** The `timer: triggered` event contains the key of the created process instance. **Describe alternatives you've considered** Add a reference of the timer to the created process instance. But this option seems not practical because a process instance can be created by different triggers, for example, a message. **Additional context** In a monitoring application, I want to track that a process instance was created by a triggered timer start event.
process
writer timer triggered event with process instance key is your feature request related to a problem please describe i deployed a process with a timer start event if the timer is triggered then a new process instance is created and an event of type timer with intent triggered is written as a monitoring application i want to track which process instance was created by this timer but currently the timer triggered event doesn t contain the key of the created process instance describe the solution you d like the timer triggered event contains the key of the created process instance describe alternatives you ve considered add a reference of the timer to the created process instance but this option seems not practical because a process instance can be created by different triggers for example a message additional context in a monitoring application i want to track that a process instance was created by a triggered timer start event
1
52,836
13,064,101,982
IssuesEvent
2020-07-30 17:32:50
GoogleCloudPlatform/golang-samples
https://api.github.com/repos/GoogleCloudPlatform/golang-samples
closed
asset/quickstart/batch-get-assets-history: TestMain failed
api: cloudasset buildcop: issue priority: p1 sample type: bug
This test failed! To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop). If I'm commenting on this issue too often, add the `buildcop: quiet` label and I will stop commenting. --- commit: 9d203dfe6a2bf97041383afc7c74880cdfbec364 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/f337f766-bb8e-4406-a4cb-9c2a7048957c), [Sponge](http://sponge2/f337f766-bb8e-4406-a4cb-9c2a7048957c) status: failed
1.0
asset/quickstart/batch-get-assets-history: TestMain failed - This test failed! To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop). If I'm commenting on this issue too often, add the `buildcop: quiet` label and I will stop commenting. --- commit: 9d203dfe6a2bf97041383afc7c74880cdfbec364 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/f337f766-bb8e-4406-a4cb-9c2a7048957c), [Sponge](http://sponge2/f337f766-bb8e-4406-a4cb-9c2a7048957c) status: failed
non_process
asset quickstart batch get assets history testmain failed this test failed to configure my behavior see if i m commenting on this issue too often add the buildcop quiet label and i will stop commenting commit buildurl status failed
0
226,832
18,044,208,715
IssuesEvent
2021-09-18 15:50:14
logicmoo/logicmoo_workspace
https://api.github.com/repos/logicmoo/logicmoo_workspace
opened
logicmoo.pfc.test.sanity_base.NEG_01V JUnit
Test_9999 logicmoo.pfc.test.sanity_base unit_test NEG_01V
(cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc) GH_MASTER_ISSUE_FINFO= ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ANEG_01V GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3 https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/pfc/t/sanity_base/neg_01v.pfc Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/NEG_01V/logicmoo_pfc_test_sanity_base_NEG_01V_JUnit/ This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/67/testReport/logicmoo.pfc.test.sanity_base/NEG_01V/logicmoo_pfc_test_sanity_base_NEG_01V_JUnit/ GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3 https://github.com/logicmoo/logicmoo_workspace/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/pfc/t/sanity_base/neg_01v.pfc ``` % running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/neg_01v.pfc'), %~ /var/lib/jenkins/.local/share/swi-prolog/pack/logicmoo_utils/prolog/logicmoo_test_header.pl:92 %~ this_test_might_need( :-( use_module( library(logicmoo_plarkc)))) :- use_module(library(statistics)). %:- mpred_notrace_exec. % reset runtime counter %:- mpred_notrace_exec. % reset runtime counter :- statistics(runtime,_Secs). ~path(X,X). path(1,2). path(2,3). path(3,4). path(1,1). :- mpred_test(path(3, 4)). %~ mpred_test("Test_0001_Line_0000__path_3",baseKB:path(3,4)) %~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L20 /*~ %~ mpred_test("Test_0001_Line_0000__path_3",baseKB:path(3,4)) passed=info(why_was_true(baseKB:path(3,4))) Justifications for path(3,4):  1.1 mfl4(_,baseKB,'* https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L17 ',17)  name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0001_Line_0000__path_3'. JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'. JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'. % saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0001_Line_0000__path_3-junit.xml ~*/ :- mpred_test(path(2, 3)). %~ mpred_test("Test_0002_Line_0000__path_2",baseKB:path(2,3)) %~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L21 /*~ %~ mpred_test("Test_0002_Line_0000__path_2",baseKB:path(2,3)) passed=info(why_was_true(baseKB:path(2,3))) Justifications for path(2,3):  1.1 mfl4(_,baseKB,'* https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L16 ',16)  name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0002_Line_0000__path_2'. JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'. JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'. % saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0002_Line_0000__path_2-junit.xml ~*/ :- mpred_test(path(1, 2)). %~ mpred_test("Test_0003_Line_0000__path_1",baseKB:path(1,2)) %~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L22 /*~ %~ mpred_test("Test_0003_Line_0000__path_1",baseKB:path(1,2)) passed=info(why_was_true(baseKB:path(1,2))) Justifications for path(1,2):  1.1 mfl4(_,baseKB,'* https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L15 ',15)  name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0003_Line_0000__path_1'. JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'. JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'. % saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0003_Line_0000__path_1-junit.xml ~*/ :- mpred_test(~path(1,1)). %~ mpred_test("Test_0004_Line_0000__path_1",baseKB: ~path(1,1)) /*~ %~ mpred_test("Test_0004_Line_0000__path_1",baseKB: ~path(1,1)) ^ Call: (68) [baseKB] ~path(1, 1) ^ Unify: (68) [baseKB] ~ (baseKB:path(1, 1)) ^ Call: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(1, 1)), info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1320, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(1, 1))))) ^ Unify: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(1, 1)), info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1320, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(1, 1))))) Call: (76) [system] set_prolog_flag(last_call_optimisation, false) Exit: (76) [system] set_prolog_flag(last_call_optimisation, false) ^ Call: (76) [loop_check] prolog_frame_attribute(1320, parent_goal, loop_check_term_frame(_87158, info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, _87164, _87166)) ^ Fail: (76) [loop_check] prolog_frame_attribute(1320, parent_goal, loop_check_term_frame(_87158, info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, _87164, _87166)) ^ Redo: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(1, 1)), info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1320, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(1, 1))))) Call: (76) [pfc_lib] neg_in_code0(baseKB:path(1, 1)) Unify: (76) [pfc_lib] neg_in_code0(baseKB:path(1, 1)) ^ Call: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(1, 1)), info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1459, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(1, 1))))) ^ Unify: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(1, 1)), info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1459, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(1, 1))))) Call: (83) [system] set_prolog_flag(last_call_optimisation, false) Exit: (83) [system] set_prolog_flag(last_call_optimisation, false) ^ Call: (83) [loop_check] prolog_frame_attribute(1459, parent_goal, loop_check_term_frame(_92882, info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, _92888, _92890)) ^ Fail: (83) [loop_check] prolog_frame_attribute(1459, parent_goal, loop_check_term_frame(_92882, info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, _92888, _92890)) ^ Redo: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(1, 1)), info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1459, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(1, 1))))) Call: (83) [pfc_lib] neg_may_naf(baseKB:path(1, 1)) Unify: (83) [pfc_lib] neg_may_naf(baseKB:path(1, 1)) ^ Call: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _96214) ^ Unify: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _96214) ^ Call: (88) [system] clause(pfc_lib:prologNegByFailure(path), true, _96214) ^ Fail: (88) [system] clause(pfc_lib:prologNegByFailure(path), true, _96214) ^ Fail: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _96214) Unify: (83) [pfc_lib] neg_may_naf(baseKB:path(1, 1)) ^ Call: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(1, 1)) ^ Unify: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(1, 1)) ^ Call: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(1, 1)) ^ Unify: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(1, 1)) ^ Fail: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(1, 1)) ^ Redo: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(1, 1)) ^ Exit: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(1, 1)) ^ Call: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384), call(_104384)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)))) ^ Unify: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384), call(_104384)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)))) ^ Call: (90) [hook_database] clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384) ^ Fail: (90) [hook_database] clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384) Call: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)) Unify: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)) ^ Call: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170)) ^ Unify: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170)) ^ Call: (92) [baseKB] clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170) ^ Fail: (92) [baseKB] clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170) ^ Fail: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170)) Fail: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)) ^ Fail: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384), call(_104384)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)))) ^ Call: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _112590) ^ Unify: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, syntaxic(_113160)) ^ Call: (88) [pfc_lib] mpred_database_term_syntax(path, 2, _113160) ^ Fail: (88) [pfc_lib] mpred_database_term_syntax(path, 2, _113160) ^ Redo: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _115162) ^ Unify: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _115790) ^ Call: (88) [pfc_lib] mpred_core_database_term(path, 2, _116358) ^ Fail: (88) [pfc_lib] mpred_core_database_term(path, 2, _116358) ^ Fail: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _117732) ^ Call: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336), call(_118336)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)))) ^ Unify: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336), call(_118336)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)))) ^ Call: (88) [hook_database] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336) ^ Fail: (88) [hook_database] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336) Call: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)) Unify: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)) ^ Call: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122)) ^ Unify: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122)) ^ Call: (90) [baseKB] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122) ^ Fail: (90) [baseKB] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122) ^ Fail: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122)) Fail: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)) ^ Fail: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336), call(_118336)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)))) ^ Call: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_126488, path, 2, prologHybrid), _126518), call(_126518)*->true;clause_b(baseKB:mpred_prop(_126488, path, 2, prologHybrid)))) ^ Unify: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_126488, path, 2, prologHybrid), _126518), call(_126518)*->true;clause_b(baseKB:mpred_prop(_126488, path, 2, prologHybrid)))) ^ Call: (91) [hook_database] clause(mpred_prop(_126488, path, 2, prologHybrid), _126518) ^ Fail: (91) [hook_database] clause(mpred_prop(_126488, path, 2, prologHybrid), _126518) Call: (91) [hook_database] clause_b(baseKB:mpred_prop(_126488, path, 2, prologHybrid)) Unify: (91) [hook_database] clause_b(baseKB:mpred_prop(_126488, path, 2, prologHybrid)) ^ Call: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_126488, path, 2, prologHybrid), _130304)) ^ Unify: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_1434, path, 2, prologHybrid), _1490)) ^ Call: (93) [baseKB] clause(mpred_prop(_1434, path, 2, prologHybrid), _1490) ^ Fail: (93) [baseKB] clause(mpred_prop(_1434, path, 2, prologHybrid), _1490) ^ Fail: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_1434, path, 2, prologHybrid), _1490)) Fail: (91) [hook_database] clause_b(baseKB:mpred_prop(_1434, path, 2, prologHybrid)) ^ Fail: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_1434, path, 2, prologHybrid), _1464), call(_1464)*->true;clause_b(baseKB:mpred_prop(_1434, path, 2, prologHybrid)))) ^ Call: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _5272) ^ Unify: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, syntaxic(_5842)) ^ Call: (89) [pfc_lib] mpred_database_term_syntax(path, 2, _5842) ^ Fail: (89) [pfc_lib] mpred_database_term_syntax(path, 2, _5842) ^ Redo: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _7844) ^ Unify: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _8472) ^ Call: (89) [pfc_lib] mpred_core_database_term(path, 2, _9040) ^ Fail: (89) [pfc_lib] mpred_core_database_term(path, 2, _9040) ^ Fail: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _10414) Call: (98) [$autoload] leave_sandbox(_11012) Unify: (98) [$autoload] leave_sandbox(_11012) Exit: (98) [$autoload] leave_sandbox(false) Call: (97) [$autoload] restore_sandbox(false) Unify: (97) [$autoload] restore_sandbox(false) Exit: (97) [$autoload] restore_sandbox(false) Fail: (83) [pfc_lib] neg_may_naf(baseKB:path(1, 1)) ^ Fail: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(1, 1)), info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1459, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(1, 1))))) Fail: (76) [pfc_lib] neg_in_code0(baseKB:path(1, 1)) ^ Fail: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(1, 1)), info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1320, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(1, 1))))) ^ Fail: (68) [baseKB] ~ (baseKB:path(1, 1)) ^ Call: (68) [must_sanity] must_sanity:mquietly_if(true, rtrace:tAt_normal) ^ Unify: (68) [must_sanity] must_sanity:mquietly_if(true, rtrace:tAt_normal) failure=info((why_was_true(baseKB:(\+ ~path(1,1))),rtrace(baseKB: ~path(1,1)))) no_proof_for(\+ ~path(1,1)). no_proof_for(\+ ~path(1,1)). no_proof_for(\+ ~path(1,1)). name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0004_Line_0000__path_1'. JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'. JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'. % saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0004_Line_0000__path_1-junit.xml ~*/ :- mpred_test(~path(2,2)). %~ mpred_test("Test_0005_Line_0000__path_2",baseKB: ~path(2,2)) /*~ %~ mpred_test("Test_0005_Line_0000__path_2",baseKB: ~path(2,2)) ^ Call: (68) [baseKB] ~path(2, 2) ^ Unify: (68) [baseKB] ~ (baseKB:path(2, 2)) ^ Call: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(2, 2)), info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1189, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(2, 2))))) ^ Unify: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(2, 2)), info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1189, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(2, 2))))) Call: (76) [system] set_prolog_flag(last_call_optimisation, false) Exit: (76) [system] set_prolog_flag(last_call_optimisation, false) ^ Call: (76) [loop_check] prolog_frame_attribute(1189, parent_goal, loop_check_term_frame(_121764, info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, _121770, _121772)) ^ Fail: (76) [loop_check] prolog_frame_attribute(1189, parent_goal, loop_check_term_frame(_121764, info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, _121770, _121772)) ^ Redo: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(2, 2)), info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1189, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(2, 2))))) Call: (76) [pfc_lib] neg_in_code0(baseKB:path(2, 2)) Unify: (76) [pfc_lib] neg_in_code0(baseKB:path(2, 2)) ^ Call: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(2, 2)), info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1328, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(2, 2))))) ^ Unify: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(2, 2)), info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1328, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(2, 2))))) Call: (83) [system] set_prolog_flag(last_call_optimisation, false) Exit: (83) [system] set_prolog_flag(last_call_optimisation, false) ^ Call: (83) [loop_check] prolog_frame_attribute(1328, parent_goal, loop_check_term_frame(_127488, info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, _127494, _127496)) ^ Fail: (83) [loop_check] prolog_frame_attribute(1328, parent_goal, loop_check_term_frame(_127488, info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, _127494, _127496)) ^ Redo: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(2, 2)), info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1328, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(2, 2))))) Call: (83) [pfc_lib] neg_may_naf(baseKB:path(2, 2)) Unify: (83) [pfc_lib] neg_may_naf(baseKB:path(2, 2)) ^ Call: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _130820) ^ Unify: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _130820) ^ Call: (88) [system] clause(pfc_lib:prologNegByFailure(path), true, _130820) ^ Fail: (88) [system] clause(pfc_lib:prologNegByFailure(path), true, _130820) ^ Fail: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _130820) Unify: (83) [pfc_lib] neg_may_naf(baseKB:path(2, 2)) ^ Call: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(2, 2)) ^ Unify: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(2, 2)) ^ Call: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(2, 2)) ^ Unify: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(2, 2)) ^ Fail: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(2, 2)) ^ Redo: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(2, 2)) ^ Exit: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(2, 2)) ^ Call: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990), call(_138990)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)))) ^ Unify: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990), call(_138990)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)))) ^ Call: (90) [hook_database] clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990) ^ Fail: (90) [hook_database] clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990) Call: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)) Unify: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)) ^ Call: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776)) ^ Unify: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776)) ^ Call: (92) [baseKB] clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776) ^ Fail: (92) [baseKB] clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776) ^ Fail: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776)) Fail: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)) ^ Fail: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990), call(_138990)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)))) ^ Call: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _147196) ^ Unify: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, syntaxic(_147766)) ^ Call: (88) [pfc_lib] mpred_database_term_syntax(path, 2, _147766) ^ Fail: (88) [pfc_lib] mpred_database_term_syntax(path, 2, _147766) ^ Redo: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _149768) ^ Unify: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _150396) ^ Call: (88) [pfc_lib] mpred_core_database_term(path, 2, _150964) ^ Fail: (88) [pfc_lib] mpred_core_database_term(path, 2, _150964) ^ Fail: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _152338) ^ Call: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942), call(_152942)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)))) ^ Unify: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942), call(_152942)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)))) ^ Call: (88) [hook_database] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942) ^ Fail: (88) [hook_database] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942) Call: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)) Unify: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)) ^ Call: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728)) ^ Unify: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728)) ^ Call: (90) [baseKB] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728) ^ Fail: (90) [baseKB] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728) ^ Fail: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728)) Fail: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)) ^ Fail: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942), call(_152942)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)))) ^ Call: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_161094, path, 2, prologHybrid), _161124), call(_161124)*->true;clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid)))) ^ Unify: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_161094, path, 2, prologHybrid), _161124), call(_161124)*->true;clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid)))) ^ Call: (91) [hook_database] clause(mpred_prop(_161094, path, 2, prologHybrid), _161124) ^ Fail: (91) [hook_database] clause(mpred_prop(_161094, path, 2, prologHybrid), _161124) Call: (91) [hook_database] clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid)) Unify: (91) [hook_database] clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid)) ^ Call: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_161094, path, 2, prologHybrid), _164910)) ^ Unify: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_161094, path, 2, prologHybrid), _164910)) ^ Call: (93) [baseKB] clause(mpred_prop(_161094, path, 2, prologHybrid), _164910) ^ Fail: (93) [baseKB] clause(mpred_prop(_161094, path, 2, prologHybrid), _164910) ^ Fail: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_161094, path, 2, prologHybrid), _164910)) Fail: (91) [hook_database] clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid)) ^ Fail: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_161094, path, 2, prologHybrid), _161124), call(_161124)*->true;clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid)))) ^ Call: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _169330) ^ Unify: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, syntaxic(_169900)) ^ Call: (89) [pfc_lib] mpred_database_term_syntax(path, 2, _169900) ^ Fail: (89) [pfc_lib] mpred_database_term_syntax(path, 2, _169900) ^ Redo: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _171902) ^ Unify: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _172530) ^ Call: (89) [pfc_lib] mpred_core_database_term(path, 2, _173098) ^ Fail: (89) [pfc_lib] mpred_core_database_term(path, 2, _173098) ^ Fail: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _174472) Call: (98) [$autoload] leave_sandbox(_175070) Unify: (98) [$autoload] leave_sandbox(_175070) Exit: (98) [$autoload] leave_sandbox(false) Call: (97) [$autoload] restore_sandbox(false) Unify: (97) [$autoload] restore_sandbox(false) Exit: (97) [$autoload] restore_sandbox(false) Fail: (83) [pfc_lib] neg_may_naf(baseKB:path(2, 2)) ^ Fail: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(2, 2)), info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1328, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(2, 2))))) Fail: (76) [pfc_lib] neg_in_code0(baseKB:path(2, 2)) ^ Fail: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(2, 2)), info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1189, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(2, 2))))) ^ Fail: (68) [baseKB] ~ (baseKB:path(2, 2)) ^ Call: (68) [must_sanity] must_sanity:mquietly_if(true, rtrace:tAt_normal) ^ Unify: (68) [must_sanity] must_sanity:mquietly_if(true, rtrace:tAt_normal) failure=info((why_was_true(baseKB:(\+ ~path(2,2))),rtrace(baseKB: ~path(2,2)))) no_proof_for(\+ ~path(2,2)). no_proof_for(\+ ~path(2,2)). no_proof_for(\+ ~path(2,2)). name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0005_Line_0000__path_2'. JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'. JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'. % saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0005_Line_0000__path_2-junit.xml ~*/ /* :- mpred_test(\+path(1,1)). :- mpred_test(\+path(2,2)). */ :- mpred_test(\+path(3,1)). %~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/neg_01v.pfc:32 %~ mpred_test("Test_0006_Line_0000__naf_path_3",baseKB:(\+path(3,1))) /*~ %~ mpred_test("Test_0006_Line_0000__naf_path_3",baseKB:(\+path(3,1))) passed=info(why_was_true(baseKB:(\+path(3,1)))) no_proof_for(\+path(3,1)). no_proof_for(\+path(3,1)). no_proof_for(\+path(3,1)). name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0006_Line_0000__naf_path_3'. JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'. JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'. % saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0006_Line_0000__naf_path_3-junit.xml ~*/ :- mpred_test(\+ ~path(1,5)). %~ mpred_test("Test_0007_Line_0000__naf_path_1",baseKB:(\+ ~path(1,5))) /*~ %~ mpred_test("Test_0007_Line_0000__naf_path_1",baseKB:(\+ ~path(1,5))) passed=info(why_was_true(baseKB:(\+ ~path(1,5)))) no_proof_for(\+ ~path(1,5)). no_proof_for(\+ ~path(1,5)). no_proof_for(\+ ~path(1,5)). name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0007_Line_0000__naf_path_1'. JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'. JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'. % saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0007_Line_0000__naf_path_1-junit.xml ~*/ :- break. %~ skipped(blocks_on_input,break) %~ unused(save_junit_results) %~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/neg_01v.pfc:36 %~ test_completed_exit(6) :- dynamic junit_prop/3. :- dynamic junit_prop/3. :- dynamic junit_prop/3. ``` totalTime=1 ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ANEG_01V GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3 https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/pfc/t/sanity_base/neg_01v.pfc Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/NEG_01V/logicmoo_pfc_test_sanity_base_NEG_01V_JUnit/ This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/67/testReport/logicmoo.pfc.test.sanity_base/NEG_01V/logicmoo_pfc_test_sanity_base_NEG_01V_JUnit/ GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3 https://github.com/logicmoo/logicmoo_workspace/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/pfc/t/sanity_base/neg_01v.pfc FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k neg_01v.pfc (returned 6)
3.0
logicmoo.pfc.test.sanity_base.NEG_01V JUnit - (cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc) GH_MASTER_ISSUE_FINFO= ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ANEG_01V GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3 https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/pfc/t/sanity_base/neg_01v.pfc Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/NEG_01V/logicmoo_pfc_test_sanity_base_NEG_01V_JUnit/ This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/67/testReport/logicmoo.pfc.test.sanity_base/NEG_01V/logicmoo_pfc_test_sanity_base_NEG_01V_JUnit/ GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3 https://github.com/logicmoo/logicmoo_workspace/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/pfc/t/sanity_base/neg_01v.pfc ``` % running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/neg_01v.pfc'), %~ /var/lib/jenkins/.local/share/swi-prolog/pack/logicmoo_utils/prolog/logicmoo_test_header.pl:92 %~ this_test_might_need( :-( use_module( library(logicmoo_plarkc)))) :- use_module(library(statistics)). %:- mpred_notrace_exec. % reset runtime counter %:- mpred_notrace_exec. % reset runtime counter :- statistics(runtime,_Secs). ~path(X,X). path(1,2). path(2,3). path(3,4). path(1,1). :- mpred_test(path(3, 4)). %~ mpred_test("Test_0001_Line_0000__path_3",baseKB:path(3,4)) %~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L20 /*~ %~ mpred_test("Test_0001_Line_0000__path_3",baseKB:path(3,4)) passed=info(why_was_true(baseKB:path(3,4))) Justifications for path(3,4):  1.1 mfl4(_,baseKB,'* https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L17 ',17)  name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0001_Line_0000__path_3'. JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'. JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'. % saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0001_Line_0000__path_3-junit.xml ~*/ :- mpred_test(path(2, 3)). %~ mpred_test("Test_0002_Line_0000__path_2",baseKB:path(2,3)) %~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L21 /*~ %~ mpred_test("Test_0002_Line_0000__path_2",baseKB:path(2,3)) passed=info(why_was_true(baseKB:path(2,3))) Justifications for path(2,3):  1.1 mfl4(_,baseKB,'* https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L16 ',16)  name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0002_Line_0000__path_2'. JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'. JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'. % saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0002_Line_0000__path_2-junit.xml ~*/ :- mpred_test(path(1, 2)). %~ mpred_test("Test_0003_Line_0000__path_1",baseKB:path(1,2)) %~ FIlE: * https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L22 /*~ %~ mpred_test("Test_0003_Line_0000__path_1",baseKB:path(1,2)) passed=info(why_was_true(baseKB:path(1,2))) Justifications for path(1,2):  1.1 mfl4(_,baseKB,'* https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/blob/master/packs_sys/pfc/t/sanity_base/neg_01v.pfc#L15 ',15)  name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0003_Line_0000__path_1'. JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'. JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'. % saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0003_Line_0000__path_1-junit.xml ~*/ :- mpred_test(~path(1,1)). %~ mpred_test("Test_0004_Line_0000__path_1",baseKB: ~path(1,1)) /*~ %~ mpred_test("Test_0004_Line_0000__path_1",baseKB: ~path(1,1)) ^ Call: (68) [baseKB] ~path(1, 1) ^ Unify: (68) [baseKB] ~ (baseKB:path(1, 1)) ^ Call: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(1, 1)), info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1320, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(1, 1))))) ^ Unify: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(1, 1)), info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1320, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(1, 1))))) Call: (76) [system] set_prolog_flag(last_call_optimisation, false) Exit: (76) [system] set_prolog_flag(last_call_optimisation, false) ^ Call: (76) [loop_check] prolog_frame_attribute(1320, parent_goal, loop_check_term_frame(_87158, info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, _87164, _87166)) ^ Fail: (76) [loop_check] prolog_frame_attribute(1320, parent_goal, loop_check_term_frame(_87158, info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, _87164, _87166)) ^ Redo: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(1, 1)), info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1320, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(1, 1))))) Call: (76) [pfc_lib] neg_in_code0(baseKB:path(1, 1)) Unify: (76) [pfc_lib] neg_in_code0(baseKB:path(1, 1)) ^ Call: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(1, 1)), info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1459, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(1, 1))))) ^ Unify: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(1, 1)), info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1459, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(1, 1))))) Call: (83) [system] set_prolog_flag(last_call_optimisation, false) Exit: (83) [system] set_prolog_flag(last_call_optimisation, false) ^ Call: (83) [loop_check] prolog_frame_attribute(1459, parent_goal, loop_check_term_frame(_92882, info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, _92888, _92890)) ^ Fail: (83) [loop_check] prolog_frame_attribute(1459, parent_goal, loop_check_term_frame(_92882, info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, _92888, _92890)) ^ Redo: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(1, 1)), info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1459, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(1, 1))))) Call: (83) [pfc_lib] neg_may_naf(baseKB:path(1, 1)) Unify: (83) [pfc_lib] neg_may_naf(baseKB:path(1, 1)) ^ Call: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _96214) ^ Unify: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _96214) ^ Call: (88) [system] clause(pfc_lib:prologNegByFailure(path), true, _96214) ^ Fail: (88) [system] clause(pfc_lib:prologNegByFailure(path), true, _96214) ^ Fail: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _96214) Unify: (83) [pfc_lib] neg_may_naf(baseKB:path(1, 1)) ^ Call: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(1, 1)) ^ Unify: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(1, 1)) ^ Call: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(1, 1)) ^ Unify: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(1, 1)) ^ Fail: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(1, 1)) ^ Redo: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(1, 1)) ^ Exit: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(1, 1)) ^ Call: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384), call(_104384)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)))) ^ Unify: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384), call(_104384)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)))) ^ Call: (90) [hook_database] clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384) ^ Fail: (90) [hook_database] clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384) Call: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)) Unify: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)) ^ Call: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170)) ^ Unify: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170)) ^ Call: (92) [baseKB] clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170) ^ Fail: (92) [baseKB] clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170) ^ Fail: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _108170)) Fail: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)) ^ Fail: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _104384), call(_104384)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)))) ^ Call: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _112590) ^ Unify: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, syntaxic(_113160)) ^ Call: (88) [pfc_lib] mpred_database_term_syntax(path, 2, _113160) ^ Fail: (88) [pfc_lib] mpred_database_term_syntax(path, 2, _113160) ^ Redo: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _115162) ^ Unify: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _115790) ^ Call: (88) [pfc_lib] mpred_core_database_term(path, 2, _116358) ^ Fail: (88) [pfc_lib] mpred_core_database_term(path, 2, _116358) ^ Fail: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _117732) ^ Call: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336), call(_118336)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)))) ^ Unify: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336), call(_118336)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)))) ^ Call: (88) [hook_database] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336) ^ Fail: (88) [hook_database] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336) Call: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)) Unify: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)) ^ Call: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122)) ^ Unify: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122)) ^ Call: (90) [baseKB] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122) ^ Fail: (90) [baseKB] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122) ^ Fail: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _122122)) Fail: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)) ^ Fail: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _118336), call(_118336)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)))) ^ Call: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_126488, path, 2, prologHybrid), _126518), call(_126518)*->true;clause_b(baseKB:mpred_prop(_126488, path, 2, prologHybrid)))) ^ Unify: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_126488, path, 2, prologHybrid), _126518), call(_126518)*->true;clause_b(baseKB:mpred_prop(_126488, path, 2, prologHybrid)))) ^ Call: (91) [hook_database] clause(mpred_prop(_126488, path, 2, prologHybrid), _126518) ^ Fail: (91) [hook_database] clause(mpred_prop(_126488, path, 2, prologHybrid), _126518) Call: (91) [hook_database] clause_b(baseKB:mpred_prop(_126488, path, 2, prologHybrid)) Unify: (91) [hook_database] clause_b(baseKB:mpred_prop(_126488, path, 2, prologHybrid)) ^ Call: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_126488, path, 2, prologHybrid), _130304)) ^ Unify: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_1434, path, 2, prologHybrid), _1490)) ^ Call: (93) [baseKB] clause(mpred_prop(_1434, path, 2, prologHybrid), _1490) ^ Fail: (93) [baseKB] clause(mpred_prop(_1434, path, 2, prologHybrid), _1490) ^ Fail: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_1434, path, 2, prologHybrid), _1490)) Fail: (91) [hook_database] clause_b(baseKB:mpred_prop(_1434, path, 2, prologHybrid)) ^ Fail: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_1434, path, 2, prologHybrid), _1464), call(_1464)*->true;clause_b(baseKB:mpred_prop(_1434, path, 2, prologHybrid)))) ^ Call: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _5272) ^ Unify: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, syntaxic(_5842)) ^ Call: (89) [pfc_lib] mpred_database_term_syntax(path, 2, _5842) ^ Fail: (89) [pfc_lib] mpred_database_term_syntax(path, 2, _5842) ^ Redo: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _7844) ^ Unify: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _8472) ^ Call: (89) [pfc_lib] mpred_core_database_term(path, 2, _9040) ^ Fail: (89) [pfc_lib] mpred_core_database_term(path, 2, _9040) ^ Fail: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _10414) Call: (98) [$autoload] leave_sandbox(_11012) Unify: (98) [$autoload] leave_sandbox(_11012) Exit: (98) [$autoload] leave_sandbox(false) Call: (97) [$autoload] restore_sandbox(false) Unify: (97) [$autoload] restore_sandbox(false) Exit: (97) [$autoload] restore_sandbox(false) Fail: (83) [pfc_lib] neg_may_naf(baseKB:path(1, 1)) ^ Fail: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(1, 1)), info(pfc_lib:neg_may_naf(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1459, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(1, 1))))) Fail: (76) [pfc_lib] neg_in_code0(baseKB:path(1, 1)) ^ Fail: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(1, 1)), info(pfc_lib:neg_in_code0(baseKB:path(1, 1)), 'mpred_core.pl':273), 1, 1320, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(1, 1))))) ^ Fail: (68) [baseKB] ~ (baseKB:path(1, 1)) ^ Call: (68) [must_sanity] must_sanity:mquietly_if(true, rtrace:tAt_normal) ^ Unify: (68) [must_sanity] must_sanity:mquietly_if(true, rtrace:tAt_normal) failure=info((why_was_true(baseKB:(\+ ~path(1,1))),rtrace(baseKB: ~path(1,1)))) no_proof_for(\+ ~path(1,1)). no_proof_for(\+ ~path(1,1)). no_proof_for(\+ ~path(1,1)). name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0004_Line_0000__path_1'. JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'. JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'. % saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0004_Line_0000__path_1-junit.xml ~*/ :- mpred_test(~path(2,2)). %~ mpred_test("Test_0005_Line_0000__path_2",baseKB: ~path(2,2)) /*~ %~ mpred_test("Test_0005_Line_0000__path_2",baseKB: ~path(2,2)) ^ Call: (68) [baseKB] ~path(2, 2) ^ Unify: (68) [baseKB] ~ (baseKB:path(2, 2)) ^ Call: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(2, 2)), info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1189, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(2, 2))))) ^ Unify: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(2, 2)), info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1189, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(2, 2))))) Call: (76) [system] set_prolog_flag(last_call_optimisation, false) Exit: (76) [system] set_prolog_flag(last_call_optimisation, false) ^ Call: (76) [loop_check] prolog_frame_attribute(1189, parent_goal, loop_check_term_frame(_121764, info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, _121770, _121772)) ^ Fail: (76) [loop_check] prolog_frame_attribute(1189, parent_goal, loop_check_term_frame(_121764, info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, _121770, _121772)) ^ Redo: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(2, 2)), info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1189, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(2, 2))))) Call: (76) [pfc_lib] neg_in_code0(baseKB:path(2, 2)) Unify: (76) [pfc_lib] neg_in_code0(baseKB:path(2, 2)) ^ Call: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(2, 2)), info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1328, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(2, 2))))) ^ Unify: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(2, 2)), info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1328, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(2, 2))))) Call: (83) [system] set_prolog_flag(last_call_optimisation, false) Exit: (83) [system] set_prolog_flag(last_call_optimisation, false) ^ Call: (83) [loop_check] prolog_frame_attribute(1328, parent_goal, loop_check_term_frame(_127488, info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, _127494, _127496)) ^ Fail: (83) [loop_check] prolog_frame_attribute(1328, parent_goal, loop_check_term_frame(_127488, info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, _127494, _127496)) ^ Redo: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(2, 2)), info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1328, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(2, 2))))) Call: (83) [pfc_lib] neg_may_naf(baseKB:path(2, 2)) Unify: (83) [pfc_lib] neg_may_naf(baseKB:path(2, 2)) ^ Call: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _130820) ^ Unify: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _130820) ^ Call: (88) [system] clause(pfc_lib:prologNegByFailure(path), true, _130820) ^ Fail: (88) [system] clause(pfc_lib:prologNegByFailure(path), true, _130820) ^ Fail: (87) [pfc_lib] hook_database:clause_i(pfc_lib:prologNegByFailure(path), true, _130820) Unify: (83) [pfc_lib] neg_may_naf(baseKB:path(2, 2)) ^ Call: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(2, 2)) ^ Unify: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(2, 2)) ^ Call: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(2, 2)) ^ Unify: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(2, 2)) ^ Fail: (85) [pfc_lib] ucatch:is_ftVar(baseKB:path(2, 2)) ^ Redo: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(2, 2)) ^ Exit: (84) [pfc_lib] ucatch:is_ftCompound(baseKB:path(2, 2)) ^ Call: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990), call(_138990)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)))) ^ Unify: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990), call(_138990)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)))) ^ Call: (90) [hook_database] clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990) ^ Fail: (90) [hook_database] clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990) Call: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)) Unify: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)) ^ Call: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776)) ^ Unify: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776)) ^ Call: (92) [baseKB] clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776) ^ Fail: (92) [baseKB] clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776) ^ Fail: (91) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologHybrid), _142776)) Fail: (90) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)) ^ Fail: (88) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologHybrid), _138990), call(_138990)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologHybrid)))) ^ Call: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _147196) ^ Unify: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, syntaxic(_147766)) ^ Call: (88) [pfc_lib] mpred_database_term_syntax(path, 2, _147766) ^ Fail: (88) [pfc_lib] mpred_database_term_syntax(path, 2, _147766) ^ Redo: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _149768) ^ Unify: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _150396) ^ Call: (88) [pfc_lib] mpred_core_database_term(path, 2, _150964) ^ Fail: (88) [pfc_lib] mpred_core_database_term(path, 2, _150964) ^ Fail: (87) [pfc_lib] baseKB:mpred_database_term(path, 2, _152338) ^ Call: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942), call(_152942)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)))) ^ Unify: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942), call(_152942)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)))) ^ Call: (88) [hook_database] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942) ^ Fail: (88) [hook_database] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942) Call: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)) Unify: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)) ^ Call: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728)) ^ Unify: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728)) ^ Call: (90) [baseKB] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728) ^ Fail: (90) [baseKB] clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728) ^ Fail: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(baseKB, path, 2, prologBuiltin), _156728)) Fail: (88) [hook_database] clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)) ^ Fail: (86) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(baseKB, path, 2, prologBuiltin), _152942), call(_152942)*->true;clause_b(baseKB:mpred_prop(baseKB, path, 2, prologBuiltin)))) ^ Call: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_161094, path, 2, prologHybrid), _161124), call(_161124)*->true;clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid)))) ^ Unify: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_161094, path, 2, prologHybrid), _161124), call(_161124)*->true;clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid)))) ^ Call: (91) [hook_database] clause(mpred_prop(_161094, path, 2, prologHybrid), _161124) ^ Fail: (91) [hook_database] clause(mpred_prop(_161094, path, 2, prologHybrid), _161124) Call: (91) [hook_database] clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid)) Unify: (91) [hook_database] clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid)) ^ Call: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_161094, path, 2, prologHybrid), _164910)) ^ Unify: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_161094, path, 2, prologHybrid), _164910)) ^ Call: (93) [baseKB] clause(mpred_prop(_161094, path, 2, prologHybrid), _164910) ^ Fail: (93) [baseKB] clause(mpred_prop(_161094, path, 2, prologHybrid), _164910) ^ Fail: (92) [hook_database] hook_database:pfc_with_quiet_vars_lock(baseKB:clause(mpred_prop(_161094, path, 2, prologHybrid), _164910)) Fail: (91) [hook_database] clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid)) ^ Fail: (89) [hook_database] hook_database:pfc_with_quiet_vars_lock((clause(mpred_prop(_161094, path, 2, prologHybrid), _161124), call(_161124)*->true;clause_b(baseKB:mpred_prop(_161094, path, 2, prologHybrid)))) ^ Call: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _169330) ^ Unify: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, syntaxic(_169900)) ^ Call: (89) [pfc_lib] mpred_database_term_syntax(path, 2, _169900) ^ Fail: (89) [pfc_lib] mpred_database_term_syntax(path, 2, _169900) ^ Redo: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _171902) ^ Unify: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _172530) ^ Call: (89) [pfc_lib] mpred_core_database_term(path, 2, _173098) ^ Fail: (89) [pfc_lib] mpred_core_database_term(path, 2, _173098) ^ Fail: (88) [pfc_lib] baseKB:mpred_database_term(path, 2, _174472) Call: (98) [$autoload] leave_sandbox(_175070) Unify: (98) [$autoload] leave_sandbox(_175070) Exit: (98) [$autoload] leave_sandbox(false) Call: (97) [$autoload] restore_sandbox(false) Unify: (97) [$autoload] restore_sandbox(false) Exit: (97) [$autoload] restore_sandbox(false) Fail: (83) [pfc_lib] neg_may_naf(baseKB:path(2, 2)) ^ Fail: (82) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_may_naf(baseKB:path(2, 2)), info(pfc_lib:neg_may_naf(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1328, pfc_lib:trace_or_throw(looped(pfc_lib:neg_may_naf(baseKB:path(2, 2))))) Fail: (76) [pfc_lib] neg_in_code0(baseKB:path(2, 2)) ^ Fail: (75) [loop_check] loop_check:loop_check_term_frame(pfc_lib:neg_in_code0(baseKB:path(2, 2)), info(pfc_lib:neg_in_code0(baseKB:path(2, 2)), 'mpred_core.pl':273), 1, 1189, pfc_lib:trace_or_throw(looped(pfc_lib:neg_in_code0(baseKB:path(2, 2))))) ^ Fail: (68) [baseKB] ~ (baseKB:path(2, 2)) ^ Call: (68) [must_sanity] must_sanity:mquietly_if(true, rtrace:tAt_normal) ^ Unify: (68) [must_sanity] must_sanity:mquietly_if(true, rtrace:tAt_normal) failure=info((why_was_true(baseKB:(\+ ~path(2,2))),rtrace(baseKB: ~path(2,2)))) no_proof_for(\+ ~path(2,2)). no_proof_for(\+ ~path(2,2)). no_proof_for(\+ ~path(2,2)). name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0005_Line_0000__path_2'. JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'. JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'. % saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0005_Line_0000__path_2-junit.xml ~*/ /* :- mpred_test(\+path(1,1)). :- mpred_test(\+path(2,2)). */ :- mpred_test(\+path(3,1)). %~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/neg_01v.pfc:32 %~ mpred_test("Test_0006_Line_0000__naf_path_3",baseKB:(\+path(3,1))) /*~ %~ mpred_test("Test_0006_Line_0000__naf_path_3",baseKB:(\+path(3,1))) passed=info(why_was_true(baseKB:(\+path(3,1)))) no_proof_for(\+path(3,1)). no_proof_for(\+path(3,1)). no_proof_for(\+path(3,1)). name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0006_Line_0000__naf_path_3'. JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'. JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'. % saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0006_Line_0000__naf_path_3-junit.xml ~*/ :- mpred_test(\+ ~path(1,5)). %~ mpred_test("Test_0007_Line_0000__naf_path_1",baseKB:(\+ ~path(1,5))) /*~ %~ mpred_test("Test_0007_Line_0000__naf_path_1",baseKB:(\+ ~path(1,5))) passed=info(why_was_true(baseKB:(\+ ~path(1,5)))) no_proof_for(\+ ~path(1,5)). no_proof_for(\+ ~path(1,5)). no_proof_for(\+ ~path(1,5)). name ='logicmoo.pfc.test.sanity_base.NEG_01V-Test_0007_Line_0000__naf_path_1'. JUNIT_CLASSNAME ='logicmoo.pfc.test.sanity_base.NEG_01V'. JUNIT_CMD ='timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif neg_01v.pfc'. % saving_junit: /var/lib/jenkins/workspace/logicmoo_workspace/test_results/jenkins/Report-logicmoo-junit-test-sanity_base-vSTARv0vSTARvvDOTvvSTARv-Units-logicmoo.pfc.test.sanity_base.NEG_01V-Test_0007_Line_0000__naf_path_1-junit.xml ~*/ :- break. %~ skipped(blocks_on_input,break) %~ unused(save_junit_results) %~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/pfc/t/sanity_base/neg_01v.pfc:36 %~ test_completed_exit(6) :- dynamic junit_prop/3. :- dynamic junit_prop/3. :- dynamic junit_prop/3. ``` totalTime=1 ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ANEG_01V GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3 https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/pfc/t/sanity_base/neg_01v.pfc Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.pfc.test.sanity_base/NEG_01V/logicmoo_pfc_test_sanity_base_NEG_01V_JUnit/ This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/67/testReport/logicmoo.pfc.test.sanity_base/NEG_01V/logicmoo_pfc_test_sanity_base_NEG_01V_JUnit/ GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3 https://github.com/logicmoo/logicmoo_workspace/blob/869479bc8cf913ee2df5ebbe49363a2dad9c9fb3/packs_sys/pfc/t/sanity_base/neg_01v.pfc FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k neg_01v.pfc (returned 6)
non_process
logicmoo pfc test sanity base neg junit cd var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base timeout foreground preserve status s sigkill k lmoo clif neg pfc gh master issue finfo issue search gitlab latest this build github running var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base neg pfc var lib jenkins local share swi prolog pack logicmoo utils prolog logicmoo test header pl this test might need use module library logicmoo plarkc use module library statistics mpred notrace exec reset runtime counter mpred notrace exec reset runtime counter statistics runtime secs path x x path path path path mpred test path mpred test test line path basekb path file mpred test test line path basekb path passed info why was true basekb path justifications for path  basekb  name logicmoo pfc test sanity base neg test line path junit classname logicmoo pfc test sanity base neg junit cmd timeout foreground preserve status s sigkill k lmoo clif neg pfc saving junit var lib jenkins workspace logicmoo workspace test results jenkins report logicmoo junit test sanity base units logicmoo pfc test sanity base neg test line path junit xml mpred test path mpred test test line path basekb path file mpred test test line path basekb path passed info why was true basekb path justifications for path  basekb  name logicmoo pfc test sanity base neg test line path junit classname logicmoo pfc test sanity base neg junit cmd timeout foreground preserve status s sigkill k lmoo clif neg pfc saving junit var lib jenkins workspace logicmoo workspace test results jenkins report logicmoo junit test sanity base units logicmoo pfc test sanity base neg test line path junit xml mpred test path mpred test test line path basekb path file mpred test test line path basekb path passed info why was true basekb path justifications for path  basekb  name logicmoo pfc test sanity base neg test line path junit classname logicmoo pfc test sanity base neg junit cmd timeout foreground preserve status s sigkill k lmoo clif neg pfc saving junit var lib jenkins workspace logicmoo workspace test results jenkins report logicmoo junit test sanity base units logicmoo pfc test sanity base neg test line path junit xml mpred test path mpred test test line path basekb path mpred test test line path basekb path call path unify basekb path call loop check loop check term frame pfc lib neg in basekb path info pfc lib neg in basekb path mpred core pl pfc lib trace or throw looped pfc lib neg in basekb path unify loop check loop check term frame pfc lib neg in basekb path info pfc lib neg in basekb path mpred core pl pfc lib trace or throw looped pfc lib neg in basekb path call set prolog flag last call optimisation false exit set prolog flag last call optimisation false call prolog frame attribute parent goal loop check term frame info pfc lib neg in basekb path mpred core pl fail prolog frame attribute parent goal loop check term frame info pfc lib neg in basekb path mpred core pl redo loop check loop check term frame pfc lib neg in basekb path info pfc lib neg in basekb path mpred core pl pfc lib trace or throw looped pfc lib neg in basekb path call neg in basekb path unify neg in basekb path call loop check loop check term frame pfc lib neg may naf basekb path info pfc lib neg may naf basekb path mpred core pl pfc lib trace or throw looped pfc lib neg may naf basekb path unify loop check loop check term frame pfc lib neg may naf basekb path info pfc lib neg may naf basekb path mpred core pl pfc lib trace or throw looped pfc lib neg may naf basekb path call set prolog flag last call optimisation false exit set prolog flag last call optimisation false call prolog frame attribute parent goal loop check term frame info pfc lib neg may naf basekb path mpred core pl fail prolog frame attribute parent goal loop check term frame info pfc lib neg may naf basekb path mpred core pl redo loop check loop check term frame pfc lib neg may naf basekb path info pfc lib neg may naf basekb path mpred core pl pfc lib trace or throw looped pfc lib neg may naf basekb path call neg may naf basekb path unify neg may naf basekb path call hook database clause i pfc lib prolognegbyfailure path true unify hook database clause i pfc lib prolognegbyfailure path true call clause pfc lib prolognegbyfailure path true fail clause pfc lib prolognegbyfailure path true fail hook database clause i pfc lib prolognegbyfailure path true unify neg may naf basekb path call ucatch is ftcompound basekb path unify ucatch is ftcompound basekb path call ucatch is ftvar basekb path unify ucatch is ftvar basekb path fail ucatch is ftvar basekb path redo ucatch is ftcompound basekb path exit ucatch is ftcompound basekb path call hook database pfc with quiet vars lock clause mpred prop basekb path prologhybrid call true clause b basekb mpred prop basekb path prologhybrid unify hook database pfc with quiet vars lock clause mpred prop basekb path prologhybrid call true clause b basekb mpred prop basekb path prologhybrid call clause mpred prop basekb path prologhybrid fail clause mpred prop basekb path prologhybrid call clause b basekb mpred prop basekb path prologhybrid unify clause b basekb mpred prop basekb path prologhybrid call hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologhybrid unify hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologhybrid call clause mpred prop basekb path prologhybrid fail clause mpred prop basekb path prologhybrid fail hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologhybrid fail clause b basekb mpred prop basekb path prologhybrid fail hook database pfc with quiet vars lock clause mpred prop basekb path prologhybrid call true clause b basekb mpred prop basekb path prologhybrid call basekb mpred database term path unify basekb mpred database term path syntaxic call mpred database term syntax path fail mpred database term syntax path redo basekb mpred database term path unify basekb mpred database term path call mpred core database term path fail mpred core database term path fail basekb mpred database term path call hook database pfc with quiet vars lock clause mpred prop basekb path prologbuiltin call true clause b basekb mpred prop basekb path prologbuiltin unify hook database pfc with quiet vars lock clause mpred prop basekb path prologbuiltin call true clause b basekb mpred prop basekb path prologbuiltin call clause mpred prop basekb path prologbuiltin fail clause mpred prop basekb path prologbuiltin call clause b basekb mpred prop basekb path prologbuiltin unify clause b basekb mpred prop basekb path prologbuiltin call hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologbuiltin unify hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologbuiltin call clause mpred prop basekb path prologbuiltin fail clause mpred prop basekb path prologbuiltin fail hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologbuiltin fail clause b basekb mpred prop basekb path prologbuiltin fail hook database pfc with quiet vars lock clause mpred prop basekb path prologbuiltin call true clause b basekb mpred prop basekb path prologbuiltin call hook database pfc with quiet vars lock clause mpred prop path prologhybrid call true clause b basekb mpred prop path prologhybrid unify hook database pfc with quiet vars lock clause mpred prop path prologhybrid call true clause b basekb mpred prop path prologhybrid call clause mpred prop path prologhybrid fail clause mpred prop path prologhybrid call clause b basekb mpred prop path prologhybrid unify clause b basekb mpred prop path prologhybrid call hook database pfc with quiet vars lock basekb clause mpred prop path prologhybrid unify hook database pfc with quiet vars lock basekb clause mpred prop path prologhybrid call clause mpred prop path prologhybrid fail clause mpred prop path prologhybrid fail hook database pfc with quiet vars lock basekb clause mpred prop path prologhybrid fail clause b basekb mpred prop path prologhybrid fail hook database pfc with quiet vars lock clause mpred prop path prologhybrid call true clause b basekb mpred prop path prologhybrid call basekb mpred database term path unify basekb mpred database term path syntaxic call mpred database term syntax path fail mpred database term syntax path redo basekb mpred database term path unify basekb mpred database term path call mpred core database term path fail mpred core database term path fail basekb mpred database term path call leave sandbox unify leave sandbox exit leave sandbox false call restore sandbox false unify restore sandbox false exit restore sandbox false fail neg may naf basekb path fail loop check loop check term frame pfc lib neg may naf basekb path info pfc lib neg may naf basekb path mpred core pl pfc lib trace or throw looped pfc lib neg may naf basekb path fail neg in basekb path fail loop check loop check term frame pfc lib neg in basekb path info pfc lib neg in basekb path mpred core pl pfc lib trace or throw looped pfc lib neg in basekb path fail basekb path call must sanity mquietly if true rtrace tat normal unify must sanity mquietly if true rtrace tat normal failure info why was true basekb path rtrace basekb path no proof for path no proof for path no proof for path name logicmoo pfc test sanity base neg test line path junit classname logicmoo pfc test sanity base neg junit cmd timeout foreground preserve status s sigkill k lmoo clif neg pfc saving junit var lib jenkins workspace logicmoo workspace test results jenkins report logicmoo junit test sanity base units logicmoo pfc test sanity base neg test line path junit xml mpred test path mpred test test line path basekb path mpred test test line path basekb path call path unify basekb path call loop check loop check term frame pfc lib neg in basekb path info pfc lib neg in basekb path mpred core pl pfc lib trace or throw looped pfc lib neg in basekb path unify loop check loop check term frame pfc lib neg in basekb path info pfc lib neg in basekb path mpred core pl pfc lib trace or throw looped pfc lib neg in basekb path call set prolog flag last call optimisation false exit set prolog flag last call optimisation false call prolog frame attribute parent goal loop check term frame info pfc lib neg in basekb path mpred core pl fail prolog frame attribute parent goal loop check term frame info pfc lib neg in basekb path mpred core pl redo loop check loop check term frame pfc lib neg in basekb path info pfc lib neg in basekb path mpred core pl pfc lib trace or throw looped pfc lib neg in basekb path call neg in basekb path unify neg in basekb path call loop check loop check term frame pfc lib neg may naf basekb path info pfc lib neg may naf basekb path mpred core pl pfc lib trace or throw looped pfc lib neg may naf basekb path unify loop check loop check term frame pfc lib neg may naf basekb path info pfc lib neg may naf basekb path mpred core pl pfc lib trace or throw looped pfc lib neg may naf basekb path call set prolog flag last call optimisation false exit set prolog flag last call optimisation false call prolog frame attribute parent goal loop check term frame info pfc lib neg may naf basekb path mpred core pl fail prolog frame attribute parent goal loop check term frame info pfc lib neg may naf basekb path mpred core pl redo loop check loop check term frame pfc lib neg may naf basekb path info pfc lib neg may naf basekb path mpred core pl pfc lib trace or throw looped pfc lib neg may naf basekb path call neg may naf basekb path unify neg may naf basekb path call hook database clause i pfc lib prolognegbyfailure path true unify hook database clause i pfc lib prolognegbyfailure path true call clause pfc lib prolognegbyfailure path true fail clause pfc lib prolognegbyfailure path true fail hook database clause i pfc lib prolognegbyfailure path true unify neg may naf basekb path call ucatch is ftcompound basekb path unify ucatch is ftcompound basekb path call ucatch is ftvar basekb path unify ucatch is ftvar basekb path fail ucatch is ftvar basekb path redo ucatch is ftcompound basekb path exit ucatch is ftcompound basekb path call hook database pfc with quiet vars lock clause mpred prop basekb path prologhybrid call true clause b basekb mpred prop basekb path prologhybrid unify hook database pfc with quiet vars lock clause mpred prop basekb path prologhybrid call true clause b basekb mpred prop basekb path prologhybrid call clause mpred prop basekb path prologhybrid fail clause mpred prop basekb path prologhybrid call clause b basekb mpred prop basekb path prologhybrid unify clause b basekb mpred prop basekb path prologhybrid call hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologhybrid unify hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologhybrid call clause mpred prop basekb path prologhybrid fail clause mpred prop basekb path prologhybrid fail hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologhybrid fail clause b basekb mpred prop basekb path prologhybrid fail hook database pfc with quiet vars lock clause mpred prop basekb path prologhybrid call true clause b basekb mpred prop basekb path prologhybrid call basekb mpred database term path unify basekb mpred database term path syntaxic call mpred database term syntax path fail mpred database term syntax path redo basekb mpred database term path unify basekb mpred database term path call mpred core database term path fail mpred core database term path fail basekb mpred database term path call hook database pfc with quiet vars lock clause mpred prop basekb path prologbuiltin call true clause b basekb mpred prop basekb path prologbuiltin unify hook database pfc with quiet vars lock clause mpred prop basekb path prologbuiltin call true clause b basekb mpred prop basekb path prologbuiltin call clause mpred prop basekb path prologbuiltin fail clause mpred prop basekb path prologbuiltin call clause b basekb mpred prop basekb path prologbuiltin unify clause b basekb mpred prop basekb path prologbuiltin call hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologbuiltin unify hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologbuiltin call clause mpred prop basekb path prologbuiltin fail clause mpred prop basekb path prologbuiltin fail hook database pfc with quiet vars lock basekb clause mpred prop basekb path prologbuiltin fail clause b basekb mpred prop basekb path prologbuiltin fail hook database pfc with quiet vars lock clause mpred prop basekb path prologbuiltin call true clause b basekb mpred prop basekb path prologbuiltin call hook database pfc with quiet vars lock clause mpred prop path prologhybrid call true clause b basekb mpred prop path prologhybrid unify hook database pfc with quiet vars lock clause mpred prop path prologhybrid call true clause b basekb mpred prop path prologhybrid call clause mpred prop path prologhybrid fail clause mpred prop path prologhybrid call clause b basekb mpred prop path prologhybrid unify clause b basekb mpred prop path prologhybrid call hook database pfc with quiet vars lock basekb clause mpred prop path prologhybrid unify hook database pfc with quiet vars lock basekb clause mpred prop path prologhybrid call clause mpred prop path prologhybrid fail clause mpred prop path prologhybrid fail hook database pfc with quiet vars lock basekb clause mpred prop path prologhybrid fail clause b basekb mpred prop path prologhybrid fail hook database pfc with quiet vars lock clause mpred prop path prologhybrid call true clause b basekb mpred prop path prologhybrid call basekb mpred database term path unify basekb mpred database term path syntaxic call mpred database term syntax path fail mpred database term syntax path redo basekb mpred database term path unify basekb mpred database term path call mpred core database term path fail mpred core database term path fail basekb mpred database term path call leave sandbox unify leave sandbox exit leave sandbox false call restore sandbox false unify restore sandbox false exit restore sandbox false fail neg may naf basekb path fail loop check loop check term frame pfc lib neg may naf basekb path info pfc lib neg may naf basekb path mpred core pl pfc lib trace or throw looped pfc lib neg may naf basekb path fail neg in basekb path fail loop check loop check term frame pfc lib neg in basekb path info pfc lib neg in basekb path mpred core pl pfc lib trace or throw looped pfc lib neg in basekb path fail basekb path call must sanity mquietly if true rtrace tat normal unify must sanity mquietly if true rtrace tat normal failure info why was true basekb path rtrace basekb path no proof for path no proof for path no proof for path name logicmoo pfc test sanity base neg test line path junit classname logicmoo pfc test sanity base neg junit cmd timeout foreground preserve status s sigkill k lmoo clif neg pfc saving junit var lib jenkins workspace logicmoo workspace test results jenkins report logicmoo junit test sanity base units logicmoo pfc test sanity base neg test line path junit xml mpred test path mpred test path mpred test path var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base neg pfc mpred test test line naf path basekb path mpred test test line naf path basekb path passed info why was true basekb path no proof for path no proof for path no proof for path name logicmoo pfc test sanity base neg test line naf path junit classname logicmoo pfc test sanity base neg junit cmd timeout foreground preserve status s sigkill k lmoo clif neg pfc saving junit var lib jenkins workspace logicmoo workspace test results jenkins report logicmoo junit test sanity base units logicmoo pfc test sanity base neg test line naf path junit xml mpred test path mpred test test line naf path basekb path mpred test test line naf path basekb path passed info why was true basekb path no proof for path no proof for path no proof for path name logicmoo pfc test sanity base neg test line naf path junit classname logicmoo pfc test sanity base neg junit cmd timeout foreground preserve status s sigkill k lmoo clif neg pfc saving junit var lib jenkins workspace logicmoo workspace test results jenkins report logicmoo junit test sanity base units logicmoo pfc test sanity base neg test line naf path junit xml break skipped blocks on input break unused save junit results var lib jenkins workspace logicmoo workspace packs sys pfc t sanity base neg pfc test completed exit dynamic junit prop dynamic junit prop dynamic junit prop totaltime issue search gitlab latest this build github failed var lib jenkins workspace logicmoo workspace bin lmoo junit minor k neg pfc returned
0
317
2,762,978,072
IssuesEvent
2015-04-29 04:42:05
sysown/proxysql-0.2
https://api.github.com/repos/sysown/proxysql-0.2
opened
Enabling fast_forward at runtime can cause crashes
ADMIN MYSQL PROTOCOL QUERY PROCESSOR
Enabling fast_forward at runtime can cause crashes if there are already established connections
1.0
Enabling fast_forward at runtime can cause crashes - Enabling fast_forward at runtime can cause crashes if there are already established connections
process
enabling fast forward at runtime can cause crashes enabling fast forward at runtime can cause crashes if there are already established connections
1
726
3,213,368,795
IssuesEvent
2015-10-06 19:34:09
nationalparkservice/places-data
https://api.github.com/repos/nationalparkservice/places-data
closed
Add highway=steps
pending-other-process schema
http://wiki.openstreetmap.org/wiki/Tag:highway%3Dsteps There are a lot of staircases in parks, and this would be useful to have in our map
1.0
Add highway=steps - http://wiki.openstreetmap.org/wiki/Tag:highway%3Dsteps There are a lot of staircases in parks, and this would be useful to have in our map
process
add highway steps there are a lot of staircases in parks and this would be useful to have in our map
1
692,133
23,723,871,713
IssuesEvent
2022-08-30 17:41:03
project-pareto/project-pareto
https://api.github.com/repos/project-pareto/project-pareto
closed
Network errors in toy_case_study cause occasional CI failures
bug Priority:Normal
- The error message seems to occur for requests to `router.project-osrm.org - However, it's not immediately clear if this is due to rate limiting enforced by that server (which is triggered by multiple CI jobs running in parallel), or other network issues ## Traceback From https://github.com/project-pareto/project-pareto/runs/6984942337?check_suite_focus=true ```txt Traceback (most recent call last): File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connection.py", line 174, in _new_conn conn = connection.create_connection( File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\util\connection.py", line 72, in create_connection for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\socket.py", line 954, in getaddrinfo for res in _socket.getaddrinfo(host, port, family, type, proto, flags): socket.gaierror: [Errno 11001] getaddrinfo failed During handling of the above exception, another exception occurred: Traceback (most recent call last): File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connectionpool.py", line 703, in urlopen httplib_response = self._make_request( File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connectionpool.py", line 386, in _make_request self._validate_conn(conn) File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connectionpool.py", line 1040, in _validate_conn conn.connect() File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connection.py", line 358, in connect self.sock = conn = self._new_conn() File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connection.py", line 186, in _new_conn raise NewConnectionError( urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x0000013C6C5A44C0>: Failed to establish a new connection: [Errno 11001] getaddrinfo failed During handling of the above exception, another exception occurred: Traceback (most recent call last): File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\adapters.py", line 489, in send resp = conn.urlopen( File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connectionpool.py", line 785, in urlopen retries = retries.increment( File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\util\retry.py", line 592, in increment raise MaxRetryError(_pool, url, error or ResponseError(cause)) urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='router.project-osrm.org', port=443): Max retries exceeded with url: /table/v1/driving/-122.36685609717435,47.62390651082141;-122.39902392056128,47.544942555956496;-122.2875801822827,47.54628135125307;-122.2901468966058,47.59922966761389;-122.3345,47.6044;-122.38514810411333,47.6785514001581;-122.31941699831951,47.671605016754775;-122.37256636119979,47.51373173275276;-122.30833430642372,47.5982109494788;-122.27578180785954,47.54501307370556;-122.25392487570376,47.53198073978911;-122.23615203748196,47.56473303789118;-122.38490267916536,47.52979263294095;-122.43178036972577,47.63168538306254;-122.3748467760494,47.66220329642665;-122.33832847843796,47.650576946946856;-122.24953147273567,47.655834380297506;-122.27548462416144,47.68078709171784;-122.41679135415569,47.53282219470006;-122.29739693081503,47.63214139655833;-122.39872536852423,47.606697845554585;-122.26887716808142,47.61641535939208;-122.27286360239516,47.60037158057717;-122.27032458981431,47.53461603738988;-122.36685609717435,47.62390651082141;-122.39902392056128,47.544942555956496;-122.2875801822827,47.54628135125307;-122.2901468966058,47.59922966761389;-122.3345,47.6044;-122.38514810411333,47.6785514001581;-122.31941699831951,47.671605016754775;-122.37256636119979,47.51373173275276;-122.30833430642372,47.5982109494788;-122.27578180785954,47.54501307370556;-122.25392487570376,47.53198073978911;-122.23615203748196,47.56473303789118;-122.38490267916536,47.52979263294095;-122.43178036972577,47.63168538306254;-122.3748467760494,47.66220329642665;-122.33832847843796,47.650576946946856;-122.24953147273567,47.655834380297506;-122.27548462416144,47.68078709171784;-122.41679135415569,47.53282219470006;-122.29739693081503,47.63214139655833;-122.39872536852423,47.606697845554585;-122.26887716808142,47.61641535939208;-122.27286360239516,47.60037158057717;-122.27032458981431,47.53461603738988?sources=0;1;2;3;4;5;6;7;8;9;10;11;12;13;14;15;16;17;18;19;20;21;22;23&destinations=24;25;26;27;28;29;30;31;32;33;34;35;36;37;38;39;40;41;42;43;44;45;46;47&annotations=duration,distance (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x0000013C6C5A44C0>: Failed to establish a new connection: [Errno 11001] getaddrinfo failed')) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\runpy.py", line 197, in _run_module_as_main return _run_code(code, main_globals, None, File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\runpy.py", line 87, in _run_code exec(code, run_globals) File "D:\a\project-pareto\project-pareto\pareto\case_studies\toy_case_study.py", line 203, in <module> df_parameters["DriveTimes"] = od_matrix(od_matrix_input) File "D:\a\project-pareto\project-pareto\pareto\utilities\get_data.py", line 462, in od_matrix response = requests.get( File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\api.py", line 73, in get return request("get", url, params=params, **kwargs) File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\api.py", line 59, in request return session.request(method=method, url=url, **kwargs) File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\sessions.py", line 587, in request resp = self.send(prep, **send_kwargs) File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\sessions.py", line 701, in send r = adapter.send(request, **kwargs) File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\adapters.py", line 565, in send raise ConnectionError(e, request=request) requests.exceptions.ConnectionError: HTTPSConnectionPool(host='router.project-osrm.org', port=443): Max retries exceeded with url: /table/v1/driving/-122.36685609717435,47.62390651082141;-122.39902392056128,47.544942555956496;-122.2875801822827,47.54628135125307;-122.2901468966058,47.59922966761389;-122.3345,47.6044;-122.38514810411333,47.6785514001581;-122.31941699831951,47.671605016754775;-122.37256636119979,47.51373173275276;-122.30833430642372,47.5982109494788;-122.27578180785954,47.54501307370556;-122.25392487570376,47.53198073978911;-122.23615203748196,47.56473303789118;-122.38490267916536,47.52979263294095;-122.43178036972577,47.63168538306254;-122.3748467760494,47.66220329642665;-122.33832847843796,47.650576946946856;-122.24953147273567,47.655834380297506;-122.27548462416144,47.68078709171784;-122.41679135415569,47.53282219470006;-122.29739693081503,47.63214139655833;-122.39872536852423,47.606697845554585;-122.26887716808142,47.61641535939208;-122.27286360239516,47.60037158057717;-122.27032458981431,47.53461603738988;-122.36685609717435,47.62390651082141;-122.39902392056128,47.544942555956496;-122.2875801822827,47.54628135125307;-122.2901468966058,47.59922966761389;-122.3345,47.6044;-122.38514810411333,47.6785514001581;-122.31941699831951,47.671605016754775;-122.37256636119979,47.51373173275276;-122.30833430642372,47.5982109494788;-122.27578180785954,47.54501307370556;-122.25392487570376,47.53198073978911;-122.23615203748196,47.56473303789118;-122.38490267916536,47.52979263294095;-122.43178036972577,47.63168538306254;-122.3748467760494,47.66220329642665;-122.33832847843796,47.650576946946856;-122.24953147273567,47.655834380297506;-122.27548462416144,47.68078709171784;-122.41679135415569,47.53282219470006;-122.29739693081503,47.63214139655833;-122.39872536852423,47.606697845554585;-122.26887716808142,47.61641535939208;-122.27286360239516,47.60037158057717;-122.27032458981431,47.53461603738988?sources=0;1;2;3;4;5;6;7;8;9;10;11;12;13;14;15;16;17;18;19;20;21;22;23&destinations=24;25;26;27;28;29;30;31;32;33;34;35;36;37;38;39;40;41;42;43;44;45;46;47&annotations=duration,distance (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x0000013C6C5A44C0>: Failed to establish a new connection: [Errno 11001] getaddrinfo failed')) ```
1.0
Network errors in toy_case_study cause occasional CI failures - - The error message seems to occur for requests to `router.project-osrm.org - However, it's not immediately clear if this is due to rate limiting enforced by that server (which is triggered by multiple CI jobs running in parallel), or other network issues ## Traceback From https://github.com/project-pareto/project-pareto/runs/6984942337?check_suite_focus=true ```txt Traceback (most recent call last): File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connection.py", line 174, in _new_conn conn = connection.create_connection( File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\util\connection.py", line 72, in create_connection for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\socket.py", line 954, in getaddrinfo for res in _socket.getaddrinfo(host, port, family, type, proto, flags): socket.gaierror: [Errno 11001] getaddrinfo failed During handling of the above exception, another exception occurred: Traceback (most recent call last): File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connectionpool.py", line 703, in urlopen httplib_response = self._make_request( File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connectionpool.py", line 386, in _make_request self._validate_conn(conn) File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connectionpool.py", line 1040, in _validate_conn conn.connect() File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connection.py", line 358, in connect self.sock = conn = self._new_conn() File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connection.py", line 186, in _new_conn raise NewConnectionError( urllib3.exceptions.NewConnectionError: <urllib3.connection.HTTPSConnection object at 0x0000013C6C5A44C0>: Failed to establish a new connection: [Errno 11001] getaddrinfo failed During handling of the above exception, another exception occurred: Traceback (most recent call last): File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\adapters.py", line 489, in send resp = conn.urlopen( File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\connectionpool.py", line 785, in urlopen retries = retries.increment( File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\urllib3\util\retry.py", line 592, in increment raise MaxRetryError(_pool, url, error or ResponseError(cause)) urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='router.project-osrm.org', port=443): Max retries exceeded with url: /table/v1/driving/-122.36685609717435,47.62390651082141;-122.39902392056128,47.544942555956496;-122.2875801822827,47.54628135125307;-122.2901468966058,47.59922966761389;-122.3345,47.6044;-122.38514810411333,47.6785514001581;-122.31941699831951,47.671605016754775;-122.37256636119979,47.51373173275276;-122.30833430642372,47.5982109494788;-122.27578180785954,47.54501307370556;-122.25392487570376,47.53198073978911;-122.23615203748196,47.56473303789118;-122.38490267916536,47.52979263294095;-122.43178036972577,47.63168538306254;-122.3748467760494,47.66220329642665;-122.33832847843796,47.650576946946856;-122.24953147273567,47.655834380297506;-122.27548462416144,47.68078709171784;-122.41679135415569,47.53282219470006;-122.29739693081503,47.63214139655833;-122.39872536852423,47.606697845554585;-122.26887716808142,47.61641535939208;-122.27286360239516,47.60037158057717;-122.27032458981431,47.53461603738988;-122.36685609717435,47.62390651082141;-122.39902392056128,47.544942555956496;-122.2875801822827,47.54628135125307;-122.2901468966058,47.59922966761389;-122.3345,47.6044;-122.38514810411333,47.6785514001581;-122.31941699831951,47.671605016754775;-122.37256636119979,47.51373173275276;-122.30833430642372,47.5982109494788;-122.27578180785954,47.54501307370556;-122.25392487570376,47.53198073978911;-122.23615203748196,47.56473303789118;-122.38490267916536,47.52979263294095;-122.43178036972577,47.63168538306254;-122.3748467760494,47.66220329642665;-122.33832847843796,47.650576946946856;-122.24953147273567,47.655834380297506;-122.27548462416144,47.68078709171784;-122.41679135415569,47.53282219470006;-122.29739693081503,47.63214139655833;-122.39872536852423,47.606697845554585;-122.26887716808142,47.61641535939208;-122.27286360239516,47.60037158057717;-122.27032458981431,47.53461603738988?sources=0;1;2;3;4;5;6;7;8;9;10;11;12;13;14;15;16;17;18;19;20;21;22;23&destinations=24;25;26;27;28;29;30;31;32;33;34;35;36;37;38;39;40;41;42;43;44;45;46;47&annotations=duration,distance (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x0000013C6C5A44C0>: Failed to establish a new connection: [Errno 11001] getaddrinfo failed')) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\runpy.py", line 197, in _run_module_as_main return _run_code(code, main_globals, None, File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\runpy.py", line 87, in _run_code exec(code, run_globals) File "D:\a\project-pareto\project-pareto\pareto\case_studies\toy_case_study.py", line 203, in <module> df_parameters["DriveTimes"] = od_matrix(od_matrix_input) File "D:\a\project-pareto\project-pareto\pareto\utilities\get_data.py", line 462, in od_matrix response = requests.get( File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\api.py", line 73, in get return request("get", url, params=params, **kwargs) File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\api.py", line 59, in request return session.request(method=method, url=url, **kwargs) File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\sessions.py", line 587, in request resp = self.send(prep, **send_kwargs) File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\sessions.py", line 701, in send r = adapter.send(request, **kwargs) File "C:\hostedtoolcache\windows\Python\3.9.13\x64\lib\site-packages\requests\adapters.py", line 565, in send raise ConnectionError(e, request=request) requests.exceptions.ConnectionError: HTTPSConnectionPool(host='router.project-osrm.org', port=443): Max retries exceeded with url: /table/v1/driving/-122.36685609717435,47.62390651082141;-122.39902392056128,47.544942555956496;-122.2875801822827,47.54628135125307;-122.2901468966058,47.59922966761389;-122.3345,47.6044;-122.38514810411333,47.6785514001581;-122.31941699831951,47.671605016754775;-122.37256636119979,47.51373173275276;-122.30833430642372,47.5982109494788;-122.27578180785954,47.54501307370556;-122.25392487570376,47.53198073978911;-122.23615203748196,47.56473303789118;-122.38490267916536,47.52979263294095;-122.43178036972577,47.63168538306254;-122.3748467760494,47.66220329642665;-122.33832847843796,47.650576946946856;-122.24953147273567,47.655834380297506;-122.27548462416144,47.68078709171784;-122.41679135415569,47.53282219470006;-122.29739693081503,47.63214139655833;-122.39872536852423,47.606697845554585;-122.26887716808142,47.61641535939208;-122.27286360239516,47.60037158057717;-122.27032458981431,47.53461603738988;-122.36685609717435,47.62390651082141;-122.39902392056128,47.544942555956496;-122.2875801822827,47.54628135125307;-122.2901468966058,47.59922966761389;-122.3345,47.6044;-122.38514810411333,47.6785514001581;-122.31941699831951,47.671605016754775;-122.37256636119979,47.51373173275276;-122.30833430642372,47.5982109494788;-122.27578180785954,47.54501307370556;-122.25392487570376,47.53198073978911;-122.23615203748196,47.56473303789118;-122.38490267916536,47.52979263294095;-122.43178036972577,47.63168538306254;-122.3748467760494,47.66220329642665;-122.33832847843796,47.650576946946856;-122.24953147273567,47.655834380297506;-122.27548462416144,47.68078709171784;-122.41679135415569,47.53282219470006;-122.29739693081503,47.63214139655833;-122.39872536852423,47.606697845554585;-122.26887716808142,47.61641535939208;-122.27286360239516,47.60037158057717;-122.27032458981431,47.53461603738988?sources=0;1;2;3;4;5;6;7;8;9;10;11;12;13;14;15;16;17;18;19;20;21;22;23&destinations=24;25;26;27;28;29;30;31;32;33;34;35;36;37;38;39;40;41;42;43;44;45;46;47&annotations=duration,distance (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x0000013C6C5A44C0>: Failed to establish a new connection: [Errno 11001] getaddrinfo failed')) ```
non_process
network errors in toy case study cause occasional ci failures the error message seems to occur for requests to router project osrm org however it s not immediately clear if this is due to rate limiting enforced by that server which is triggered by multiple ci jobs running in parallel or other network issues traceback from txt traceback most recent call last file c hostedtoolcache windows python lib site packages connection py line in new conn conn connection create connection file c hostedtoolcache windows python lib site packages util connection py line in create connection for res in socket getaddrinfo host port family socket sock stream file c hostedtoolcache windows python lib socket py line in getaddrinfo for res in socket getaddrinfo host port family type proto flags socket gaierror getaddrinfo failed during handling of the above exception another exception occurred traceback most recent call last file c hostedtoolcache windows python lib site packages connectionpool py line in urlopen httplib response self make request file c hostedtoolcache windows python lib site packages connectionpool py line in make request self validate conn conn file c hostedtoolcache windows python lib site packages connectionpool py line in validate conn conn connect file c hostedtoolcache windows python lib site packages connection py line in connect self sock conn self new conn file c hostedtoolcache windows python lib site packages connection py line in new conn raise newconnectionerror exceptions newconnectionerror failed to establish a new connection getaddrinfo failed during handling of the above exception another exception occurred traceback most recent call last file c hostedtoolcache windows python lib site packages requests adapters py line in send resp conn urlopen file c hostedtoolcache windows python lib site packages connectionpool py line in urlopen retries retries increment file c hostedtoolcache windows python lib site packages util retry py line in increment raise maxretryerror pool url error or responseerror cause exceptions maxretryerror httpsconnectionpool host router project osrm org port max retries exceeded with url table driving sources destinations annotations duration distance caused by newconnectionerror failed to establish a new connection getaddrinfo failed during handling of the above exception another exception occurred traceback most recent call last file c hostedtoolcache windows python lib runpy py line in run module as main return run code code main globals none file c hostedtoolcache windows python lib runpy py line in run code exec code run globals file d a project pareto project pareto pareto case studies toy case study py line in df parameters od matrix od matrix input file d a project pareto project pareto pareto utilities get data py line in od matrix response requests get file c hostedtoolcache windows python lib site packages requests api py line in get return request get url params params kwargs file c hostedtoolcache windows python lib site packages requests api py line in request return session request method method url url kwargs file c hostedtoolcache windows python lib site packages requests sessions py line in request resp self send prep send kwargs file c hostedtoolcache windows python lib site packages requests sessions py line in send r adapter send request kwargs file c hostedtoolcache windows python lib site packages requests adapters py line in send raise connectionerror e request request requests exceptions connectionerror httpsconnectionpool host router project osrm org port max retries exceeded with url table driving sources destinations annotations duration distance caused by newconnectionerror failed to establish a new connection getaddrinfo failed
0
93,947
15,946,447,026
IssuesEvent
2021-04-15 01:05:29
jgeraigery/core
https://api.github.com/repos/jgeraigery/core
opened
CVE-2021-25122 (High) detected in tomcat-embed-core-8.5.34.jar
security vulnerability
## CVE-2021-25122 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.34.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p> <p>Path to dependency file: core/nimbus-entity-dsl/pom.xml</p> <p>Path to vulnerable library: core/8.5.34/tomcat-embed-core-8.5.34.jar,core/8.5.34/tomcat-embed-core-8.5.34.jar,core/8.5.34/tomcat-embed-core-8.5.34.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.0.5.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-2.0.5.RELEASE.jar - :x: **tomcat-embed-core-8.5.34.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When responding to new h2c connection requests, Apache Tomcat versions 10.0.0-M1 to 10.0.0, 9.0.0.M1 to 9.0.41 and 8.5.0 to 8.5.61 could duplicate request headers and a limited amount of request body from one request to another meaning user A and user B could both see the results of user A's request. <p>Publish Date: 2021-03-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25122>CVE-2021-25122</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://lists.apache.org/thread.html/r7b95bc248603360501f18c8eb03bb6001ec0ee3296205b34b07105b7%40%3Cannounce.tomcat.apache.org%3E">https://lists.apache.org/thread.html/r7b95bc248603360501f18c8eb03bb6001ec0ee3296205b34b07105b7%40%3Cannounce.tomcat.apache.org%3E</a></p> <p>Release Date: 2021-03-01</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:8.5.62,9.0.42,10.0.2;org.apache.tomcat:tomcat-coyote:8.5.62,9.0.42,10.0.2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"8.5.34","packageFilePaths":["/nimbus-entity-dsl/pom.xml","/nimbus-test/pom.xml","/nimbus-core/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.0.5.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:2.0.5.RELEASE;org.apache.tomcat.embed:tomcat-embed-core:8.5.34","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:8.5.62,9.0.42,10.0.2;org.apache.tomcat:tomcat-coyote:8.5.62,9.0.42,10.0.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-25122","vulnerabilityDetails":"When responding to new h2c connection requests, Apache Tomcat versions 10.0.0-M1 to 10.0.0, 9.0.0.M1 to 9.0.41 and 8.5.0 to 8.5.61 could duplicate request headers and a limited amount of request body from one request to another meaning user A and user B could both see the results of user A\u0027s request.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25122","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-25122 (High) detected in tomcat-embed-core-8.5.34.jar - ## CVE-2021-25122 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.34.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p> <p>Path to dependency file: core/nimbus-entity-dsl/pom.xml</p> <p>Path to vulnerable library: core/8.5.34/tomcat-embed-core-8.5.34.jar,core/8.5.34/tomcat-embed-core-8.5.34.jar,core/8.5.34/tomcat-embed-core-8.5.34.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.0.5.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-2.0.5.RELEASE.jar - :x: **tomcat-embed-core-8.5.34.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When responding to new h2c connection requests, Apache Tomcat versions 10.0.0-M1 to 10.0.0, 9.0.0.M1 to 9.0.41 and 8.5.0 to 8.5.61 could duplicate request headers and a limited amount of request body from one request to another meaning user A and user B could both see the results of user A's request. <p>Publish Date: 2021-03-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25122>CVE-2021-25122</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://lists.apache.org/thread.html/r7b95bc248603360501f18c8eb03bb6001ec0ee3296205b34b07105b7%40%3Cannounce.tomcat.apache.org%3E">https://lists.apache.org/thread.html/r7b95bc248603360501f18c8eb03bb6001ec0ee3296205b34b07105b7%40%3Cannounce.tomcat.apache.org%3E</a></p> <p>Release Date: 2021-03-01</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:8.5.62,9.0.42,10.0.2;org.apache.tomcat:tomcat-coyote:8.5.62,9.0.42,10.0.2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"8.5.34","packageFilePaths":["/nimbus-entity-dsl/pom.xml","/nimbus-test/pom.xml","/nimbus-core/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.0.5.RELEASE;org.springframework.boot:spring-boot-starter-tomcat:2.0.5.RELEASE;org.apache.tomcat.embed:tomcat-embed-core:8.5.34","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:8.5.62,9.0.42,10.0.2;org.apache.tomcat:tomcat-coyote:8.5.62,9.0.42,10.0.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-25122","vulnerabilityDetails":"When responding to new h2c connection requests, Apache Tomcat versions 10.0.0-M1 to 10.0.0, 9.0.0.M1 to 9.0.41 and 8.5.0 to 8.5.61 could duplicate request headers and a limited amount of request body from one request to another meaning user A and user B could both see the results of user A\u0027s request.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25122","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in tomcat embed core jar cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation library home page a href path to dependency file core nimbus entity dsl pom xml path to vulnerable library core tomcat embed core jar core tomcat embed core jar core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library found in base branch master vulnerability details when responding to new connection requests apache tomcat versions to to and to could duplicate request headers and a limited amount of request body from one request to another meaning user a and user b could both see the results of user a s request publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat coyote isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org springframework boot spring boot starter web release org springframework boot spring boot starter tomcat release org apache tomcat embed tomcat embed core isminimumfixversionavailable true minimumfixversion org apache tomcat embed tomcat embed core org apache tomcat tomcat coyote basebranches vulnerabilityidentifier cve vulnerabilitydetails when responding to new connection requests apache tomcat versions to to and to could duplicate request headers and a limited amount of request body from one request to another meaning user a and user b could both see the results of user a request vulnerabilityurl
0
6,452
9,546,480,214
IssuesEvent
2019-05-01 20:04:17
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Department of State: Experience Questions- Add character limit
Apply Process Approved Requirements Ready State Dept.
Who: Student What:Character Account Why: As an applicant I would like to know the character limit to a text box A/C - On the Experience & References Page Under the "Other" selection - the "If you chose "other", please specify - provide a character limit (250 characters remaining) - Please indicate the total length of your overseas experience(s) provide a character limit (250 characters remaining) - The character limit will change as the user types in the box and reflect the number of remaining characters.
1.0
Department of State: Experience Questions- Add character limit - Who: Student What:Character Account Why: As an applicant I would like to know the character limit to a text box A/C - On the Experience & References Page Under the "Other" selection - the "If you chose "other", please specify - provide a character limit (250 characters remaining) - Please indicate the total length of your overseas experience(s) provide a character limit (250 characters remaining) - The character limit will change as the user types in the box and reflect the number of remaining characters.
process
department of state experience questions add character limit who student what character account why as an applicant i would like to know the character limit to a text box a c on the experience references page under the other selection the if you chose other please specify provide a character limit characters remaining please indicate the total length of your overseas experience s provide a character limit characters remaining the character limit will change as the user types in the box and reflect the number of remaining characters
1
14,310
17,316,261,846
IssuesEvent
2021-07-27 06:36:44
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
System.Diagnostics.Process.Tests fails on iOS/tvOS
area-System.Diagnostics.Process in pr os-ios os-tvos
System.Diagnostics.Process.Tests System.Diagnostics.Tests.ProcessStartInfoTests.TestEnvironmentVariablesPropertyUnix System.Diagnostics.Tests.ProcessStartInfoTests.TestEnvironmentProperty System.Diagnostics.Tests.ProcessTests.MaxWorkingSet_GetNotStarted_ThrowsInvalidOperationException System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes System.Diagnostics.Tests.ProcessTests.Kill_NotStarted_ThrowsInvalidOperationException System.Diagnostics.Tests.ProcessTests.ProcessStart_TryOpenFolder_UseShellExecuteIsFalse_ThrowsWin32Exception System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_WorkingDirectory System.Diagnostics.Tests.ProcessTests.TestStartOnUnixWithBadPermissions System.Diagnostics.Tests.ProcessTests.ProcessNameMatchesScriptName System.Diagnostics.Tests.ProcessTests.TestStartWithMissingFile System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecuteFalse_FilenameIsUrl_ThrowsWin32Exception System.Diagnostics.Tests.ProcessTests.TestStartOnUnixWithBadFormat System.Diagnostics.Tests.ProcessTests.MinWorkingSet_GetNotStarted_ThrowsInvalidOperationException System.Diagnostics.Tests.ProcessTests.ProcessStart_TryExitCommandAsFileName_ThrowsWin32Exception ``` System.Diagnostics.Process.Tests.dll Failed: 16 Test collection for System.Diagnostics.Process.Tests.dll System.Diagnostics.Tests.ProcessStartInfoTests.TestEnvironmentVariablesPropertyUnix Assert.NotEqual() Failure\nExpected: Not 0\nActual: 0 System.Diagnostics.Tests.ProcessStartInfoTests.TestEnvironmentProperty Assert.NotEqual() Failure\nExpected: Not 0\nActual: 0 System.Diagnostics.Tests.ProcessTests.MaxWorkingSet_GetNotStarted_ThrowsInvalidOperationException Assert.Throws() Failure\nExpected: typeof(System.InvalidOperationException)\nActual: typeof(System.PlatformNotSupportedException): Getting or setting the working set limits on other processes is not supported on this platform.\n---- System.PlatformNotSupportedException : Getting or setting the working set limits on other processes is not supported on this platform. System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes(filenameAsUrl: False) System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes(filenameAsUrl: True) System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.Kill_NotStarted_ThrowsInvalidOperationException Assert.Throws() Failure\nExpected: typeof(System.InvalidOperationException)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.ProcessStart_TryOpenFolder_UseShellExecuteIsFalse_ThrowsWin32Exception Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_WorkingDirectory System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.TestStartOnUnixWithBadPermissions Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.ProcessNameMatchesScriptName System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.TestStartWithMissingFile(fullPath: True) Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.TestStartWithMissingFile(fullPath: False) Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecuteFalse_FilenameIsUrl_ThrowsWin32Exception Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.TestStartOnUnixWithBadFormat Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.MinWorkingSet_GetNotStarted_ThrowsInvalidOperationException Assert.Throws() Failure\nExpected: typeof(System.InvalidOperationException)\nActual: typeof(System.PlatformNotSupportedException): Getting or setting the working set limits on other processes is not supported on this platform.\n---- System.PlatformNotSupportedException : Getting or setting the working set limits on other processes is not supported on this platform. System.Diagnostics.Tests.ProcessTests.ProcessStart_TryExitCommandAsFileName_ThrowsWin32Exception Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
1.0
System.Diagnostics.Process.Tests fails on iOS/tvOS - System.Diagnostics.Process.Tests System.Diagnostics.Tests.ProcessStartInfoTests.TestEnvironmentVariablesPropertyUnix System.Diagnostics.Tests.ProcessStartInfoTests.TestEnvironmentProperty System.Diagnostics.Tests.ProcessTests.MaxWorkingSet_GetNotStarted_ThrowsInvalidOperationException System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes System.Diagnostics.Tests.ProcessTests.Kill_NotStarted_ThrowsInvalidOperationException System.Diagnostics.Tests.ProcessTests.ProcessStart_TryOpenFolder_UseShellExecuteIsFalse_ThrowsWin32Exception System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_WorkingDirectory System.Diagnostics.Tests.ProcessTests.TestStartOnUnixWithBadPermissions System.Diagnostics.Tests.ProcessTests.ProcessNameMatchesScriptName System.Diagnostics.Tests.ProcessTests.TestStartWithMissingFile System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecuteFalse_FilenameIsUrl_ThrowsWin32Exception System.Diagnostics.Tests.ProcessTests.TestStartOnUnixWithBadFormat System.Diagnostics.Tests.ProcessTests.MinWorkingSet_GetNotStarted_ThrowsInvalidOperationException System.Diagnostics.Tests.ProcessTests.ProcessStart_TryExitCommandAsFileName_ThrowsWin32Exception ``` System.Diagnostics.Process.Tests.dll Failed: 16 Test collection for System.Diagnostics.Process.Tests.dll System.Diagnostics.Tests.ProcessStartInfoTests.TestEnvironmentVariablesPropertyUnix Assert.NotEqual() Failure\nExpected: Not 0\nActual: 0 System.Diagnostics.Tests.ProcessStartInfoTests.TestEnvironmentProperty Assert.NotEqual() Failure\nExpected: Not 0\nActual: 0 System.Diagnostics.Tests.ProcessTests.MaxWorkingSet_GetNotStarted_ThrowsInvalidOperationException Assert.Throws() Failure\nExpected: typeof(System.InvalidOperationException)\nActual: typeof(System.PlatformNotSupportedException): Getting or setting the working set limits on other processes is not supported on this platform.\n---- System.PlatformNotSupportedException : Getting or setting the working set limits on other processes is not supported on this platform. System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes(filenameAsUrl: False) System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_Executes(filenameAsUrl: True) System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.Kill_NotStarted_ThrowsInvalidOperationException Assert.Throws() Failure\nExpected: typeof(System.InvalidOperationException)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.ProcessStart_TryOpenFolder_UseShellExecuteIsFalse_ThrowsWin32Exception Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecute_WorkingDirectory System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.TestStartOnUnixWithBadPermissions Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.ProcessNameMatchesScriptName System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.TestStartWithMissingFile(fullPath: True) Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.TestStartWithMissingFile(fullPath: False) Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.ProcessStart_UseShellExecuteFalse_FilenameIsUrl_ThrowsWin32Exception Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.TestStartOnUnixWithBadFormat Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform. System.Diagnostics.Tests.ProcessTests.MinWorkingSet_GetNotStarted_ThrowsInvalidOperationException Assert.Throws() Failure\nExpected: typeof(System.InvalidOperationException)\nActual: typeof(System.PlatformNotSupportedException): Getting or setting the working set limits on other processes is not supported on this platform.\n---- System.PlatformNotSupportedException : Getting or setting the working set limits on other processes is not supported on this platform. System.Diagnostics.Tests.ProcessTests.ProcessStart_TryExitCommandAsFileName_ThrowsWin32Exception Assert.Throws() Failure\nExpected: typeof(System.ComponentModel.Win32Exception)\nActual: typeof(System.PlatformNotSupportedException): Operation is not supported on this platform.\n---- System.PlatformNotSupportedException : Operation is not supported on this platform.
process
system diagnostics process tests fails on ios tvos system diagnostics process tests system diagnostics tests processstartinfotests testenvironmentvariablespropertyunix system diagnostics tests processstartinfotests testenvironmentproperty system diagnostics tests processtests maxworkingset getnotstarted throwsinvalidoperationexception system diagnostics tests processtests processstart useshellexecute executes system diagnostics tests processtests kill notstarted throwsinvalidoperationexception system diagnostics tests processtests processstart tryopenfolder useshellexecuteisfalse system diagnostics tests processtests processstart useshellexecute workingdirectory system diagnostics tests processtests teststartonunixwithbadpermissions system diagnostics tests processtests processnamematchesscriptname system diagnostics tests processtests teststartwithmissingfile system diagnostics tests processtests processstart useshellexecutefalse filenameisurl system diagnostics tests processtests teststartonunixwithbadformat system diagnostics tests processtests minworkingset getnotstarted throwsinvalidoperationexception system diagnostics tests processtests processstart tryexitcommandasfilename system diagnostics process tests dll failed test collection for system diagnostics process tests dll system diagnostics tests processstartinfotests testenvironmentvariablespropertyunix assert notequal failure nexpected not nactual system diagnostics tests processstartinfotests testenvironmentproperty assert notequal failure nexpected not nactual system diagnostics tests processtests maxworkingset getnotstarted throwsinvalidoperationexception assert throws failure nexpected typeof system invalidoperationexception nactual typeof system platformnotsupportedexception getting or setting the working set limits on other processes is not supported on this platform n system platformnotsupportedexception getting or setting the working set limits on other processes is not supported on this platform system diagnostics tests processtests processstart useshellexecute executes filenameasurl false system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests processstart useshellexecute executes filenameasurl true system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests kill notstarted throwsinvalidoperationexception assert throws failure nexpected typeof system invalidoperationexception nactual typeof system platformnotsupportedexception operation is not supported on this platform n system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests processstart tryopenfolder useshellexecuteisfalse assert throws failure nexpected typeof system componentmodel nactual typeof system platformnotsupportedexception operation is not supported on this platform n system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests processstart useshellexecute workingdirectory system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests teststartonunixwithbadpermissions assert throws failure nexpected typeof system componentmodel nactual typeof system platformnotsupportedexception operation is not supported on this platform n system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests processnamematchesscriptname system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests teststartwithmissingfile fullpath true assert throws failure nexpected typeof system componentmodel nactual typeof system platformnotsupportedexception operation is not supported on this platform n system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests teststartwithmissingfile fullpath false assert throws failure nexpected typeof system componentmodel nactual typeof system platformnotsupportedexception operation is not supported on this platform n system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests processstart useshellexecutefalse filenameisurl assert throws failure nexpected typeof system componentmodel nactual typeof system platformnotsupportedexception operation is not supported on this platform n system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests teststartonunixwithbadformat assert throws failure nexpected typeof system componentmodel nactual typeof system platformnotsupportedexception operation is not supported on this platform n system platformnotsupportedexception operation is not supported on this platform system diagnostics tests processtests minworkingset getnotstarted throwsinvalidoperationexception assert throws failure nexpected typeof system invalidoperationexception nactual typeof system platformnotsupportedexception getting or setting the working set limits on other processes is not supported on this platform n system platformnotsupportedexception getting or setting the working set limits on other processes is not supported on this platform system diagnostics tests processtests processstart tryexitcommandasfilename assert throws failure nexpected typeof system componentmodel nactual typeof system platformnotsupportedexception operation is not supported on this platform n system platformnotsupportedexception operation is not supported on this platform
1
56,611
32,076,419,892
IssuesEvent
2023-09-25 11:19:19
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
opt: slower plans chosen due to poor stats for TPC-H queries
C-performance A-sql-optimizer X-stale no-issue-activity
With join reordering, slower plans are chose for TPC-H queries 16 and 20. Q16 goes from ~600ms to ~900ms, and Q20 goes from ~13s to ~18s. Q16 fast-plan cost: `1090503.24` Q16 slow-plan cost: `342084.224` Q20 false-plan cost: `5790840.88` Q20 slow-plan cost: `5782416.86` Both queries have a `LIKE` filter for which selectivity is severely underestimated; this is because the default conjunct selectivity estimate of 1/3 is used for `LIKE` filters. This leads to row-count overestimation. Q20 has an additional problem with stats estimation; selectivity for the `l_partkey:31 = ps_partkey:14 AND l_suppkey:32 = ps_suppkey:15` join filter is overestimated (as opposed to the `LIKE` filter, the selectivity of which is underestimated). This leads to row-count underestimation. Jira issue: CRDB-3976
True
opt: slower plans chosen due to poor stats for TPC-H queries - With join reordering, slower plans are chose for TPC-H queries 16 and 20. Q16 goes from ~600ms to ~900ms, and Q20 goes from ~13s to ~18s. Q16 fast-plan cost: `1090503.24` Q16 slow-plan cost: `342084.224` Q20 false-plan cost: `5790840.88` Q20 slow-plan cost: `5782416.86` Both queries have a `LIKE` filter for which selectivity is severely underestimated; this is because the default conjunct selectivity estimate of 1/3 is used for `LIKE` filters. This leads to row-count overestimation. Q20 has an additional problem with stats estimation; selectivity for the `l_partkey:31 = ps_partkey:14 AND l_suppkey:32 = ps_suppkey:15` join filter is overestimated (as opposed to the `LIKE` filter, the selectivity of which is underestimated). This leads to row-count underestimation. Jira issue: CRDB-3976
non_process
opt slower plans chosen due to poor stats for tpc h queries with join reordering slower plans are chose for tpc h queries and goes from to and goes from to fast plan cost slow plan cost false plan cost slow plan cost both queries have a like filter for which selectivity is severely underestimated this is because the default conjunct selectivity estimate of is used for like filters this leads to row count overestimation has an additional problem with stats estimation selectivity for the l partkey ps partkey and l suppkey ps suppkey join filter is overestimated as opposed to the like filter the selectivity of which is underestimated this leads to row count underestimation jira issue crdb
0
75,919
7,496,217,559
IssuesEvent
2018-04-08 07:00:31
MajkiIT/polish-ads-filter
https://api.github.com/repos/MajkiIT/polish-ads-filter
closed
hrubie.pl
reguły gotowe/testowanie reklama
Ad empty container ### Zrzut ekranu ![capture_114_07042018_135804](https://user-images.githubusercontent.com/37474031/38454673-cc822894-3a6b-11e8-9406-703e46ed9a77.jpg) ### Link bezpośredni http://www.hrubie.pl/ ### Co trzeba zrobić, aby pojawił się element, reklama albo błąd? 1. Visit the page ### Moja konfiguracja Win 7 SP1 [64bit] Firefox 59.0.2 [64bit] uBlock Origin 1.15.24 ### Filtry: <details> <summary>Click to expand</summary> Nano Defender 13.48 Nano Defender filter Nano Base filter Nano Whitelist filter uBlock filters uBlock filters – Annoyances uBlock filters – Badware risks uBlock filters – Privacy uBlock filters – Resource abuse​​​​​ uBlock filters – Unbreak Adguard’s Annoyance List Adguard Base Filters Adguard Spyware Filters Adblock & uBlock polish filter - AdGuard supplement Adblock Warning Removal List EasyList EasyList Polish EasyPrivacy Fanboy’s Annoyance List Fanboy’s Anti-Thirdparty Social (see warning inside list) Fanboy’s Cookiemonster List Fanboy’s Enhanced Tracking List Fanboy's Polish Fanboy’s Social Blocking List I'm OK with cookies​ I don't care about cookies hpHosts’ Ad and tracking servers Peter Lowe’s Ad and tracking server list POL: polskie filtry do Adblocka i uBlocka Polish Privacy Filters Polskie Filtry Anty-Donacyjne Polskie Filtry Ciasteczkowe​ Polskie Filtry Elementów Irytujących Polskie Filtry RSS​ Polskie Filtry Społecznościowe​ Polskie Filtry Wewnętrzne AlleBlock​ KAD - Przekręty Web Annoyances Ultralist​ </details>
1.0
hrubie.pl - Ad empty container ### Zrzut ekranu ![capture_114_07042018_135804](https://user-images.githubusercontent.com/37474031/38454673-cc822894-3a6b-11e8-9406-703e46ed9a77.jpg) ### Link bezpośredni http://www.hrubie.pl/ ### Co trzeba zrobić, aby pojawił się element, reklama albo błąd? 1. Visit the page ### Moja konfiguracja Win 7 SP1 [64bit] Firefox 59.0.2 [64bit] uBlock Origin 1.15.24 ### Filtry: <details> <summary>Click to expand</summary> Nano Defender 13.48 Nano Defender filter Nano Base filter Nano Whitelist filter uBlock filters uBlock filters – Annoyances uBlock filters – Badware risks uBlock filters – Privacy uBlock filters – Resource abuse​​​​​ uBlock filters – Unbreak Adguard’s Annoyance List Adguard Base Filters Adguard Spyware Filters Adblock & uBlock polish filter - AdGuard supplement Adblock Warning Removal List EasyList EasyList Polish EasyPrivacy Fanboy’s Annoyance List Fanboy’s Anti-Thirdparty Social (see warning inside list) Fanboy’s Cookiemonster List Fanboy’s Enhanced Tracking List Fanboy's Polish Fanboy’s Social Blocking List I'm OK with cookies​ I don't care about cookies hpHosts’ Ad and tracking servers Peter Lowe’s Ad and tracking server list POL: polskie filtry do Adblocka i uBlocka Polish Privacy Filters Polskie Filtry Anty-Donacyjne Polskie Filtry Ciasteczkowe​ Polskie Filtry Elementów Irytujących Polskie Filtry RSS​ Polskie Filtry Społecznościowe​ Polskie Filtry Wewnętrzne AlleBlock​ KAD - Przekręty Web Annoyances Ultralist​ </details>
non_process
hrubie pl ad empty container zrzut ekranu link bezpośredni co trzeba zrobić aby pojawił się element reklama albo błąd visit the page moja konfiguracja win firefox ublock origin filtry click to expand nano defender nano defender filter nano base filter nano whitelist filter ublock filters ublock filters – annoyances ublock filters – badware risks ublock filters – privacy ublock filters – resource abuse​​​​​ ublock filters – unbreak adguard’s annoyance list adguard base filters adguard spyware filters adblock ublock polish filter adguard supplement adblock warning removal list easylist easylist polish easyprivacy fanboy’s annoyance list fanboy’s anti thirdparty social see warning inside list fanboy’s cookiemonster list fanboy’s enhanced tracking list fanboy s polish fanboy’s social blocking list i m ok with cookies​ i don t care about cookies hphosts’ ad and tracking servers peter lowe’s ad and tracking server list pol polskie filtry do adblocka i ublocka polish privacy filters polskie filtry anty donacyjne polskie filtry ciasteczkowe​ polskie filtry elementów irytujących polskie filtry rss​ polskie filtry społecznościowe​ polskie filtry wewnętrzne alleblock​ kad przekręty web annoyances ultralist​
0
186,125
15,047,976,016
IssuesEvent
2021-02-03 09:37:01
NotOnlyStudents/Documentazione
https://api.github.com/repos/NotOnlyStudents/Documentazione
closed
UC3 Come funzionalità e non implementazione
adr bug documentation
Tullio: "UC3 non rappresenta una funzionalità, ma è già calato in una possibile implementazione che prevede l’uso di “menù”. I sotto-casi individuati soffrono del medesimo problema."
1.0
UC3 Come funzionalità e non implementazione - Tullio: "UC3 non rappresenta una funzionalità, ma è già calato in una possibile implementazione che prevede l’uso di “menù”. I sotto-casi individuati soffrono del medesimo problema."
non_process
come funzionalità e non implementazione tullio non rappresenta una funzionalità ma è già calato in una possibile implementazione che prevede l’uso di “menù” i sotto casi individuati soffrono del medesimo problema
0
330,609
10,053,250,311
IssuesEvent
2019-07-21 15:14:02
ticket721/web-app
https://api.github.com/repos/ticket721/web-app
closed
d.feat: generic price / balance displayer
[priority] [➖➖➖➖] critical [status] to do [type] feat
Find a way to display value in $ -> hover shows ether and DAI value, highlights fixed currency
1.0
d.feat: generic price / balance displayer - Find a way to display value in $ -> hover shows ether and DAI value, highlights fixed currency
non_process
d feat generic price balance displayer find a way to display value in hover shows ether and dai value highlights fixed currency
0
10,748
7,300,697,978
IssuesEvent
2018-02-27 01:00:30
dotnet/coreclr
https://api.github.com/repos/dotnet/coreclr
opened
[Perf] Investigate CscBench Regression between release/2.0.0 and release/2.1
area-Benchmarks tenet-performance
We saw a regression of 5% between 2.0 and 2.1
True
[Perf] Investigate CscBench Regression between release/2.0.0 and release/2.1 - We saw a regression of 5% between 2.0 and 2.1
non_process
investigate cscbench regression between release and release we saw a regression of between and
0
1,068
3,536,075,422
IssuesEvent
2016-01-17 00:28:31
MaretEngineering/MROV
https://api.github.com/repos/MaretEngineering/MROV
closed
Add constants to replace numbers
Arduino enhancement Processing
There should be almost no numbers left on their own in the main body of the program. They should all be stored in a constant with a good name to add clarity.
1.0
Add constants to replace numbers - There should be almost no numbers left on their own in the main body of the program. They should all be stored in a constant with a good name to add clarity.
process
add constants to replace numbers there should be almost no numbers left on their own in the main body of the program they should all be stored in a constant with a good name to add clarity
1
826,352
31,592,092,172
IssuesEvent
2023-09-05 00:03:33
Selody-project/Backend
https://api.github.com/repos/Selody-project/Backend
closed
[BACK-TASK] 개인 일정 공유 그룹 or 비공유 그룹
Priority 2
### 설명 - 공유 그룹인 경우에는 기존과 같은 방식으로 개인 일정을 사용 - 일정 비공유 그룹인 경우에는 해당 유저의 개인 일정을 그룹에서 사용하지 않도록 수정
1.0
[BACK-TASK] 개인 일정 공유 그룹 or 비공유 그룹 - ### 설명 - 공유 그룹인 경우에는 기존과 같은 방식으로 개인 일정을 사용 - 일정 비공유 그룹인 경우에는 해당 유저의 개인 일정을 그룹에서 사용하지 않도록 수정
non_process
개인 일정 공유 그룹 or 비공유 그룹 설명 공유 그룹인 경우에는 기존과 같은 방식으로 개인 일정을 사용 일정 비공유 그룹인 경우에는 해당 유저의 개인 일정을 그룹에서 사용하지 않도록 수정
0
17,030
22,406,824,612
IssuesEvent
2022-06-18 04:46:58
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
[metricstransformprocessor] Summary datapoint merging is not handled.
bug proc: metricstransformprocessor
**Describe the bug** In `datapoint_aggregation.go`, the `MetricDescriptor_SUMMARY` type is not handled. **Steps to reproduce** 1. Get a Summary metric into the pipeline. For example, the `http.server.requests` metric emitted from Spring Boot + `micrometer-registry-statsd` is a "timer" type, and becomes a "summary" when received by the `statsdreceiver`, with this config: ``` statsd: timer_histogram_mapping: - statsd_type: histogram observer_type: summary - statsd_type: timer observer_type: summary ``` 2. Apply a `metricstransform` processor with a "sum" aggregation type to the metric, such as with this config: ``` metricstransform: transforms: - include: http.server.requests action: update operations: - action: aggregate_labels label_set: [ uri ] aggregation_type: sum ``` **What did you expect to see?** I expect the labels of the `http.server.requests` metric besides `uri` to be removed, and the summaries for those datapoints to be aggregated. **What did you see instead?** The `http.server.requests` metric is dropped entirely. **What version did you use?** Version: aws-opentelemetry-collector v0.15 **Additional context** The `datapoint_aggregation.go` file already has logic for aggregating histogram datapoints, but is missing logic for summary datapoints (specifically, it has logic for `MetricDescriptor_GAUGE_DISTRIBUTION`, but not `MetricDescriptor_SUMMARY`)
1.0
[metricstransformprocessor] Summary datapoint merging is not handled. - **Describe the bug** In `datapoint_aggregation.go`, the `MetricDescriptor_SUMMARY` type is not handled. **Steps to reproduce** 1. Get a Summary metric into the pipeline. For example, the `http.server.requests` metric emitted from Spring Boot + `micrometer-registry-statsd` is a "timer" type, and becomes a "summary" when received by the `statsdreceiver`, with this config: ``` statsd: timer_histogram_mapping: - statsd_type: histogram observer_type: summary - statsd_type: timer observer_type: summary ``` 2. Apply a `metricstransform` processor with a "sum" aggregation type to the metric, such as with this config: ``` metricstransform: transforms: - include: http.server.requests action: update operations: - action: aggregate_labels label_set: [ uri ] aggregation_type: sum ``` **What did you expect to see?** I expect the labels of the `http.server.requests` metric besides `uri` to be removed, and the summaries for those datapoints to be aggregated. **What did you see instead?** The `http.server.requests` metric is dropped entirely. **What version did you use?** Version: aws-opentelemetry-collector v0.15 **Additional context** The `datapoint_aggregation.go` file already has logic for aggregating histogram datapoints, but is missing logic for summary datapoints (specifically, it has logic for `MetricDescriptor_GAUGE_DISTRIBUTION`, but not `MetricDescriptor_SUMMARY`)
process
summary datapoint merging is not handled describe the bug in datapoint aggregation go the metricdescriptor summary type is not handled steps to reproduce get a summary metric into the pipeline for example the http server requests metric emitted from spring boot micrometer registry statsd is a timer type and becomes a summary when received by the statsdreceiver with this config statsd timer histogram mapping statsd type histogram observer type summary statsd type timer observer type summary apply a metricstransform processor with a sum aggregation type to the metric such as with this config metricstransform transforms include http server requests action update operations action aggregate labels label set aggregation type sum what did you expect to see i expect the labels of the http server requests metric besides uri to be removed and the summaries for those datapoints to be aggregated what did you see instead the http server requests metric is dropped entirely what version did you use version aws opentelemetry collector additional context the datapoint aggregation go file already has logic for aggregating histogram datapoints but is missing logic for summary datapoints specifically it has logic for metricdescriptor gauge distribution but not metricdescriptor summary
1
18,546
24,555,302,205
IssuesEvent
2022-10-12 15:25:41
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[iOS] Formstep > text choice > user is able to select multiple answer options even though 'Single select' option is selected in the SB
Bug P1 iOS Process: Fixed Process: Tested dev
Steps: 1. In SB, create a study with a 'Text choice' response type by selecting single select option 2. Launch the study 3. Sign up or sign in to the mobile app 4. Enroll to the study 5. Click on text choice activity 6. Try to select multiple answer options AR: Participant is able to select multiple answer options ER: Participant should not able to select multiple answer options when the 'single select' option is selected in the SB SB: ![image](https://user-images.githubusercontent.com/71445210/185600408-eb9c9796-db7a-455c-a2c3-226d64bba8c4.png) ER: ![image](https://user-images.githubusercontent.com/71445210/185600787-80a511c9-6acd-4f8a-80c9-dadc496e7452.png)
2.0
[iOS] Formstep > text choice > user is able to select multiple answer options even though 'Single select' option is selected in the SB - Steps: 1. In SB, create a study with a 'Text choice' response type by selecting single select option 2. Launch the study 3. Sign up or sign in to the mobile app 4. Enroll to the study 5. Click on text choice activity 6. Try to select multiple answer options AR: Participant is able to select multiple answer options ER: Participant should not able to select multiple answer options when the 'single select' option is selected in the SB SB: ![image](https://user-images.githubusercontent.com/71445210/185600408-eb9c9796-db7a-455c-a2c3-226d64bba8c4.png) ER: ![image](https://user-images.githubusercontent.com/71445210/185600787-80a511c9-6acd-4f8a-80c9-dadc496e7452.png)
process
formstep text choice user is able to select multiple answer options even though single select option is selected in the sb steps in sb create a study with a text choice response type by selecting single select option launch the study sign up or sign in to the mobile app enroll to the study click on text choice activity try to select multiple answer options ar participant is able to select multiple answer options er participant should not able to select multiple answer options when the single select option is selected in the sb sb er
1
597,295
18,160,590,165
IssuesEvent
2021-09-27 09:09:38
CaptureCoop/SnipSniper
https://api.github.com/repos/CaptureCoop/SnipSniper
opened
ConfigWindow -> Better tabs
enhancement Medium Priority
Make each tab its own class for cleaner code. Also implement that when you close the window it prompts you if window is dirty. Also ask the same when switching from a dirty tab to another one.
1.0
ConfigWindow -> Better tabs - Make each tab its own class for cleaner code. Also implement that when you close the window it prompts you if window is dirty. Also ask the same when switching from a dirty tab to another one.
non_process
configwindow better tabs make each tab its own class for cleaner code also implement that when you close the window it prompts you if window is dirty also ask the same when switching from a dirty tab to another one
0
334,063
10,135,605,353
IssuesEvent
2019-08-02 10:35:31
aiidateam/aiida-quantumespresso
https://api.github.com/repos/aiidateam/aiida-quantumespresso
closed
Import error in `BasePwCpInputGenerator.prepare_for_submission`
priority/important topic/calculations type/bug
``` also_bands = settings.pop('ALSO_BANDS', None) if also_bands: import logging from aiida.common.log import get_dblogger_extra ```
1.0
Import error in `BasePwCpInputGenerator.prepare_for_submission` - ``` also_bands = settings.pop('ALSO_BANDS', None) if also_bands: import logging from aiida.common.log import get_dblogger_extra ```
non_process
import error in basepwcpinputgenerator prepare for submission also bands settings pop also bands none if also bands import logging from aiida common log import get dblogger extra
0
707,642
24,312,757,552
IssuesEvent
2022-09-30 01:16:13
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.amazon.co.jp - site is not usable
priority-important status-needsinfo-karlcow browser-fenix engine-gecko
<!-- @browser: Firefox Mobile 107.0 --> <!-- @ua_header: Mozilla/5.0 (Android 12; Mobile; rv:107.0) Gecko/107.0 Firefox/107.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/111454 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.amazon.co.jp/dp/B09P83BLHW?binding=kindle_edition&ref=dbs_dp_rwt_sb_pc_tukn **Browser / Version**: Firefox Mobile 107.0 **Operating System**: Android 12 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Buttons or links not working **Steps to Reproduce**: I can not purchase by "1-Click to buy now"(1-clickで今すぐ買う)button. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2022/9/d362e4c9-aabb-4b21-86e7-44fc0a572f94.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220926093803</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/9/225da95d-fd1f-4939-91ff-03ff4fc2cd01) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.amazon.co.jp - site is not usable - <!-- @browser: Firefox Mobile 107.0 --> <!-- @ua_header: Mozilla/5.0 (Android 12; Mobile; rv:107.0) Gecko/107.0 Firefox/107.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/111454 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.amazon.co.jp/dp/B09P83BLHW?binding=kindle_edition&ref=dbs_dp_rwt_sb_pc_tukn **Browser / Version**: Firefox Mobile 107.0 **Operating System**: Android 12 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Buttons or links not working **Steps to Reproduce**: I can not purchase by "1-Click to buy now"(1-clickで今すぐ買う)button. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2022/9/d362e4c9-aabb-4b21-86e7-44fc0a572f94.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220926093803</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/9/225da95d-fd1f-4939-91ff-03ff4fc2cd01) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
site is not usable url browser version firefox mobile operating system android tested another browser yes chrome problem type site is not usable description buttons or links not working steps to reproduce i can not purchase by click to buy now ( clickで今すぐ買う)button view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
159,791
25,046,457,495
IssuesEvent
2022-11-05 10:07:34
hypha-dao/dho-web-client
https://api.github.com/repos/hypha-dao/dho-web-client
closed
Proposal Cards should always be the same height
Bug Design
Reproduce: 1. Go to https://dao.hypha.earth/hypha/proposals/ 2. Check height of cards: <img width="666" alt="image" src="https://user-images.githubusercontent.com/75991832/192541781-8d579d85-c6aa-4251-83b7-a2c70550f31b.png"> Note: probably because 1 card has a comment?
1.0
Proposal Cards should always be the same height - Reproduce: 1. Go to https://dao.hypha.earth/hypha/proposals/ 2. Check height of cards: <img width="666" alt="image" src="https://user-images.githubusercontent.com/75991832/192541781-8d579d85-c6aa-4251-83b7-a2c70550f31b.png"> Note: probably because 1 card has a comment?
non_process
proposal cards should always be the same height reproduce go to check height of cards img width alt image src note probably because card has a comment
0
71,645
18,819,987,248
IssuesEvent
2021-11-10 06:55:04
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
closed
[Bug]: Toast should be having a max width
Bug High Release UI Building Pod regression UI Builders Pod hotfix
### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior Toast messages are too lengthy ![image (1)](https://user-images.githubusercontent.com/10436935/140890417-22692f1b-7ddb-4777-86b8-bfa68693efb3.png) ### Steps To Reproduce Use `showAlert` to show a large message ### Environment Release ### Version Self-Hosted
2.0
[Bug]: Toast should be having a max width - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior Toast messages are too lengthy ![image (1)](https://user-images.githubusercontent.com/10436935/140890417-22692f1b-7ddb-4777-86b8-bfa68693efb3.png) ### Steps To Reproduce Use `showAlert` to show a large message ### Environment Release ### Version Self-Hosted
non_process
toast should be having a max width is there an existing issue for this i have searched the existing issues current behavior toast messages are too lengthy steps to reproduce use showalert to show a large message environment release version self hosted
0
53,295
13,795,765,869
IssuesEvent
2020-10-09 18:36:13
fsjs-academy/aroa-ui
https://api.github.com/repos/fsjs-academy/aroa-ui
closed
CVE-2012-6708 (Medium) detected in jquery-1.8.1.min.js
security vulnerability
## CVE-2012-6708 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: aroa-ui/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: aroa-ui/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fsjs-academy/aroa-ui/commit/86ab73fa8c8dcf1da44f79b5fa99aebd57adcd1d">86ab73fa8c8dcf1da44f79b5fa99aebd57adcd1d</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2012-6708 (Medium) detected in jquery-1.8.1.min.js - ## CVE-2012-6708 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: aroa-ui/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: aroa-ui/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fsjs-academy/aroa-ui/commit/86ab73fa8c8dcf1da44f79b5fa99aebd57adcd1d">86ab73fa8c8dcf1da44f79b5fa99aebd57adcd1d</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file aroa ui node modules redeyed examples browser index html path to vulnerable library aroa ui node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch main vulnerability details jquery before is vulnerable to cross site scripting xss attacks the jquery strinput function does not differentiate selectors from html in a reliable fashion in vulnerable versions jquery determined whether the input was html by looking for the character anywhere in the string giving attackers more flexibility when attempting to construct a malicious payload in fixed versions jquery only deems the input to be html if it explicitly starts with the character limiting exploitability only to attackers who can control the beginning of a string which is far less common publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
9,319
12,338,228,672
IssuesEvent
2020-05-14 16:07:51
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Runtime parameters - out of date banner
Pri1 devops-cicd-process/tech devops/prod doc-bug
On the page with runtime parameters there is this banner: ![image](https://user-images.githubusercontent.com/17653648/80288083-c79a0e00-8735-11ea-9df5-ec72e90a933c.png) I think this banner is out of date right now. This feature is probably available to almost all organizations. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 790318bb-8220-3241-4ca7-73351074492f * Version Independent ID: db1da9db-3694-779b-17aa-1ed67fcecf86 * Content: [Use runtime and type-safe parameters - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/runtime-parameters?view=azure-devops&tabs=script) * Content Source: [docs/pipelines/process/runtime-parameters.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/runtime-parameters.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Runtime parameters - out of date banner - On the page with runtime parameters there is this banner: ![image](https://user-images.githubusercontent.com/17653648/80288083-c79a0e00-8735-11ea-9df5-ec72e90a933c.png) I think this banner is out of date right now. This feature is probably available to almost all organizations. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 790318bb-8220-3241-4ca7-73351074492f * Version Independent ID: db1da9db-3694-779b-17aa-1ed67fcecf86 * Content: [Use runtime and type-safe parameters - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/runtime-parameters?view=azure-devops&tabs=script) * Content Source: [docs/pipelines/process/runtime-parameters.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/runtime-parameters.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
runtime parameters out of date banner on the page with runtime parameters there is this banner i think this banner is out of date right now this feature is probably available to almost all organizations document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1