Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
717
3,206,328,337
IssuesEvent
2015-10-04 22:19:31
GFUCABAM/statler
https://api.github.com/repos/GFUCABAM/statler
opened
Understand integration options for external library
Language processing
Determine how we should handle the integration of external open-source code, specifically https://github.com/vivekn/sentiment. Determine if we should run the app alongside our own, as a service we call, or incldue the code in our app.Investigate using a package manager (such as Pip), versus simply pulling the raw files.
1.0
Understand integration options for external library - Determine how we should handle the integration of external open-source code, specifically https://github.com/vivekn/sentiment. Determine if we should run the app alongside our own, as a service we call, or incldue the code in our app.Investigate using a package manager (such as Pip), versus simply pulling the raw files.
process
understand integration options for external library determine how we should handle the integration of external open source code specifically determine if we should run the app alongside our own as a service we call or incldue the code in our app investigate using a package manager such as pip versus simply pulling the raw files
1
262,925
8,272,612,777
IssuesEvent
2018-09-16 22:06:57
javaee/glassfish
https://api.github.com/repos/javaee/glassfish
closed
container-common includes connector runtime classes
3_1-next 3_1_1-scrubbed 3_1_x-exclude Component: naming ERR: Assignee Priority: Major Type: Improvement
While experimenting the idea of a light weight web container distribution for GF, I noticed that container common depends on connector runtime. I see no reason for the same. Pl. fix it. #### Affected Versions [3.1]
1.0
container-common includes connector runtime classes - While experimenting the idea of a light weight web container distribution for GF, I noticed that container common depends on connector runtime. I see no reason for the same. Pl. fix it. #### Affected Versions [3.1]
non_process
container common includes connector runtime classes while experimenting the idea of a light weight web container distribution for gf i noticed that container common depends on connector runtime i see no reason for the same pl fix it affected versions
0
99,187
11,135,023,471
IssuesEvent
2019-12-20 13:24:14
gbv/cocoda
https://api.github.com/repos/gbv/cocoda
opened
PDF manual differs between travis, esx-206 and local machine
bug documentation
`npm run manual-pdf` should result in the same PDF but on esx-206 it's 13 pages and on my local machine 12 pages. Paper sizes and fonts are the same.
1.0
PDF manual differs between travis, esx-206 and local machine - `npm run manual-pdf` should result in the same PDF but on esx-206 it's 13 pages and on my local machine 12 pages. Paper sizes and fonts are the same.
non_process
pdf manual differs between travis esx and local machine npm run manual pdf should result in the same pdf but on esx it s pages and on my local machine pages paper sizes and fonts are the same
0
1,717
4,366,246,372
IssuesEvent
2016-08-03 13:49:15
SIMEXP/niak
https://api.github.com/repos/SIMEXP/niak
closed
Report for the fMRI preprocessing pipeline
enhancement preprocessing quality control
We would need an interactive html report for the preprocessing pipeline, presenting in a friendly way the following information: * QC summary measures on registration T1-to-template and BOLD-to-T1, as well as motion levels. * group maps, average group T1 vs template, average group BOLD vs average group T1, average BOLD mask vs average group T1, group BOLD mask vs average group T1. * small montage of individual T1 and BOLD volume to assess registration quality between them and with the template. * small movies (on a couple of slices) of BOLD time series, before and after motion correction synced up with motion parameters. Also a visu reference volume of each run vs reference volume of a subject, to check inter-run, within-subject registration.
1.0
Report for the fMRI preprocessing pipeline - We would need an interactive html report for the preprocessing pipeline, presenting in a friendly way the following information: * QC summary measures on registration T1-to-template and BOLD-to-T1, as well as motion levels. * group maps, average group T1 vs template, average group BOLD vs average group T1, average BOLD mask vs average group T1, group BOLD mask vs average group T1. * small montage of individual T1 and BOLD volume to assess registration quality between them and with the template. * small movies (on a couple of slices) of BOLD time series, before and after motion correction synced up with motion parameters. Also a visu reference volume of each run vs reference volume of a subject, to check inter-run, within-subject registration.
process
report for the fmri preprocessing pipeline we would need an interactive html report for the preprocessing pipeline presenting in a friendly way the following information qc summary measures on registration to template and bold to as well as motion levels group maps average group vs template average group bold vs average group average bold mask vs average group group bold mask vs average group small montage of individual and bold volume to assess registration quality between them and with the template small movies on a couple of slices of bold time series before and after motion correction synced up with motion parameters also a visu reference volume of each run vs reference volume of a subject to check inter run within subject registration
1
138,584
18,793,961,645
IssuesEvent
2021-11-08 19:55:45
Dima2022/hygieia-workflow-github-collector
https://api.github.com/repos/Dima2022/hygieia-workflow-github-collector
opened
CVE-2018-1271 (Medium) detected in spring-webmvc-4.2.5.RELEASE.jar
security vulnerability
## CVE-2018-1271 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-webmvc-4.2.5.RELEASE.jar</b></p></summary> <p>Spring Web MVC</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: hygieia-workflow-github-collector/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-webmvc/4.2.5.RELEASE/spring-webmvc-4.2.5.RELEASE.jar</p> <p> Dependency Hierarchy: - core-3.9.7.jar (Root Library) - :x: **spring-webmvc-4.2.5.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Dima2022/hygieia-workflow-github-collector/commit/236baaa856b74774f7b43ecb1eeade5a8d1d0496">236baaa856b74774f7b43ecb1eeade5a8d1d0496</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Spring Framework, versions 5.0 prior to 5.0.5 and versions 4.3 prior to 4.3.15 and older unsupported versions, allow applications to configure Spring MVC to serve static resources (e.g. CSS, JS, images). When static resources are served from a file system on Windows (as opposed to the classpath, or the ServletContext), a malicious user can send a request using a specially crafted URL that can lead a directory traversal attack. <p>Publish Date: 2018-04-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1271>CVE-2018-1271</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1271">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1271</a></p> <p>Release Date: 2018-04-06</p> <p>Fix Resolution: org.springframework:spring-webflux:5.0.5.RELEASE,org.springframework:spring-webmvc:4.3.15.RELEASE,5.0.5.RELEASE</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-webmvc","packageVersion":"4.2.5.RELEASE","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.capitalone.dashboard:core:3.9.7;org.springframework:spring-webmvc:4.2.5.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework:spring-webflux:5.0.5.RELEASE,org.springframework:spring-webmvc:4.3.15.RELEASE,5.0.5.RELEASE"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-1271","vulnerabilityDetails":"Spring Framework, versions 5.0 prior to 5.0.5 and versions 4.3 prior to 4.3.15 and older unsupported versions, allow applications to configure Spring MVC to serve static resources (e.g. CSS, JS, images). When static resources are served from a file system on Windows (as opposed to the classpath, or the ServletContext), a malicious user can send a request using a specially crafted URL that can lead a directory traversal attack.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1271","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-1271 (Medium) detected in spring-webmvc-4.2.5.RELEASE.jar - ## CVE-2018-1271 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-webmvc-4.2.5.RELEASE.jar</b></p></summary> <p>Spring Web MVC</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: hygieia-workflow-github-collector/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-webmvc/4.2.5.RELEASE/spring-webmvc-4.2.5.RELEASE.jar</p> <p> Dependency Hierarchy: - core-3.9.7.jar (Root Library) - :x: **spring-webmvc-4.2.5.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Dima2022/hygieia-workflow-github-collector/commit/236baaa856b74774f7b43ecb1eeade5a8d1d0496">236baaa856b74774f7b43ecb1eeade5a8d1d0496</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Spring Framework, versions 5.0 prior to 5.0.5 and versions 4.3 prior to 4.3.15 and older unsupported versions, allow applications to configure Spring MVC to serve static resources (e.g. CSS, JS, images). When static resources are served from a file system on Windows (as opposed to the classpath, or the ServletContext), a malicious user can send a request using a specially crafted URL that can lead a directory traversal attack. <p>Publish Date: 2018-04-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1271>CVE-2018-1271</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1271">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1271</a></p> <p>Release Date: 2018-04-06</p> <p>Fix Resolution: org.springframework:spring-webflux:5.0.5.RELEASE,org.springframework:spring-webmvc:4.3.15.RELEASE,5.0.5.RELEASE</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-webmvc","packageVersion":"4.2.5.RELEASE","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.capitalone.dashboard:core:3.9.7;org.springframework:spring-webmvc:4.2.5.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework:spring-webflux:5.0.5.RELEASE,org.springframework:spring-webmvc:4.3.15.RELEASE,5.0.5.RELEASE"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-1271","vulnerabilityDetails":"Spring Framework, versions 5.0 prior to 5.0.5 and versions 4.3 prior to 4.3.15 and older unsupported versions, allow applications to configure Spring MVC to serve static resources (e.g. CSS, JS, images). When static resources are served from a file system on Windows (as opposed to the classpath, or the ServletContext), a malicious user can send a request using a specially crafted URL that can lead a directory traversal attack.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1271","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in spring webmvc release jar cve medium severity vulnerability vulnerable library spring webmvc release jar spring web mvc library home page a href path to dependency file hygieia workflow github collector pom xml path to vulnerable library home wss scanner repository org springframework spring webmvc release spring webmvc release jar dependency hierarchy core jar root library x spring webmvc release jar vulnerable library found in head commit a href found in base branch main vulnerability details spring framework versions prior to and versions prior to and older unsupported versions allow applications to configure spring mvc to serve static resources e g css js images when static resources are served from a file system on windows as opposed to the classpath or the servletcontext a malicious user can send a request using a specially crafted url that can lead a directory traversal attack publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring webflux release org springframework spring webmvc release release isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree com capitalone dashboard core org springframework spring webmvc release isminimumfixversionavailable true minimumfixversion org springframework spring webflux release org springframework spring webmvc release release basebranches vulnerabilityidentifier cve vulnerabilitydetails spring framework versions prior to and versions prior to and older unsupported versions allow applications to configure spring mvc to serve static resources e g css js images when static resources are served from a file system on windows as opposed to the classpath or the servletcontext a malicious user can send a request using a specially crafted url that can lead a directory traversal attack vulnerabilityurl
0
1,796
4,538,040,264
IssuesEvent
2016-09-09 04:07:58
MModel/MetaModel
https://api.github.com/repos/MModel/MetaModel
closed
Use Metafile instead of a bunch of meta folder
[Difficulty] Easy [Status] In Process
`meta` folder seems too heavy under current circumstance. A better way is to use a Metafile instead of it which located in project root path `./Metafile`. Metafile may include all the information `MetaModel` needs to build a framework from sketch. ```ruby metamodel_version '0.1.1' define :User do end define :Person do end ``` It is much easier to handle all the thing in this one file rather than a bunch of ruby scripts in meta folder.
1.0
Use Metafile instead of a bunch of meta folder - `meta` folder seems too heavy under current circumstance. A better way is to use a Metafile instead of it which located in project root path `./Metafile`. Metafile may include all the information `MetaModel` needs to build a framework from sketch. ```ruby metamodel_version '0.1.1' define :User do end define :Person do end ``` It is much easier to handle all the thing in this one file rather than a bunch of ruby scripts in meta folder.
process
use metafile instead of a bunch of meta folder meta folder seems too heavy under current circumstance a better way is to use a metafile instead of it which located in project root path metafile metafile may include all the information metamodel needs to build a framework from sketch ruby metamodel version define user do end define person do end it is much easier to handle all the thing in this one file rather than a bunch of ruby scripts in meta folder
1
319,165
27,353,820,549
IssuesEvent
2023-02-27 11:25:52
oras-project/oras
https://api.github.com/repos/oras-project/oras
closed
add e2e specs for `oras discover`
E2E testing
Happy paths: - should discover direct referrers of a subject in remote registry - should discover via both tag schema and referrer API - should discover referrers recursively in tree view - should discover referrers and output annotations - should discover referrers with certain artifact type Bad paths: - should fail directly if not logged in - should fail if no subject tag or digest provided
1.0
add e2e specs for `oras discover` - Happy paths: - should discover direct referrers of a subject in remote registry - should discover via both tag schema and referrer API - should discover referrers recursively in tree view - should discover referrers and output annotations - should discover referrers with certain artifact type Bad paths: - should fail directly if not logged in - should fail if no subject tag or digest provided
non_process
add specs for oras discover happy paths should discover direct referrers of a subject in remote registry should discover via both tag schema and referrer api should discover referrers recursively in tree view should discover referrers and output annotations should discover referrers with certain artifact type bad paths should fail directly if not logged in should fail if no subject tag or digest provided
0
2,575
5,329,990,989
IssuesEvent
2017-02-15 16:05:29
SvePu/MyBB-PageManager
https://api.github.com/repos/SvePu/MyBB-PageManager
opened
Option to show page only specified groups
enhancement in process
Add an option to show the page only specified groups
1.0
Option to show page only specified groups - Add an option to show the page only specified groups
process
option to show page only specified groups add an option to show the page only specified groups
1
7,851
11,027,076,251
IssuesEvent
2019-12-06 08:36:33
wirecard/shop-systems-coding-guidelines
https://api.github.com/repos/wirecard/shop-systems-coding-guidelines
opened
How to do code reviews
processes
A tutorial on how to do them, how to navigate code, what to watch out for, where to draw a line, etc.
1.0
How to do code reviews - A tutorial on how to do them, how to navigate code, what to watch out for, where to draw a line, etc.
process
how to do code reviews a tutorial on how to do them how to navigate code what to watch out for where to draw a line etc
1
68,502
13,147,992,014
IssuesEvent
2020-08-08 18:45:33
NCAR/MET
https://api.github.com/repos/NCAR/MET
closed
Additional testing and refinement of the Grid-Diag tool.
component: application code priority: high requestor: NCAR type: enhancement
## Describe the Enhancement ## Bri Nelson and Michelle Harrold ran into issues running the Grid-Diag tool. I ran their commands and agree that some cleanup is needed. This task is to do that cleanup, add more error checking, and make the tool more user-friendly with better log messages. Bri originally reported this error message: ERROR : DataPlane::two_to_one() -> range check error: (Nx, Ny) = (0, 0), (x, y) = (0, 0) The tool was unable to read the data that was requested, didn't check whether the read was successfully, and tried to regrid an empty dataset to a new grid. That resulted in this error message. The tasks for this work include, but are not limited to: (1) Check that each attempt to read gridded data is successful. If not, print a warning message and continue. (2) Enhance the NetCDF output file to make it more self-describing, adding global and variable attributes as needed. (3) Make sure the output variable names do not include the use of stars and comma's, which happens when reading NetCDF input files. (4) Review existing log messages at all verbosity levels and tweak them to make them more user friendly. (5) Remove any unnecessary development code. (6) Review existing documentation and update as needed. (7) Review existing unit tests and update as needed. ### Time Estimate ### 3 days. ### Sub-Issues ### Consider breaking the enhancement down into sub-issues. No clear sub-issues yet. I hope to complete this is less than 1 week. ### Relevant Deadlines ### None ### Funding Source ### 7790971 UIUC S2S ## Define the Metadata ## ### Assignee ### - [X] Select **engineer(s)** or **no engineer** required: John Halley Gotway - [X] Select **scientist(s)** or **no scientist** required: Michelle Harrold ### Labels ### - [X] Select **component(s)** - [X] Select **priority** - [X] Select **requestor(s)** ### Projects and Milestone ### - [X] Review **projects** and select relevant **Repository** and **Organization** ones - [X] Select **milestone** ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [X] [METplus](https://github.com/NCAR/METplus/issues/new/choose), [MET](https://github.com/NCAR/MET/issues/new/choose), [METdb](https://github.com/NCAR/METdb/issues/new/choose), [METviewer](https://github.com/NCAR/METviewer/issues/new/choose), [METexpress](https://github.com/NCAR/METexpress/issues/new/choose), [METcalcpy](https://github.com/NCAR/METcalcpy/issues/new/choose), [METplotpy](https://github.com/NCAR/METplotpy/issues/new/choose) No impacts. ## Enhancement Checklist ## See the [METplus Workflow](https://ncar.github.io/METplus/Contributors_Guide/github_workflow.html) for details. - [x] Complete the issue definition above. - [x] Fork this repository or create a branch of **develop**. Branch name: `feature_<Issue Number>_<Description>` - [x] Complete the development and test your changes. - [x] Add/update unit tests. - [x] Add/update documentation. - [x] Push local changes to GitHub. - [x] Submit a pull request to merge into **develop**. Pull request: `feature <Issue Number> <Description>` - [x] Iterate until the reviewer(s) accept and merge your changes. - [x] Delete your fork or branch. - [x] Close this issue.
1.0
Additional testing and refinement of the Grid-Diag tool. - ## Describe the Enhancement ## Bri Nelson and Michelle Harrold ran into issues running the Grid-Diag tool. I ran their commands and agree that some cleanup is needed. This task is to do that cleanup, add more error checking, and make the tool more user-friendly with better log messages. Bri originally reported this error message: ERROR : DataPlane::two_to_one() -> range check error: (Nx, Ny) = (0, 0), (x, y) = (0, 0) The tool was unable to read the data that was requested, didn't check whether the read was successfully, and tried to regrid an empty dataset to a new grid. That resulted in this error message. The tasks for this work include, but are not limited to: (1) Check that each attempt to read gridded data is successful. If not, print a warning message and continue. (2) Enhance the NetCDF output file to make it more self-describing, adding global and variable attributes as needed. (3) Make sure the output variable names do not include the use of stars and comma's, which happens when reading NetCDF input files. (4) Review existing log messages at all verbosity levels and tweak them to make them more user friendly. (5) Remove any unnecessary development code. (6) Review existing documentation and update as needed. (7) Review existing unit tests and update as needed. ### Time Estimate ### 3 days. ### Sub-Issues ### Consider breaking the enhancement down into sub-issues. No clear sub-issues yet. I hope to complete this is less than 1 week. ### Relevant Deadlines ### None ### Funding Source ### 7790971 UIUC S2S ## Define the Metadata ## ### Assignee ### - [X] Select **engineer(s)** or **no engineer** required: John Halley Gotway - [X] Select **scientist(s)** or **no scientist** required: Michelle Harrold ### Labels ### - [X] Select **component(s)** - [X] Select **priority** - [X] Select **requestor(s)** ### Projects and Milestone ### - [X] Review **projects** and select relevant **Repository** and **Organization** ones - [X] Select **milestone** ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [X] [METplus](https://github.com/NCAR/METplus/issues/new/choose), [MET](https://github.com/NCAR/MET/issues/new/choose), [METdb](https://github.com/NCAR/METdb/issues/new/choose), [METviewer](https://github.com/NCAR/METviewer/issues/new/choose), [METexpress](https://github.com/NCAR/METexpress/issues/new/choose), [METcalcpy](https://github.com/NCAR/METcalcpy/issues/new/choose), [METplotpy](https://github.com/NCAR/METplotpy/issues/new/choose) No impacts. ## Enhancement Checklist ## See the [METplus Workflow](https://ncar.github.io/METplus/Contributors_Guide/github_workflow.html) for details. - [x] Complete the issue definition above. - [x] Fork this repository or create a branch of **develop**. Branch name: `feature_<Issue Number>_<Description>` - [x] Complete the development and test your changes. - [x] Add/update unit tests. - [x] Add/update documentation. - [x] Push local changes to GitHub. - [x] Submit a pull request to merge into **develop**. Pull request: `feature <Issue Number> <Description>` - [x] Iterate until the reviewer(s) accept and merge your changes. - [x] Delete your fork or branch. - [x] Close this issue.
non_process
additional testing and refinement of the grid diag tool describe the enhancement bri nelson and michelle harrold ran into issues running the grid diag tool i ran their commands and agree that some cleanup is needed this task is to do that cleanup add more error checking and make the tool more user friendly with better log messages bri originally reported this error message error dataplane two to one range check error nx ny x y the tool was unable to read the data that was requested didn t check whether the read was successfully and tried to regrid an empty dataset to a new grid that resulted in this error message the tasks for this work include but are not limited to check that each attempt to read gridded data is successful if not print a warning message and continue enhance the netcdf output file to make it more self describing adding global and variable attributes as needed make sure the output variable names do not include the use of stars and comma s which happens when reading netcdf input files review existing log messages at all verbosity levels and tweak them to make them more user friendly remove any unnecessary development code review existing documentation and update as needed review existing unit tests and update as needed time estimate days sub issues consider breaking the enhancement down into sub issues no clear sub issues yet i hope to complete this is less than week relevant deadlines none funding source uiuc define the metadata assignee select engineer s or no engineer required john halley gotway select scientist s or no scientist required michelle harrold labels select component s select priority select requestor s projects and milestone review projects and select relevant repository and organization ones select milestone define related issue s consider the impact to the other metplus components no impacts enhancement checklist see the for details complete the issue definition above fork this repository or create a branch of develop branch name feature complete the development and test your changes add update unit tests add update documentation push local changes to github submit a pull request to merge into develop pull request feature iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue
0
572,719
17,023,524,654
IssuesEvent
2021-07-03 02:28:13
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Render multiple names
Component: osmarender Priority: minor Resolution: wontfix Type: enhancement
**[Submitted to the original trac issue database at 1.34pm, Friday, 11th December 2009]** Multiple values for a key are (supposed to be) delimited by semicolons. Where a "name" key has multiple value, the results are ugly. Many people have (incorrectly) applied the approach of mapping for the renderer, resulting in names of the forms "a (b)", "c - d", "e / f" etc. This is particularly problematic in the case of "c - d", since it causes confusion with hyphenated names of the form "g-h". Where the name is "x;y", rather than rendering the name as "x;y", put in something like a newline or some other separator to make the correct form less ugly, and to discourage format proliferation. See also #2549 for mapnik
1.0
Render multiple names - **[Submitted to the original trac issue database at 1.34pm, Friday, 11th December 2009]** Multiple values for a key are (supposed to be) delimited by semicolons. Where a "name" key has multiple value, the results are ugly. Many people have (incorrectly) applied the approach of mapping for the renderer, resulting in names of the forms "a (b)", "c - d", "e / f" etc. This is particularly problematic in the case of "c - d", since it causes confusion with hyphenated names of the form "g-h". Where the name is "x;y", rather than rendering the name as "x;y", put in something like a newline or some other separator to make the correct form less ugly, and to discourage format proliferation. See also #2549 for mapnik
non_process
render multiple names multiple values for a key are supposed to be delimited by semicolons where a name key has multiple value the results are ugly many people have incorrectly applied the approach of mapping for the renderer resulting in names of the forms a b c d e f etc this is particularly problematic in the case of c d since it causes confusion with hyphenated names of the form g h where the name is x y rather than rendering the name as x y put in something like a newline or some other separator to make the correct form less ugly and to discourage format proliferation see also for mapnik
0
42,159
2,869,101,034
IssuesEvent
2015-06-05 23:20:40
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
pkg/yaml: format-preserving editing of parsed yaml
Area-Pkg Pkg-Yaml Priority-Low Triaged Type-Enhancement
*This issue was originally filed by @seaneagan* _____ This would be useful for issue #18911 and issue #20997.
1.0
pkg/yaml: format-preserving editing of parsed yaml - *This issue was originally filed by @seaneagan* _____ This would be useful for issue #18911 and issue #20997.
non_process
pkg yaml format preserving editing of parsed yaml this issue was originally filed by seaneagan this would be useful for issue and issue
0
177,837
13,749,825,764
IssuesEvent
2020-10-06 11:05:55
enonic/app-contentstudio
https://api.github.com/repos/enonic/app-contentstudio
closed
Add ui test to verify issue#2089
Test
Project selector button should not be clickable if current user has access to only one project #2089
1.0
Add ui test to verify issue#2089 - Project selector button should not be clickable if current user has access to only one project #2089
non_process
add ui test to verify issue project selector button should not be clickable if current user has access to only one project
0
17,089
9,602,254,012
IssuesEvent
2019-05-10 14:11:15
nim-lang/Nim
https://api.github.com/repos/nim-lang/Nim
closed
Using newSeqWith causes out of memory while newSeq doesn't
Performance Won't fix
Same sequence length, but newSeqWith fails with OOM. ``` nim import sequtils const N = 500000000 # var a = newSeqWith(N, 1) # out of memory var b = newSeq[int](N) for i in 1..<N: b[i] = 1 ``` I use `ulimit -S -v 6291456` for all applications and this is likely causes OOM, but pure newSeq works and that's wierd.
True
Using newSeqWith causes out of memory while newSeq doesn't - Same sequence length, but newSeqWith fails with OOM. ``` nim import sequtils const N = 500000000 # var a = newSeqWith(N, 1) # out of memory var b = newSeq[int](N) for i in 1..<N: b[i] = 1 ``` I use `ulimit -S -v 6291456` for all applications and this is likely causes OOM, but pure newSeq works and that's wierd.
non_process
using newseqwith causes out of memory while newseq doesn t same sequence length but newseqwith fails with oom nim import sequtils const n var a newseqwith n out of memory var b newseq n for i in n b i use ulimit s v for all applications and this is likely causes oom but pure newseq works and that s wierd
0
246,591
26,611,853,112
IssuesEvent
2023-01-24 01:15:51
turkdevops/training-kit
https://api.github.com/repos/turkdevops/training-kit
opened
CVE-2023-22796 (High) detected in activesupport-6.0.6.gem
security vulnerability
## CVE-2023-22796 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>activesupport-6.0.6.gem</b></p></summary> <p>A toolkit of support libraries and Ruby core extensions extracted from the Rails framework. Rich support for multibyte strings, internationalization, time zones, and testing.</p> <p>Library home page: <a href="https://rubygems.org/gems/activesupport-6.0.6.gem">https://rubygems.org/gems/activesupport-6.0.6.gem</a></p> <p>Path to dependency file: /Gemfile.lock</p> <p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/activesupport-6.0.6.gem</p> <p> Dependency Hierarchy: - github-pages-227.gem (Root Library) - jemoji-0.12.0.gem - html-pipeline-2.14.3.gem - :x: **activesupport-6.0.6.gem** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> There is a possible regular expression based DoS vulnerability in Active Support. A specially crafted string passed to the underscore method can cause the regular expression engine to enter a state of catastrophic backtracking. This can cause the process to use large amounts of CPU and memory, leading to a possible DoS vulnerability. The issue is patched in versions 6.1.7.1 and 7.0.4.1. <p>Publish Date: 2023-01-06 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-22796>CVE-2023-22796</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-j6gc-792m-qgm2">https://github.com/advisories/GHSA-j6gc-792m-qgm2</a></p> <p>Release Date: 2023-01-06</p> <p>Fix Resolution: activesupport - 6.1.7.1,7.0.4.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-22796 (High) detected in activesupport-6.0.6.gem - ## CVE-2023-22796 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>activesupport-6.0.6.gem</b></p></summary> <p>A toolkit of support libraries and Ruby core extensions extracted from the Rails framework. Rich support for multibyte strings, internationalization, time zones, and testing.</p> <p>Library home page: <a href="https://rubygems.org/gems/activesupport-6.0.6.gem">https://rubygems.org/gems/activesupport-6.0.6.gem</a></p> <p>Path to dependency file: /Gemfile.lock</p> <p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/activesupport-6.0.6.gem</p> <p> Dependency Hierarchy: - github-pages-227.gem (Root Library) - jemoji-0.12.0.gem - html-pipeline-2.14.3.gem - :x: **activesupport-6.0.6.gem** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> There is a possible regular expression based DoS vulnerability in Active Support. A specially crafted string passed to the underscore method can cause the regular expression engine to enter a state of catastrophic backtracking. This can cause the process to use large amounts of CPU and memory, leading to a possible DoS vulnerability. The issue is patched in versions 6.1.7.1 and 7.0.4.1. <p>Publish Date: 2023-01-06 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-22796>CVE-2023-22796</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-j6gc-792m-qgm2">https://github.com/advisories/GHSA-j6gc-792m-qgm2</a></p> <p>Release Date: 2023-01-06</p> <p>Fix Resolution: activesupport - 6.1.7.1,7.0.4.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in activesupport gem cve high severity vulnerability vulnerable library activesupport gem a toolkit of support libraries and ruby core extensions extracted from the rails framework rich support for multibyte strings internationalization time zones and testing library home page a href path to dependency file gemfile lock path to vulnerable library home wss scanner gem ruby cache activesupport gem dependency hierarchy github pages gem root library jemoji gem html pipeline gem x activesupport gem vulnerable library found in base branch master vulnerability details there is a possible regular expression based dos vulnerability in active support a specially crafted string passed to the underscore method can cause the regular expression engine to enter a state of catastrophic backtracking this can cause the process to use large amounts of cpu and memory leading to a possible dos vulnerability the issue is patched in versions and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution activesupport step up your open source security game with mend
0
736,509
25,476,821,687
IssuesEvent
2022-11-25 15:15:07
lbryio/lbry-docker
https://api.github.com/repos/lbryio/lbry-docker
closed
preliminary container for Reflector
type: improvement help wanted type: new feature priority: low on hold needs: exploration Epic
Reflector.go is an internally utilized appliance at lbryio, however there's no reason why it can't be publicly usable so I'm aiming to make it accessible via Docker. The following is a list of objects in the way. - [ ] Dockerfile - [ ] Healthcheck - [ ] Start.sh - [ ] docker-compose.yml - [ ] docker-entrypoint.sh [If needed] - [ ] hub.docker.com/r/lbryio/
1.0
preliminary container for Reflector - Reflector.go is an internally utilized appliance at lbryio, however there's no reason why it can't be publicly usable so I'm aiming to make it accessible via Docker. The following is a list of objects in the way. - [ ] Dockerfile - [ ] Healthcheck - [ ] Start.sh - [ ] docker-compose.yml - [ ] docker-entrypoint.sh [If needed] - [ ] hub.docker.com/r/lbryio/
non_process
preliminary container for reflector reflector go is an internally utilized appliance at lbryio however there s no reason why it can t be publicly usable so i m aiming to make it accessible via docker the following is a list of objects in the way dockerfile healthcheck start sh docker compose yml docker entrypoint sh hub docker com r lbryio
0
6,992
10,142,813,939
IssuesEvent
2019-08-04 05:40:40
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Gdal2xyz do not allow choose the band in batch mode
Bug Processing
- [x] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [x] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue **Describe the bug** The tool gdal2xyz give an error when we try to do the batch mode. **How to Reproduce QGIS 3.4.9** ![gdal2xyzQGIS349](https://user-images.githubusercontent.com/4480601/62367854-b283ac00-b522-11e9-9749-dfb9a10c4bb1.gif) 1. Open QGIS and load some raster layers; 2. Open GDAl2xyz in Processing and toggle batch mode. 3. Define outputs and click Run. 4. See error error QGIS 3.4.9. **Error in QGIS 3.4.9** ```python 2019-08-02T12:16:10 WARNING Traceback (most recent call last): File "C:/OSGEO4~1/apps/qgis-ltr/./python/plugins\processing\gui\BatchPanel.py", line 224, in save widget = wrapper.wrappedWidget() AttributeError: 'NoneType' object has no attribute 'wrappedWidget' 2019-08-02T12:16:13 WARNING Traceback (most recent call last): File "C:/OSGEO4~1/apps/qgis-ltr/./python/plugins\processing\gui\BatchAlgorithmDialog.py", line 88, in runAlgorithm parameters[param.name()] = wrapper.parameterValue() AttributeError: 'NoneType' object has no attribute 'parameterValue' 2019-08-02T12:16:13 WARNING Traceback (most recent call last): File "C:/OSGEO4~1/apps/qgis-ltr/./python/plugins\processing\gui\BatchAlgorithmDialog.py", line 88, in runAlgorithm parameters[param.name()] = wrapper.parameterValue() AttributeError: 'NoneType' object has no attribute 'parameterValue' ``` **How to Reproduce in QGIS 3.8.0** ![gdal2xyzQGIS380](https://user-images.githubusercontent.com/4480601/62367832-a26bcc80-b522-11e9-8124-5be1560ad6d9.gif) 1. Open QGIS and load some raster layers; 2. Open GDAl2xyz in Processing and toggle batch mode. 3. See error error QGIS 3.8.0. **Error in QGIS 3.8.0** ```python AttributeError: 'BatchInputSelectionPanel' object has no attribute 'value' Traceback (most recent call last): File "C:/OSGEO4~1/apps/qgis/./python/plugins\processing\gui\wrappers.py", line 209, in widgetValue return self.value() File "C:/OSGEO4~1/apps/qgis/./python/plugins\processing\gui\wrappers.py", line 1012, in value return self.widget.value() AttributeError: 'BatchInputSelectionPanel' object has no attribute 'value' ``` **QGIS and OS versions** Tested on 3.4.9 and 3.8.0 64-bits QGIS version | 3.4.9-Madeira | QGIS code revision | 54585165bb -- | -- | -- | -- Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2 Compiled against GDAL/OGR | 2.4.1 | Running against GDAL/OGR | 2.4.1 Compiled against GEOS | 3.7.2-CAPI-1.11.0 | Running against GEOS | 3.7.2-CAPI-1.11.0 b55d2125 PostgreSQL Client Version | 9.2.4 | SpatiaLite Version | 4.3.0 QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8 Compiled against PROJ | 520 | Running against PROJ | 5.2.0 ----------------------------------------------------------------------- QGIS version | 3.8.0-Zanzibar | QGIS code revision | 11aff65f10 -- | -- | -- | -- Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2 Compiled against GDAL/OGR | 2.4.1 | Running against GDAL/OGR | 2.4.1 Compiled against GEOS | 3.7.2-CAPI-1.11.0 | Running against GEOS | 3.7.2-CAPI-1.11.0 b55d2125 PostgreSQL Client Version | 9.2.4 | SpatiaLite Version | 4.3.0 QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8 Compiled against PROJ | 5.2.0 | Running against PROJ | Rel. 5.2.0, September 15th, 2018 OS Version | Windows 7 SP 1 (6.1) |   |   <!-- In the QGIS menu help/about, click in the dialog, Ctrl+A and then Ctrl+C. Finally paste here --> **Additional context** I will try to test in QGIS 3.4.10 if this have some kind of error. Please let me know if I'm doing some kind of mistake. Regards
1.0
Gdal2xyz do not allow choose the band in batch mode - - [x] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [x] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue **Describe the bug** The tool gdal2xyz give an error when we try to do the batch mode. **How to Reproduce QGIS 3.4.9** ![gdal2xyzQGIS349](https://user-images.githubusercontent.com/4480601/62367854-b283ac00-b522-11e9-9749-dfb9a10c4bb1.gif) 1. Open QGIS and load some raster layers; 2. Open GDAl2xyz in Processing and toggle batch mode. 3. Define outputs and click Run. 4. See error error QGIS 3.4.9. **Error in QGIS 3.4.9** ```python 2019-08-02T12:16:10 WARNING Traceback (most recent call last): File "C:/OSGEO4~1/apps/qgis-ltr/./python/plugins\processing\gui\BatchPanel.py", line 224, in save widget = wrapper.wrappedWidget() AttributeError: 'NoneType' object has no attribute 'wrappedWidget' 2019-08-02T12:16:13 WARNING Traceback (most recent call last): File "C:/OSGEO4~1/apps/qgis-ltr/./python/plugins\processing\gui\BatchAlgorithmDialog.py", line 88, in runAlgorithm parameters[param.name()] = wrapper.parameterValue() AttributeError: 'NoneType' object has no attribute 'parameterValue' 2019-08-02T12:16:13 WARNING Traceback (most recent call last): File "C:/OSGEO4~1/apps/qgis-ltr/./python/plugins\processing\gui\BatchAlgorithmDialog.py", line 88, in runAlgorithm parameters[param.name()] = wrapper.parameterValue() AttributeError: 'NoneType' object has no attribute 'parameterValue' ``` **How to Reproduce in QGIS 3.8.0** ![gdal2xyzQGIS380](https://user-images.githubusercontent.com/4480601/62367832-a26bcc80-b522-11e9-8124-5be1560ad6d9.gif) 1. Open QGIS and load some raster layers; 2. Open GDAl2xyz in Processing and toggle batch mode. 3. See error error QGIS 3.8.0. **Error in QGIS 3.8.0** ```python AttributeError: 'BatchInputSelectionPanel' object has no attribute 'value' Traceback (most recent call last): File "C:/OSGEO4~1/apps/qgis/./python/plugins\processing\gui\wrappers.py", line 209, in widgetValue return self.value() File "C:/OSGEO4~1/apps/qgis/./python/plugins\processing\gui\wrappers.py", line 1012, in value return self.widget.value() AttributeError: 'BatchInputSelectionPanel' object has no attribute 'value' ``` **QGIS and OS versions** Tested on 3.4.9 and 3.8.0 64-bits QGIS version | 3.4.9-Madeira | QGIS code revision | 54585165bb -- | -- | -- | -- Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2 Compiled against GDAL/OGR | 2.4.1 | Running against GDAL/OGR | 2.4.1 Compiled against GEOS | 3.7.2-CAPI-1.11.0 | Running against GEOS | 3.7.2-CAPI-1.11.0 b55d2125 PostgreSQL Client Version | 9.2.4 | SpatiaLite Version | 4.3.0 QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8 Compiled against PROJ | 520 | Running against PROJ | 5.2.0 ----------------------------------------------------------------------- QGIS version | 3.8.0-Zanzibar | QGIS code revision | 11aff65f10 -- | -- | -- | -- Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2 Compiled against GDAL/OGR | 2.4.1 | Running against GDAL/OGR | 2.4.1 Compiled against GEOS | 3.7.2-CAPI-1.11.0 | Running against GEOS | 3.7.2-CAPI-1.11.0 b55d2125 PostgreSQL Client Version | 9.2.4 | SpatiaLite Version | 4.3.0 QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8 Compiled against PROJ | 5.2.0 | Running against PROJ | Rel. 5.2.0, September 15th, 2018 OS Version | Windows 7 SP 1 (6.1) |   |   <!-- In the QGIS menu help/about, click in the dialog, Ctrl+A and then Ctrl+C. Finally paste here --> **Additional context** I will try to test in QGIS 3.4.10 if this have some kind of error. Please let me know if I'm doing some kind of mistake. Regards
process
do not allow choose the band in batch mode search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue describe the bug the tool give an error when we try to do the batch mode how to reproduce qgis open qgis and load some raster layers open in processing and toggle batch mode define outputs and click run see error error qgis error in qgis python warning traceback most recent call last file c apps qgis ltr python plugins processing gui batchpanel py line in save widget wrapper wrappedwidget attributeerror nonetype object has no attribute wrappedwidget warning traceback most recent call last file c apps qgis ltr python plugins processing gui batchalgorithmdialog py line in runalgorithm parameters wrapper parametervalue attributeerror nonetype object has no attribute parametervalue warning traceback most recent call last file c apps qgis ltr python plugins processing gui batchalgorithmdialog py line in runalgorithm parameters wrapper parametervalue attributeerror nonetype object has no attribute parametervalue how to reproduce in qgis open qgis and load some raster layers open in processing and toggle batch mode see error error qgis error in qgis python attributeerror batchinputselectionpanel object has no attribute value traceback most recent call last file c apps qgis python plugins processing gui wrappers py line in widgetvalue return self value file c apps qgis python plugins processing gui wrappers py line in value return self widget value attributeerror batchinputselectionpanel object has no attribute value qgis and os versions tested on and bits qgis version madeira qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi postgresql client version spatialite version qwt version version compiled against proj running against proj qgis version zanzibar qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi postgresql client version spatialite version qwt version version compiled against proj running against proj rel september os version windows sp     additional context i will try to test in qgis if this have some kind of error please let me know if i m doing some kind of mistake regards
1
185,759
14,379,352,507
IssuesEvent
2020-12-02 00:10:05
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Test: proposed API for status item background color
testplan-item
Refs: https://github.com/microsoft/vscode/issues/110214 - [x] anyOS @sbatten - [x] anyOS @dbaeumer Complexity: 3 [Create Issue](https://github.com/microsoft/vscode/issues/new?body=Testing+%23111423%0A%0A) --- [Release Notes](https://github.com/microsoft/vscode-docs/blob/vnext/release-notes/v1_52.md#statusbar-entry-background-color-api) Write a simple extension with proposed API enabled and create a status bar entry with `backgroundColor` applied. Refer to the JSDoc comment to see which color is supported (currently only error). If you are using an invalid color, no background color should be applied. **Note:** Use the [vscode-dts](https://www.npmjs.com/package/vscode-dts)-utility to use the latest version of `vscode.d.ts` and/or `vscode.proposed.d.ts`.
1.0
Test: proposed API for status item background color - Refs: https://github.com/microsoft/vscode/issues/110214 - [x] anyOS @sbatten - [x] anyOS @dbaeumer Complexity: 3 [Create Issue](https://github.com/microsoft/vscode/issues/new?body=Testing+%23111423%0A%0A) --- [Release Notes](https://github.com/microsoft/vscode-docs/blob/vnext/release-notes/v1_52.md#statusbar-entry-background-color-api) Write a simple extension with proposed API enabled and create a status bar entry with `backgroundColor` applied. Refer to the JSDoc comment to see which color is supported (currently only error). If you are using an invalid color, no background color should be applied. **Note:** Use the [vscode-dts](https://www.npmjs.com/package/vscode-dts)-utility to use the latest version of `vscode.d.ts` and/or `vscode.proposed.d.ts`.
non_process
test proposed api for status item background color refs anyos sbatten anyos dbaeumer complexity write a simple extension with proposed api enabled and create a status bar entry with backgroundcolor applied refer to the jsdoc comment to see which color is supported currently only error if you are using an invalid color no background color should be applied note use the to use the latest version of vscode d ts and or vscode proposed d ts
0
8,028
11,209,211,775
IssuesEvent
2020-01-06 09:57:33
inasafe/inasafe-realtime
https://api.github.com/repos/inasafe/inasafe-realtime
closed
Notes from PVMBG meetings
feature request ready realtime processor volcanic ash web page
Ash Realtime: - Change disclaimer - add notes above sea level on the form field - add fields duration forecasts - add in the reports a notes: what the time is and the duration - add KRB/evacuation route info to the map - auto conversion from netCDF to GeoTiff? - swap district and province name in the autocomplete volcano name - perhaps eruption height can be taken from netCDF - thresholds in ash might be the same for landcover and population, but the action checklist might be different
1.0
Notes from PVMBG meetings - Ash Realtime: - Change disclaimer - add notes above sea level on the form field - add fields duration forecasts - add in the reports a notes: what the time is and the duration - add KRB/evacuation route info to the map - auto conversion from netCDF to GeoTiff? - swap district and province name in the autocomplete volcano name - perhaps eruption height can be taken from netCDF - thresholds in ash might be the same for landcover and population, but the action checklist might be different
process
notes from pvmbg meetings ash realtime change disclaimer add notes above sea level on the form field add fields duration forecasts add in the reports a notes what the time is and the duration add krb evacuation route info to the map auto conversion from netcdf to geotiff swap district and province name in the autocomplete volcano name perhaps eruption height can be taken from netcdf thresholds in ash might be the same for landcover and population but the action checklist might be different
1
22,515
31,564,820,558
IssuesEvent
2023-09-03 17:39:14
python/cpython
https://api.github.com/repos/python/cpython
closed
windows: pythonw cannot spawn multiprocessing.Process
type-bug OS-windows pending topic-multiprocessing
# Bug report ### Checklist - [X] I am confident this is a bug in CPython, not a bug in a third-party project - [X] I have searched the [CPython issue tracker](https://github.com/python/cpython/issues?q=is%3Aissue+sort%3Acreated-desc), and am confident this bug has not been reported before ### CPython versions tested on: 3.11 ### Operating systems tested on: Windows ### Output from running 'python -VV' on the command line: Python 3.11.4 | packaged by Anaconda, Inc. | (main, Jul 5 2023, 13:38:37) [MSC v.1916 64 bit (AMD64 )] ### A clear and concise description of the bug: The following [script](https://docs.python.org/3/library/multiprocessing.html) runs as expected under `python.exe` but **not** under `pythonw.exe`: ```python from multiprocessing import Process import os def info(title): with open(f"{__name__}-{os.getppid()}-{os.getpid()}","a") as fd: fd.write(f"{title}\n") class P(Process): def __init__(self, x): self.x = x super().__init__(name=str(os.getpid())) info(f"init {self}") self.start() def run(self): info(f"running {self}") if __name__ == '__main__': info('main line') pl = [P(x) for x in range(2)] info(f"started {pl}") for p in pl: p.join() info(f"joined {p}") ``` # `python.exe` 2 sub-processes are started and 3 files are created, exactly as expected/documented: ### `==> __main__-36968-52232 <==` ``` main line init <P name='52232' parent=52232 initial> init <P name='52232' parent=52232 initial> started [<P name='52232' pid=21116 parent=52232 started>, <P name='52232' pid=43464 parent=52232 started>] joined <P name='52232' pid=21116 parent=52232 stopped exitcode=0> joined <P name='52232' pid=43464 parent=52232 stopped exitcode=0> ``` ### `==> __mp_main__-52232-21116 <==` ``` running <P name='52232' parent=52232 started> ``` ### `==> __mp_main__-52232-43464 <==` ``` running <P name='52232' parent=52232 started> ``` # `pythonw.exe` No subprocesses are started and only 1 file is created: ### `__main__-20572-33044` ``` main line init <P name='33044' parent=33044 initial> init <P name='33044' parent=33044 initial> started [<P name='33044' pid=35276 parent=33044 started>, <P name='33044' pid=41872 parent=33044 started>] joined <P name='33044' pid=35276 parent=33044 stopped exitcode=1> joined <P name='33044' pid=41872 parent=33044 stopped exitcode=1> ```
1.0
windows: pythonw cannot spawn multiprocessing.Process - # Bug report ### Checklist - [X] I am confident this is a bug in CPython, not a bug in a third-party project - [X] I have searched the [CPython issue tracker](https://github.com/python/cpython/issues?q=is%3Aissue+sort%3Acreated-desc), and am confident this bug has not been reported before ### CPython versions tested on: 3.11 ### Operating systems tested on: Windows ### Output from running 'python -VV' on the command line: Python 3.11.4 | packaged by Anaconda, Inc. | (main, Jul 5 2023, 13:38:37) [MSC v.1916 64 bit (AMD64 )] ### A clear and concise description of the bug: The following [script](https://docs.python.org/3/library/multiprocessing.html) runs as expected under `python.exe` but **not** under `pythonw.exe`: ```python from multiprocessing import Process import os def info(title): with open(f"{__name__}-{os.getppid()}-{os.getpid()}","a") as fd: fd.write(f"{title}\n") class P(Process): def __init__(self, x): self.x = x super().__init__(name=str(os.getpid())) info(f"init {self}") self.start() def run(self): info(f"running {self}") if __name__ == '__main__': info('main line') pl = [P(x) for x in range(2)] info(f"started {pl}") for p in pl: p.join() info(f"joined {p}") ``` # `python.exe` 2 sub-processes are started and 3 files are created, exactly as expected/documented: ### `==> __main__-36968-52232 <==` ``` main line init <P name='52232' parent=52232 initial> init <P name='52232' parent=52232 initial> started [<P name='52232' pid=21116 parent=52232 started>, <P name='52232' pid=43464 parent=52232 started>] joined <P name='52232' pid=21116 parent=52232 stopped exitcode=0> joined <P name='52232' pid=43464 parent=52232 stopped exitcode=0> ``` ### `==> __mp_main__-52232-21116 <==` ``` running <P name='52232' parent=52232 started> ``` ### `==> __mp_main__-52232-43464 <==` ``` running <P name='52232' parent=52232 started> ``` # `pythonw.exe` No subprocesses are started and only 1 file is created: ### `__main__-20572-33044` ``` main line init <P name='33044' parent=33044 initial> init <P name='33044' parent=33044 initial> started [<P name='33044' pid=35276 parent=33044 started>, <P name='33044' pid=41872 parent=33044 started>] joined <P name='33044' pid=35276 parent=33044 stopped exitcode=1> joined <P name='33044' pid=41872 parent=33044 stopped exitcode=1> ```
process
windows pythonw cannot spawn multiprocessing process bug report checklist i am confident this is a bug in cpython not a bug in a third party project i have searched the and am confident this bug has not been reported before cpython versions tested on operating systems tested on windows output from running python vv on the command line python packaged by anaconda inc main jul a clear and concise description of the bug the following runs as expected under python exe but not under pythonw exe python from multiprocessing import process import os def info title with open f name os getppid os getpid a as fd fd write f title n class p process def init self x self x x super init name str os getpid info f init self self start def run self info f running self if name main info main line pl info f started pl for p in pl p join info f joined p python exe sub processes are started and files are created exactly as expected documented main main line init init started joined joined mp main running mp main running pythonw exe no subprocesses are started and only file is created main main line init init started joined joined
1
7,335
10,471,218,307
IssuesEvent
2019-09-23 07:13:08
Snoopyjoy/blog
https://api.github.com/repos/Snoopyjoy/blog
opened
实现Node.js集群
/share/2019/03/21/nodejs-multi-process.html Gitalk
https://github.hxl2lgy.top/share/2019/03/21/nodejs-multi-process.html 单线程的Node.js        Node.js的代码是在单线程的环境执行的。单线程环境下编程更简单,没有线程安全问题和线程切换的消耗问题。但是也因为单线程,它的缺点很明显:1. 无法胜任CPU密集型的任务。2. 无法充分利用服务器的资源。        单线程的Node.js性能怎么样呢?Node.js在I...
1.0
实现Node.js集群 - https://github.hxl2lgy.top/share/2019/03/21/nodejs-multi-process.html 单线程的Node.js        Node.js的代码是在单线程的环境执行的。单线程环境下编程更简单,没有线程安全问题和线程切换的消耗问题。但是也因为单线程,它的缺点很明显:1. 无法胜任CPU密集型的任务。2. 无法充分利用服务器的资源。        单线程的Node.js性能怎么样呢?Node.js在I...
process
实现node js集群 单线程的node js        node js的代码是在单线程的环境执行的。单线程环境下编程更简单,没有线程安全问题和线程切换的消耗问题。但是也因为单线程,它的缺点很明显: 无法胜任cpu密集型的任务。 无法充分利用服务器的资源。        单线程的node js性能怎么样呢?node js在i
1
257,019
27,561,756,180
IssuesEvent
2023-03-07 22:44:23
samqws-marketing/walmartlabs-concord
https://api.github.com/repos/samqws-marketing/walmartlabs-concord
closed
CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz - autoclosed
security vulnerability
## CVE-2021-3918 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary> <p>JSON Schema validation and specifications</p> <p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p> <p>Path to dependency file: /console2/package.json</p> <p>Path to vulnerable library: /console2/node_modules/json-schema/package.json</p> <p> Dependency Hierarchy: - react-scripts-4.0.3.tgz (Root Library) - jest-26.6.0.tgz - jest-cli-26.6.3.tgz - jest-config-26.6.3.tgz - jest-environment-jsdom-26.6.2.tgz - jsdom-16.5.3.tgz - request-2.88.2.tgz - http-signature-1.2.0.tgz - jsprim-1.4.1.tgz - :x: **json-schema-0.2.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/walmartlabs-concord/commit/b9420f3b9e73a9d381266ece72f7afb756f35a76">b9420f3b9e73a9d381266ece72f7afb756f35a76</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution') <p>Publish Date: 2021-11-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3918>CVE-2021-3918</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-3918">https://nvd.nist.gov/vuln/detail/CVE-2021-3918</a></p> <p>Release Date: 2021-11-13</p> <p>Fix Resolution (json-schema): 0.4.0</p> <p>Direct dependency fix Resolution (react-scripts): 5.0.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
True
CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz - autoclosed - ## CVE-2021-3918 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary> <p>JSON Schema validation and specifications</p> <p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p> <p>Path to dependency file: /console2/package.json</p> <p>Path to vulnerable library: /console2/node_modules/json-schema/package.json</p> <p> Dependency Hierarchy: - react-scripts-4.0.3.tgz (Root Library) - jest-26.6.0.tgz - jest-cli-26.6.3.tgz - jest-config-26.6.3.tgz - jest-environment-jsdom-26.6.2.tgz - jsdom-16.5.3.tgz - request-2.88.2.tgz - http-signature-1.2.0.tgz - jsprim-1.4.1.tgz - :x: **json-schema-0.2.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/walmartlabs-concord/commit/b9420f3b9e73a9d381266ece72f7afb756f35a76">b9420f3b9e73a9d381266ece72f7afb756f35a76</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution') <p>Publish Date: 2021-11-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3918>CVE-2021-3918</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-3918">https://nvd.nist.gov/vuln/detail/CVE-2021-3918</a></p> <p>Release Date: 2021-11-13</p> <p>Fix Resolution (json-schema): 0.4.0</p> <p>Direct dependency fix Resolution (react-scripts): 5.0.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
non_process
cve high detected in json schema tgz autoclosed cve high severity vulnerability vulnerable library json schema tgz json schema validation and specifications library home page a href path to dependency file package json path to vulnerable library node modules json schema package json dependency hierarchy react scripts tgz root library jest tgz jest cli tgz jest config tgz jest environment jsdom tgz jsdom tgz request tgz http signature tgz jsprim tgz x json schema tgz vulnerable library found in head commit a href found in base branch master vulnerability details json schema is vulnerable to improperly controlled modification of object prototype attributes prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution json schema direct dependency fix resolution react scripts check this box to open an automated fix pr
0
16,998
22,364,070,896
IssuesEvent
2022-06-16 00:46:04
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
opened
Add an eth_call acceptance test
enhancement P1 process
### Problem No acceptance test exists for `eth_call` ### Solution Add an `eth_call` acceptance to cover - Happy path - non matching hash - ... ### Alternatives _No response_
1.0
Add an eth_call acceptance test - ### Problem No acceptance test exists for `eth_call` ### Solution Add an `eth_call` acceptance to cover - Happy path - non matching hash - ... ### Alternatives _No response_
process
add an eth call acceptance test problem no acceptance test exists for eth call solution add an eth call acceptance to cover happy path non matching hash alternatives no response
1
20,618
27,291,102,986
IssuesEvent
2023-02-23 16:39:58
gobuffalo/.github
https://api.github.com/repos/gobuffalo/.github
closed
task: update testing version for Go (to 1.18 and 1.19)
process
I would like to do this after finishing the next release of the core.
1.0
task: update testing version for Go (to 1.18 and 1.19) - I would like to do this after finishing the next release of the core.
process
task update testing version for go to and i would like to do this after finishing the next release of the core
1
21,906
30,386,838,942
IssuesEvent
2023-07-13 02:00:08
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Thu, 13 Jul 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB ### SC-NeuS: Consistent Neural Surface Reconstruction from Sparse and Noisy Views - **Authors:** Shi-Sheng Huang, Zi-Xin Zou, Yi-Chi Zhang, Hua Huang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.05892 - **Pdf link:** https://arxiv.org/pdf/2307.05892 - **Abstract** The recent neural surface reconstruction by volume rendering approaches have made much progress by achieving impressive surface reconstruction quality, but are still limited to dense and highly accurate posed views. To overcome such drawbacks, this paper pays special attention on the consistent surface reconstruction from sparse views with noisy camera poses. Unlike previous approaches, the key difference of this paper is to exploit the multi-view constraints directly from the explicit geometry of the neural surface, which can be used as effective regularization to jointly learn the neural surface and refine the camera poses. To build effective multi-view constraints, we introduce a fast differentiable on-surface intersection to generate on-surface points, and propose view-consistent losses based on such differentiable points to regularize the neural surface learning. Based on this point, we propose a jointly learning strategy for neural surface and camera poses, named SC-NeuS, to perform geometry-consistent surface reconstruction in an end-to-end manner. With extensive evaluation on public datasets, our SC-NeuS can achieve consistently better surface reconstruction results with fine-grained details than previous state-of-the-art neural surface reconstruction approaches, especially from sparse and noisy camera views. ## Keyword: ISP ### Stochastic Light Field Holography - **Authors:** Florian Schiffers, Praneeth Chakravarthula, Nathan Matsuda, Grace Kuo, Ethan Tseng, Douglas Lanman, Felix Heide, Oliver Cossairt - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR); Image and Video Processing (eess.IV); Optics (physics.optics) - **Arxiv link:** https://arxiv.org/abs/2307.06277 - **Pdf link:** https://arxiv.org/pdf/2307.06277 - **Abstract** The Visual Turing Test is the ultimate goal to evaluate the realism of holographic displays. Previous studies have focused on addressing challenges such as limited \'etendue and image quality over a large focal volume, but they have not investigated the effect of pupil sampling on the viewing experience in full 3D holograms. In this work, we tackle this problem with a novel hologram generation algorithm motivated by matching the projection operators of incoherent Light Field and coherent Wigner Function light transport. To this end, we supervise hologram computation using synthesized photographs, which are rendered on-the-fly using Light Field refocusing from stochastically sampled pupil states during optimization. The proposed method produces holograms with correct parallax and focus cues, which are important for passing the Visual Turing Test. We validate that our approach compares favorably to state-of-the-art CGH algorithms that use Light Field and Focal Stack supervision. Our experiments demonstrate that our algorithm significantly improves the realism of the viewing experience for a variety of different pupil states. ### MMBench: Is Your Multi-modal Model an All-around Player? - **Authors:** Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL) - **Arxiv link:** https://arxiv.org/abs/2307.06281 - **Pdf link:** https://arxiv.org/pdf/2307.06281 - **Abstract** Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### AICT: An Adaptive Image Compression Transformer - **Authors:** Ahmed Ghorbel, Wassim Hamidouche, Luce Morin - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2307.06091 - **Pdf link:** https://arxiv.org/pdf/2307.06091 - **Abstract** Motivated by the efficiency investigation of the Tranformer-based transform coding framework, namely SwinT-ChARM, we propose to enhance the latter, as first, with a more straightforward yet effective Tranformer-based channel-wise auto-regressive prior model, resulting in an absolute image compression transformer (ICT). Current methods that still rely on ConvNet-based entropy coding are limited in long-range modeling dependencies due to their local connectivity and an increasing number of architectural biases and priors. On the contrary, the proposed ICT can capture both global and local contexts from the latent representations and better parameterize the distribution of the quantized latents. Further, we leverage a learnable scaling module with a sandwich ConvNeXt-based pre/post-processor to accurately extract more compact latent representation while reconstructing higher-quality images. Extensive experimental results on benchmark datasets showed that the proposed adaptive image compression transformer (AICT) framework significantly improves the trade-off between coding efficiency and decoder complexity over the versatile video coding (VVC) reference encoder (VTM-18.0) and the neural codec SwinT-ChARM. ### Learning Kernel-Modulated Neural Representation for Efficient Light Field Compression - **Authors:** Jinglei Shi, Yihong Xu, Christine Guillemot - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2307.06143 - **Pdf link:** https://arxiv.org/pdf/2307.06143 - **Abstract** Light field is a type of image data that captures the 3D scene information by recording light rays emitted from a scene at various orientations. It offers a more immersive perception than classic 2D images but at the cost of huge data volume. In this paper, we draw inspiration from the visual characteristics of Sub-Aperture Images (SAIs) of light field and design a compact neural network representation for the light field compression task. The network backbone takes randomly initialized noise as input and is supervised on the SAIs of the target light field. It is composed of two types of complementary kernels: descriptive kernels (descriptors) that store scene description information learned during training, and modulatory kernels (modulators) that control the rendering of different SAIs from the queried perspectives. To further enhance compactness of the network meanwhile retain high quality of the decoded light field, we accordingly introduce modulator allocation and kernel tensor decomposition mechanisms, followed by non-uniform quantization and lossless entropy coding techniques, to finally form an efficient compression pipeline. Extensive experiments demonstrate that our method outperforms other state-of-the-art (SOTA) methods by a significant margin in the light field compression task. Moreover, after aligning descriptors, the modulators learned from one light field can be transferred to new light fields for rendering dense views, indicating a potential solution for view synthesis task. ## Keyword: RAW ### MoP-CLIP: A Mixture of Prompt-Tuned CLIP Models for Domain Incremental Learning - **Authors:** Julien Nicolas, Florent Chiaroni, Imtiaz Ziko, Ola Ahmad, Christian Desrosiers, Jose Dolz - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2307.05707 - **Pdf link:** https://arxiv.org/pdf/2307.05707 - **Abstract** Despite the recent progress in incremental learning, addressing catastrophic forgetting under distributional drift is still an open and important problem. Indeed, while state-of-the-art domain incremental learning (DIL) methods perform satisfactorily within known domains, their performance largely degrades in the presence of novel domains. This limitation hampers their generalizability, and restricts their scalability to more realistic settings where train and test data are drawn from different distributions. To address these limitations, we present a novel DIL approach based on a mixture of prompt-tuned CLIP models (MoP-CLIP), which generalizes the paradigm of S-Prompting to handle both in-distribution and out-of-distribution data at inference. In particular, at the training stage we model the features distribution of every class in each domain, learning individual text and visual prompts to adapt to a given domain. At inference, the learned distributions allow us to identify whether a given test sample belongs to a known domain, selecting the correct prompt for the classification task, or from an unseen domain, leveraging a mixture of the prompt-tuned CLIP models. Our empirical evaluation reveals the poor performance of existing DIL methods under domain shift, and suggests that the proposed MoP-CLIP performs competitively in the standard DIL settings while outperforming state-of-the-art methods in OOD scenarios. These results demonstrate the superiority of MoP-CLIP, offering a robust and general solution to the problem of domain incremental learning. ### SC-NeuS: Consistent Neural Surface Reconstruction from Sparse and Noisy Views - **Authors:** Shi-Sheng Huang, Zi-Xin Zou, Yi-Chi Zhang, Hua Huang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.05892 - **Pdf link:** https://arxiv.org/pdf/2307.05892 - **Abstract** The recent neural surface reconstruction by volume rendering approaches have made much progress by achieving impressive surface reconstruction quality, but are still limited to dense and highly accurate posed views. To overcome such drawbacks, this paper pays special attention on the consistent surface reconstruction from sparse views with noisy camera poses. Unlike previous approaches, the key difference of this paper is to exploit the multi-view constraints directly from the explicit geometry of the neural surface, which can be used as effective regularization to jointly learn the neural surface and refine the camera poses. To build effective multi-view constraints, we introduce a fast differentiable on-surface intersection to generate on-surface points, and propose view-consistent losses based on such differentiable points to regularize the neural surface learning. Based on this point, we propose a jointly learning strategy for neural surface and camera poses, named SC-NeuS, to perform geometry-consistent surface reconstruction in an end-to-end manner. With extensive evaluation on public datasets, our SC-NeuS can achieve consistently better surface reconstruction results with fine-grained details than previous state-of-the-art neural surface reconstruction approaches, especially from sparse and noisy camera views. ### Learning Kernel-Modulated Neural Representation for Efficient Light Field Compression - **Authors:** Jinglei Shi, Yihong Xu, Christine Guillemot - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2307.06143 - **Pdf link:** https://arxiv.org/pdf/2307.06143 - **Abstract** Light field is a type of image data that captures the 3D scene information by recording light rays emitted from a scene at various orientations. It offers a more immersive perception than classic 2D images but at the cost of huge data volume. In this paper, we draw inspiration from the visual characteristics of Sub-Aperture Images (SAIs) of light field and design a compact neural network representation for the light field compression task. The network backbone takes randomly initialized noise as input and is supervised on the SAIs of the target light field. It is composed of two types of complementary kernels: descriptive kernels (descriptors) that store scene description information learned during training, and modulatory kernels (modulators) that control the rendering of different SAIs from the queried perspectives. To further enhance compactness of the network meanwhile retain high quality of the decoded light field, we accordingly introduce modulator allocation and kernel tensor decomposition mechanisms, followed by non-uniform quantization and lossless entropy coding techniques, to finally form an efficient compression pipeline. Extensive experiments demonstrate that our method outperforms other state-of-the-art (SOTA) methods by a significant margin in the light field compression task. Moreover, after aligning descriptors, the modulators learned from one light field can be transferred to new light fields for rendering dense views, indicating a potential solution for view synthesis task. ## Keyword: raw image There is no result
2.0
New submissions for Thu, 13 Jul 23 - ## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB ### SC-NeuS: Consistent Neural Surface Reconstruction from Sparse and Noisy Views - **Authors:** Shi-Sheng Huang, Zi-Xin Zou, Yi-Chi Zhang, Hua Huang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.05892 - **Pdf link:** https://arxiv.org/pdf/2307.05892 - **Abstract** The recent neural surface reconstruction by volume rendering approaches have made much progress by achieving impressive surface reconstruction quality, but are still limited to dense and highly accurate posed views. To overcome such drawbacks, this paper pays special attention on the consistent surface reconstruction from sparse views with noisy camera poses. Unlike previous approaches, the key difference of this paper is to exploit the multi-view constraints directly from the explicit geometry of the neural surface, which can be used as effective regularization to jointly learn the neural surface and refine the camera poses. To build effective multi-view constraints, we introduce a fast differentiable on-surface intersection to generate on-surface points, and propose view-consistent losses based on such differentiable points to regularize the neural surface learning. Based on this point, we propose a jointly learning strategy for neural surface and camera poses, named SC-NeuS, to perform geometry-consistent surface reconstruction in an end-to-end manner. With extensive evaluation on public datasets, our SC-NeuS can achieve consistently better surface reconstruction results with fine-grained details than previous state-of-the-art neural surface reconstruction approaches, especially from sparse and noisy camera views. ## Keyword: ISP ### Stochastic Light Field Holography - **Authors:** Florian Schiffers, Praneeth Chakravarthula, Nathan Matsuda, Grace Kuo, Ethan Tseng, Douglas Lanman, Felix Heide, Oliver Cossairt - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR); Image and Video Processing (eess.IV); Optics (physics.optics) - **Arxiv link:** https://arxiv.org/abs/2307.06277 - **Pdf link:** https://arxiv.org/pdf/2307.06277 - **Abstract** The Visual Turing Test is the ultimate goal to evaluate the realism of holographic displays. Previous studies have focused on addressing challenges such as limited \'etendue and image quality over a large focal volume, but they have not investigated the effect of pupil sampling on the viewing experience in full 3D holograms. In this work, we tackle this problem with a novel hologram generation algorithm motivated by matching the projection operators of incoherent Light Field and coherent Wigner Function light transport. To this end, we supervise hologram computation using synthesized photographs, which are rendered on-the-fly using Light Field refocusing from stochastically sampled pupil states during optimization. The proposed method produces holograms with correct parallax and focus cues, which are important for passing the Visual Turing Test. We validate that our approach compares favorably to state-of-the-art CGH algorithms that use Light Field and Focal Stack supervision. Our experiments demonstrate that our algorithm significantly improves the realism of the viewing experience for a variety of different pupil states. ### MMBench: Is Your Multi-modal Model an All-around Player? - **Authors:** Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computation and Language (cs.CL) - **Arxiv link:** https://arxiv.org/abs/2307.06281 - **Pdf link:** https://arxiv.org/pdf/2307.06281 - **Abstract** Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### AICT: An Adaptive Image Compression Transformer - **Authors:** Ahmed Ghorbel, Wassim Hamidouche, Luce Morin - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2307.06091 - **Pdf link:** https://arxiv.org/pdf/2307.06091 - **Abstract** Motivated by the efficiency investigation of the Tranformer-based transform coding framework, namely SwinT-ChARM, we propose to enhance the latter, as first, with a more straightforward yet effective Tranformer-based channel-wise auto-regressive prior model, resulting in an absolute image compression transformer (ICT). Current methods that still rely on ConvNet-based entropy coding are limited in long-range modeling dependencies due to their local connectivity and an increasing number of architectural biases and priors. On the contrary, the proposed ICT can capture both global and local contexts from the latent representations and better parameterize the distribution of the quantized latents. Further, we leverage a learnable scaling module with a sandwich ConvNeXt-based pre/post-processor to accurately extract more compact latent representation while reconstructing higher-quality images. Extensive experimental results on benchmark datasets showed that the proposed adaptive image compression transformer (AICT) framework significantly improves the trade-off between coding efficiency and decoder complexity over the versatile video coding (VVC) reference encoder (VTM-18.0) and the neural codec SwinT-ChARM. ### Learning Kernel-Modulated Neural Representation for Efficient Light Field Compression - **Authors:** Jinglei Shi, Yihong Xu, Christine Guillemot - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2307.06143 - **Pdf link:** https://arxiv.org/pdf/2307.06143 - **Abstract** Light field is a type of image data that captures the 3D scene information by recording light rays emitted from a scene at various orientations. It offers a more immersive perception than classic 2D images but at the cost of huge data volume. In this paper, we draw inspiration from the visual characteristics of Sub-Aperture Images (SAIs) of light field and design a compact neural network representation for the light field compression task. The network backbone takes randomly initialized noise as input and is supervised on the SAIs of the target light field. It is composed of two types of complementary kernels: descriptive kernels (descriptors) that store scene description information learned during training, and modulatory kernels (modulators) that control the rendering of different SAIs from the queried perspectives. To further enhance compactness of the network meanwhile retain high quality of the decoded light field, we accordingly introduce modulator allocation and kernel tensor decomposition mechanisms, followed by non-uniform quantization and lossless entropy coding techniques, to finally form an efficient compression pipeline. Extensive experiments demonstrate that our method outperforms other state-of-the-art (SOTA) methods by a significant margin in the light field compression task. Moreover, after aligning descriptors, the modulators learned from one light field can be transferred to new light fields for rendering dense views, indicating a potential solution for view synthesis task. ## Keyword: RAW ### MoP-CLIP: A Mixture of Prompt-Tuned CLIP Models for Domain Incremental Learning - **Authors:** Julien Nicolas, Florent Chiaroni, Imtiaz Ziko, Ola Ahmad, Christian Desrosiers, Jose Dolz - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2307.05707 - **Pdf link:** https://arxiv.org/pdf/2307.05707 - **Abstract** Despite the recent progress in incremental learning, addressing catastrophic forgetting under distributional drift is still an open and important problem. Indeed, while state-of-the-art domain incremental learning (DIL) methods perform satisfactorily within known domains, their performance largely degrades in the presence of novel domains. This limitation hampers their generalizability, and restricts their scalability to more realistic settings where train and test data are drawn from different distributions. To address these limitations, we present a novel DIL approach based on a mixture of prompt-tuned CLIP models (MoP-CLIP), which generalizes the paradigm of S-Prompting to handle both in-distribution and out-of-distribution data at inference. In particular, at the training stage we model the features distribution of every class in each domain, learning individual text and visual prompts to adapt to a given domain. At inference, the learned distributions allow us to identify whether a given test sample belongs to a known domain, selecting the correct prompt for the classification task, or from an unseen domain, leveraging a mixture of the prompt-tuned CLIP models. Our empirical evaluation reveals the poor performance of existing DIL methods under domain shift, and suggests that the proposed MoP-CLIP performs competitively in the standard DIL settings while outperforming state-of-the-art methods in OOD scenarios. These results demonstrate the superiority of MoP-CLIP, offering a robust and general solution to the problem of domain incremental learning. ### SC-NeuS: Consistent Neural Surface Reconstruction from Sparse and Noisy Views - **Authors:** Shi-Sheng Huang, Zi-Xin Zou, Yi-Chi Zhang, Hua Huang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.05892 - **Pdf link:** https://arxiv.org/pdf/2307.05892 - **Abstract** The recent neural surface reconstruction by volume rendering approaches have made much progress by achieving impressive surface reconstruction quality, but are still limited to dense and highly accurate posed views. To overcome such drawbacks, this paper pays special attention on the consistent surface reconstruction from sparse views with noisy camera poses. Unlike previous approaches, the key difference of this paper is to exploit the multi-view constraints directly from the explicit geometry of the neural surface, which can be used as effective regularization to jointly learn the neural surface and refine the camera poses. To build effective multi-view constraints, we introduce a fast differentiable on-surface intersection to generate on-surface points, and propose view-consistent losses based on such differentiable points to regularize the neural surface learning. Based on this point, we propose a jointly learning strategy for neural surface and camera poses, named SC-NeuS, to perform geometry-consistent surface reconstruction in an end-to-end manner. With extensive evaluation on public datasets, our SC-NeuS can achieve consistently better surface reconstruction results with fine-grained details than previous state-of-the-art neural surface reconstruction approaches, especially from sparse and noisy camera views. ### Learning Kernel-Modulated Neural Representation for Efficient Light Field Compression - **Authors:** Jinglei Shi, Yihong Xu, Christine Guillemot - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2307.06143 - **Pdf link:** https://arxiv.org/pdf/2307.06143 - **Abstract** Light field is a type of image data that captures the 3D scene information by recording light rays emitted from a scene at various orientations. It offers a more immersive perception than classic 2D images but at the cost of huge data volume. In this paper, we draw inspiration from the visual characteristics of Sub-Aperture Images (SAIs) of light field and design a compact neural network representation for the light field compression task. The network backbone takes randomly initialized noise as input and is supervised on the SAIs of the target light field. It is composed of two types of complementary kernels: descriptive kernels (descriptors) that store scene description information learned during training, and modulatory kernels (modulators) that control the rendering of different SAIs from the queried perspectives. To further enhance compactness of the network meanwhile retain high quality of the decoded light field, we accordingly introduce modulator allocation and kernel tensor decomposition mechanisms, followed by non-uniform quantization and lossless entropy coding techniques, to finally form an efficient compression pipeline. Extensive experiments demonstrate that our method outperforms other state-of-the-art (SOTA) methods by a significant margin in the light field compression task. Moreover, after aligning descriptors, the modulators learned from one light field can be transferred to new light fields for rendering dense views, indicating a potential solution for view synthesis task. ## Keyword: raw image There is no result
process
new submissions for thu jul keyword events there is no result keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb sc neus consistent neural surface reconstruction from sparse and noisy views authors shi sheng huang zi xin zou yi chi zhang hua huang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract the recent neural surface reconstruction by volume rendering approaches have made much progress by achieving impressive surface reconstruction quality but are still limited to dense and highly accurate posed views to overcome such drawbacks this paper pays special attention on the consistent surface reconstruction from sparse views with noisy camera poses unlike previous approaches the key difference of this paper is to exploit the multi view constraints directly from the explicit geometry of the neural surface which can be used as effective regularization to jointly learn the neural surface and refine the camera poses to build effective multi view constraints we introduce a fast differentiable on surface intersection to generate on surface points and propose view consistent losses based on such differentiable points to regularize the neural surface learning based on this point we propose a jointly learning strategy for neural surface and camera poses named sc neus to perform geometry consistent surface reconstruction in an end to end manner with extensive evaluation on public datasets our sc neus can achieve consistently better surface reconstruction results with fine grained details than previous state of the art neural surface reconstruction approaches especially from sparse and noisy camera views keyword isp stochastic light field holography authors florian schiffers praneeth chakravarthula nathan matsuda grace kuo ethan tseng douglas lanman felix heide oliver cossairt subjects computer vision and pattern recognition cs cv graphics cs gr image and video processing eess iv optics physics optics arxiv link pdf link abstract the visual turing test is the ultimate goal to evaluate the realism of holographic displays previous studies have focused on addressing challenges such as limited etendue and image quality over a large focal volume but they have not investigated the effect of pupil sampling on the viewing experience in full holograms in this work we tackle this problem with a novel hologram generation algorithm motivated by matching the projection operators of incoherent light field and coherent wigner function light transport to this end we supervise hologram computation using synthesized photographs which are rendered on the fly using light field refocusing from stochastically sampled pupil states during optimization the proposed method produces holograms with correct parallax and focus cues which are important for passing the visual turing test we validate that our approach compares favorably to state of the art cgh algorithms that use light field and focal stack supervision our experiments demonstrate that our algorithm significantly improves the realism of the viewing experience for a variety of different pupil states mmbench is your multi modal model an all around player authors yuan liu haodong duan yuanhan zhang bo li songyang zhang wangbo zhao yike yuan jiaqi wang conghui he ziwei liu kai chen dahua lin subjects computer vision and pattern recognition cs cv computation and language cs cl arxiv link pdf link abstract large vision language models have recently achieved remarkable progress exhibiting great perception and reasoning abilities concerning visual information however how to effectively evaluate these large vision language models remains a major obstacle hindering future model development traditional benchmarks like or coco caption provide quantitative performance measurements but suffer from a lack of fine grained ability assessment and non robust evaluation metrics recent subjective benchmarks such as owleval offer comprehensive evaluations of a model s abilities by incorporating human labor but they are not scalable and display significant bias in response to these challenges we propose mmbench a novel multi modality benchmark mmbench methodically develops a comprehensive evaluation pipeline primarily comprised of two elements the first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities the second element introduces a novel circulareval strategy and incorporates the use of chatgpt this implementation is designed to convert free form predictions into pre defined choices thereby facilitating a more robust evaluation of the model s predictions mmbench is a systematically designed objective benchmark for robustly evaluating the various abilities of vision language models we hope mmbench will assist the research community in better evaluating their models and encourage future advancements in this domain project page keyword image signal processing there is no result keyword image signal process there is no result keyword compression aict an adaptive image compression transformer authors ahmed ghorbel wassim hamidouche luce morin subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract motivated by the efficiency investigation of the tranformer based transform coding framework namely swint charm we propose to enhance the latter as first with a more straightforward yet effective tranformer based channel wise auto regressive prior model resulting in an absolute image compression transformer ict current methods that still rely on convnet based entropy coding are limited in long range modeling dependencies due to their local connectivity and an increasing number of architectural biases and priors on the contrary the proposed ict can capture both global and local contexts from the latent representations and better parameterize the distribution of the quantized latents further we leverage a learnable scaling module with a sandwich convnext based pre post processor to accurately extract more compact latent representation while reconstructing higher quality images extensive experimental results on benchmark datasets showed that the proposed adaptive image compression transformer aict framework significantly improves the trade off between coding efficiency and decoder complexity over the versatile video coding vvc reference encoder vtm and the neural codec swint charm learning kernel modulated neural representation for efficient light field compression authors jinglei shi yihong xu christine guillemot subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract light field is a type of image data that captures the scene information by recording light rays emitted from a scene at various orientations it offers a more immersive perception than classic images but at the cost of huge data volume in this paper we draw inspiration from the visual characteristics of sub aperture images sais of light field and design a compact neural network representation for the light field compression task the network backbone takes randomly initialized noise as input and is supervised on the sais of the target light field it is composed of two types of complementary kernels descriptive kernels descriptors that store scene description information learned during training and modulatory kernels modulators that control the rendering of different sais from the queried perspectives to further enhance compactness of the network meanwhile retain high quality of the decoded light field we accordingly introduce modulator allocation and kernel tensor decomposition mechanisms followed by non uniform quantization and lossless entropy coding techniques to finally form an efficient compression pipeline extensive experiments demonstrate that our method outperforms other state of the art sota methods by a significant margin in the light field compression task moreover after aligning descriptors the modulators learned from one light field can be transferred to new light fields for rendering dense views indicating a potential solution for view synthesis task keyword raw mop clip a mixture of prompt tuned clip models for domain incremental learning authors julien nicolas florent chiaroni imtiaz ziko ola ahmad christian desrosiers jose dolz subjects computer vision and pattern recognition cs cv machine learning cs lg arxiv link pdf link abstract despite the recent progress in incremental learning addressing catastrophic forgetting under distributional drift is still an open and important problem indeed while state of the art domain incremental learning dil methods perform satisfactorily within known domains their performance largely degrades in the presence of novel domains this limitation hampers their generalizability and restricts their scalability to more realistic settings where train and test data are drawn from different distributions to address these limitations we present a novel dil approach based on a mixture of prompt tuned clip models mop clip which generalizes the paradigm of s prompting to handle both in distribution and out of distribution data at inference in particular at the training stage we model the features distribution of every class in each domain learning individual text and visual prompts to adapt to a given domain at inference the learned distributions allow us to identify whether a given test sample belongs to a known domain selecting the correct prompt for the classification task or from an unseen domain leveraging a mixture of the prompt tuned clip models our empirical evaluation reveals the poor performance of existing dil methods under domain shift and suggests that the proposed mop clip performs competitively in the standard dil settings while outperforming state of the art methods in ood scenarios these results demonstrate the superiority of mop clip offering a robust and general solution to the problem of domain incremental learning sc neus consistent neural surface reconstruction from sparse and noisy views authors shi sheng huang zi xin zou yi chi zhang hua huang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract the recent neural surface reconstruction by volume rendering approaches have made much progress by achieving impressive surface reconstruction quality but are still limited to dense and highly accurate posed views to overcome such drawbacks this paper pays special attention on the consistent surface reconstruction from sparse views with noisy camera poses unlike previous approaches the key difference of this paper is to exploit the multi view constraints directly from the explicit geometry of the neural surface which can be used as effective regularization to jointly learn the neural surface and refine the camera poses to build effective multi view constraints we introduce a fast differentiable on surface intersection to generate on surface points and propose view consistent losses based on such differentiable points to regularize the neural surface learning based on this point we propose a jointly learning strategy for neural surface and camera poses named sc neus to perform geometry consistent surface reconstruction in an end to end manner with extensive evaluation on public datasets our sc neus can achieve consistently better surface reconstruction results with fine grained details than previous state of the art neural surface reconstruction approaches especially from sparse and noisy camera views learning kernel modulated neural representation for efficient light field compression authors jinglei shi yihong xu christine guillemot subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract light field is a type of image data that captures the scene information by recording light rays emitted from a scene at various orientations it offers a more immersive perception than classic images but at the cost of huge data volume in this paper we draw inspiration from the visual characteristics of sub aperture images sais of light field and design a compact neural network representation for the light field compression task the network backbone takes randomly initialized noise as input and is supervised on the sais of the target light field it is composed of two types of complementary kernels descriptive kernels descriptors that store scene description information learned during training and modulatory kernels modulators that control the rendering of different sais from the queried perspectives to further enhance compactness of the network meanwhile retain high quality of the decoded light field we accordingly introduce modulator allocation and kernel tensor decomposition mechanisms followed by non uniform quantization and lossless entropy coding techniques to finally form an efficient compression pipeline extensive experiments demonstrate that our method outperforms other state of the art sota methods by a significant margin in the light field compression task moreover after aligning descriptors the modulators learned from one light field can be transferred to new light fields for rendering dense views indicating a potential solution for view synthesis task keyword raw image there is no result
1
19,671
26,030,084,209
IssuesEvent
2022-12-21 20:13:32
googleapis/google-cloud-go
https://api.github.com/repos/googleapis/google-cloud-go
closed
all: have a deprecated release status
type: process
This would be nice for gapic versioning. We could generate extra docs.
1.0
all: have a deprecated release status - This would be nice for gapic versioning. We could generate extra docs.
process
all have a deprecated release status this would be nice for gapic versioning we could generate extra docs
1
237,191
7,757,182,290
IssuesEvent
2018-05-31 15:36:33
jonas747/yagpdb
https://api.github.com/repos/jonas747/yagpdb
opened
Control panel redesign
Priority: High enhancement
- [ ] Convert all pages - [ ] Add async form submitting where appropiate - [ ] General work Should be done with this around the weekend of the 8th
1.0
Control panel redesign - - [ ] Convert all pages - [ ] Add async form submitting where appropiate - [ ] General work Should be done with this around the weekend of the 8th
non_process
control panel redesign convert all pages add async form submitting where appropiate general work should be done with this around the weekend of the
0
9,347
11,386,359,406
IssuesEvent
2020-01-29 13:07:34
luisfabib/DeerAnalysis2
https://api.github.com/repos/luisfabib/DeerAnalysis2
opened
fitbackground (free-version) not compatible with MacOS and Linux
compatibility
The ``fitbackground`` function uses the ``nlsqbnd`` as default for a toolbox-free use. However, the solver requires a mex file only available for WinOS. Adapt this in a similar way as in ``fitparamodel``.
True
fitbackground (free-version) not compatible with MacOS and Linux - The ``fitbackground`` function uses the ``nlsqbnd`` as default for a toolbox-free use. However, the solver requires a mex file only available for WinOS. Adapt this in a similar way as in ``fitparamodel``.
non_process
fitbackground free version not compatible with macos and linux the fitbackground function uses the nlsqbnd as default for a toolbox free use however the solver requires a mex file only available for winos adapt this in a similar way as in fitparamodel
0
115,423
14,742,873,132
IssuesEvent
2021-01-07 13:02:21
kubermatic/dashboard
https://api.github.com/repos/kubermatic/dashboard
closed
Improve event handling after user input
kind/design team/ui
As a user, I want prominently displayed events when I did not achieve my goal, so that adapting to the situation becomes effortless. **What is the problem, where is it and why is it an issue?** Create a cluster on hetzner cloud with limited quota available. Different event logs where shown, but I had to scroll down and extend the events in order to see what happened. The information displayed was helpful and correct. **Suggest how to fix it, if possible** If events occur, events should be displayed prominently within in users view so that he does not need to scroll / search for the information. This is especially important if he just initiated some action.
1.0
Improve event handling after user input - As a user, I want prominently displayed events when I did not achieve my goal, so that adapting to the situation becomes effortless. **What is the problem, where is it and why is it an issue?** Create a cluster on hetzner cloud with limited quota available. Different event logs where shown, but I had to scroll down and extend the events in order to see what happened. The information displayed was helpful and correct. **Suggest how to fix it, if possible** If events occur, events should be displayed prominently within in users view so that he does not need to scroll / search for the information. This is especially important if he just initiated some action.
non_process
improve event handling after user input as a user i want prominently displayed events when i did not achieve my goal so that adapting to the situation becomes effortless what is the problem where is it and why is it an issue create a cluster on hetzner cloud with limited quota available different event logs where shown but i had to scroll down and extend the events in order to see what happened the information displayed was helpful and correct suggest how to fix it if possible if events occur events should be displayed prominently within in users view so that he does not need to scroll search for the information this is especially important if he just initiated some action
0
202,610
7,050,276,704
IssuesEvent
2018-01-03 04:42:43
squizlabs/PHP_CodeSniffer
https://api.github.com/repos/squizlabs/PHP_CodeSniffer
closed
Option to cancel word wrapping in reports
Enhancement Low Priority
Good morning! Here's what `phpcs --no-colors` produces: ``` ---------------------------------------------------------------------- FOUND 69 ERRORS AFFECTING 66 LINES ---------------------------------------------------------------------- 35 | ERROR | [x] Spaces must be used to indent lines; tabs are not | | allowed 37 | ERROR | [x] Spaces must be used to indent lines; tabs are not | | allowed 47 | ERROR | [x] Spaces must be used to indent lines; tabs are not | | allowed 56 | ERROR | [x] Spaces must be used to indent lines; tabs are not | | allowed ``` Is there an option to cancel those word wraps at the end of lines? I usually send these reports into an internal bug tracker, so, it's really not for reading in console. Thank you in advance.
1.0
Option to cancel word wrapping in reports - Good morning! Here's what `phpcs --no-colors` produces: ``` ---------------------------------------------------------------------- FOUND 69 ERRORS AFFECTING 66 LINES ---------------------------------------------------------------------- 35 | ERROR | [x] Spaces must be used to indent lines; tabs are not | | allowed 37 | ERROR | [x] Spaces must be used to indent lines; tabs are not | | allowed 47 | ERROR | [x] Spaces must be used to indent lines; tabs are not | | allowed 56 | ERROR | [x] Spaces must be used to indent lines; tabs are not | | allowed ``` Is there an option to cancel those word wraps at the end of lines? I usually send these reports into an internal bug tracker, so, it's really not for reading in console. Thank you in advance.
non_process
option to cancel word wrapping in reports good morning here s what phpcs no colors produces found errors affecting lines error spaces must be used to indent lines tabs are not allowed error spaces must be used to indent lines tabs are not allowed error spaces must be used to indent lines tabs are not allowed error spaces must be used to indent lines tabs are not allowed is there an option to cancel those word wraps at the end of lines i usually send these reports into an internal bug tracker so it s really not for reading in console thank you in advance
0
316,323
9,645,649,577
IssuesEvent
2019-05-17 09:11:48
MTG/andalusian-corpus-notebooks
https://api.github.com/repos/MTG/andalusian-corpus-notebooks
opened
Use pitch distribution with step size 7.5
medium priority improvement
The methods which use pitch distribution works with 1200 value (one value for each cent). This is not necessary. It could be used a step size of 7.5. This will strongly improve the performance, reducing the computation time which are too long. Pay attention! It is not enough to change only this parameter, but it is necessary to change several methods starting from the pitch distribution creation!!! https://github.com/MTG/andalusian-corpus-notebooks/blob/c634b9aa44ff3da24c8e3115f5ea7fbddf4c7916/utilities/recordingcomputation.py#L29
1.0
Use pitch distribution with step size 7.5 - The methods which use pitch distribution works with 1200 value (one value for each cent). This is not necessary. It could be used a step size of 7.5. This will strongly improve the performance, reducing the computation time which are too long. Pay attention! It is not enough to change only this parameter, but it is necessary to change several methods starting from the pitch distribution creation!!! https://github.com/MTG/andalusian-corpus-notebooks/blob/c634b9aa44ff3da24c8e3115f5ea7fbddf4c7916/utilities/recordingcomputation.py#L29
non_process
use pitch distribution with step size the methods which use pitch distribution works with value one value for each cent this is not necessary it could be used a step size of this will strongly improve the performance reducing the computation time which are too long pay attention it is not enough to change only this parameter but it is necessary to change several methods starting from the pitch distribution creation
0
29,925
24,396,116,918
IssuesEvent
2022-10-04 19:25:48
ProjectPythiaCookbooks/cookbook-template
https://api.github.com/repos/ProjectPythiaCookbooks/cookbook-template
closed
Preview does not deploy if the link-checker fails
infrastructure
As I found out over at https://github.com/ProjectPythiaTutorials/physical-oceanography-cookbook/issues/9, if there is a link-checker failure for a PR, the action that deploys the preview does not run. This is a bug in the workflow. The preview should be independent of the link checker.
1.0
Preview does not deploy if the link-checker fails - As I found out over at https://github.com/ProjectPythiaTutorials/physical-oceanography-cookbook/issues/9, if there is a link-checker failure for a PR, the action that deploys the preview does not run. This is a bug in the workflow. The preview should be independent of the link checker.
non_process
preview does not deploy if the link checker fails as i found out over at if there is a link checker failure for a pr the action that deploys the preview does not run this is a bug in the workflow the preview should be independent of the link checker
0
454,067
13,094,272,087
IssuesEvent
2020-08-03 12:05:24
geosolutions-it/ckanext-faoclh
https://api.github.com/repos/geosolutions-it/ckanext-faoclh
opened
Topic translation missing in ES and FR
Priority: Blocker
For French it should be **Sujets** For Spanish it should be **Temas** **Home** ![image.png](https://images.zenhubusercontent.com/5c9ca89d6607b6793c66ec35/34d6134e-56ac-4baf-b0ca-56c11974b157) **Datasets search page** ![2020_08_03_14_02_20_Jeux_de_données_Climate_and_Land_Hub.png](https://images.zenhubusercontent.com/5c9ca89d6607b6793c66ec35/36726901-d3b1-4606-9143-b34ba1d48db3) **Dataset page** ![2020_08_03_14_03_19_Guidelines_Implementing_Practices_for_Resilient_Livelihoods_and_Risk_Management.png](https://images.zenhubusercontent.com/5c9ca89d6607b6793c66ec35/d2ea53a5-bf41-498f-83f4-4efca4ebf2a4)
1.0
Topic translation missing in ES and FR - For French it should be **Sujets** For Spanish it should be **Temas** **Home** ![image.png](https://images.zenhubusercontent.com/5c9ca89d6607b6793c66ec35/34d6134e-56ac-4baf-b0ca-56c11974b157) **Datasets search page** ![2020_08_03_14_02_20_Jeux_de_données_Climate_and_Land_Hub.png](https://images.zenhubusercontent.com/5c9ca89d6607b6793c66ec35/36726901-d3b1-4606-9143-b34ba1d48db3) **Dataset page** ![2020_08_03_14_03_19_Guidelines_Implementing_Practices_for_Resilient_Livelihoods_and_Risk_Management.png](https://images.zenhubusercontent.com/5c9ca89d6607b6793c66ec35/d2ea53a5-bf41-498f-83f4-4efca4ebf2a4)
non_process
topic translation missing in es and fr for french it should be sujets for spanish it should be temas home datasets search page dataset page
0
7,776
10,918,657,542
IssuesEvent
2019-11-21 17:19:01
googleapis/google-cloud-go
https://api.github.com/repos/googleapis/google-cloud-go
closed
bump dialogflow to stable
api: dialogflow type: process
Package name: **dialogflow/apiv2** Current release: **alpha** Proposed release: **GA** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [x] 28 days elapsed since last beta release with new API surface - [x] Server API is GA - [x] Package API is stable, and we can commit to backward compatibility - [x] All dependencies are GA ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [ ] Manual code has been reviewed by API producer - [ ] Manual code has been reviewed by a DPE responsible for samples - [x] 'Client LIbraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
1.0
bump dialogflow to stable - Package name: **dialogflow/apiv2** Current release: **alpha** Proposed release: **GA** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [x] 28 days elapsed since last beta release with new API surface - [x] Server API is GA - [x] Package API is stable, and we can commit to backward compatibility - [x] All dependencies are GA ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [ ] Manual code has been reviewed by API producer - [ ] Manual code has been reviewed by a DPE responsible for samples - [x] 'Client LIbraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
process
bump dialogflow to stable package name dialogflow current release alpha proposed release ga instructions check the lists below adding tests documentation as required once all the required boxes are ticked please create a release and close this issue required days elapsed since last beta release with new api surface server api is ga package api is stable and we can commit to backward compatibility all dependencies are ga optional most common important scenarios have descriptive samples public manual methods have at least one usage sample each excluding overloads per api readme includes a full description of the api per api readme contains at least one “getting started” sample using the most common api scenario manual code has been reviewed by api producer manual code has been reviewed by a dpe responsible for samples client libraries page is added to the product documentation in apis reference section of the product s documentation on cloud site
1
2,133
2,666,046,375
IssuesEvent
2015-03-21 04:12:41
vsivsi/meteor-job-collection
https://api.github.com/repos/vsivsi/meteor-job-collection
closed
Connecting Meteor-based worker to collection in another app
documentation enhancement
What's the best approach to connecting workers to a queue in a different Meteor app? I'm wanting to separate the responsibilities without moving the workers out of the Meteor environment.
1.0
Connecting Meteor-based worker to collection in another app - What's the best approach to connecting workers to a queue in a different Meteor app? I'm wanting to separate the responsibilities without moving the workers out of the Meteor environment.
non_process
connecting meteor based worker to collection in another app what s the best approach to connecting workers to a queue in a different meteor app i m wanting to separate the responsibilities without moving the workers out of the meteor environment
0
10,433
13,220,046,427
IssuesEvent
2020-08-17 11:40:15
km4ack/pi-build
https://api.github.com/repos/km4ack/pi-build
closed
Remove temp dir before running update update
bug in process
The ~/pi-build/temp dir needs to be removed before running the update script. Not removing it can prevent the script from working correctly. This was seen with one of the latest updates to Chirp.
1.0
Remove temp dir before running update update - The ~/pi-build/temp dir needs to be removed before running the update script. Not removing it can prevent the script from working correctly. This was seen with one of the latest updates to Chirp.
process
remove temp dir before running update update the pi build temp dir needs to be removed before running the update script not removing it can prevent the script from working correctly this was seen with one of the latest updates to chirp
1
91,361
26,369,093,084
IssuesEvent
2023-01-11 19:02:27
tarantool/tarantool
https://api.github.com/repos/tarantool/tarantool
closed
clang-15: build fails with 'Unable to get thread stack'
build 1.10
The build works with clang 14: ```sh CC=clang-14 CXX=clang++-14 cmake . \ -DCMAKE_BUILD_TYPE=RelWithDebInfo \ -DENABLE_BACKTRACE=ON \ -DENABLE_DIST=ON \ -DENABLE_FEEDBACK_DAEMON=OFF \ -DENABLE_BUNDLED_LIBCURL=OFF && \ make -j ``` With clang 14 it shows the following cmake check results around pthread: ``` -- Looking for pthread_np.h -- Looking for pthread_np.h - not found -- Performing Test HAVE_PTHREAD_SETNAME_NP -- Performing Test HAVE_PTHREAD_SETNAME_NP - Success -- Performing Test HAVE_PTHREAD_SETNAME_NP_1 -- Performing Test HAVE_PTHREAD_SETNAME_NP_1 - Failed -- Performing Test HAVE_PTHREAD_SET_NAME_NP -- Performing Test HAVE_PTHREAD_SET_NAME_NP - Failed -- Performing Test HAVE_PTHREAD_GETATTR_NP -- Performing Test HAVE_PTHREAD_GETATTR_NP - Success -- Performing Test HAVE_PTHREAD_STACKSEG_NP -- Performing Test HAVE_PTHREAD_STACKSEG_NP - Failed -- Performing Test HAVE_PTHREAD_ATTR_GET_NP -- Performing Test HAVE_PTHREAD_ATTR_GET_NP - Failed -- Performing Test HAVE_PTHREAD_GET_STACKSIZE_NP -- Performing Test HAVE_PTHREAD_GET_STACKSIZE_NP - Failed -- Performing Test HAVE_PTHREAD_GET_STACKADDR_NP -- Performing Test HAVE_PTHREAD_GET_STACKADDR_NP - Failed ``` However the same on clang 15: ```sh CC=clang-15 CXX=clang++-15 cmake . \ -DCMAKE_BUILD_TYPE=RelWithDebInfo \ -DENABLE_BACKTRACE=ON \ -DENABLE_DIST=ON \ -DENABLE_FEEDBACK_DAEMON=OFF \ -DENABLE_BUNDLED_LIBCURL=OFF && \ make -j ``` Gives 'all failed' for the pthread related checks. ``` -- Looking for pthread_np.h -- Looking for pthread_np.h - not found -- Performing Test HAVE_PTHREAD_SETNAME_NP -- Performing Test HAVE_PTHREAD_SETNAME_NP - Failed -- Performing Test HAVE_PTHREAD_SETNAME_NP_1 -- Performing Test HAVE_PTHREAD_SETNAME_NP_1 - Failed -- Performing Test HAVE_PTHREAD_SET_NAME_NP -- Performing Test HAVE_PTHREAD_SET_NAME_NP - Failed -- Performing Test HAVE_PTHREAD_GETATTR_NP -- Performing Test HAVE_PTHREAD_GETATTR_NP - Failed -- Performing Test HAVE_PTHREAD_STACKSEG_NP -- Performing Test HAVE_PTHREAD_STACKSEG_NP - Failed -- Performing Test HAVE_PTHREAD_ATTR_GET_NP -- Performing Test HAVE_PTHREAD_ATTR_GET_NP - Failed -- Performing Test HAVE_PTHREAD_GET_STACKSIZE_NP -- Performing Test HAVE_PTHREAD_GET_STACKSIZE_NP - Failed -- Performing Test HAVE_PTHREAD_GET_STACKADDR_NP -- Performing Test HAVE_PTHREAD_GET_STACKADDR_NP - Failed ``` And the build fails later. ```c [ 9%] Building C object src/lib/core/CMakeFiles/core.dir/diag.c.o In file included from /home/alex/p/tarantool-meta/r/t-5/src/lib/core/diag.c:32: In file included from /home/alex/p/tarantool-meta/r/t-5/src/lib/core/fiber.h:37: /home/alex/p/tarantool-meta/r/t-5/src/tt_pthread.h:377:2: error: Unable to get thread stack #error Unable to get thread stack ^ /home/alex/p/tarantool-meta/r/t-5/src/tt_pthread.h:340:36: warning: unused parameter 'thread' [-Wunused-parameter] tt_pthread_attr_getstack(pthread_t thread, void **stackaddr, size_t *stacksize) ^ /home/alex/p/tarantool-meta/r/t-5/src/tt_pthread.h:340:51: warning: unused parameter 'stackaddr' [-Wunused-parameter] tt_pthread_attr_getstack(pthread_t thread, void **stackaddr, size_t *stacksize) ^ /home/alex/p/tarantool-meta/r/t-5/src/tt_pthread.h:340:70: warning: unused parameter 'stacksize' [-Wunused-parameter] tt_pthread_attr_getstack(pthread_t thread, void **stackaddr, size_t *stacksize) ^ 3 warnings and 1 error generated. ``` The reason is inaccurate CMake checks code. From `CMakeFiles/CMakeError.log`: ```c /usr/lib/llvm/15/bin/clang-15 -DHAVE_PTHREAD_SETNAME_NP -D_DARWIN_C_SOURCE -D_GNU_SOURCE -fexceptions -funwind-tables -fasynchronous-unwind-tables -fno-common -fopenmp -msse2 -pedantic-errors -MD -MT CMakeFiles/cmTC_6822f.dir/src.c.o -MF CMakeFiles/cmTC_6822f.dir/src.c.o.d -o CMakeFiles/cmTC_6822f.dir/src.c.o -c /home/alex/p/tarantool-meta/r/t-5/CMakeFiles/CMakeScratch/TryCompile-78KaOK/src.c /home/alex/p/tarantool-meta/r/t-5/CMakeFiles/CMakeScratch/TryCompile-78KaOK/src.c:4:17: error: a function declaration without a prototype is deprecated in all versions of C [-Werror,-Wstrict-prototypes] int main() { pthread_setname_np(pthread_self(), ""); } ^ void 1 error generated. ```
1.0
clang-15: build fails with 'Unable to get thread stack' - The build works with clang 14: ```sh CC=clang-14 CXX=clang++-14 cmake . \ -DCMAKE_BUILD_TYPE=RelWithDebInfo \ -DENABLE_BACKTRACE=ON \ -DENABLE_DIST=ON \ -DENABLE_FEEDBACK_DAEMON=OFF \ -DENABLE_BUNDLED_LIBCURL=OFF && \ make -j ``` With clang 14 it shows the following cmake check results around pthread: ``` -- Looking for pthread_np.h -- Looking for pthread_np.h - not found -- Performing Test HAVE_PTHREAD_SETNAME_NP -- Performing Test HAVE_PTHREAD_SETNAME_NP - Success -- Performing Test HAVE_PTHREAD_SETNAME_NP_1 -- Performing Test HAVE_PTHREAD_SETNAME_NP_1 - Failed -- Performing Test HAVE_PTHREAD_SET_NAME_NP -- Performing Test HAVE_PTHREAD_SET_NAME_NP - Failed -- Performing Test HAVE_PTHREAD_GETATTR_NP -- Performing Test HAVE_PTHREAD_GETATTR_NP - Success -- Performing Test HAVE_PTHREAD_STACKSEG_NP -- Performing Test HAVE_PTHREAD_STACKSEG_NP - Failed -- Performing Test HAVE_PTHREAD_ATTR_GET_NP -- Performing Test HAVE_PTHREAD_ATTR_GET_NP - Failed -- Performing Test HAVE_PTHREAD_GET_STACKSIZE_NP -- Performing Test HAVE_PTHREAD_GET_STACKSIZE_NP - Failed -- Performing Test HAVE_PTHREAD_GET_STACKADDR_NP -- Performing Test HAVE_PTHREAD_GET_STACKADDR_NP - Failed ``` However the same on clang 15: ```sh CC=clang-15 CXX=clang++-15 cmake . \ -DCMAKE_BUILD_TYPE=RelWithDebInfo \ -DENABLE_BACKTRACE=ON \ -DENABLE_DIST=ON \ -DENABLE_FEEDBACK_DAEMON=OFF \ -DENABLE_BUNDLED_LIBCURL=OFF && \ make -j ``` Gives 'all failed' for the pthread related checks. ``` -- Looking for pthread_np.h -- Looking for pthread_np.h - not found -- Performing Test HAVE_PTHREAD_SETNAME_NP -- Performing Test HAVE_PTHREAD_SETNAME_NP - Failed -- Performing Test HAVE_PTHREAD_SETNAME_NP_1 -- Performing Test HAVE_PTHREAD_SETNAME_NP_1 - Failed -- Performing Test HAVE_PTHREAD_SET_NAME_NP -- Performing Test HAVE_PTHREAD_SET_NAME_NP - Failed -- Performing Test HAVE_PTHREAD_GETATTR_NP -- Performing Test HAVE_PTHREAD_GETATTR_NP - Failed -- Performing Test HAVE_PTHREAD_STACKSEG_NP -- Performing Test HAVE_PTHREAD_STACKSEG_NP - Failed -- Performing Test HAVE_PTHREAD_ATTR_GET_NP -- Performing Test HAVE_PTHREAD_ATTR_GET_NP - Failed -- Performing Test HAVE_PTHREAD_GET_STACKSIZE_NP -- Performing Test HAVE_PTHREAD_GET_STACKSIZE_NP - Failed -- Performing Test HAVE_PTHREAD_GET_STACKADDR_NP -- Performing Test HAVE_PTHREAD_GET_STACKADDR_NP - Failed ``` And the build fails later. ```c [ 9%] Building C object src/lib/core/CMakeFiles/core.dir/diag.c.o In file included from /home/alex/p/tarantool-meta/r/t-5/src/lib/core/diag.c:32: In file included from /home/alex/p/tarantool-meta/r/t-5/src/lib/core/fiber.h:37: /home/alex/p/tarantool-meta/r/t-5/src/tt_pthread.h:377:2: error: Unable to get thread stack #error Unable to get thread stack ^ /home/alex/p/tarantool-meta/r/t-5/src/tt_pthread.h:340:36: warning: unused parameter 'thread' [-Wunused-parameter] tt_pthread_attr_getstack(pthread_t thread, void **stackaddr, size_t *stacksize) ^ /home/alex/p/tarantool-meta/r/t-5/src/tt_pthread.h:340:51: warning: unused parameter 'stackaddr' [-Wunused-parameter] tt_pthread_attr_getstack(pthread_t thread, void **stackaddr, size_t *stacksize) ^ /home/alex/p/tarantool-meta/r/t-5/src/tt_pthread.h:340:70: warning: unused parameter 'stacksize' [-Wunused-parameter] tt_pthread_attr_getstack(pthread_t thread, void **stackaddr, size_t *stacksize) ^ 3 warnings and 1 error generated. ``` The reason is inaccurate CMake checks code. From `CMakeFiles/CMakeError.log`: ```c /usr/lib/llvm/15/bin/clang-15 -DHAVE_PTHREAD_SETNAME_NP -D_DARWIN_C_SOURCE -D_GNU_SOURCE -fexceptions -funwind-tables -fasynchronous-unwind-tables -fno-common -fopenmp -msse2 -pedantic-errors -MD -MT CMakeFiles/cmTC_6822f.dir/src.c.o -MF CMakeFiles/cmTC_6822f.dir/src.c.o.d -o CMakeFiles/cmTC_6822f.dir/src.c.o -c /home/alex/p/tarantool-meta/r/t-5/CMakeFiles/CMakeScratch/TryCompile-78KaOK/src.c /home/alex/p/tarantool-meta/r/t-5/CMakeFiles/CMakeScratch/TryCompile-78KaOK/src.c:4:17: error: a function declaration without a prototype is deprecated in all versions of C [-Werror,-Wstrict-prototypes] int main() { pthread_setname_np(pthread_self(), ""); } ^ void 1 error generated. ```
non_process
clang build fails with unable to get thread stack the build works with clang sh cc clang cxx clang cmake dcmake build type relwithdebinfo denable backtrace on denable dist on denable feedback daemon off denable bundled libcurl off make j with clang it shows the following cmake check results around pthread looking for pthread np h looking for pthread np h not found performing test have pthread setname np performing test have pthread setname np success performing test have pthread setname np performing test have pthread setname np failed performing test have pthread set name np performing test have pthread set name np failed performing test have pthread getattr np performing test have pthread getattr np success performing test have pthread stackseg np performing test have pthread stackseg np failed performing test have pthread attr get np performing test have pthread attr get np failed performing test have pthread get stacksize np performing test have pthread get stacksize np failed performing test have pthread get stackaddr np performing test have pthread get stackaddr np failed however the same on clang sh cc clang cxx clang cmake dcmake build type relwithdebinfo denable backtrace on denable dist on denable feedback daemon off denable bundled libcurl off make j gives all failed for the pthread related checks looking for pthread np h looking for pthread np h not found performing test have pthread setname np performing test have pthread setname np failed performing test have pthread setname np performing test have pthread setname np failed performing test have pthread set name np performing test have pthread set name np failed performing test have pthread getattr np performing test have pthread getattr np failed performing test have pthread stackseg np performing test have pthread stackseg np failed performing test have pthread attr get np performing test have pthread attr get np failed performing test have pthread get stacksize np performing test have pthread get stacksize np failed performing test have pthread get stackaddr np performing test have pthread get stackaddr np failed and the build fails later c building c object src lib core cmakefiles core dir diag c o in file included from home alex p tarantool meta r t src lib core diag c in file included from home alex p tarantool meta r t src lib core fiber h home alex p tarantool meta r t src tt pthread h error unable to get thread stack error unable to get thread stack home alex p tarantool meta r t src tt pthread h warning unused parameter thread tt pthread attr getstack pthread t thread void stackaddr size t stacksize home alex p tarantool meta r t src tt pthread h warning unused parameter stackaddr tt pthread attr getstack pthread t thread void stackaddr size t stacksize home alex p tarantool meta r t src tt pthread h warning unused parameter stacksize tt pthread attr getstack pthread t thread void stackaddr size t stacksize warnings and error generated the reason is inaccurate cmake checks code from cmakefiles cmakeerror log c usr lib llvm bin clang dhave pthread setname np d darwin c source d gnu source fexceptions funwind tables fasynchronous unwind tables fno common fopenmp pedantic errors md mt cmakefiles cmtc dir src c o mf cmakefiles cmtc dir src c o d o cmakefiles cmtc dir src c o c home alex p tarantool meta r t cmakefiles cmakescratch trycompile src c home alex p tarantool meta r t cmakefiles cmakescratch trycompile src c error a function declaration without a prototype is deprecated in all versions of c int main pthread setname np pthread self void error generated
0
17,026
22,392,984,450
IssuesEvent
2022-06-17 09:32:25
python/cpython
https://api.github.com/repos/python/cpython
closed
test_multiprocessing_fork leaks processes on PPC64LE RHEL8 LTO + PGO 3.x
tests 3.9 expert-multiprocessing
BPO | [40500](https://bugs.python.org/issue40500) --- | :--- Nosy | @pitrou, @vstinner, @tomMoral <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2020-05-04.12:38:50.919> labels = ['tests', '3.9'] title = 'test_multiprocessing_fork leaks processes on PPC64LE RHEL8 LTO + PGO 3.x' updated_at = <Date 2020-05-11.21:48:22.349> user = 'https://github.com/vstinner' ``` bugs.python.org fields: ```python activity = <Date 2020-05-11.21:48:22.349> actor = 'vstinner' assignee = 'none' closed = False closed_date = None closer = None components = ['Tests'] creation = <Date 2020-05-04.12:38:50.919> creator = 'vstinner' dependencies = [] files = [] hgrepos = [] issue_num = 40500 keywords = [] message_count = 3.0 messages = ['368043', '368082', '368661'] nosy_count = 3.0 nosy_names = ['pitrou', 'vstinner', 'tomMoral'] pr_nums = [] priority = 'normal' resolution = None stage = None status = 'open' superseder = None type = None url = 'https://bugs.python.org/issue40500' versions = ['Python 3.9'] ``` </p></details>
1.0
test_multiprocessing_fork leaks processes on PPC64LE RHEL8 LTO + PGO 3.x - BPO | [40500](https://bugs.python.org/issue40500) --- | :--- Nosy | @pitrou, @vstinner, @tomMoral <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2020-05-04.12:38:50.919> labels = ['tests', '3.9'] title = 'test_multiprocessing_fork leaks processes on PPC64LE RHEL8 LTO + PGO 3.x' updated_at = <Date 2020-05-11.21:48:22.349> user = 'https://github.com/vstinner' ``` bugs.python.org fields: ```python activity = <Date 2020-05-11.21:48:22.349> actor = 'vstinner' assignee = 'none' closed = False closed_date = None closer = None components = ['Tests'] creation = <Date 2020-05-04.12:38:50.919> creator = 'vstinner' dependencies = [] files = [] hgrepos = [] issue_num = 40500 keywords = [] message_count = 3.0 messages = ['368043', '368082', '368661'] nosy_count = 3.0 nosy_names = ['pitrou', 'vstinner', 'tomMoral'] pr_nums = [] priority = 'normal' resolution = None stage = None status = 'open' superseder = None type = None url = 'https://bugs.python.org/issue40500' versions = ['Python 3.9'] ``` </p></details>
process
test multiprocessing fork leaks processes on lto pgo x bpo nosy pitrou vstinner tommoral note these values reflect the state of the issue at the time it was migrated and might not reflect the current state show more details github fields python assignee none closed at none created at labels title test multiprocessing fork leaks processes on lto pgo x updated at user bugs python org fields python activity actor vstinner assignee none closed false closed date none closer none components creation creator vstinner dependencies files hgrepos issue num keywords message count messages nosy count nosy names pr nums priority normal resolution none stage none status open superseder none type none url versions
1
198,060
6,969,320,591
IssuesEvent
2017-12-11 04:28:26
infoScoop/infoscoop
https://api.github.com/repos/infoScoop/infoscoop
opened
Limit of the arrangement number of possible gadget
Priority-Medium Type-Enhancement
It's restricted by a property of a management screen. A target is a gadget bag in all personalization territory.
1.0
Limit of the arrangement number of possible gadget - It's restricted by a property of a management screen. A target is a gadget bag in all personalization territory.
non_process
limit of the arrangement number of possible gadget it s restricted by a property of a management screen a target is a gadget bag in all personalization territory
0
273,864
29,831,106,236
IssuesEvent
2023-06-18 09:32:54
RG4421/ampere-centos-kernel
https://api.github.com/repos/RG4421/ampere-centos-kernel
closed
CVE-2023-1611 (Medium) detected in linuxv5.2 - autoclosed
Mend: dependency security vulnerability
## CVE-2023-1611 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/qgroup.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/qgroup.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A use-after-free flaw was found in btrfs_search_slot in fs/btrfs/ctree.c in btrfs in the Linux Kernel.This flaw allows an attacker to crash the system and possibly cause a kernel information lea <p>Publish Date: 2023-04-03 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1611>CVE-2023-1611</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1611">https://www.linuxkernelcves.com/cves/CVE-2023-1611</a></p> <p>Release Date: 2023-04-03</p> <p>Fix Resolution: v5.10.177,v5.15.106,v6.1.23,v6.2.10</p> </p> </details> <p></p>
True
CVE-2023-1611 (Medium) detected in linuxv5.2 - autoclosed - ## CVE-2023-1611 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/qgroup.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/qgroup.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A use-after-free flaw was found in btrfs_search_slot in fs/btrfs/ctree.c in btrfs in the Linux Kernel.This flaw allows an attacker to crash the system and possibly cause a kernel information lea <p>Publish Date: 2023-04-03 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1611>CVE-2023-1611</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1611">https://www.linuxkernelcves.com/cves/CVE-2023-1611</a></p> <p>Release Date: 2023-04-03</p> <p>Fix Resolution: v5.10.177,v5.15.106,v6.1.23,v6.2.10</p> </p> </details> <p></p>
non_process
cve medium detected in autoclosed cve medium severity vulnerability vulnerable library linux kernel source tree library home page a href found in base branch amp centos kernel vulnerable source files fs btrfs qgroup c fs btrfs qgroup c vulnerability details a use after free flaw was found in btrfs search slot in fs btrfs ctree c in btrfs in the linux kernel this flaw allows an attacker to crash the system and possibly cause a kernel information lea publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
45,147
9,683,718,134
IssuesEvent
2019-05-23 12:13:02
nopSolutions/nopCommerce
https://api.github.com/repos/nopSolutions/nopCommerce
closed
Support custom validation logic during plugin installation/uninstallation
refactoring / source code
nopCommerce version: 4.20 BETA We really like the new feature to be able to install/uninstall several plugins at once. But the way it works now does not allow us to prevent certain plugins to be uninstalled. Let me give you an example. We have a Core plugin that contain common libraries shared by all our plugins. In other words, we don't allow this plugin to be Uninstalled if there are any other SevenSpikes plugins already installed. Unfortunately with the new functionality we have no way to prevent our customers of uninstalling the Core plugin by mistake and thus their website will not be able to be even started as there will be missing assemblies from the Core plugin. We also have other plugins that are dependent on each other i.e Ajax Filters Pro require the Ajax Filters to be installed first as otherwise they will not work etc. So it would be great if we could tell nopCommerce if certain plugins could not be installed/unistalled and return some message i.e "You could not uninstall Core plugin as it is used by other plugins" or "You could not install Ajax Filters Pro plugin as it required the Ajax Filters plugin to be installed first" etc. Source: https://www.nopcommerce.com/boards/t/61916/nopcommerce-420-feedback-and-improvements.aspx#227022 --------------------------------- To add to my previous post how we handled this in the previous version of nopCommerce. In the Install method of all our plugins we had an additional check ShouldInstallPlugin which checks that if the Core plugin is not installed then it simply does not call the Install method of the BasePlugin (we always inherit the BasePlugin). This way, even without showing any messages to the clients, at least we didn't allow them to install a plugin that will potentially break the site. In the Uninstall method of all our plugins we had an additional check ShouldUnInstallPlugin. If we should not uninstall a plugin we simply don't call the Uninstall method of the BasePlugin and thus it was never really uninstalled. Now in nopCommerce 4.20 the logic of the actual installation/uninstallation of the plugin is not in the plugin (as it used to be in the BasePlugin) itself and we have no control over it. The bottom line is that we need a way to prevent certain plugins to be added in the list of plugins that are to be Uninstalled as once they go there and the application is restarted we could do nothing. Source: https://www.nopcommerce.com/boards/t/61916/nopcommerce-420-feedback-and-improvements.aspx#227023
1.0
Support custom validation logic during plugin installation/uninstallation - nopCommerce version: 4.20 BETA We really like the new feature to be able to install/uninstall several plugins at once. But the way it works now does not allow us to prevent certain plugins to be uninstalled. Let me give you an example. We have a Core plugin that contain common libraries shared by all our plugins. In other words, we don't allow this plugin to be Uninstalled if there are any other SevenSpikes plugins already installed. Unfortunately with the new functionality we have no way to prevent our customers of uninstalling the Core plugin by mistake and thus their website will not be able to be even started as there will be missing assemblies from the Core plugin. We also have other plugins that are dependent on each other i.e Ajax Filters Pro require the Ajax Filters to be installed first as otherwise they will not work etc. So it would be great if we could tell nopCommerce if certain plugins could not be installed/unistalled and return some message i.e "You could not uninstall Core plugin as it is used by other plugins" or "You could not install Ajax Filters Pro plugin as it required the Ajax Filters plugin to be installed first" etc. Source: https://www.nopcommerce.com/boards/t/61916/nopcommerce-420-feedback-and-improvements.aspx#227022 --------------------------------- To add to my previous post how we handled this in the previous version of nopCommerce. In the Install method of all our plugins we had an additional check ShouldInstallPlugin which checks that if the Core plugin is not installed then it simply does not call the Install method of the BasePlugin (we always inherit the BasePlugin). This way, even without showing any messages to the clients, at least we didn't allow them to install a plugin that will potentially break the site. In the Uninstall method of all our plugins we had an additional check ShouldUnInstallPlugin. If we should not uninstall a plugin we simply don't call the Uninstall method of the BasePlugin and thus it was never really uninstalled. Now in nopCommerce 4.20 the logic of the actual installation/uninstallation of the plugin is not in the plugin (as it used to be in the BasePlugin) itself and we have no control over it. The bottom line is that we need a way to prevent certain plugins to be added in the list of plugins that are to be Uninstalled as once they go there and the application is restarted we could do nothing. Source: https://www.nopcommerce.com/boards/t/61916/nopcommerce-420-feedback-and-improvements.aspx#227023
non_process
support custom validation logic during plugin installation uninstallation nopcommerce version beta we really like the new feature to be able to install uninstall several plugins at once but the way it works now does not allow us to prevent certain plugins to be uninstalled let me give you an example we have a core plugin that contain common libraries shared by all our plugins in other words we don t allow this plugin to be uninstalled if there are any other sevenspikes plugins already installed unfortunately with the new functionality we have no way to prevent our customers of uninstalling the core plugin by mistake and thus their website will not be able to be even started as there will be missing assemblies from the core plugin we also have other plugins that are dependent on each other i e ajax filters pro require the ajax filters to be installed first as otherwise they will not work etc so it would be great if we could tell nopcommerce if certain plugins could not be installed unistalled and return some message i e you could not uninstall core plugin as it is used by other plugins or you could not install ajax filters pro plugin as it required the ajax filters plugin to be installed first etc source to add to my previous post how we handled this in the previous version of nopcommerce in the install method of all our plugins we had an additional check shouldinstallplugin which checks that if the core plugin is not installed then it simply does not call the install method of the baseplugin we always inherit the baseplugin this way even without showing any messages to the clients at least we didn t allow them to install a plugin that will potentially break the site in the uninstall method of all our plugins we had an additional check shoulduninstallplugin if we should not uninstall a plugin we simply don t call the uninstall method of the baseplugin and thus it was never really uninstalled now in nopcommerce the logic of the actual installation uninstallation of the plugin is not in the plugin as it used to be in the baseplugin itself and we have no control over it the bottom line is that we need a way to prevent certain plugins to be added in the list of plugins that are to be uninstalled as once they go there and the application is restarted we could do nothing source
0
247,227
20,964,926,671
IssuesEvent
2022-03-28 05:13:20
kubeshop/testkube
https://api.github.com/repos/kubeshop/testkube
closed
Check testkube installation without MinIO - postman and Curl should work w/o it
🧪 tests
**Describe the bug** Make sure that it's working correctly > Is Minio/S3 necessary? It was mentioned that it’s used only at the moment for artifacts from Cypress, but helm installation assumes that some address will be configured. It would be good to disable Minio altogether, if it’s not planned to be used.
1.0
Check testkube installation without MinIO - postman and Curl should work w/o it - **Describe the bug** Make sure that it's working correctly > Is Minio/S3 necessary? It was mentioned that it’s used only at the moment for artifacts from Cypress, but helm installation assumes that some address will be configured. It would be good to disable Minio altogether, if it’s not planned to be used.
non_process
check testkube installation without minio postman and curl should work w o it describe the bug make sure that it s working correctly is minio necessary it was mentioned that it’s used only at the moment for artifacts from cypress but helm installation assumes that some address will be configured it would be good to disable minio altogether if it’s not planned to be used
0
578,630
17,149,199,398
IssuesEvent
2021-07-13 18:08:26
Couchers-org/couchers
https://api.github.com/repos/Couchers-org/couchers
closed
Picture upload issues
bug bug tool priority: critical
Subject: Invalid picture Description: Not able to upload pics Results: Uploaded several pics Backend version: develop-762ac12b Frontend version: develop-762ac12b User Agent: Mozilla/5.0 (iPhone; CPU iPhone OS 14_6 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/14.1.1 Mobile/15E148 Safari/604.1 Page: https://app.couchers.org/profile/edit User (spoofable): davidc (1250)
1.0
Picture upload issues - Subject: Invalid picture Description: Not able to upload pics Results: Uploaded several pics Backend version: develop-762ac12b Frontend version: develop-762ac12b User Agent: Mozilla/5.0 (iPhone; CPU iPhone OS 14_6 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/14.1.1 Mobile/15E148 Safari/604.1 Page: https://app.couchers.org/profile/edit User (spoofable): davidc (1250)
non_process
picture upload issues subject invalid picture description not able to upload pics results uploaded several pics backend version develop frontend version develop user agent mozilla iphone cpu iphone os like mac os x applewebkit khtml like gecko version mobile safari page user spoofable davidc
0
397,980
27,182,950,807
IssuesEvent
2023-02-18 21:34:47
DD2480G19/teammates
https://api.github.com/repos/DD2480G19/teammates
closed
Re-structure README to follow Report Template
documentation
Add the missing sections from the [report template](https://canvas.kth.se/courses/37918/files/6157559/download?download_frd=1) to `README`. In each section, refer to the corresponding files we have created. For example, under "complexity", add a reference to the `complexity analysis.md` file.
1.0
Re-structure README to follow Report Template - Add the missing sections from the [report template](https://canvas.kth.se/courses/37918/files/6157559/download?download_frd=1) to `README`. In each section, refer to the corresponding files we have created. For example, under "complexity", add a reference to the `complexity analysis.md` file.
non_process
re structure readme to follow report template add the missing sections from the to readme in each section refer to the corresponding files we have created for example under complexity add a reference to the complexity analysis md file
0
47,216
5,871,328,540
IssuesEvent
2017-05-15 08:24:32
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
EntryProcessorTest.testBackupMapEntryProcessorAllKeys
Team: Core Type: Test-Failure
``` junit.framework.AssertionFailedError: expected:<86> but was:<85> at junit.framework.Assert.fail(Assert.java:57) at junit.framework.Assert.failNotEquals(Assert.java:329) at junit.framework.Assert.assertEquals(Assert.java:78) at junit.framework.Assert.assertEquals(Assert.java:86) at junit.framework.TestCase.assertEquals(TestCase.java:253) at com.hazelcast.map.EntryProcessorTest.testBackupMapEntryProcessorAllKeys(EntryProcessorTest.java:481) ``` https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-IbmJDK1.7/com.hazelcast$hazelcast/627/testReport/junit/com.hazelcast.map/EntryProcessorTest/testBackupMapEntryProcessorAllKeys/
1.0
EntryProcessorTest.testBackupMapEntryProcessorAllKeys - ``` junit.framework.AssertionFailedError: expected:<86> but was:<85> at junit.framework.Assert.fail(Assert.java:57) at junit.framework.Assert.failNotEquals(Assert.java:329) at junit.framework.Assert.assertEquals(Assert.java:78) at junit.framework.Assert.assertEquals(Assert.java:86) at junit.framework.TestCase.assertEquals(TestCase.java:253) at com.hazelcast.map.EntryProcessorTest.testBackupMapEntryProcessorAllKeys(EntryProcessorTest.java:481) ``` https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-IbmJDK1.7/com.hazelcast$hazelcast/627/testReport/junit/com.hazelcast.map/EntryProcessorTest/testBackupMapEntryProcessorAllKeys/
non_process
entryprocessortest testbackupmapentryprocessorallkeys junit framework assertionfailederror expected but was at junit framework assert fail assert java at junit framework assert failnotequals assert java at junit framework assert assertequals assert java at junit framework assert assertequals assert java at junit framework testcase assertequals testcase java at com hazelcast map entryprocessortest testbackupmapentryprocessorallkeys entryprocessortest java
0
9,250
12,288,689,828
IssuesEvent
2020-05-09 17:53:19
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Unable to select "MTL.txt" files for GRASS "i.landsat.toar" processing tool
Bug Processing
**Describe the bug** Unable to select MTL.txt files for GRASS processing tool i.landsat.toar. Only "*.met" files can be selected as input. Although the description says that you can select the "MTL.txt" files. **How to Reproduce** 1. Go to 'Processing' toolbar. 2. Search for "i.landsat.toar" tool and open it. 3. Click browse button (3 dots "...") in the "Name of Landsat metadata file (.met or MTL.txt)" attribute. 4. In the new browse window try to select different (not "*.met") file format (you can't). ![2](https://user-images.githubusercontent.com/20655102/81403556-12138580-914d-11ea-8eaa-26bdcecc5cb5.jpg) ![1](https://user-images.githubusercontent.com/20655102/81403561-1b9ced80-914d-11ea-9809-e3b77faafcce.jpg) **QGIS and OS versions** QGIS version 3.12.2-București QGIS code revision 8a1fb33634 Compiled against Qt 5.11.2 Running against Qt 5.11.2 Compiled against GDAL/OGR 3.0.4 Running against GDAL/OGR 3.0.4 Compiled against GEOS 3.8.1-CAPI-1.13.3 Running against GEOS 3.8.1-CAPI-1.13.3 Compiled against SQLite 3.29.0 Running against SQLite 3.29.0 PostgreSQL Client Version 11.5 SpatiaLite Version 4.3.0 QWT Version 6.1.3 QScintilla2 Version 2.10.8 Compiled against PROJ 6.3.1 Running against PROJ Rel. 6.3.2, May 1st, 2020 OS Version Windows 10 (10.0) Active python plugins AppendFeaturesToLayer; DigitizingTools; imagefootprint_plugin; latlontools; qgsAzimuth; quick_map_services; valuetool; db_manager; MetaSearch; processing **Additional context**
1.0
Unable to select "MTL.txt" files for GRASS "i.landsat.toar" processing tool - **Describe the bug** Unable to select MTL.txt files for GRASS processing tool i.landsat.toar. Only "*.met" files can be selected as input. Although the description says that you can select the "MTL.txt" files. **How to Reproduce** 1. Go to 'Processing' toolbar. 2. Search for "i.landsat.toar" tool and open it. 3. Click browse button (3 dots "...") in the "Name of Landsat metadata file (.met or MTL.txt)" attribute. 4. In the new browse window try to select different (not "*.met") file format (you can't). ![2](https://user-images.githubusercontent.com/20655102/81403556-12138580-914d-11ea-8eaa-26bdcecc5cb5.jpg) ![1](https://user-images.githubusercontent.com/20655102/81403561-1b9ced80-914d-11ea-9809-e3b77faafcce.jpg) **QGIS and OS versions** QGIS version 3.12.2-București QGIS code revision 8a1fb33634 Compiled against Qt 5.11.2 Running against Qt 5.11.2 Compiled against GDAL/OGR 3.0.4 Running against GDAL/OGR 3.0.4 Compiled against GEOS 3.8.1-CAPI-1.13.3 Running against GEOS 3.8.1-CAPI-1.13.3 Compiled against SQLite 3.29.0 Running against SQLite 3.29.0 PostgreSQL Client Version 11.5 SpatiaLite Version 4.3.0 QWT Version 6.1.3 QScintilla2 Version 2.10.8 Compiled against PROJ 6.3.1 Running against PROJ Rel. 6.3.2, May 1st, 2020 OS Version Windows 10 (10.0) Active python plugins AppendFeaturesToLayer; DigitizingTools; imagefootprint_plugin; latlontools; qgsAzimuth; quick_map_services; valuetool; db_manager; MetaSearch; processing **Additional context**
process
unable to select mtl txt files for grass i landsat toar processing tool describe the bug unable to select mtl txt files for grass processing tool i landsat toar only met files can be selected as input although the description says that you can select the mtl txt files how to reproduce go to processing toolbar search for i landsat toar tool and open it click browse button dots in the name of landsat metadata file met or mtl txt attribute in the new browse window try to select different not met file format you can t qgis and os versions qgis version bucurești qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel may os version windows active python plugins appendfeaturestolayer digitizingtools imagefootprint plugin latlontools qgsazimuth quick map services valuetool db manager metasearch processing additional context
1
22,370
31,141,279,354
IssuesEvent
2023-08-16 00:21:00
googleapis/python-firestore
https://api.github.com/repos/googleapis/python-firestore
closed
Add GitHub action to inspect conformance test deltas
api: firestore type: process priority: p2
Add linting-style Action to block PRs if running the Makefile generates new diffs
1.0
Add GitHub action to inspect conformance test deltas - Add linting-style Action to block PRs if running the Makefile generates new diffs
process
add github action to inspect conformance test deltas add linting style action to block prs if running the makefile generates new diffs
1
121,246
25,943,006,991
IssuesEvent
2022-12-16 20:32:26
aws/aws-toolkit-vscode
https://api.github.com/repos/aws/aws-toolkit-vscode
closed
Nagging popup for CodeWhisperer appears on every VSCode window launch - even if not in use, cannot be permanently dismissed
bug pending-release codewhisperer
## Problem The AWS Toolkit extension has a nagging popup that occurs even when CodeWhisperer is disabled and cannot be permanently dismissed / disabled without removing the extension. ![SCR-20221128-lar](https://user-images.githubusercontent.com/862951/204382138-1e294b6d-0587-44df-8aa5-d75992bd8181.png) ## Steps to reproduce the issue 1. Install the AWS Toolkit extension 2. Ensure CodeWhisperer is disabled 3. Open a VSCode window 4. Get popup 5. Try to disable pop - there's no option ## Expected behaviour Any one, or all of: - Dismiss the popup once and never see it again. - If CodeWhipserer is not in use, the popup should never appear. - There should be an option to never show the popup again. ## System details (run the `AWS: About Toolkit` command) - - OS: Darwin arm64 22.1.0 - Visual Studio Code extension host: 1.73.1 - AWS Toolkit: 1.56.0 - node: 16.14.2 - electron: 19.0.17 --- ![image](https://user-images.githubusercontent.com/862951/204382979-b1562304-0816-446d-acd7-b09394881893.png)
1.0
Nagging popup for CodeWhisperer appears on every VSCode window launch - even if not in use, cannot be permanently dismissed - ## Problem The AWS Toolkit extension has a nagging popup that occurs even when CodeWhisperer is disabled and cannot be permanently dismissed / disabled without removing the extension. ![SCR-20221128-lar](https://user-images.githubusercontent.com/862951/204382138-1e294b6d-0587-44df-8aa5-d75992bd8181.png) ## Steps to reproduce the issue 1. Install the AWS Toolkit extension 2. Ensure CodeWhisperer is disabled 3. Open a VSCode window 4. Get popup 5. Try to disable pop - there's no option ## Expected behaviour Any one, or all of: - Dismiss the popup once and never see it again. - If CodeWhipserer is not in use, the popup should never appear. - There should be an option to never show the popup again. ## System details (run the `AWS: About Toolkit` command) - - OS: Darwin arm64 22.1.0 - Visual Studio Code extension host: 1.73.1 - AWS Toolkit: 1.56.0 - node: 16.14.2 - electron: 19.0.17 --- ![image](https://user-images.githubusercontent.com/862951/204382979-b1562304-0816-446d-acd7-b09394881893.png)
non_process
nagging popup for codewhisperer appears on every vscode window launch even if not in use cannot be permanently dismissed problem the aws toolkit extension has a nagging popup that occurs even when codewhisperer is disabled and cannot be permanently dismissed disabled without removing the extension steps to reproduce the issue install the aws toolkit extension ensure codewhisperer is disabled open a vscode window get popup try to disable pop there s no option expected behaviour any one or all of dismiss the popup once and never see it again if codewhipserer is not in use the popup should never appear there should be an option to never show the popup again system details run the aws about toolkit command os darwin visual studio code extension host aws toolkit node electron
0
31,599
8,717,608,724
IssuesEvent
2018-12-07 17:41:05
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
TST: Azure CI for Windows testing
Build issues CI
Since our Appveyor runs are rather slow I've been working on getting Azure CI working for SciPy Windows testing. To reduce noise / CI burden I've been doing this in a [PR in my fork](https://github.com/tylerjereddy/scipy/pull/20), where you can see the current diff & Azure CI results. I've struggled for several hours trying to get openblas to work properly--I note that [our instructions for installing openblas](https://scipy.github.io/devdocs/building/windows.html#installing-openblas) point to a `FOUND` output with `language = c`, while `language = f77` in current appveyor builds, but I see `c` in my current build attempts on Azure (and the library name gets duplicated in the list on Azure too). There are also apparent differences in the files / static libraries used for our master branch appveyor runs vs. macpython wheel building on appveyor & I've tried copying the static libraries from both URLs without full success yet. The other question is whether the pre-installed mingw / gfortran on the Azure images is compatible with this "just copy the static library(ies)" from openblas workflow. The Appveyor file suggests that we should `install OpenBLAS.dll to MINGW\bin`, but actually doesn't appear to do this in our master branch CI. I'm also not entirely clear on why we need to replace NumPy distutils in our builds using the master branch--is this still true / helpful? Presumably the changes to distutils will eventually show up in a release and we can remove that stuff?
1.0
TST: Azure CI for Windows testing - Since our Appveyor runs are rather slow I've been working on getting Azure CI working for SciPy Windows testing. To reduce noise / CI burden I've been doing this in a [PR in my fork](https://github.com/tylerjereddy/scipy/pull/20), where you can see the current diff & Azure CI results. I've struggled for several hours trying to get openblas to work properly--I note that [our instructions for installing openblas](https://scipy.github.io/devdocs/building/windows.html#installing-openblas) point to a `FOUND` output with `language = c`, while `language = f77` in current appveyor builds, but I see `c` in my current build attempts on Azure (and the library name gets duplicated in the list on Azure too). There are also apparent differences in the files / static libraries used for our master branch appveyor runs vs. macpython wheel building on appveyor & I've tried copying the static libraries from both URLs without full success yet. The other question is whether the pre-installed mingw / gfortran on the Azure images is compatible with this "just copy the static library(ies)" from openblas workflow. The Appveyor file suggests that we should `install OpenBLAS.dll to MINGW\bin`, but actually doesn't appear to do this in our master branch CI. I'm also not entirely clear on why we need to replace NumPy distutils in our builds using the master branch--is this still true / helpful? Presumably the changes to distutils will eventually show up in a release and we can remove that stuff?
non_process
tst azure ci for windows testing since our appveyor runs are rather slow i ve been working on getting azure ci working for scipy windows testing to reduce noise ci burden i ve been doing this in a where you can see the current diff azure ci results i ve struggled for several hours trying to get openblas to work properly i note that point to a found output with language c while language in current appveyor builds but i see c in my current build attempts on azure and the library name gets duplicated in the list on azure too there are also apparent differences in the files static libraries used for our master branch appveyor runs vs macpython wheel building on appveyor i ve tried copying the static libraries from both urls without full success yet the other question is whether the pre installed mingw gfortran on the azure images is compatible with this just copy the static library ies from openblas workflow the appveyor file suggests that we should install openblas dll to mingw bin but actually doesn t appear to do this in our master branch ci i m also not entirely clear on why we need to replace numpy distutils in our builds using the master branch is this still true helpful presumably the changes to distutils will eventually show up in a release and we can remove that stuff
0
714,540
24,565,689,580
IssuesEvent
2022-10-13 02:38:26
autowarefoundation/autoware.universe
https://api.github.com/repos/autowarefoundation/autoware.universe
closed
Sometimes the sensing module doesn't publish concatenated pointcloud
bug help wanted low priority sensing
### Checklist - [X] I've read the [contribution guidelines](https://github.com/autowarefoundation/autoware/blob/main/CONTRIBUTING.md). - [X] I've searched other issues and no duplicate issues were found. - [X] I'm convinced that this is not my fault but a bug. ### Description sometimes the sensing module cannot launch concatenated and cropbox filter with the following error ``` [ERROR] [launch_ros.actions.load_composable_nodes]: Failed to load node 'concatenate_data' of type 'pointcloud_preprocessor::PointCloudConcatenateDataSynchronizerComponent' in container 'pointcloud_container': Failed to find class with the requested plugin name. [ERROR] [launch_ros.actions.load_composable_nodes]: Failed to load node 'crop_box_filter' of type 'pointcloud_preprocessor::CropBoxFilterComponent' in container 'pointcloud_container': Failed to find class with the requested plugin name. ``` ### Expected behavior launch concatenated and cropbox filter without error ### Actual behavior launch concatenated and cropbox filter without error ### Steps to reproduce play rosbag replay simulator in tutorials ### Versions _No response_ ### Possible causes _No response_ ### Additional context _No response_ ### Temporary fix - [ ] #1797
1.0
Sometimes the sensing module doesn't publish concatenated pointcloud - ### Checklist - [X] I've read the [contribution guidelines](https://github.com/autowarefoundation/autoware/blob/main/CONTRIBUTING.md). - [X] I've searched other issues and no duplicate issues were found. - [X] I'm convinced that this is not my fault but a bug. ### Description sometimes the sensing module cannot launch concatenated and cropbox filter with the following error ``` [ERROR] [launch_ros.actions.load_composable_nodes]: Failed to load node 'concatenate_data' of type 'pointcloud_preprocessor::PointCloudConcatenateDataSynchronizerComponent' in container 'pointcloud_container': Failed to find class with the requested plugin name. [ERROR] [launch_ros.actions.load_composable_nodes]: Failed to load node 'crop_box_filter' of type 'pointcloud_preprocessor::CropBoxFilterComponent' in container 'pointcloud_container': Failed to find class with the requested plugin name. ``` ### Expected behavior launch concatenated and cropbox filter without error ### Actual behavior launch concatenated and cropbox filter without error ### Steps to reproduce play rosbag replay simulator in tutorials ### Versions _No response_ ### Possible causes _No response_ ### Additional context _No response_ ### Temporary fix - [ ] #1797
non_process
sometimes the sensing module doesn t publish concatenated pointcloud checklist i ve read the i ve searched other issues and no duplicate issues were found i m convinced that this is not my fault but a bug description sometimes the sensing module cannot launch concatenated and cropbox filter with the following error failed to load node concatenate data of type pointcloud preprocessor pointcloudconcatenatedatasynchronizercomponent in container pointcloud container failed to find class with the requested plugin name failed to load node crop box filter of type pointcloud preprocessor cropboxfiltercomponent in container pointcloud container failed to find class with the requested plugin name expected behavior launch concatenated and cropbox filter without error actual behavior launch concatenated and cropbox filter without error steps to reproduce play rosbag replay simulator in tutorials versions no response possible causes no response additional context no response temporary fix
0
182,541
6,671,086,586
IssuesEvent
2017-10-04 04:45:58
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.cinecalidad.to - see bug description
browser-firefox priority-normal status-needstriage
<!-- @browser: Firefox 57.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:57.0) Gecko/20100101 Firefox/57.0 --> <!-- @reported_with: web --> **URL**: http://www.cinecalidad.to **Browser / Version**: Firefox 57.0 **Operating System**: Windows 10 **Tested Another Browser**: No **Problem type**: Something else **Description**: it's difficult to play videos **Steps to Reproduce**: with https://www.alexa.com/topsites/countries/GT _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.cinecalidad.to - see bug description - <!-- @browser: Firefox 57.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:57.0) Gecko/20100101 Firefox/57.0 --> <!-- @reported_with: web --> **URL**: http://www.cinecalidad.to **Browser / Version**: Firefox 57.0 **Operating System**: Windows 10 **Tested Another Browser**: No **Problem type**: Something else **Description**: it's difficult to play videos **Steps to Reproduce**: with https://www.alexa.com/topsites/countries/GT _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
see bug description url browser version firefox operating system windows tested another browser no problem type something else description it s difficult to play videos steps to reproduce with from with ❤️
0
804
3,283,343,467
IssuesEvent
2015-10-28 12:10:30
superroma/testcafe-hammerhead
https://api.github.com/repos/superroma/testcafe-hammerhead
closed
Create js-processor modifiers filter based on node type.
AREA: server SYSTEM: resource processing TYPE: enhancement
Currently each modifier condition is called for each node in the AST (see: https://github.com/churkin/testcafe-hammerhead/blob/master/src/processing/js/modifiers.js). We can significantly reduce complexity of the modifier search if we split modifiers into the groups by node type. So using this kind of filter will reduce calls to `condition()`
1.0
Create js-processor modifiers filter based on node type. - Currently each modifier condition is called for each node in the AST (see: https://github.com/churkin/testcafe-hammerhead/blob/master/src/processing/js/modifiers.js). We can significantly reduce complexity of the modifier search if we split modifiers into the groups by node type. So using this kind of filter will reduce calls to `condition()`
process
create js processor modifiers filter based on node type currently each modifier condition is called for each node in the ast see we can significantly reduce complexity of the modifier search if we split modifiers into the groups by node type so using this kind of filter will reduce calls to condition
1
808,691
30,107,657,952
IssuesEvent
2023-06-30 03:51:52
XRPLF/rippled
https://api.github.com/repos/XRPLF/rippled
closed
Proposal: SetFreezeTransferRate
Feature Request Reviewed Low Priority
Please see https://forum.ripple.com/viewtopic.php?f=1&t=18307 Please discuss why specifically this may not be feasible due to technical & legal reasoning.
1.0
Proposal: SetFreezeTransferRate - Please see https://forum.ripple.com/viewtopic.php?f=1&t=18307 Please discuss why specifically this may not be feasible due to technical & legal reasoning.
non_process
proposal setfreezetransferrate please see please discuss why specifically this may not be feasible due to technical legal reasoning
0
138,564
20,608,802,609
IssuesEvent
2022-03-07 05:42:10
chemins-de-la-transition/semapps
https://api.github.com/repos/chemins-de-la-transition/semapps
opened
Création page profil acteur : mode édition / lecture
back-office front-office design frontend
CF issue #156 qui évoque déjà mon point de vue sur la nécessité de guider avec un tutoriel les acteurs pour créer leurs Lieux/évènements/voyages
1.0
Création page profil acteur : mode édition / lecture - CF issue #156 qui évoque déjà mon point de vue sur la nécessité de guider avec un tutoriel les acteurs pour créer leurs Lieux/évènements/voyages
non_process
création page profil acteur mode édition lecture cf issue qui évoque déjà mon point de vue sur la nécessité de guider avec un tutoriel les acteurs pour créer leurs lieux évènements voyages
0
6,357
9,415,880,480
IssuesEvent
2019-04-10 13:36:11
brandon1roadgears/Interpreter-of-programming-language-of-Turing-Machine
https://api.github.com/repos/brandon1roadgears/Interpreter-of-programming-language-of-Turing-Machine
opened
Написать функцию для ввода проавил
C++ Work in process
#### Требуется реализовать функцию, которая позволит пользователю вводить правила. Думаю стоит написать структуру, состоящую из 5 полей. Далее создать массив структур. И вводить данные пока next_state != "##" `struct action` `{` `char state[3];` `char expected_symbol[2];` `char new_symbol[2];` `char move[2];` `char next_state;` `};`
1.0
Написать функцию для ввода проавил - #### Требуется реализовать функцию, которая позволит пользователю вводить правила. Думаю стоит написать структуру, состоящую из 5 полей. Далее создать массив структур. И вводить данные пока next_state != "##" `struct action` `{` `char state[3];` `char expected_symbol[2];` `char new_symbol[2];` `char move[2];` `char next_state;` `};`
process
написать функцию для ввода проавил требуется реализовать функцию которая позволит пользователю вводить правила думаю стоит написать структуру состоящую из полей далее создать массив структур и вводить данные пока next state struct action char state char expected symbol char new symbol char move char next state
1
169,274
6,398,022,178
IssuesEvent
2017-08-04 19:29:54
python/mypy
https://api.github.com/repos/python/mypy
closed
error: "type" has no attribute "var"
bug priority-1-normal
test.py: ```python class A(object): var = 2 class B(A): var = 5 for cls in [A, B]: print(cls.var) ``` ``` $ mypy test.py test.py:9: error: "type" has no attribute "var" ``` ``` $ python --version Python 3.5.2 ``` ``` $ mypy --version mypy 0.470 ``` test.py: ```python class A(object): var = 2 # type: int class B(A): var = 5 # type: int for cls in [A, B]: print(cls.var) ``` ``` test.py:9: error: "type" has no attribute "var" ```
1.0
error: "type" has no attribute "var" - test.py: ```python class A(object): var = 2 class B(A): var = 5 for cls in [A, B]: print(cls.var) ``` ``` $ mypy test.py test.py:9: error: "type" has no attribute "var" ``` ``` $ python --version Python 3.5.2 ``` ``` $ mypy --version mypy 0.470 ``` test.py: ```python class A(object): var = 2 # type: int class B(A): var = 5 # type: int for cls in [A, B]: print(cls.var) ``` ``` test.py:9: error: "type" has no attribute "var" ```
non_process
error type has no attribute var test py python class a object var class b a var for cls in print cls var mypy test py test py error type has no attribute var python version python mypy version mypy test py python class a object var type int class b a var type int for cls in print cls var test py error type has no attribute var
0
204,539
15,501,060,583
IssuesEvent
2021-03-11 10:04:32
openmsupply/mobile
https://api.github.com/repos/openmsupply/mobile
closed
TESTING: Supplier Invoice Page
Needs testing Test Plan Template
# TESTING This is an issue for testing `SupplierInvoicePage`. Please copy and paste this list of tests and post your results in a comment in this issue. ## Supplier Invoice Page (SI) #### TESTER NOTES: - Have an item name that is very long (e.g. `~50` characters) # TESTING ### Special cases - [ ] Adjustments made to stock in this page are NOT reflected instantly in total stock on hand (create supplier invoice -> add some quantities -> DO NOT finalise -> check current stock) - [ ] Adjustments made to stock in this page are reflected in total stock in hand if changes are finalised (create supplier invoice -> add some quantities -> finalise -> check current stock) - [ ] If transaction is a supplier invoice, top left displays '"Invoice X"' where `X` is the current invoice number - [ ] If transaction is a supplier credit, top left displays '"Credit X"' where `X` is the current credit number ### Navigation to/from - [x] Can navigate to this page through the supplier invoice page (by clicking on a row) - [x] Can navigate to this page through creating a new supplier invoice ### Sync - [] Changes made to this supplier invoice are synced to desktop (finalise or add a comment to an existing supplier invoice -> trigger manual sync -> changes should be reflected in desktop) ### Table #### NOTE: LINES IN THIS TABLE ARE BY BATCH - NOT ITEM - [ ] Lists all batches related to this supplier invoice - [ ] Initial sort is by item name in ascending order #### Row Press - [ ] Pressing on a row has no effect #### Filtering - [ ] Batches are filtered by item name - [ ] Filtering is cleared when the search term is cleared #### Columns - [ ] **ITEM CODE** - [ ] Sortable (and sorts correctly) - [ ] Batches have the correct item code - [ ] Sorting icon updates correctly to indicate ascending/descending and no sorting - [ ] Cells NOT editable. - [ ] Left aligned - [ ] **ITEM NAME** - [ ] Sortable (and sorts correctly) - [ ] Batches have the correct item name - [ ] Sorting icon updates correctly to indicate ascending/descending and no sorting - [ ] Cells NOT editable - [ ] Text overflow handled (create an item name that is very long) - [ ] Left aligned - [ ] **BATCH NAME** - [ ] Sortable (and sorts correctly) - [ ] Sorting icon updates correctly to indicate ascending/descending and no sorting - [ ] If supplier invoice/credit is NOT finalised, is editable - [ ] If supplier invoice/credit is NOT, is NOT editable - [ ] Left aligned - [ ] **QUANTITY** - [ ] Sortable (and sorts correctly) - [ ] Editable - [ ] NOT editable when finalised - [ ] Finalising this invoice increases stock on hand for all batches and quantities entered correctly - [ ] Cannot enter negative numbers - [ ] Right aligned - [ ] **EXPIRY** (Note: field is updated after submitting/losing focus) - [ ] NOT sortable. - [ ] If not focussed, valued is always in the format `MM/YYYY` - [ ] Cannot enter invalid values (e.g. `13/2000`) - [ ] When date is invalid, unfocusing resets the value - [ ] Center aligned - [ ] Is DISABLED when finalised - [ ] Placeholder text is `mm/yyyy` - [ ] Placeholder text is distinguishable from values (lighter colour) - [ ] **COST PRICE** - [ ] NOT sortable - [ ] Batches have the correct cost price - [ ] NOT editable - [ ] Right aligned - [ ] ** SELL PRICE** - [ ] NOT sortable - [ ] Editable - [ ] NOT editable if finalised - [ ] Right aligned - [ ] **Remove** - [ ] NOT sortable - [ ] Cells NOT editable - [ ] Can select any row - [ ] Can select multiple rows - [ ] Selecting a row opens a confirmation dialog at the bottom of the page - [ ] Unselecting a row closes the confirmation dialog at the bottom of the page - [ ] Confirming the confirmation dialog deletes the batches - [ ] Cancelling the confirmation dialog removes the current selection and closes the confirmation dialog - [ ] Is DISABLED when finalised ### Buttons - [ ] `"New line"` - [ ] Button is displayed - [ ] On press, `"Select an item"` dialog is displayed ### Page Info - [ ] **Entry Date** - [ ] If supplier invoice, shows the correct entry date for the invoice - [ ] If supplier credit, shows the correct entry date for the credit - [ ] **Confirm Date** - [ ] If supplier invoice, shows the correct date for confirming the invoice (on mobile this is equivalent to finalising) - [ ] If supplier credit, shows the correct date for confirming the credit (on mobile this is equivalent to finalising) - [ ] **Entered By** - [ ] If supplier invoice, shows the correct user who created the invoice - [ ] If supplier credit, shows the correct user who created the credit - [ ] **Supplier** - [ ] If supplier invoice, shows the correct supplier for the invoice - [ ] If supplier credit, shows the correct supplier for the invoice - [ ] **Their Ref** - [ ] If supplier invoice, displays the correct `Their Ref` field for the invoice - [ ] If supplier credit, displays the correct `Their Ref` field for the credit - [ ] On press, `"Their ref"` modal (see modals) is displayed - [ ] Is DISABLED when finalised. - [ ] **Comment** - [ ] If supplier invoice, displays the correct comment field for the invoice - [ ] If supplier credit, displays the correct comment field for the credit - [ ] Clicking opens the comment modal (see modals) - [ ] Is disabled when finalised. ### Modals - [ ] **New line** - [ ] Title: `"Search for an Item to add"` - [ ] List includes ALL items - [ ] Search bar is auto focused on opening - [ ] Closing the modal works, i.e. closes the modal - [ ] List is filtered by item name - [ ] List is filtered by Item code - [ ] Clearing the search term clears filtering - [ ] Right hand side column of the list of options lists the correct available quantity - [ ] Clicking an item adds a new batch to the invoice for that item - [ ] **Comment** - [ ] Title: `"Enter the invoice comment"` - [ ] If there is already a comment, it is entered already. - [ ] Empty if no comment - [ ] Entering a comment and pressing `"OK"` saves the comment correctly - [ ] Closing the modal works, i.e. closes the modal - [ ] **Their ref** - [ ] Title: `"Enter their reference"` - [ ] If reference already entered, reference is displayed correctly - [ ] If no reference, no reference is displayed - [ ] Entering a reference and pressing `"OK"` saves the reference correctly - [ ] Closing the modal works, i.e. closes the modal ### Finalising - [ ] Pressing finalise opens a confirmation dialog warning about locking the invoice - [ ] Cancelling the confirmation cancels finalizing the invoice - [ ] Confirming the confirmation dialog finalises the invoice, locking all edits (as listed above) - [ ] Finalising the invoice correctly sets the status of the invoice to finalised - [ ] All `0` quantity items are pruned from the invoice, such that only items with a quantity more than `0` remain
2.0
TESTING: Supplier Invoice Page - # TESTING This is an issue for testing `SupplierInvoicePage`. Please copy and paste this list of tests and post your results in a comment in this issue. ## Supplier Invoice Page (SI) #### TESTER NOTES: - Have an item name that is very long (e.g. `~50` characters) # TESTING ### Special cases - [ ] Adjustments made to stock in this page are NOT reflected instantly in total stock on hand (create supplier invoice -> add some quantities -> DO NOT finalise -> check current stock) - [ ] Adjustments made to stock in this page are reflected in total stock in hand if changes are finalised (create supplier invoice -> add some quantities -> finalise -> check current stock) - [ ] If transaction is a supplier invoice, top left displays '"Invoice X"' where `X` is the current invoice number - [ ] If transaction is a supplier credit, top left displays '"Credit X"' where `X` is the current credit number ### Navigation to/from - [x] Can navigate to this page through the supplier invoice page (by clicking on a row) - [x] Can navigate to this page through creating a new supplier invoice ### Sync - [] Changes made to this supplier invoice are synced to desktop (finalise or add a comment to an existing supplier invoice -> trigger manual sync -> changes should be reflected in desktop) ### Table #### NOTE: LINES IN THIS TABLE ARE BY BATCH - NOT ITEM - [ ] Lists all batches related to this supplier invoice - [ ] Initial sort is by item name in ascending order #### Row Press - [ ] Pressing on a row has no effect #### Filtering - [ ] Batches are filtered by item name - [ ] Filtering is cleared when the search term is cleared #### Columns - [ ] **ITEM CODE** - [ ] Sortable (and sorts correctly) - [ ] Batches have the correct item code - [ ] Sorting icon updates correctly to indicate ascending/descending and no sorting - [ ] Cells NOT editable. - [ ] Left aligned - [ ] **ITEM NAME** - [ ] Sortable (and sorts correctly) - [ ] Batches have the correct item name - [ ] Sorting icon updates correctly to indicate ascending/descending and no sorting - [ ] Cells NOT editable - [ ] Text overflow handled (create an item name that is very long) - [ ] Left aligned - [ ] **BATCH NAME** - [ ] Sortable (and sorts correctly) - [ ] Sorting icon updates correctly to indicate ascending/descending and no sorting - [ ] If supplier invoice/credit is NOT finalised, is editable - [ ] If supplier invoice/credit is NOT, is NOT editable - [ ] Left aligned - [ ] **QUANTITY** - [ ] Sortable (and sorts correctly) - [ ] Editable - [ ] NOT editable when finalised - [ ] Finalising this invoice increases stock on hand for all batches and quantities entered correctly - [ ] Cannot enter negative numbers - [ ] Right aligned - [ ] **EXPIRY** (Note: field is updated after submitting/losing focus) - [ ] NOT sortable. - [ ] If not focussed, valued is always in the format `MM/YYYY` - [ ] Cannot enter invalid values (e.g. `13/2000`) - [ ] When date is invalid, unfocusing resets the value - [ ] Center aligned - [ ] Is DISABLED when finalised - [ ] Placeholder text is `mm/yyyy` - [ ] Placeholder text is distinguishable from values (lighter colour) - [ ] **COST PRICE** - [ ] NOT sortable - [ ] Batches have the correct cost price - [ ] NOT editable - [ ] Right aligned - [ ] ** SELL PRICE** - [ ] NOT sortable - [ ] Editable - [ ] NOT editable if finalised - [ ] Right aligned - [ ] **Remove** - [ ] NOT sortable - [ ] Cells NOT editable - [ ] Can select any row - [ ] Can select multiple rows - [ ] Selecting a row opens a confirmation dialog at the bottom of the page - [ ] Unselecting a row closes the confirmation dialog at the bottom of the page - [ ] Confirming the confirmation dialog deletes the batches - [ ] Cancelling the confirmation dialog removes the current selection and closes the confirmation dialog - [ ] Is DISABLED when finalised ### Buttons - [ ] `"New line"` - [ ] Button is displayed - [ ] On press, `"Select an item"` dialog is displayed ### Page Info - [ ] **Entry Date** - [ ] If supplier invoice, shows the correct entry date for the invoice - [ ] If supplier credit, shows the correct entry date for the credit - [ ] **Confirm Date** - [ ] If supplier invoice, shows the correct date for confirming the invoice (on mobile this is equivalent to finalising) - [ ] If supplier credit, shows the correct date for confirming the credit (on mobile this is equivalent to finalising) - [ ] **Entered By** - [ ] If supplier invoice, shows the correct user who created the invoice - [ ] If supplier credit, shows the correct user who created the credit - [ ] **Supplier** - [ ] If supplier invoice, shows the correct supplier for the invoice - [ ] If supplier credit, shows the correct supplier for the invoice - [ ] **Their Ref** - [ ] If supplier invoice, displays the correct `Their Ref` field for the invoice - [ ] If supplier credit, displays the correct `Their Ref` field for the credit - [ ] On press, `"Their ref"` modal (see modals) is displayed - [ ] Is DISABLED when finalised. - [ ] **Comment** - [ ] If supplier invoice, displays the correct comment field for the invoice - [ ] If supplier credit, displays the correct comment field for the credit - [ ] Clicking opens the comment modal (see modals) - [ ] Is disabled when finalised. ### Modals - [ ] **New line** - [ ] Title: `"Search for an Item to add"` - [ ] List includes ALL items - [ ] Search bar is auto focused on opening - [ ] Closing the modal works, i.e. closes the modal - [ ] List is filtered by item name - [ ] List is filtered by Item code - [ ] Clearing the search term clears filtering - [ ] Right hand side column of the list of options lists the correct available quantity - [ ] Clicking an item adds a new batch to the invoice for that item - [ ] **Comment** - [ ] Title: `"Enter the invoice comment"` - [ ] If there is already a comment, it is entered already. - [ ] Empty if no comment - [ ] Entering a comment and pressing `"OK"` saves the comment correctly - [ ] Closing the modal works, i.e. closes the modal - [ ] **Their ref** - [ ] Title: `"Enter their reference"` - [ ] If reference already entered, reference is displayed correctly - [ ] If no reference, no reference is displayed - [ ] Entering a reference and pressing `"OK"` saves the reference correctly - [ ] Closing the modal works, i.e. closes the modal ### Finalising - [ ] Pressing finalise opens a confirmation dialog warning about locking the invoice - [ ] Cancelling the confirmation cancels finalizing the invoice - [ ] Confirming the confirmation dialog finalises the invoice, locking all edits (as listed above) - [ ] Finalising the invoice correctly sets the status of the invoice to finalised - [ ] All `0` quantity items are pruned from the invoice, such that only items with a quantity more than `0` remain
non_process
testing supplier invoice page testing this is an issue for testing supplierinvoicepage please copy and paste this list of tests and post your results in a comment in this issue supplier invoice page si tester notes have an item name that is very long e g characters testing special cases adjustments made to stock in this page are not reflected instantly in total stock on hand create supplier invoice add some quantities do not finalise check current stock adjustments made to stock in this page are reflected in total stock in hand if changes are finalised create supplier invoice add some quantities finalise check current stock if transaction is a supplier invoice top left displays invoice x where x is the current invoice number if transaction is a supplier credit top left displays credit x where x is the current credit number navigation to from can navigate to this page through the supplier invoice page by clicking on a row can navigate to this page through creating a new supplier invoice sync changes made to this supplier invoice are synced to desktop finalise or add a comment to an existing supplier invoice trigger manual sync changes should be reflected in desktop table note lines in this table are by batch not item lists all batches related to this supplier invoice initial sort is by item name in ascending order row press pressing on a row has no effect filtering batches are filtered by item name filtering is cleared when the search term is cleared columns item code sortable and sorts correctly batches have the correct item code sorting icon updates correctly to indicate ascending descending and no sorting cells not editable left aligned item name sortable and sorts correctly batches have the correct item name sorting icon updates correctly to indicate ascending descending and no sorting cells not editable text overflow handled create an item name that is very long left aligned batch name sortable and sorts correctly sorting icon updates correctly to indicate ascending descending and no sorting if supplier invoice credit is not finalised is editable if supplier invoice credit is not is not editable left aligned quantity sortable and sorts correctly editable not editable when finalised finalising this invoice increases stock on hand for all batches and quantities entered correctly cannot enter negative numbers right aligned expiry note field is updated after submitting losing focus not sortable if not focussed valued is always in the format mm yyyy cannot enter invalid values e g when date is invalid unfocusing resets the value center aligned is disabled when finalised placeholder text is mm yyyy placeholder text is distinguishable from values lighter colour cost price not sortable batches have the correct cost price not editable right aligned sell price not sortable editable not editable if finalised right aligned remove not sortable cells not editable can select any row can select multiple rows selecting a row opens a confirmation dialog at the bottom of the page unselecting a row closes the confirmation dialog at the bottom of the page confirming the confirmation dialog deletes the batches cancelling the confirmation dialog removes the current selection and closes the confirmation dialog is disabled when finalised buttons new line button is displayed on press select an item dialog is displayed page info entry date if supplier invoice shows the correct entry date for the invoice if supplier credit shows the correct entry date for the credit confirm date if supplier invoice shows the correct date for confirming the invoice on mobile this is equivalent to finalising if supplier credit shows the correct date for confirming the credit on mobile this is equivalent to finalising entered by if supplier invoice shows the correct user who created the invoice if supplier credit shows the correct user who created the credit supplier if supplier invoice shows the correct supplier for the invoice if supplier credit shows the correct supplier for the invoice their ref if supplier invoice displays the correct their ref field for the invoice if supplier credit displays the correct their ref field for the credit on press their ref modal see modals is displayed is disabled when finalised comment if supplier invoice displays the correct comment field for the invoice if supplier credit displays the correct comment field for the credit clicking opens the comment modal see modals is disabled when finalised modals new line title search for an item to add list includes all items search bar is auto focused on opening closing the modal works i e closes the modal list is filtered by item name list is filtered by item code clearing the search term clears filtering right hand side column of the list of options lists the correct available quantity clicking an item adds a new batch to the invoice for that item comment title enter the invoice comment if there is already a comment it is entered already empty if no comment entering a comment and pressing ok saves the comment correctly closing the modal works i e closes the modal their ref title enter their reference if reference already entered reference is displayed correctly if no reference no reference is displayed entering a reference and pressing ok saves the reference correctly closing the modal works i e closes the modal finalising pressing finalise opens a confirmation dialog warning about locking the invoice cancelling the confirmation cancels finalizing the invoice confirming the confirmation dialog finalises the invoice locking all edits as listed above finalising the invoice correctly sets the status of the invoice to finalised all quantity items are pruned from the invoice such that only items with a quantity more than remain
0
16,132
6,108,275,903
IssuesEvent
2017-06-21 10:05:16
systemd/systemd
https://api.github.com/repos/systemd/systemd
closed
build with meson fails with ld.bfd (using -Dlink-udev-shared=false -Db_lto=true)
build-system has-pr meson
### Submission type - [X] Bug report - [ ] Request for enhancement (RFE) ### systemd version the issue has been seen with > git master I reconfigured the xenial-s390x CI build to use meson instead of autotools. This currently fails to build ([example log](https://objectstorage.prodstack4-5.canonical.com/v1/AUTH_77e2ada1e7a84929a74ba3b87153c0ac/autopkgtest-xenial-pitti-systemd-semaphore/xenial/s390x/s/systemd-upstream/20170427_101741_d8343@/log.gz)), mostly on missing libpthread and libselinux libs, but also a rather weird size check failure: ``` In function ‘__read_alias’, inlined from ‘compress_stream_xz’ at ../src/journal/compress.c:398:27: /usr/include/s390x-linux-gnu/bits/unistd.h:39:9: warning: call to ‘__read_chk_warn’ declared with attribute warning: read called with bigger length than size of the destination buffer return __read_chk (__fd, __buf, __nbytes, __bos0 (__buf)); ^ In function ‘__ppoll_alias’, inlined from ‘bus_poll’ at ../src/libsystemd/sd-bus/sd-bus.c:2915:11: /usr/include/s390x-linux-gnu/bits/poll2.h:71:9: warning: call to ‘__ppoll_chk_warn’ declared with attribute warning: ppoll called with fds buffer too small file nfds entries return __ppoll_chk (__fds, __nfds, __timeout, __ss, __bos (__fds)); FAILED: systemd-getty-generator cc -o systemd-getty-generator 'systemd-getty-generator@exe/src_getty-generator_getty-generator.c.o' '-flto' '-Wl,--no-undefined' '-Wl,--as-needed' '-Wl,-z,relro' '-Wl,-z,now' '-pie' '-Wl,--gc-sections' '-g' '-O2' '-fstack-protector-strong' '-Wformat' '-Werror=format-security' '-Wl,-Bsymbolic-functions' '-Wl,-z,relro' 'src/shared/libsystemd-shared-233.so' 'src/libudev/libudev.so.1.6.6' 'src/basic/libbasic.a' 'src/libsystemd/libsystemd.a' 'src/basic/libbasic.a' 'src/basic/libbasic.a' '-Wl,-rpath,/data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/src/shared:/data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/src/libudev' /tmp/cclLTKce.ltrans3.ltrans.o: In function `mac_selinux_create_file_clear': /data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/../src/basic/selinux-util.c:374: undefined reference to `setfscreatecon_raw' /tmp/cclLTKce.ltrans3.ltrans.o: In function `mac_selinux_have': /data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/../src/basic/selinux-util.c:59: undefined reference to `is_selinux_enabled' collect2: error: ld returned 1 exit status FAILED: systemd-sysusers cc -o systemd-sysusers 'systemd-sysusers@exe/src_sysusers_sysusers.c.o' '-flto' '-Wl,--no-undefined' '-Wl,--as-needed' '-Wl,-z,relro' '-Wl,-z,now' '-pie' '-Wl,--gc-sections' '-g' '-O2' '-fstack-protector-strong' '-Wformat' '-Werror=format-security' '-Wl,-Bsymbolic-functions' '-Wl,-z,relro' 'src/shared/libsystemd-shared-233.so' 'src/libudev/libudev.so.1.6.6' 'src/basic/libbasic.a' 'src/libsystemd/libsystemd.a' 'src/basic/libbasic.a' 'src/basic/libbasic.a' '-Wl,-rpath,/data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/src/shared:/data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/src/libudev' /usr/bin/ld: /tmp/ccpz1SmY.ltrans0.ltrans.o: undefined reference to symbol 'selabel_open' //lib/s390x-linux-gnu/libselinux.so.1: error adding symbols: DSO missing from command line collect2: error: ld returned 1 exit status FAILED: busctl cc -o busctl 'busctl@exe/src_busctl_busctl.c.o' 'busctl@exe/src_busctl_busctl-introspect.c.o' '-flto' '-Wl,--no-undefined' '-Wl,--as-needed' '-Wl,-z,relro' '-Wl,-z,now' '-pie' '-Wl,--gc-sections' '-g' '-O2' '-fstack-protector-strong' '-Wformat' '-Werror=format-security' '-Wl,-Bsymbolic-functions' '-Wl,-z,relro' 'src/shared/libsystemd-shared-233.so' 'src/libudev/libudev.so.1.6.6' 'src/basic/libbasic.a' 'src/libsystemd/libsystemd.a' 'src/basic/libbasic.a' 'src/basic/libbasic.a' '-Wl,-rpath,/data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/src/shared:/data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/src/libudev' In function ‘__ppoll_alias’, inlined from ‘bus_poll’ at ../src/libsystemd/sd-bus/sd-bus.c:2915:11: /usr/include/s390x-linux-gnu/bits/poll2.h:71:9: warning: call to ‘__ppoll_chk_warn’ declared with attribute warning: ppoll called with fds buffer too small file nfds entries return __ppoll_chk (__fds, __nfds, __timeout, __ss, __bos (__fds)); ^ /usr/bin/ld: /tmp/ccXySaHU.ltrans17.ltrans.o: undefined reference to symbol 'is_selinux_enabled' //lib/s390x-linux-gnu/libselinux.so.1: error adding symbols: DSO missing from command line AILED: systemd-run cc -o systemd-run 'systemd-run@exe/src_run_run.c.o' '-flto' '-Wl,--no-undefined' '-Wl,--as-needed' '-Wl,-z,relro' '-Wl,-z,now' '-pie' '-Wl,--gc-sections' '-g' '-O2' '-fstack-protector-strong' '-Wformat' '-Werror=format-security' '-Wl,-Bsymbolic-functions' '-Wl,-z,relro' 'src/shared/libsystemd-shared-233.so' 'src/libudev/libudev.so.1.6.6' 'src/basic/libbasic.a' 'src/libsystemd/libsystemd.a' 'src/basic/libbasic.a' 'src/basic/libbasic.a' '-Wl,-rpath,/data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/src/shared:/data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/src/libudev' In function ‘__ppoll_alias’, inlined from ‘bus_poll’ at ../src/libsystemd/sd-bus/sd-bus.c:2915:11: /usr/include/s390x-linux-gnu/bits/poll2.h:71:9: warning: call to ‘__ppoll_chk_warn’ declared with attribute warning: ppoll called with fds buffer too small file nfds entries return __ppoll_chk (__fds, __nfds, __timeout, __ss, __bos (__fds)); ^ /usr/bin/ld: /tmp/ccYYGrIZ.ltrans7.ltrans.o: undefined reference to symbol 'pthread_sigmask@@GLIBC_2.2' //lib/s390x-linux-gnu/libpthread.so.0: error adding symbols: DSO missing from command line collect2: error: ld returned 1 exit status ```
1.0
build with meson fails with ld.bfd (using -Dlink-udev-shared=false -Db_lto=true) - ### Submission type - [X] Bug report - [ ] Request for enhancement (RFE) ### systemd version the issue has been seen with > git master I reconfigured the xenial-s390x CI build to use meson instead of autotools. This currently fails to build ([example log](https://objectstorage.prodstack4-5.canonical.com/v1/AUTH_77e2ada1e7a84929a74ba3b87153c0ac/autopkgtest-xenial-pitti-systemd-semaphore/xenial/s390x/s/systemd-upstream/20170427_101741_d8343@/log.gz)), mostly on missing libpthread and libselinux libs, but also a rather weird size check failure: ``` In function ‘__read_alias’, inlined from ‘compress_stream_xz’ at ../src/journal/compress.c:398:27: /usr/include/s390x-linux-gnu/bits/unistd.h:39:9: warning: call to ‘__read_chk_warn’ declared with attribute warning: read called with bigger length than size of the destination buffer return __read_chk (__fd, __buf, __nbytes, __bos0 (__buf)); ^ In function ‘__ppoll_alias’, inlined from ‘bus_poll’ at ../src/libsystemd/sd-bus/sd-bus.c:2915:11: /usr/include/s390x-linux-gnu/bits/poll2.h:71:9: warning: call to ‘__ppoll_chk_warn’ declared with attribute warning: ppoll called with fds buffer too small file nfds entries return __ppoll_chk (__fds, __nfds, __timeout, __ss, __bos (__fds)); FAILED: systemd-getty-generator cc -o systemd-getty-generator 'systemd-getty-generator@exe/src_getty-generator_getty-generator.c.o' '-flto' '-Wl,--no-undefined' '-Wl,--as-needed' '-Wl,-z,relro' '-Wl,-z,now' '-pie' '-Wl,--gc-sections' '-g' '-O2' '-fstack-protector-strong' '-Wformat' '-Werror=format-security' '-Wl,-Bsymbolic-functions' '-Wl,-z,relro' 'src/shared/libsystemd-shared-233.so' 'src/libudev/libudev.so.1.6.6' 'src/basic/libbasic.a' 'src/libsystemd/libsystemd.a' 'src/basic/libbasic.a' 'src/basic/libbasic.a' '-Wl,-rpath,/data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/src/shared:/data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/src/libudev' /tmp/cclLTKce.ltrans3.ltrans.o: In function `mac_selinux_create_file_clear': /data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/../src/basic/selinux-util.c:374: undefined reference to `setfscreatecon_raw' /tmp/cclLTKce.ltrans3.ltrans.o: In function `mac_selinux_have': /data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/../src/basic/selinux-util.c:59: undefined reference to `is_selinux_enabled' collect2: error: ld returned 1 exit status FAILED: systemd-sysusers cc -o systemd-sysusers 'systemd-sysusers@exe/src_sysusers_sysusers.c.o' '-flto' '-Wl,--no-undefined' '-Wl,--as-needed' '-Wl,-z,relro' '-Wl,-z,now' '-pie' '-Wl,--gc-sections' '-g' '-O2' '-fstack-protector-strong' '-Wformat' '-Werror=format-security' '-Wl,-Bsymbolic-functions' '-Wl,-z,relro' 'src/shared/libsystemd-shared-233.so' 'src/libudev/libudev.so.1.6.6' 'src/basic/libbasic.a' 'src/libsystemd/libsystemd.a' 'src/basic/libbasic.a' 'src/basic/libbasic.a' '-Wl,-rpath,/data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/src/shared:/data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/src/libudev' /usr/bin/ld: /tmp/ccpz1SmY.ltrans0.ltrans.o: undefined reference to symbol 'selabel_open' //lib/s390x-linux-gnu/libselinux.so.1: error adding symbols: DSO missing from command line collect2: error: ld returned 1 exit status FAILED: busctl cc -o busctl 'busctl@exe/src_busctl_busctl.c.o' 'busctl@exe/src_busctl_busctl-introspect.c.o' '-flto' '-Wl,--no-undefined' '-Wl,--as-needed' '-Wl,-z,relro' '-Wl,-z,now' '-pie' '-Wl,--gc-sections' '-g' '-O2' '-fstack-protector-strong' '-Wformat' '-Werror=format-security' '-Wl,-Bsymbolic-functions' '-Wl,-z,relro' 'src/shared/libsystemd-shared-233.so' 'src/libudev/libudev.so.1.6.6' 'src/basic/libbasic.a' 'src/libsystemd/libsystemd.a' 'src/basic/libbasic.a' 'src/basic/libbasic.a' '-Wl,-rpath,/data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/src/shared:/data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/src/libudev' In function ‘__ppoll_alias’, inlined from ‘bus_poll’ at ../src/libsystemd/sd-bus/sd-bus.c:2915:11: /usr/include/s390x-linux-gnu/bits/poll2.h:71:9: warning: call to ‘__ppoll_chk_warn’ declared with attribute warning: ppoll called with fds buffer too small file nfds entries return __ppoll_chk (__fds, __nfds, __timeout, __ss, __bos (__fds)); ^ /usr/bin/ld: /tmp/ccXySaHU.ltrans17.ltrans.o: undefined reference to symbol 'is_selinux_enabled' //lib/s390x-linux-gnu/libselinux.so.1: error adding symbols: DSO missing from command line AILED: systemd-run cc -o systemd-run 'systemd-run@exe/src_run_run.c.o' '-flto' '-Wl,--no-undefined' '-Wl,--as-needed' '-Wl,-z,relro' '-Wl,-z,now' '-pie' '-Wl,--gc-sections' '-g' '-O2' '-fstack-protector-strong' '-Wformat' '-Werror=format-security' '-Wl,-Bsymbolic-functions' '-Wl,-z,relro' 'src/shared/libsystemd-shared-233.so' 'src/libudev/libudev.so.1.6.6' 'src/basic/libbasic.a' 'src/libsystemd/libsystemd.a' 'src/basic/libbasic.a' 'src/basic/libbasic.a' '-Wl,-rpath,/data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/src/shared:/data/adttmp/autopkgtest-virt-lxc.shared.lbwgsqlu/downtmp/build.I3N/debian/build-deb/src/libudev' In function ‘__ppoll_alias’, inlined from ‘bus_poll’ at ../src/libsystemd/sd-bus/sd-bus.c:2915:11: /usr/include/s390x-linux-gnu/bits/poll2.h:71:9: warning: call to ‘__ppoll_chk_warn’ declared with attribute warning: ppoll called with fds buffer too small file nfds entries return __ppoll_chk (__fds, __nfds, __timeout, __ss, __bos (__fds)); ^ /usr/bin/ld: /tmp/ccYYGrIZ.ltrans7.ltrans.o: undefined reference to symbol 'pthread_sigmask@@GLIBC_2.2' //lib/s390x-linux-gnu/libpthread.so.0: error adding symbols: DSO missing from command line collect2: error: ld returned 1 exit status ```
non_process
build with meson fails with ld bfd using dlink udev shared false db lto true submission type bug report request for enhancement rfe systemd version the issue has been seen with git master i reconfigured the xenial ci build to use meson instead of autotools this currently fails to build mostly on missing libpthread and libselinux libs but also a rather weird size check failure in function ‘ read alias’ inlined from ‘compress stream xz’ at src journal compress c usr include linux gnu bits unistd h warning call to ‘ read chk warn’ declared with attribute warning read called with bigger length than size of the destination buffer return read chk fd buf nbytes buf in function ‘ ppoll alias’ inlined from ‘bus poll’ at src libsystemd sd bus sd bus c usr include linux gnu bits h warning call to ‘ ppoll chk warn’ declared with attribute warning ppoll called with fds buffer too small file nfds entries return ppoll chk fds nfds timeout ss bos fds failed systemd getty generator cc o systemd getty generator systemd getty generator exe src getty generator getty generator c o flto wl no undefined wl as needed wl z relro wl z now pie wl gc sections g fstack protector strong wformat werror format security wl bsymbolic functions wl z relro src shared libsystemd shared so src libudev libudev so src basic libbasic a src libsystemd libsystemd a src basic libbasic a src basic libbasic a wl rpath data adttmp autopkgtest virt lxc shared lbwgsqlu downtmp build debian build deb src shared data adttmp autopkgtest virt lxc shared lbwgsqlu downtmp build debian build deb src libudev tmp cclltkce ltrans o in function mac selinux create file clear data adttmp autopkgtest virt lxc shared lbwgsqlu downtmp build debian build deb src basic selinux util c undefined reference to setfscreatecon raw tmp cclltkce ltrans o in function mac selinux have data adttmp autopkgtest virt lxc shared lbwgsqlu downtmp build debian build deb src basic selinux util c undefined reference to is selinux enabled error ld returned exit status failed systemd sysusers cc o systemd sysusers systemd sysusers exe src sysusers sysusers c o flto wl no undefined wl as needed wl z relro wl z now pie wl gc sections g fstack protector strong wformat werror format security wl bsymbolic functions wl z relro src shared libsystemd shared so src libudev libudev so src basic libbasic a src libsystemd libsystemd a src basic libbasic a src basic libbasic a wl rpath data adttmp autopkgtest virt lxc shared lbwgsqlu downtmp build debian build deb src shared data adttmp autopkgtest virt lxc shared lbwgsqlu downtmp build debian build deb src libudev usr bin ld tmp ltrans o undefined reference to symbol selabel open lib linux gnu libselinux so error adding symbols dso missing from command line error ld returned exit status failed busctl cc o busctl busctl exe src busctl busctl c o busctl exe src busctl busctl introspect c o flto wl no undefined wl as needed wl z relro wl z now pie wl gc sections g fstack protector strong wformat werror format security wl bsymbolic functions wl z relro src shared libsystemd shared so src libudev libudev so src basic libbasic a src libsystemd libsystemd a src basic libbasic a src basic libbasic a wl rpath data adttmp autopkgtest virt lxc shared lbwgsqlu downtmp build debian build deb src shared data adttmp autopkgtest virt lxc shared lbwgsqlu downtmp build debian build deb src libudev in function ‘ ppoll alias’ inlined from ‘bus poll’ at src libsystemd sd bus sd bus c usr include linux gnu bits h warning call to ‘ ppoll chk warn’ declared with attribute warning ppoll called with fds buffer too small file nfds entries return ppoll chk fds nfds timeout ss bos fds usr bin ld tmp ccxysahu ltrans o undefined reference to symbol is selinux enabled lib linux gnu libselinux so error adding symbols dso missing from command line ailed systemd run cc o systemd run systemd run exe src run run c o flto wl no undefined wl as needed wl z relro wl z now pie wl gc sections g fstack protector strong wformat werror format security wl bsymbolic functions wl z relro src shared libsystemd shared so src libudev libudev so src basic libbasic a src libsystemd libsystemd a src basic libbasic a src basic libbasic a wl rpath data adttmp autopkgtest virt lxc shared lbwgsqlu downtmp build debian build deb src shared data adttmp autopkgtest virt lxc shared lbwgsqlu downtmp build debian build deb src libudev in function ‘ ppoll alias’ inlined from ‘bus poll’ at src libsystemd sd bus sd bus c usr include linux gnu bits h warning call to ‘ ppoll chk warn’ declared with attribute warning ppoll called with fds buffer too small file nfds entries return ppoll chk fds nfds timeout ss bos fds usr bin ld tmp ccyygriz ltrans o undefined reference to symbol pthread sigmask glibc lib linux gnu libpthread so error adding symbols dso missing from command line error ld returned exit status
0
188,788
6,782,018,441
IssuesEvent
2017-10-30 05:28:25
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.msn.com - video or audio doesn't play
browser-firefox priority-critical status-duplicate type-stylo
<!-- @browser: Firefox 57.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:57.0) Gecko/20100101 Firefox/57.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://www.msn.com/en-us/weather/maps/Leesburg,FL,United-States/we-city?iso=US&el=JJ86J%2Fm16gcgLgOffuDFcA%3D%3D **Browser / Version**: Firefox 57.0 **Operating System**: Windows 10 **Tested Another Browser**: No **Problem type**: Video or audio doesn't play **Description**: No video or audio **Steps to Reproduce**: It worked yesterday [![Screenshot Description](https://webcompat.com/uploads/2017/10/9dea135a-7859-4188-ae81-94ddb336f9d1-thumb.jpg)](https://webcompat.com/uploads/2017/10/9dea135a-7859-4188-ae81-94ddb336f9d1.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.msn.com - video or audio doesn't play - <!-- @browser: Firefox 57.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:57.0) Gecko/20100101 Firefox/57.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://www.msn.com/en-us/weather/maps/Leesburg,FL,United-States/we-city?iso=US&el=JJ86J%2Fm16gcgLgOffuDFcA%3D%3D **Browser / Version**: Firefox 57.0 **Operating System**: Windows 10 **Tested Another Browser**: No **Problem type**: Video or audio doesn't play **Description**: No video or audio **Steps to Reproduce**: It worked yesterday [![Screenshot Description](https://webcompat.com/uploads/2017/10/9dea135a-7859-4188-ae81-94ddb336f9d1-thumb.jpg)](https://webcompat.com/uploads/2017/10/9dea135a-7859-4188-ae81-94ddb336f9d1.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
video or audio doesn t play url browser version firefox operating system windows tested another browser no problem type video or audio doesn t play description no video or audio steps to reproduce it worked yesterday from with ❤️
0
605,272
18,727,935,654
IssuesEvent
2021-11-03 18:13:10
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
mac m1 minikube start error
kind/bug os/macos priority/awaiting-more-evidence triage/needs-information co/docker-driver
<!--- Please include the "minikube start" command you used in your reproduction steps ---> **Steps to reproduce the issue:** 1. brew install minikube 2. minikube start 3. **Full output of `minikube logs` command:** <details> Exiting due to GUEST_PROVISION: Failed to start host: recreate: creating host: create: creating: setting up container node: preparing volume for minikube container: docker run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.20@sha256:0250dab3644403384bd54f566921c6b57138eecffbb861f9392feef9b2ec44f6 -d /var/lib: exit status 125 stdout: </details> <!--- TIP: Add the "--alsologtostderr" flag to the command-line for more logs ---> **Full output of failed command:** <details> docker: Error: error contacting notary server: unknown: Project 'project:gcr.io:k8s-minikube' not found or deleted. See 'docker run --help'. </details>
1.0
mac m1 minikube start error - <!--- Please include the "minikube start" command you used in your reproduction steps ---> **Steps to reproduce the issue:** 1. brew install minikube 2. minikube start 3. **Full output of `minikube logs` command:** <details> Exiting due to GUEST_PROVISION: Failed to start host: recreate: creating host: create: creating: setting up container node: preparing volume for minikube container: docker run --rm --name minikube-preload-sidecar --label created_by.minikube.sigs.k8s.io=true --label name.minikube.sigs.k8s.io=minikube --entrypoint /usr/bin/test -v minikube:/var gcr.io/k8s-minikube/kicbase:v0.0.20@sha256:0250dab3644403384bd54f566921c6b57138eecffbb861f9392feef9b2ec44f6 -d /var/lib: exit status 125 stdout: </details> <!--- TIP: Add the "--alsologtostderr" flag to the command-line for more logs ---> **Full output of failed command:** <details> docker: Error: error contacting notary server: unknown: Project 'project:gcr.io:k8s-minikube' not found or deleted. See 'docker run --help'. </details>
non_process
mac minikube start error steps to reproduce the issue brew install minikube minikube start full output of minikube logs command exiting due to guest provision failed to start host recreate creating host create creating setting up container node preparing volume for minikube container docker run rm name minikube preload sidecar label created by minikube sigs io true label name minikube sigs io minikube entrypoint usr bin test v minikube var gcr io minikube kicbase d var lib exit status stdout full output of failed command docker error error contacting notary server unknown project project gcr io minikube not found or deleted see docker run help
0
17,234
22,950,735,527
IssuesEvent
2022-07-19 07:10:09
Tencent/tdesign-miniprogram
https://api.github.com/repos/Tencent/tdesign-miniprogram
closed
[Tabs] 标签数量很多,切换标签自动滚动到当前激活项
enhancement good first issue processing
### 这个功能解决了什么问题 当 t-tab-panel 数量有十几个的时候,在用户点击对应项的标签,标签并未展示到可视区域中间,而是需要手动去滑动 ![动画1](https://user-images.githubusercontent.com/45986102/179441282-60653768-775f-4b15-9e9e-f3f1d541f229.gif) ### 你建议的方案是什么 增加自动滚动的功能,同时可以提供外部获取标签偏移距离的的api ![动画](https://user-images.githubusercontent.com/45986102/179440650-b419de19-74b6-4e26-911d-160580ccdb15.gif)
1.0
[Tabs] 标签数量很多,切换标签自动滚动到当前激活项 - ### 这个功能解决了什么问题 当 t-tab-panel 数量有十几个的时候,在用户点击对应项的标签,标签并未展示到可视区域中间,而是需要手动去滑动 ![动画1](https://user-images.githubusercontent.com/45986102/179441282-60653768-775f-4b15-9e9e-f3f1d541f229.gif) ### 你建议的方案是什么 增加自动滚动的功能,同时可以提供外部获取标签偏移距离的的api ![动画](https://user-images.githubusercontent.com/45986102/179440650-b419de19-74b6-4e26-911d-160580ccdb15.gif)
process
标签数量很多,切换标签自动滚动到当前激活项 这个功能解决了什么问题 当 t tab panel 数量有十几个的时候,在用户点击对应项的标签,标签并未展示到可视区域中间,而是需要手动去滑动 你建议的方案是什么 增加自动滚动的功能,同时可以提供外部获取标签偏移距离的的api
1
16,143
20,405,703,307
IssuesEvent
2022-02-23 04:59:04
fmnas/fmnas-site
https://api.github.com/repos/fmnas/fmnas-site
closed
Delete serialized in confirm()
public form processor x-small (<1h)
--- _This issue has been automatically created by [todo-actions](https://github.com/apps/todo-actions) based on a TODO comment found in [public/application/index.php:79](https://github.com/fmnas/fmnas-site/blob/main/public/application/index.php#L79). It will automatically be closed when the TODO comment is removed from the default branch (main)._
1.0
Delete serialized in confirm() - --- _This issue has been automatically created by [todo-actions](https://github.com/apps/todo-actions) based on a TODO comment found in [public/application/index.php:79](https://github.com/fmnas/fmnas-site/blob/main/public/application/index.php#L79). It will automatically be closed when the TODO comment is removed from the default branch (main)._
process
delete serialized in confirm this issue has been automatically created by based on a todo comment found in it will automatically be closed when the todo comment is removed from the default branch main
1
18,088
24,111,555,124
IssuesEvent
2022-09-20 11:47:43
saibrotech/mentoria
https://api.github.com/repos/saibrotech/mentoria
closed
Inscrever-se no "Upskill your Talent 2022"
processo seletivo
https://app.becas-santander.com/pt/program/upskill-your-talent-2022 Inscrição Aberta de 06/09/22 ao 15/11/22 Avaliação das candidaturas de 15/11/22 ao 05/01/23 Seleção de Candidatos em 05/01/23 - [x] Inscrição - [x] Realizar teste: Penn Santander Personality Questionnaire - [x] Realizar teste: B2C Verbal Reasoning Test (Enhanced)
1.0
Inscrever-se no "Upskill your Talent 2022" - https://app.becas-santander.com/pt/program/upskill-your-talent-2022 Inscrição Aberta de 06/09/22 ao 15/11/22 Avaliação das candidaturas de 15/11/22 ao 05/01/23 Seleção de Candidatos em 05/01/23 - [x] Inscrição - [x] Realizar teste: Penn Santander Personality Questionnaire - [x] Realizar teste: B2C Verbal Reasoning Test (Enhanced)
process
inscrever se no upskill your talent inscrição aberta de ao avaliação das candidaturas de ao seleção de candidatos em inscrição realizar teste penn santander personality questionnaire realizar teste verbal reasoning test enhanced
1
6,788
9,921,574,666
IssuesEvent
2019-06-30 19:28:10
ION28/BLUESPAWN
https://api.github.com/repos/ION28/BLUESPAWN
opened
Create Processes Base
enhancement epic hard high priority processes
This covers creating the .h/.cpp files and some basic functions that the corresponding HuntT####/other analysis functions will utilize. For example, there might be functions for retrieving a list of all processes running, getting all threads of a process, analyzing its code stack, or generating the checksum of a process' memory image. Feel free to decompose into smaller issues as needed
1.0
Create Processes Base - This covers creating the .h/.cpp files and some basic functions that the corresponding HuntT####/other analysis functions will utilize. For example, there might be functions for retrieving a list of all processes running, getting all threads of a process, analyzing its code stack, or generating the checksum of a process' memory image. Feel free to decompose into smaller issues as needed
process
create processes base this covers creating the h cpp files and some basic functions that the corresponding huntt other analysis functions will utilize for example there might be functions for retrieving a list of all processes running getting all threads of a process analyzing its code stack or generating the checksum of a process memory image feel free to decompose into smaller issues as needed
1
10,428
13,218,618,775
IssuesEvent
2020-08-17 09:03:13
pystatgen/sgkit
https://api.github.com/repos/pystatgen/sgkit
closed
mypy support for hypothesis decorators
process + tools
@ravwojdyla do you know of a way to avoid doing this everywhere in tests: ```python @given(st.integers(0, 25)) # type: ignore[misc] @settings(max_examples=10) # type: ignore[misc] def test_something(n: int) -> None: ... ``` I get mypy errors w/o a return type (only for decorated functions) and I'm adding `ignore[misc]` for `Untyped decorator makes function "test_something" untyped`. Is there some way to tell mypy to ignore type checking for any function wrapped by hypothesis decorators globally? _Originally posted by @eric-czech in https://github.com/pystatgen/sgkit/pull/69#issuecomment-665541975_
1.0
mypy support for hypothesis decorators - @ravwojdyla do you know of a way to avoid doing this everywhere in tests: ```python @given(st.integers(0, 25)) # type: ignore[misc] @settings(max_examples=10) # type: ignore[misc] def test_something(n: int) -> None: ... ``` I get mypy errors w/o a return type (only for decorated functions) and I'm adding `ignore[misc]` for `Untyped decorator makes function "test_something" untyped`. Is there some way to tell mypy to ignore type checking for any function wrapped by hypothesis decorators globally? _Originally posted by @eric-czech in https://github.com/pystatgen/sgkit/pull/69#issuecomment-665541975_
process
mypy support for hypothesis decorators ravwojdyla do you know of a way to avoid doing this everywhere in tests python given st integers type ignore settings max examples type ignore def test something n int none i get mypy errors w o a return type only for decorated functions and i m adding ignore for untyped decorator makes function test something untyped is there some way to tell mypy to ignore type checking for any function wrapped by hypothesis decorators globally originally posted by eric czech in
1
326,282
24,077,162,879
IssuesEvent
2022-09-18 23:36:55
JaredDobry/enpm613
https://api.github.com/repos/JaredDobry/enpm613
opened
Non-Functional Requirements Derivation
documentation
- [ ] Derive non-functional requirements and scenarios for the system For each identified scenario: 1. Provide a brief description (one-two sentences) 2. Assign a number from 1 to 3 (where Low = 1, Medium = 2, High = 3) for its utility (importance) to the users 3. Assign a number from 1 to 3 (where Difficult or high risk = 1, Medium difficulty or risk = 2, Rather easy and low risk = 3) for the estimated difficulty or risk to implement that scenario 4. Compute the priority score (utility * difficulty). This score will help you decide what scenario will be within the scope of your software for design and implementation. See the provided Requirements Analysis template. - [ ] Check them into source control in enpm613/docs
1.0
Non-Functional Requirements Derivation - - [ ] Derive non-functional requirements and scenarios for the system For each identified scenario: 1. Provide a brief description (one-two sentences) 2. Assign a number from 1 to 3 (where Low = 1, Medium = 2, High = 3) for its utility (importance) to the users 3. Assign a number from 1 to 3 (where Difficult or high risk = 1, Medium difficulty or risk = 2, Rather easy and low risk = 3) for the estimated difficulty or risk to implement that scenario 4. Compute the priority score (utility * difficulty). This score will help you decide what scenario will be within the scope of your software for design and implementation. See the provided Requirements Analysis template. - [ ] Check them into source control in enpm613/docs
non_process
non functional requirements derivation derive non functional requirements and scenarios for the system for each identified scenario provide a brief description one two sentences assign a number from to where low medium high for its utility importance to the users assign a number from to where difficult or high risk medium difficulty or risk rather easy and low risk for the estimated difficulty or risk to implement that scenario compute the priority score utility difficulty this score will help you decide what scenario will be within the scope of your software for design and implementation see the provided requirements analysis template check them into source control in docs
0
208,482
7,154,969,044
IssuesEvent
2018-01-26 10:43:23
Mandiklopper/People-Connect
https://api.github.com/repos/Mandiklopper/People-Connect
opened
Add function to print letters
Normal Priority
We need to look into adding a function where the client can print letters for Work ethics as example directly from the Disciplinary screen for that specific transaction. Currently you need to go into reports and then remember the Sanction number which can cause issues. Ideally if we have a button that says "Print Letter" on the Disciplinary screen this will work.
1.0
Add function to print letters - We need to look into adding a function where the client can print letters for Work ethics as example directly from the Disciplinary screen for that specific transaction. Currently you need to go into reports and then remember the Sanction number which can cause issues. Ideally if we have a button that says "Print Letter" on the Disciplinary screen this will work.
non_process
add function to print letters we need to look into adding a function where the client can print letters for work ethics as example directly from the disciplinary screen for that specific transaction currently you need to go into reports and then remember the sanction number which can cause issues ideally if we have a button that says print letter on the disciplinary screen this will work
0
606,991
18,771,481,027
IssuesEvent
2021-11-06 22:54:38
Lissy93/dashy
https://api.github.com/repos/Lissy93/dashy
closed
[DOCS] Broken Links found in Documentation
👩‍💻 Good First Issue 📕 Docs 💤 Low Priority 🕸️ Potential Duplicate 🤖 Auto
Errors were reported while checking the availability of links. ``` 📝 Summary --------------------- 🔍 Total.........1056 ✅ Successful....1008 ⏳ Timeouts.........9 🔀 Redirected.......0 👻 Excluded.........0 ⚠ Unknown..........0 🚫 Errors..........39 Errors in docs/deployment.md ✗ https://developers.cloudflare.com/workers/platform/sites (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/platform/sites)) Errors in docs/credits.md ✗ https://developers.cloudflare.com/workers/runtime-apis/web-crypto (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/runtime-apis/web-crypto)) ⧖ https://play-with-docker.com/ ✗ https://developers.cloudflare.com/workers/runtime-apis/kv (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/runtime-apis/kv)) Errors in docs/troubleshooting.md ✗ file:///github/workspace/docs/http:/[dashy-location] (Cannot find file file:///github/workspace/docs/http:/[dashy-location]) ✗ https://location-of-dashy/ (error sending request for url (https://location-of-dashy/): error trying to connect: dns error: failed to lookup address information: Name or service not known) Errors in README.md ✗ https://developers.cloudflare.com/workers/learning/how-workers-works (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/learning/how-workers-works)) ✗ https://plex.lab.local/ (error sending request for url (https://plex.lab.local/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://developers.cloudflare.com/workers/learning/how-kv-works (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/learning/how-kv-works)) ✗ https://whoogle.local/search?q= (error sending request for url (https://whoogle.local/search?q=): error trying to connect: dns error: failed to lookup address information: Name or service not known) ⧖ https://play-with-docker.com/ ✗ https://developers.cloudflare.com/workers/runtime-apis/kv (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/runtime-apis/kv)) ✗ https://bookstack.local/ (error sending request for url (https://bookstack.local/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://server-start.local/ (error sending request for url (https://server-start.local/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://developers.cloudflare.com/workers/runtime-apis/web-crypto (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/runtime-apis/web-crypto)) ✗ https://start-page.local/ (error sending request for url (https://start-page.local/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://tldrlegal.com/license/mit-license (HTTP status server error (503 Service Unavailable) for url (https://tldrlegal.com/license/mit-license)) Errors in docs/backup-restore.md ✗ https://developers.cloudflare.com/workers/learning/how-workers-works (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/learning/how-workers-works)) ✗ https://developers.cloudflare.com/workers/learning/how-kv-works (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/learning/how-kv-works)) Errors in docs/authentication.md ✗ http://localhost:8081/ (error sending request for url (http://localhost:8081/): error trying to connect: tcp connect error: Cannot assign requested address (os error 99)) Errors in public/default.html ✗ https://fonts.gstatic.com/ (HTTP status client error (404 Not Found) for url (https://fonts.gstatic.com/)) Errors in docs/developing.md ✗ http://localhost:8080/ (error sending request for url (http://localhost:8080/): error trying to connect: tcp connect error: Cannot assign requested address (os error 99)) ✗ file:///github/workspace/docs/docs/development-guides.md (Cannot find file file:///github/workspace/docs/docs/development-guides.md) Errors in docs/theming.md ✗ https://example.com/my-stylesheet.css (HTTP status client error (404 Not Found) for url (https://example.com/my-stylesheet.css)) Errors in docs/management.md ✗ http://dashy/ (error sending request for url (http://dashy/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://www.cloudflare.com/en-gb/learning/ssl/what-is-an-ssl-certificate/ (HTTP status client error (403 Forbidden) for url (https://www.cloudflare.com/en-gb/learning/ssl/what-is-an-ssl-certificate/)) Errors in docs/contributing.md ✗ file:///github/workspace/docs/mailto:info@example.com (Cannot find file file:///github/workspace/docs/mailto:info@example.com) ✗ https://www.slant.co/improve/topics/27783/viewpoints/1/~self-hosted-homelab-startpage~dashy (HTTP status client error (403 Forbidden) for url (https://www.slant.co/improve/topics/27783/viewpoints/1/~self-hosted-homelab-startpage~dashy)) Errors in docs/configuring.md ✗ file:///github/workspace/docs/docs/backup-restore.md (Cannot find file file:///github/workspace/docs/docs/backup-restore.md) ⧖ http://192.168.1.1/ ⧖ http://192.168.130.1/ Errors in public/index.html ✗ file:///github/workspace/public/%3C%25=%20BASE_URL%20%25%3E/web-icons/favicon-64x64.png (Cannot find file file:///github/workspace/public/%3C%25=%20BASE_URL%20%25%3E/web-icons/favicon-64x64.png) ✗ file:///github/workspace/public/%3C%25=%20BASE_URL%20%25%3Efavicon.ico (Cannot find file file:///github/workspace/public/%3C%25=%20BASE_URL%20%25%3Efavicon.ico) Errors in docs/quick-start.md ⧖ http://192.168.1.1/ ✗ file:///github/workspace/docs/app.netlify.com/login (Cannot find file file:///github/workspace/docs/app.netlify.com/login) ⧖ http://192.168.130.1/ ✗ file:///github/workspace/docs/docs/deployment.md (Cannot find file file:///github/workspace/docs/docs/deployment.md) Errors in docs/status-indicators.md ⧖ https://192.168.1.1/ ⧖ http://192.168.1.1:8338/ ✗ https://location-of-dashy/ (error sending request for url (https://location-of-dashy/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ⧖ http://192.168.1.1:3001/ Errors in docs/searching.md ✗ https://freshrss.lab.local/ (error sending request for url (https://freshrss.lab.local/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://metager.org/meta (HTTP status client error (404 Not Found) for url (https://metager.org/meta)) ✗ https://whoogle.local/search?q= (error sending request for url (https://whoogle.local/search?q=): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://plex.lab.local/ (error sending request for url (https://plex.lab.local/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://git.lab.local/ (error sending request for url (https://git.lab.local/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://bookstack.lab.local/ (error sending request for url (https://bookstack.lab.local/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://searx.local/search?q= (error sending request for url (https://searx.local/search?q=): error trying to connect: dns error: failed to lookup address information: Name or service not known) ``` [Full Github Actions output](https://github.com/Lissy93/dashy/actions/runs/1406066566?check_suite_focus=true)
1.0
[DOCS] Broken Links found in Documentation - Errors were reported while checking the availability of links. ``` 📝 Summary --------------------- 🔍 Total.........1056 ✅ Successful....1008 ⏳ Timeouts.........9 🔀 Redirected.......0 👻 Excluded.........0 ⚠ Unknown..........0 🚫 Errors..........39 Errors in docs/deployment.md ✗ https://developers.cloudflare.com/workers/platform/sites (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/platform/sites)) Errors in docs/credits.md ✗ https://developers.cloudflare.com/workers/runtime-apis/web-crypto (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/runtime-apis/web-crypto)) ⧖ https://play-with-docker.com/ ✗ https://developers.cloudflare.com/workers/runtime-apis/kv (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/runtime-apis/kv)) Errors in docs/troubleshooting.md ✗ file:///github/workspace/docs/http:/[dashy-location] (Cannot find file file:///github/workspace/docs/http:/[dashy-location]) ✗ https://location-of-dashy/ (error sending request for url (https://location-of-dashy/): error trying to connect: dns error: failed to lookup address information: Name or service not known) Errors in README.md ✗ https://developers.cloudflare.com/workers/learning/how-workers-works (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/learning/how-workers-works)) ✗ https://plex.lab.local/ (error sending request for url (https://plex.lab.local/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://developers.cloudflare.com/workers/learning/how-kv-works (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/learning/how-kv-works)) ✗ https://whoogle.local/search?q= (error sending request for url (https://whoogle.local/search?q=): error trying to connect: dns error: failed to lookup address information: Name or service not known) ⧖ https://play-with-docker.com/ ✗ https://developers.cloudflare.com/workers/runtime-apis/kv (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/runtime-apis/kv)) ✗ https://bookstack.local/ (error sending request for url (https://bookstack.local/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://server-start.local/ (error sending request for url (https://server-start.local/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://developers.cloudflare.com/workers/runtime-apis/web-crypto (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/runtime-apis/web-crypto)) ✗ https://start-page.local/ (error sending request for url (https://start-page.local/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://tldrlegal.com/license/mit-license (HTTP status server error (503 Service Unavailable) for url (https://tldrlegal.com/license/mit-license)) Errors in docs/backup-restore.md ✗ https://developers.cloudflare.com/workers/learning/how-workers-works (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/learning/how-workers-works)) ✗ https://developers.cloudflare.com/workers/learning/how-kv-works (HTTP status server error (503 Service Unavailable) for url (https://developers.cloudflare.com/workers/learning/how-kv-works)) Errors in docs/authentication.md ✗ http://localhost:8081/ (error sending request for url (http://localhost:8081/): error trying to connect: tcp connect error: Cannot assign requested address (os error 99)) Errors in public/default.html ✗ https://fonts.gstatic.com/ (HTTP status client error (404 Not Found) for url (https://fonts.gstatic.com/)) Errors in docs/developing.md ✗ http://localhost:8080/ (error sending request for url (http://localhost:8080/): error trying to connect: tcp connect error: Cannot assign requested address (os error 99)) ✗ file:///github/workspace/docs/docs/development-guides.md (Cannot find file file:///github/workspace/docs/docs/development-guides.md) Errors in docs/theming.md ✗ https://example.com/my-stylesheet.css (HTTP status client error (404 Not Found) for url (https://example.com/my-stylesheet.css)) Errors in docs/management.md ✗ http://dashy/ (error sending request for url (http://dashy/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://www.cloudflare.com/en-gb/learning/ssl/what-is-an-ssl-certificate/ (HTTP status client error (403 Forbidden) for url (https://www.cloudflare.com/en-gb/learning/ssl/what-is-an-ssl-certificate/)) Errors in docs/contributing.md ✗ file:///github/workspace/docs/mailto:info@example.com (Cannot find file file:///github/workspace/docs/mailto:info@example.com) ✗ https://www.slant.co/improve/topics/27783/viewpoints/1/~self-hosted-homelab-startpage~dashy (HTTP status client error (403 Forbidden) for url (https://www.slant.co/improve/topics/27783/viewpoints/1/~self-hosted-homelab-startpage~dashy)) Errors in docs/configuring.md ✗ file:///github/workspace/docs/docs/backup-restore.md (Cannot find file file:///github/workspace/docs/docs/backup-restore.md) ⧖ http://192.168.1.1/ ⧖ http://192.168.130.1/ Errors in public/index.html ✗ file:///github/workspace/public/%3C%25=%20BASE_URL%20%25%3E/web-icons/favicon-64x64.png (Cannot find file file:///github/workspace/public/%3C%25=%20BASE_URL%20%25%3E/web-icons/favicon-64x64.png) ✗ file:///github/workspace/public/%3C%25=%20BASE_URL%20%25%3Efavicon.ico (Cannot find file file:///github/workspace/public/%3C%25=%20BASE_URL%20%25%3Efavicon.ico) Errors in docs/quick-start.md ⧖ http://192.168.1.1/ ✗ file:///github/workspace/docs/app.netlify.com/login (Cannot find file file:///github/workspace/docs/app.netlify.com/login) ⧖ http://192.168.130.1/ ✗ file:///github/workspace/docs/docs/deployment.md (Cannot find file file:///github/workspace/docs/docs/deployment.md) Errors in docs/status-indicators.md ⧖ https://192.168.1.1/ ⧖ http://192.168.1.1:8338/ ✗ https://location-of-dashy/ (error sending request for url (https://location-of-dashy/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ⧖ http://192.168.1.1:3001/ Errors in docs/searching.md ✗ https://freshrss.lab.local/ (error sending request for url (https://freshrss.lab.local/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://metager.org/meta (HTTP status client error (404 Not Found) for url (https://metager.org/meta)) ✗ https://whoogle.local/search?q= (error sending request for url (https://whoogle.local/search?q=): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://plex.lab.local/ (error sending request for url (https://plex.lab.local/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://git.lab.local/ (error sending request for url (https://git.lab.local/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://bookstack.lab.local/ (error sending request for url (https://bookstack.lab.local/): error trying to connect: dns error: failed to lookup address information: Name or service not known) ✗ https://searx.local/search?q= (error sending request for url (https://searx.local/search?q=): error trying to connect: dns error: failed to lookup address information: Name or service not known) ``` [Full Github Actions output](https://github.com/Lissy93/dashy/actions/runs/1406066566?check_suite_focus=true)
non_process
broken links found in documentation errors were reported while checking the availability of links 📝 summary 🔍 total ✅ successful ⏳ timeouts 🔀 redirected 👻 excluded ⚠ unknown 🚫 errors errors in docs deployment md ✗ http status server error service unavailable for url errors in docs credits md ✗ http status server error service unavailable for url ⧖ ✗ http status server error service unavailable for url errors in docs troubleshooting md ✗ file github workspace docs http cannot find file file github workspace docs http ✗ error sending request for url error trying to connect dns error failed to lookup address information name or service not known errors in readme md ✗ http status server error service unavailable for url ✗ error sending request for url error trying to connect dns error failed to lookup address information name or service not known ✗ http status server error service unavailable for url ✗ error sending request for url error trying to connect dns error failed to lookup address information name or service not known ⧖ ✗ http status server error service unavailable for url ✗ error sending request for url error trying to connect dns error failed to lookup address information name or service not known ✗ error sending request for url error trying to connect dns error failed to lookup address information name or service not known ✗ http status server error service unavailable for url ✗ error sending request for url error trying to connect dns error failed to lookup address information name or service not known ✗ http status server error service unavailable for url errors in docs backup restore md ✗ http status server error service unavailable for url ✗ http status server error service unavailable for url errors in docs authentication md ✗ error sending request for url error trying to connect tcp connect error cannot assign requested address os error errors in public default html ✗ http status client error not found for url errors in docs developing md ✗ error sending request for url error trying to connect tcp connect error cannot assign requested address os error ✗ file github workspace docs docs development guides md cannot find file file github workspace docs docs development guides md errors in docs theming md ✗ http status client error not found for url errors in docs management md ✗ error sending request for url error trying to connect dns error failed to lookup address information name or service not known ✗ http status client error forbidden for url errors in docs contributing md ✗ file github workspace docs mailto info example com cannot find file file github workspace docs mailto info example com ✗ http status client error forbidden for url errors in docs configuring md ✗ file github workspace docs docs backup restore md cannot find file file github workspace docs docs backup restore md ⧖ ⧖ errors in public index html ✗ file github workspace public url web icons favicon png cannot find file file github workspace public url web icons favicon png ✗ file github workspace public url ico cannot find file file github workspace public url ico errors in docs quick start md ⧖ ✗ file github workspace docs app netlify com login cannot find file file github workspace docs app netlify com login ⧖ ✗ file github workspace docs docs deployment md cannot find file file github workspace docs docs deployment md errors in docs status indicators md ⧖ ⧖ ✗ error sending request for url error trying to connect dns error failed to lookup address information name or service not known ⧖ errors in docs searching md ✗ error sending request for url error trying to connect dns error failed to lookup address information name or service not known ✗ http status client error not found for url ✗ error sending request for url error trying to connect dns error failed to lookup address information name or service not known ✗ error sending request for url error trying to connect dns error failed to lookup address information name or service not known ✗ error sending request for url error trying to connect dns error failed to lookup address information name or service not known ✗ error sending request for url error trying to connect dns error failed to lookup address information name or service not known ✗ error sending request for url error trying to connect dns error failed to lookup address information name or service not known
0
61,824
14,641,722,792
IssuesEvent
2020-12-25 08:07:14
srivatsamarichi/pip-kotlin
https://api.github.com/repos/srivatsamarichi/pip-kotlin
opened
CVE-2020-15250 (Medium) detected in junit-4.12.jar
security vulnerability
## CVE-2020-15250 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>junit-4.12.jar</b></p></summary> <p>JUnit is a unit testing framework for Java, created by Erich Gamma and Kent Beck.</p> <p>Library home page: <a href="http://junit.org">http://junit.org</a></p> <p>Path to dependency file: pip-kotlin/app/build.gradle</p> <p>Path to vulnerable library: 20201225080542_KKSKKS/downloadResource_XGIZMN/20201225080612/junit-4.12.jar,/tmp/ws-ua_20201225080542_KKSKKS/downloadResource_XGIZMN/20201225080612/junit-4.12.jar</p> <p> Dependency Hierarchy: - :x: **junit-4.12.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/srivatsamarichi/pip-kotlin/commit/35d64e7f8134bf0c52e80c29038763207bf63c1b">35d64e7f8134bf0c52e80c29038763207bf63c1b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In JUnit4 from version 4.7 and before 4.13.1, the test rule TemporaryFolder contains a local information disclosure vulnerability. On Unix like systems, the system's temporary directory is shared between all users on that system. Because of this, when files and directories are written into this directory they are, by default, readable by other users on that same system. This vulnerability does not allow other users to overwrite the contents of these directories or files. This is purely an information disclosure vulnerability. This vulnerability impacts you if the JUnit tests write sensitive information, like API keys or passwords, into the temporary folder, and the JUnit tests execute in an environment where the OS has other untrusted users. Because certain JDK file system APIs were only added in JDK 1.7, this this fix is dependent upon the version of the JDK you are using. For Java 1.7 and higher users: this vulnerability is fixed in 4.13.1. For Java 1.6 and lower users: no patch is available, you must use the workaround below. If you are unable to patch, or are stuck running on Java 1.6, specifying the `java.io.tmpdir` system environment variable to a directory that is exclusively owned by the executing user will fix this vulnerability. For more information, including an example of vulnerable code, see the referenced GitHub Security Advisory. <p>Publish Date: 2020-10-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15250>CVE-2020-15250</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/junit-team/junit4/security/advisories/GHSA-269g-pwp5-87pp">https://github.com/junit-team/junit4/security/advisories/GHSA-269g-pwp5-87pp</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: junit:junit:4.13.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-15250 (Medium) detected in junit-4.12.jar - ## CVE-2020-15250 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>junit-4.12.jar</b></p></summary> <p>JUnit is a unit testing framework for Java, created by Erich Gamma and Kent Beck.</p> <p>Library home page: <a href="http://junit.org">http://junit.org</a></p> <p>Path to dependency file: pip-kotlin/app/build.gradle</p> <p>Path to vulnerable library: 20201225080542_KKSKKS/downloadResource_XGIZMN/20201225080612/junit-4.12.jar,/tmp/ws-ua_20201225080542_KKSKKS/downloadResource_XGIZMN/20201225080612/junit-4.12.jar</p> <p> Dependency Hierarchy: - :x: **junit-4.12.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/srivatsamarichi/pip-kotlin/commit/35d64e7f8134bf0c52e80c29038763207bf63c1b">35d64e7f8134bf0c52e80c29038763207bf63c1b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In JUnit4 from version 4.7 and before 4.13.1, the test rule TemporaryFolder contains a local information disclosure vulnerability. On Unix like systems, the system's temporary directory is shared between all users on that system. Because of this, when files and directories are written into this directory they are, by default, readable by other users on that same system. This vulnerability does not allow other users to overwrite the contents of these directories or files. This is purely an information disclosure vulnerability. This vulnerability impacts you if the JUnit tests write sensitive information, like API keys or passwords, into the temporary folder, and the JUnit tests execute in an environment where the OS has other untrusted users. Because certain JDK file system APIs were only added in JDK 1.7, this this fix is dependent upon the version of the JDK you are using. For Java 1.7 and higher users: this vulnerability is fixed in 4.13.1. For Java 1.6 and lower users: no patch is available, you must use the workaround below. If you are unable to patch, or are stuck running on Java 1.6, specifying the `java.io.tmpdir` system environment variable to a directory that is exclusively owned by the executing user will fix this vulnerability. For more information, including an example of vulnerable code, see the referenced GitHub Security Advisory. <p>Publish Date: 2020-10-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15250>CVE-2020-15250</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/junit-team/junit4/security/advisories/GHSA-269g-pwp5-87pp">https://github.com/junit-team/junit4/security/advisories/GHSA-269g-pwp5-87pp</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: junit:junit:4.13.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in junit jar cve medium severity vulnerability vulnerable library junit jar junit is a unit testing framework for java created by erich gamma and kent beck library home page a href path to dependency file pip kotlin app build gradle path to vulnerable library kkskks downloadresource xgizmn junit jar tmp ws ua kkskks downloadresource xgizmn junit jar dependency hierarchy x junit jar vulnerable library found in head commit a href found in base branch master vulnerability details in from version and before the test rule temporaryfolder contains a local information disclosure vulnerability on unix like systems the system s temporary directory is shared between all users on that system because of this when files and directories are written into this directory they are by default readable by other users on that same system this vulnerability does not allow other users to overwrite the contents of these directories or files this is purely an information disclosure vulnerability this vulnerability impacts you if the junit tests write sensitive information like api keys or passwords into the temporary folder and the junit tests execute in an environment where the os has other untrusted users because certain jdk file system apis were only added in jdk this this fix is dependent upon the version of the jdk you are using for java and higher users this vulnerability is fixed in for java and lower users no patch is available you must use the workaround below if you are unable to patch or are stuck running on java specifying the java io tmpdir system environment variable to a directory that is exclusively owned by the executing user will fix this vulnerability for more information including an example of vulnerable code see the referenced github security advisory publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution junit junit step up your open source security game with whitesource
0
660,450
21,966,602,817
IssuesEvent
2022-05-24 21:00:52
meower-media-co/Meower-Vanilla
https://api.github.com/repos/meower-media-co/Meower-Vanilla
closed
UI Bugs
Bug Low Priority Scratch
Here are some UI Bugs within Meower: 1) If you are in any menu apart from home, the create post button on the sidebar doesn't work 2) After entering a username/IP in the moderator panel, it continues to accept input after pressing 'Enter', it's just not shown in the confirmation screen but is still sent like that 3) As a level 1 moderator choosing some options in the moderator panel bring you to the wrong option, such as 'Ban User' leading to 'Pardon User IP' -- I know I caused this bug by rushing the level 1 moderator panel, sorry
1.0
UI Bugs - Here are some UI Bugs within Meower: 1) If you are in any menu apart from home, the create post button on the sidebar doesn't work 2) After entering a username/IP in the moderator panel, it continues to accept input after pressing 'Enter', it's just not shown in the confirmation screen but is still sent like that 3) As a level 1 moderator choosing some options in the moderator panel bring you to the wrong option, such as 'Ban User' leading to 'Pardon User IP' -- I know I caused this bug by rushing the level 1 moderator panel, sorry
non_process
ui bugs here are some ui bugs within meower if you are in any menu apart from home the create post button on the sidebar doesn t work after entering a username ip in the moderator panel it continues to accept input after pressing enter it s just not shown in the confirmation screen but is still sent like that as a level moderator choosing some options in the moderator panel bring you to the wrong option such as ban user leading to pardon user ip i know i caused this bug by rushing the level moderator panel sorry
0
402,445
27,369,543,865
IssuesEvent
2023-02-27 22:04:56
varikvalefor/agda-truthbrary
https://api.github.com/repos/varikvalefor/agda-truthbrary
closed
ni'o xu cadga fa lo nu basygau fi zo co'e
documentation preti
``` \newcommand\eqq[1]{ga jonai ga je la'o zoi.\ \B a .zoi.\ du la'o zoi.\ \B b .zoi.\ gi ga je ko'a goi la'o zoi.\ \F{isYes} \Sym \$ {#1} .zoi.\ du la'oi .\F{true}.\ gi co'e gi ko'a du la'o zoi.\ \F{nothing}.} ```
1.0
ni'o xu cadga fa lo nu basygau fi zo co'e - ``` \newcommand\eqq[1]{ga jonai ga je la'o zoi.\ \B a .zoi.\ du la'o zoi.\ \B b .zoi.\ gi ga je ko'a goi la'o zoi.\ \F{isYes} \Sym \$ {#1} .zoi.\ du la'oi .\F{true}.\ gi co'e gi ko'a du la'o zoi.\ \F{nothing}.} ```
non_process
ni o xu cadga fa lo nu basygau fi zo co e newcommand eqq ga jonai ga je la o zoi b a zoi du la o zoi b b zoi gi ga je ko a goi la o zoi f isyes sym zoi du la oi f true gi co e gi ko a du la o zoi f nothing
0
662
3,130,906,920
IssuesEvent
2015-09-09 12:15:09
e-government-ua/i
https://api.github.com/repos/e-government-ua/i
closed
Доработать систему вспомогательной технической авторизации.
active hi priority In process of testing test
Доработать класс: \i\wf-base\src\main\java\org\activity\rest\security\AccessKeyAuthFilter.java и связанные с ним классы: \i\wf-base\src\main\java\org\activity\rest\security\*.java Чтоб по мимо пары параметров(для авторизации кроме фиксированного логина/пароля энкоденного в хеадере): sAccessKey (ключ, по которому подтягивается сверяемый контент) nID_Subject (сам контент, с которым происходит сверка) Поддерживалась комбинация: sAccessKey (ключ, по которому подтягивается сверяемый контент) sAccessContract=Request (параметр, говорящий, что сверяемым контентом является сама строка запроса) При которой вместо сверки совпадения контента nID_Subject с контентом, полученным по ключу(sAccessKey), в качестве контента будет выступать строка самого текущего запроса. НО при этом, из строки реквеста важно будет вырезать сам параметр "&sAccessKey=(тот код, что пришел)" Например: "/wf-region/service/setPaymentStatus_TaskActiviti?sID_Order=TaskActiviti_105123&sID_PaymentSystem=Liqpay&sData=&sAccessContract=Request" (хотя по логике там еще должен был быть параметр &sAccessKey=5d1184c7-2d84-4931-a8f5-1032b50ed699) ВАЖНО: - Если по ключу sAccessKey вытягивается контент, но он не совпадет со "сверяемым" выдавать эксепшин с сообщением: "Contract data by sAccessKey is not equal !!! (sAccessContract=Request, sAccessKey=5d1184c7-2d84-4931-a8f5-1032b50ed699)" (а в логах выводить один и другой контент, для сравнения, чтоб можно было дебажить) - Если по ключу sAccessKey вообще не найден контента, выдавать эксепшин с сообщением: "Content not found by sAccessKey !!! (sAccessContract=Request, sAccessKey=5d1184c7-2d84-4931-a8f5-1032b50ed699)" П.С.: Это логическое продолжение развития предыдущей таски: https://github.com/e-government-ua/i/issues/502
1.0
Доработать систему вспомогательной технической авторизации. - Доработать класс: \i\wf-base\src\main\java\org\activity\rest\security\AccessKeyAuthFilter.java и связанные с ним классы: \i\wf-base\src\main\java\org\activity\rest\security\*.java Чтоб по мимо пары параметров(для авторизации кроме фиксированного логина/пароля энкоденного в хеадере): sAccessKey (ключ, по которому подтягивается сверяемый контент) nID_Subject (сам контент, с которым происходит сверка) Поддерживалась комбинация: sAccessKey (ключ, по которому подтягивается сверяемый контент) sAccessContract=Request (параметр, говорящий, что сверяемым контентом является сама строка запроса) При которой вместо сверки совпадения контента nID_Subject с контентом, полученным по ключу(sAccessKey), в качестве контента будет выступать строка самого текущего запроса. НО при этом, из строки реквеста важно будет вырезать сам параметр "&sAccessKey=(тот код, что пришел)" Например: "/wf-region/service/setPaymentStatus_TaskActiviti?sID_Order=TaskActiviti_105123&sID_PaymentSystem=Liqpay&sData=&sAccessContract=Request" (хотя по логике там еще должен был быть параметр &sAccessKey=5d1184c7-2d84-4931-a8f5-1032b50ed699) ВАЖНО: - Если по ключу sAccessKey вытягивается контент, но он не совпадет со "сверяемым" выдавать эксепшин с сообщением: "Contract data by sAccessKey is not equal !!! (sAccessContract=Request, sAccessKey=5d1184c7-2d84-4931-a8f5-1032b50ed699)" (а в логах выводить один и другой контент, для сравнения, чтоб можно было дебажить) - Если по ключу sAccessKey вообще не найден контента, выдавать эксепшин с сообщением: "Content not found by sAccessKey !!! (sAccessContract=Request, sAccessKey=5d1184c7-2d84-4931-a8f5-1032b50ed699)" П.С.: Это логическое продолжение развития предыдущей таски: https://github.com/e-government-ua/i/issues/502
process
доработать систему вспомогательной технической авторизации доработать класс i wf base src main java org activity rest security accesskeyauthfilter java и связанные с ним классы i wf base src main java org activity rest security java чтоб по мимо пары параметров для авторизации кроме фиксированного логина пароля энкоденного в хеадере saccesskey ключ по которому подтягивается сверяемый контент nid subject сам контент с которым происходит сверка поддерживалась комбинация saccesskey ключ по которому подтягивается сверяемый контент saccesscontract request параметр говорящий что сверяемым контентом является сама строка запроса при которой вместо сверки совпадения контента nid subject с контентом полученным по ключу saccesskey в качестве контента будет выступать строка самого текущего запроса но при этом из строки реквеста важно будет вырезать сам параметр saccesskey тот код что пришел например wf region service setpaymentstatus taskactiviti sid order taskactiviti sid paymentsystem liqpay sdata saccesscontract request хотя по логике там еще должен был быть параметр saccesskey важно если по ключу saccesskey вытягивается контент но он не совпадет со сверяемым выдавать эксепшин с сообщением contract data by saccesskey is not equal saccesscontract request saccesskey а в логах выводить один и другой контент для сравнения чтоб можно было дебажить если по ключу saccesskey вообще не найден контента выдавать эксепшин с сообщением content not found by saccesskey saccesscontract request saccesskey п с это логическое продолжение развития предыдущей таски
1
475,574
13,723,163,034
IssuesEvent
2020-10-03 07:51:22
VoxAndrews/Penquin-Thief
https://api.github.com/repos/VoxAndrews/Penquin-Thief
opened
Collision Issues - Walking through Walls (Player & AI)
bug priority (High)
**Describe the bug** Player's have encountered an issue where they are able to not only walk through walls but certian objects as well. This issue also carrys over to the Night Guard AI where they will also walk through certian objects. **Version Found** v0.3a **To Reproduce** Steps to reproduce the behavior: 1. Start the game 2. Begin moving around 3. Collide with an Object 4. Some objects will have no collision, and can be walked through **Expected behavior** When the player walks into an objects, they should not be able to walk through it. The same goes for the Night Guard, so when he walks into an objects, he should not be able to walk through it. **Screenshots/Videos** If applicable, add screenshots/videos to help explain your problem. **Desktop (please complete the following information):** - OS: Windows 10 **Additional context** Some objects did not have proper collision in the first place. I think the best way to tackle this will be to remove the objects currently in the scene, create a prefab of each individual obejct, and then place them back in. That way, we should be able to ensure that every object has proper collisions each time they are added as opposed to adding collision on a case by case basis.
1.0
Collision Issues - Walking through Walls (Player & AI) - **Describe the bug** Player's have encountered an issue where they are able to not only walk through walls but certian objects as well. This issue also carrys over to the Night Guard AI where they will also walk through certian objects. **Version Found** v0.3a **To Reproduce** Steps to reproduce the behavior: 1. Start the game 2. Begin moving around 3. Collide with an Object 4. Some objects will have no collision, and can be walked through **Expected behavior** When the player walks into an objects, they should not be able to walk through it. The same goes for the Night Guard, so when he walks into an objects, he should not be able to walk through it. **Screenshots/Videos** If applicable, add screenshots/videos to help explain your problem. **Desktop (please complete the following information):** - OS: Windows 10 **Additional context** Some objects did not have proper collision in the first place. I think the best way to tackle this will be to remove the objects currently in the scene, create a prefab of each individual obejct, and then place them back in. That way, we should be able to ensure that every object has proper collisions each time they are added as opposed to adding collision on a case by case basis.
non_process
collision issues walking through walls player ai describe the bug player s have encountered an issue where they are able to not only walk through walls but certian objects as well this issue also carrys over to the night guard ai where they will also walk through certian objects version found to reproduce steps to reproduce the behavior start the game begin moving around collide with an object some objects will have no collision and can be walked through expected behavior when the player walks into an objects they should not be able to walk through it the same goes for the night guard so when he walks into an objects he should not be able to walk through it screenshots videos if applicable add screenshots videos to help explain your problem desktop please complete the following information os windows additional context some objects did not have proper collision in the first place i think the best way to tackle this will be to remove the objects currently in the scene create a prefab of each individual obejct and then place them back in that way we should be able to ensure that every object has proper collisions each time they are added as opposed to adding collision on a case by case basis
0
203,657
15,886,580,043
IssuesEvent
2021-04-09 23:01:54
anitab-org/stem-diverse-tv
https://api.github.com/repos/anitab-org/stem-diverse-tv
opened
Create test cases documentation table for `POST /videos`
Category: Documentation/Training Category: Quality Assurance Open Source Hack Status: Available
## Description As a developer or tester, I need to know some test cases examples, so that I can have an idea of how I can test the app. Creating test cases will help everyone understand how to use the API. References: - [Quality Assurance test cases](https://github.com/anitab-org/mentorship-backend/blob/cf6df094e4fef735e135674e4d5697ded5060d7d/docs/quality-assurance-test-cases.md) for mentorship-backend project - Check out the API Swagger definition here: https://stem-diverse-tv.herokuapp.com/ ## Acceptance Criteria ### Update [Required] - [ ] Create quality assurance documentation on `/docs` folder - [ ] Create test cases documentation table for `POST /videos` API endpoint ## Definition of Done - [ ] All of the required items are completed. - [ ] Approval by 1 mentor. ## Estimation 30 min
1.0
Create test cases documentation table for `POST /videos` - ## Description As a developer or tester, I need to know some test cases examples, so that I can have an idea of how I can test the app. Creating test cases will help everyone understand how to use the API. References: - [Quality Assurance test cases](https://github.com/anitab-org/mentorship-backend/blob/cf6df094e4fef735e135674e4d5697ded5060d7d/docs/quality-assurance-test-cases.md) for mentorship-backend project - Check out the API Swagger definition here: https://stem-diverse-tv.herokuapp.com/ ## Acceptance Criteria ### Update [Required] - [ ] Create quality assurance documentation on `/docs` folder - [ ] Create test cases documentation table for `POST /videos` API endpoint ## Definition of Done - [ ] All of the required items are completed. - [ ] Approval by 1 mentor. ## Estimation 30 min
non_process
create test cases documentation table for post videos description as a developer or tester i need to know some test cases examples so that i can have an idea of how i can test the app creating test cases will help everyone understand how to use the api references for mentorship backend project check out the api swagger definition here acceptance criteria update create quality assurance documentation on docs folder create test cases documentation table for post videos api endpoint definition of done all of the required items are completed approval by mentor estimation min
0
6,279
9,256,517,891
IssuesEvent
2019-03-16 19:45:05
neuropoly/spinalcordtoolbox
https://api.github.com/repos/neuropoly/spinalcordtoolbox
closed
Optimize default radius parameter
card:WORK_IN_PROCESS enhancement sct_propseg
### Description Will check quality of segmentation in large testing with radius 4 and 5. Will keep the best. This test should be done once https://github.com/neuropoly/spinalcordtoolbox/issues/1454 is fixed. ### State of spinalcordtoolbox Spinal Cord Toolbox (master/a0617b16c5d7776a91ff8b8e2bc03c95fb869ba3) ### Additional Information Screen-shots, error logs and output files.
1.0
Optimize default radius parameter - ### Description Will check quality of segmentation in large testing with radius 4 and 5. Will keep the best. This test should be done once https://github.com/neuropoly/spinalcordtoolbox/issues/1454 is fixed. ### State of spinalcordtoolbox Spinal Cord Toolbox (master/a0617b16c5d7776a91ff8b8e2bc03c95fb869ba3) ### Additional Information Screen-shots, error logs and output files.
process
optimize default radius parameter description will check quality of segmentation in large testing with radius and will keep the best this test should be done once is fixed state of spinalcordtoolbox spinal cord toolbox master additional information screen shots error logs and output files
1
166,537
26,370,996,240
IssuesEvent
2023-01-11 20:40:08
team-peekabook/Peekabook-iOS
https://api.github.com/repos/team-peekabook/Peekabook-iOS
opened
[Feat] 책 삭제 뷰 구현 및 서버 연결
인영 🧸 Design 🎨 API 📡
## 🛠 Issue <!-- 이슈에 대해 간략하게 설명해주세요 --> 책장에 있는 책 삭제 뷰 구현 및 서버 연결 ## 📝 To-do <!-- 진행할 작업에 대해 적어주세요 --> - [ ] 팝업 뷰 구현 - [ ] 서버 연결
1.0
[Feat] 책 삭제 뷰 구현 및 서버 연결 - ## 🛠 Issue <!-- 이슈에 대해 간략하게 설명해주세요 --> 책장에 있는 책 삭제 뷰 구현 및 서버 연결 ## 📝 To-do <!-- 진행할 작업에 대해 적어주세요 --> - [ ] 팝업 뷰 구현 - [ ] 서버 연결
non_process
책 삭제 뷰 구현 및 서버 연결 🛠 issue 책장에 있는 책 삭제 뷰 구현 및 서버 연결 📝 to do 팝업 뷰 구현 서버 연결
0
165,360
12,838,708,878
IssuesEvent
2020-07-07 17:57:38
connext/indra
https://api.github.com/repos/connext/indra
opened
[all] SQLite -- SequelizeTimeoutError: SQLITE_BUSY: database is locked
Bug p1 Bugs/Tests/Blockers
**Describe the bug** During tps tests, SQLite sometimes throws the following: ``` SequelizeTimeoutError: SQLITE_BUSY: database is locked ``` It looks like this is called by deadlocking during concurrent reads/writes which SQLite doesn't handle very well by default. See: https://github.com/sequelize/sequelize/issues/10304 **To Reproduce** Run the tps tests a bunch until they error -- this will typically show up when one or many clients are syncing and erroring repeatedly.
1.0
[all] SQLite -- SequelizeTimeoutError: SQLITE_BUSY: database is locked - **Describe the bug** During tps tests, SQLite sometimes throws the following: ``` SequelizeTimeoutError: SQLITE_BUSY: database is locked ``` It looks like this is called by deadlocking during concurrent reads/writes which SQLite doesn't handle very well by default. See: https://github.com/sequelize/sequelize/issues/10304 **To Reproduce** Run the tps tests a bunch until they error -- this will typically show up when one or many clients are syncing and erroring repeatedly.
non_process
sqlite sequelizetimeouterror sqlite busy database is locked describe the bug during tps tests sqlite sometimes throws the following sequelizetimeouterror sqlite busy database is locked it looks like this is called by deadlocking during concurrent reads writes which sqlite doesn t handle very well by default see to reproduce run the tps tests a bunch until they error this will typically show up when one or many clients are syncing and erroring repeatedly
0
156,240
24,586,757,606
IssuesEvent
2022-10-13 20:28:50
hackforla/expunge-assist
https://api.github.com/repos/hackforla/expunge-assist
closed
Create wireframes for "Something else"
role: design priority: high size: 1pt feature: letter generator
### Overview Per previous research, Content found that our letter generator overlooked experiences and challenges that people may. have faced due to their records. To cover other areas, Content decided to create a general flow called "Something else" ### Action Items - [x] Update the mobile wireframe's layout - [x] Create the flow (screen with activated buttons and responses) - [x] Create a desktop version - [x] Create the desktop version flow (screen with activated buttons and responses) - [x] Share with content for approval - [x] Work with content to make issues for Dev ### Resources/Instructions You can find the mobile wireframe under the letter generator section or [click here](https://www.figma.com/file/hYqRxmBVtJbDv9DJXV6nra/Expunge-Assist-Main-Figma?node-id=5526%3A50959). Ask @SamHyler if you have any question regarding the content
1.0
Create wireframes for "Something else" - ### Overview Per previous research, Content found that our letter generator overlooked experiences and challenges that people may. have faced due to their records. To cover other areas, Content decided to create a general flow called "Something else" ### Action Items - [x] Update the mobile wireframe's layout - [x] Create the flow (screen with activated buttons and responses) - [x] Create a desktop version - [x] Create the desktop version flow (screen with activated buttons and responses) - [x] Share with content for approval - [x] Work with content to make issues for Dev ### Resources/Instructions You can find the mobile wireframe under the letter generator section or [click here](https://www.figma.com/file/hYqRxmBVtJbDv9DJXV6nra/Expunge-Assist-Main-Figma?node-id=5526%3A50959). Ask @SamHyler if you have any question regarding the content
non_process
create wireframes for something else overview per previous research content found that our letter generator overlooked experiences and challenges that people may have faced due to their records to cover other areas content decided to create a general flow called something else action items update the mobile wireframe s layout create the flow screen with activated buttons and responses create a desktop version create the desktop version flow screen with activated buttons and responses share with content for approval work with content to make issues for dev resources instructions you can find the mobile wireframe under the letter generator section or ask samhyler if you have any question regarding the content
0
2,586
5,345,096,067
IssuesEvent
2017-02-17 16:08:38
jlm2017/jlm-video-subtitles
https://api.github.com/repos/jlm2017/jlm-video-subtitles
opened
[subtitles] [fr] Les Français en ont par-dessus la tête des querelles religieuses - Mélenchon
Language: French Process: Someone is working on this issue Process: [1] Writing in progress
# Video title Les Français en ont par-dessus la tête des querelles religieuses - Mélenchon # URL https://www.youtube.com/watch?v=kozDd85bN_M # Youtube subtitles language French # Duration 7:55 # Subtitles URL https://www.youtube.com/timedtext_editor?ui=hd&bl=vmp&action_mde_edit_form=1&lang=fr&ref=player&tab=captions&v=kozDd85bN_M
2.0
[subtitles] [fr] Les Français en ont par-dessus la tête des querelles religieuses - Mélenchon - # Video title Les Français en ont par-dessus la tête des querelles religieuses - Mélenchon # URL https://www.youtube.com/watch?v=kozDd85bN_M # Youtube subtitles language French # Duration 7:55 # Subtitles URL https://www.youtube.com/timedtext_editor?ui=hd&bl=vmp&action_mde_edit_form=1&lang=fr&ref=player&tab=captions&v=kozDd85bN_M
process
les français en ont par dessus la tête des querelles religieuses mélenchon video title les français en ont par dessus la tête des querelles religieuses mélenchon url youtube subtitles language french duration subtitles url
1
711,489
24,465,516,195
IssuesEvent
2022-10-07 14:42:02
WebXDAO/WebXDAO.github.io
https://api.github.com/repos/WebXDAO/WebXDAO.github.io
closed
[OTHER] Use a custom domain
🟪 priority: none 🔒 staff-only
### What would you like to share? It looks more professional when the website is on a custom domain, instead of github pages. Maybe yall can try to get a custom domain after the revamp of the site? The issue is that webxdao.com is for a massive massive price on any domain buying service, just look at this: Namecheap: ₹740,209.17/yr (9,856.91 USD) Godaddy: ₹17,30,526.32/yr (23044.35 USD) But, as this is a organization maybe yall can try to get .org, cause those prices seem reasonable: Namecheap: ₹678.27/yr (9.03 USD) Godaddy: It's usually ₹ 1,399.00 but it's on discount now, but for the first year it's only ₹ 99.00 which comes to about (1.32 USD) So yeah, I think consider getting a .org domain, cause I don't think it's worth paying such a PREMIUM for .com 😆 ### Additional information Anyone know's why the .com domain is so expensive? Is it cause the domain as web in it or something? They call the domain webxdao.com, a premium domain. @EddieHubCommunity uses DigitalOcean for domains I believe, I don't really know their prices but yall can have a look. ### Would you like to work on this issue? - [ ] Yes, I want to work on this issue! ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
1.0
[OTHER] Use a custom domain - ### What would you like to share? It looks more professional when the website is on a custom domain, instead of github pages. Maybe yall can try to get a custom domain after the revamp of the site? The issue is that webxdao.com is for a massive massive price on any domain buying service, just look at this: Namecheap: ₹740,209.17/yr (9,856.91 USD) Godaddy: ₹17,30,526.32/yr (23044.35 USD) But, as this is a organization maybe yall can try to get .org, cause those prices seem reasonable: Namecheap: ₹678.27/yr (9.03 USD) Godaddy: It's usually ₹ 1,399.00 but it's on discount now, but for the first year it's only ₹ 99.00 which comes to about (1.32 USD) So yeah, I think consider getting a .org domain, cause I don't think it's worth paying such a PREMIUM for .com 😆 ### Additional information Anyone know's why the .com domain is so expensive? Is it cause the domain as web in it or something? They call the domain webxdao.com, a premium domain. @EddieHubCommunity uses DigitalOcean for domains I believe, I don't really know their prices but yall can have a look. ### Would you like to work on this issue? - [ ] Yes, I want to work on this issue! ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
non_process
use a custom domain what would you like to share it looks more professional when the website is on a custom domain instead of github pages maybe yall can try to get a custom domain after the revamp of the site the issue is that webxdao com is for a massive massive price on any domain buying service just look at this namecheap ₹ yr usd godaddy ₹ yr usd but as this is a organization maybe yall can try to get org cause those prices seem reasonable namecheap ₹ yr usd godaddy it s usually ₹ but it s on discount now but for the first year it s only ₹ which comes to about usd so yeah i think consider getting a org domain cause i don t think it s worth paying such a premium for com 😆 additional information anyone know s why the com domain is so expensive is it cause the domain as web in it or something they call the domain webxdao com a premium domain eddiehubcommunity uses digitalocean for domains i believe i don t really know their prices but yall can have a look would you like to work on this issue yes i want to work on this issue code of conduct i agree to follow this project s code of conduct
0
84,762
10,560,478,468
IssuesEvent
2019-10-04 13:58:39
iiCe89/Lynk_Social
https://api.github.com/repos/iiCe89/Lynk_Social
opened
Design new site in Figma
Design
Create Figma board which allows user's to view/work on so everybody knows what it looks like.
1.0
Design new site in Figma - Create Figma board which allows user's to view/work on so everybody knows what it looks like.
non_process
design new site in figma create figma board which allows user s to view work on so everybody knows what it looks like
0
1,358
3,918,118,568
IssuesEvent
2016-04-21 11:05:29
SpongePowered/Mixin
https://api.github.com/repos/SpongePowered/Mixin
closed
MirrorUtils crashing on bounded type generics
annotation processor bug
Similar to the issues with `TYPEVAR` with Java 7+, the following is producing the issue with MirrorUtils: ```java public <T extends Comparable<T>, V extends T> IBlockState withProperty(IProperty<T> property, V value) { ``` ``` Caused by: java.lang.IllegalArgumentException: Unable to compute upper bound of type symbol V at org.spongepowered.tools.MirrorUtils.getUpperBound(MirrorUtils.java:254) at org.spongepowered.tools.MirrorUtils.getTypeName(MirrorUtils.java:196) at org.spongepowered.tools.MirrorUtils.getJavaSignature(MirrorUtils.java:167) at org.spongepowered.tools.obfuscation.AnnotatedMixinElementHandler.validateTargetMethod(AnnotatedMixinElementHandler.java:325) at org.spongepowered.tools.obfuscation.AnnotatedMixinOverwriteHandler.registerOverwrite(AnnotatedMixinOverwriteHandler.java:48) at org.spongepowered.tools.obfuscation.AnnotatedMixin.registerOverwrite(AnnotatedMixin.java:342) at org.spongepowered.tools.obfuscation.AnnotatedMixins.registerOverwrite(AnnotatedMixins.java:352) at org.spongepowered.tools.obfuscation.TargetObfuscationProcessor.processOverwrites(TargetObfuscationProcessor.java:117) at org.spongepowered.tools.obfuscation.TargetObfuscationProcessor.process(TargetObfuscationProcessor.java:63) at com.sun.tools.javac.processing.JavacProcessingEnvironment.callProcessor(JavacProcessingEnvironment.java:794) at com.sun.tools.javac.processing.JavacProcessingEnvironment.discoverAndRunProcs(JavacProcessingEnvironment.java:705) at com.sun.tools.javac.processing.JavacProcessingEnvironment.access$1800(JavacProcessingEnvironment.java:91) at com.sun.tools.javac.processing.JavacProcessingEnvironment$Round.run(JavacProcessingEnvironment.java:1035) at com.sun.tools.javac.processing.JavacProcessingEnvironment.doProcessing(JavacProcessingEnvironment.java:1176) at com.sun.tools.javac.main.JavaCompiler.processAnnotations(JavaCompiler.java:1170) at com.sun.tools.javac.main.JavaCompiler.compile(JavaCompiler.java:856) at com.sun.tools.javac.main.Main.compile(Main.java:523) ... 76 more ```
1.0
MirrorUtils crashing on bounded type generics - Similar to the issues with `TYPEVAR` with Java 7+, the following is producing the issue with MirrorUtils: ```java public <T extends Comparable<T>, V extends T> IBlockState withProperty(IProperty<T> property, V value) { ``` ``` Caused by: java.lang.IllegalArgumentException: Unable to compute upper bound of type symbol V at org.spongepowered.tools.MirrorUtils.getUpperBound(MirrorUtils.java:254) at org.spongepowered.tools.MirrorUtils.getTypeName(MirrorUtils.java:196) at org.spongepowered.tools.MirrorUtils.getJavaSignature(MirrorUtils.java:167) at org.spongepowered.tools.obfuscation.AnnotatedMixinElementHandler.validateTargetMethod(AnnotatedMixinElementHandler.java:325) at org.spongepowered.tools.obfuscation.AnnotatedMixinOverwriteHandler.registerOverwrite(AnnotatedMixinOverwriteHandler.java:48) at org.spongepowered.tools.obfuscation.AnnotatedMixin.registerOverwrite(AnnotatedMixin.java:342) at org.spongepowered.tools.obfuscation.AnnotatedMixins.registerOverwrite(AnnotatedMixins.java:352) at org.spongepowered.tools.obfuscation.TargetObfuscationProcessor.processOverwrites(TargetObfuscationProcessor.java:117) at org.spongepowered.tools.obfuscation.TargetObfuscationProcessor.process(TargetObfuscationProcessor.java:63) at com.sun.tools.javac.processing.JavacProcessingEnvironment.callProcessor(JavacProcessingEnvironment.java:794) at com.sun.tools.javac.processing.JavacProcessingEnvironment.discoverAndRunProcs(JavacProcessingEnvironment.java:705) at com.sun.tools.javac.processing.JavacProcessingEnvironment.access$1800(JavacProcessingEnvironment.java:91) at com.sun.tools.javac.processing.JavacProcessingEnvironment$Round.run(JavacProcessingEnvironment.java:1035) at com.sun.tools.javac.processing.JavacProcessingEnvironment.doProcessing(JavacProcessingEnvironment.java:1176) at com.sun.tools.javac.main.JavaCompiler.processAnnotations(JavaCompiler.java:1170) at com.sun.tools.javac.main.JavaCompiler.compile(JavaCompiler.java:856) at com.sun.tools.javac.main.Main.compile(Main.java:523) ... 76 more ```
process
mirrorutils crashing on bounded type generics similar to the issues with typevar with java the following is producing the issue with mirrorutils java public v extends t iblockstate withproperty iproperty property v value caused by java lang illegalargumentexception unable to compute upper bound of type symbol v at org spongepowered tools mirrorutils getupperbound mirrorutils java at org spongepowered tools mirrorutils gettypename mirrorutils java at org spongepowered tools mirrorutils getjavasignature mirrorutils java at org spongepowered tools obfuscation annotatedmixinelementhandler validatetargetmethod annotatedmixinelementhandler java at org spongepowered tools obfuscation annotatedmixinoverwritehandler registeroverwrite annotatedmixinoverwritehandler java at org spongepowered tools obfuscation annotatedmixin registeroverwrite annotatedmixin java at org spongepowered tools obfuscation annotatedmixins registeroverwrite annotatedmixins java at org spongepowered tools obfuscation targetobfuscationprocessor processoverwrites targetobfuscationprocessor java at org spongepowered tools obfuscation targetobfuscationprocessor process targetobfuscationprocessor java at com sun tools javac processing javacprocessingenvironment callprocessor javacprocessingenvironment java at com sun tools javac processing javacprocessingenvironment discoverandrunprocs javacprocessingenvironment java at com sun tools javac processing javacprocessingenvironment access javacprocessingenvironment java at com sun tools javac processing javacprocessingenvironment round run javacprocessingenvironment java at com sun tools javac processing javacprocessingenvironment doprocessing javacprocessingenvironment java at com sun tools javac main javacompiler processannotations javacompiler java at com sun tools javac main javacompiler compile javacompiler java at com sun tools javac main main compile main java more
1
132,435
5,185,985,889
IssuesEvent
2017-01-20 12:27:01
openshift/origin
https://api.github.com/repos/openshift/origin
closed
prune images of schema 2 - should prune old image with config
component/imageregistry kind/bug kind/post-rebase priority/P1
I'm seeing this extended test failure both in current master and in the current rebase: ``` • Failure [56.993 seconds] [images] prune images /data/src/github.com/openshift/origin/test/extended/images/prune.go:85 of schema 2 /data/src/github.com/openshift/origin/test/extended/images/prune.go:84 should prune old image with config [It] /data/src/github.com/openshift/origin/test/extended/images/prune.go:83 Expected <string>: Dry run enabled - no modifications will be made. Add --confirm to remove images Deleting references from image streams to images ... STREAM IMAGE TAGS extended-test-prune-images-3lzp8-dvg65/prune sha256:989d62758d6a0b8a0c326d2ed84e5b9eaa4975f8e9a7880f7b1b55a5f77930d1 latest Deleting registry repository layer links ... REPO LAYER LINK extended-test-prune-images-3lzp8-dvg65/prune sha256:f1041d65727d2d7ee29d8f3f8975d2782b34246b100579b978ffaa9338816b7a extended-test-prune-images-3lzp8-dvg65/prune sha256:09ec8a0721608045992660137db1cc817b0fd945214b2452adad00bc5e12e7e8 Deleting registry layer blobs ... BLOB sha256:f1041d65727d2d7ee29d8f3f8975d2782b34246b100579b978ffaa9338816b7a sha256:e7e405bfcd8e5ae1e1b162c1f0a8f51ccfd76e5b04f55f4cc0b4ade19ea920e8 sha256:09ec8a0721608045992660137db1cc817b0fd945214b2452adad00bc5e12e7e8 sha256:cbd820c371f7c529183be0b318fc8763769817b5a229f911d9aa5064ee80f59a sha256:729b1393f8be141b88c8d7e220f12fb24121feabe6dbe5d291b6a6cc02bbff7c sha256:6d5981658b9e8728da196504c0818138a86d96e8ccafefa5e01672800d1a0f1f sha256:01875e248676e4a572ebc1659cef0effa5939d3e11130fa514d6220ec2bada20 sha256:1606620ef7ebe33d68dc819980b2ae7e0ad43a0c33ece9833a9fc468307286c3 sha256:2a88dc48f0522f21ec6889dd6177ef1b6132cedf10cad95b461759cc66343a1c sha256:d6d897d19b25daf0149bb3b1f70f9eb3c0b35df82db95e8999a8802e7136630b sha256:a3ed95caeb02ffe68cdd9fd84406680ae93d633cb16422d00e8a7c22955b46d4 sha256:45a2e645736c4c66ef34acce2407ded21f7a9b231199d3b92d6c9776df264729 sha256:b74941a5e64da5fe7f91e8e3e40529661b18b3e54ecacabf7b82a920a606921e sha256:23f0135afcd024032f3bba55e857222dd5a4932758ca22e93caa3f10945e18eb sha256:f7d81f468b937c89cd7803f82eb0441988fd31ae90429f33a731d0b656cfe26a sha256:50d0864be036eb14696d1c889d7884b863e4bc65bb79ddfb3160f3727cd1845d sha256:3f92963ff97b7ecef976010b5e1d257ca91ecf8a6bc70542c6dc4540b587ed7b Deleting registry repository manifest data ... REPO IMAGE extended-test-prune-images-3lzp8-dvg65/prune sha256:989d62758d6a0b8a0c326d2ed84e5b9eaa4975f8e9a7880f7b1b55a5f77930d1 Deleting images from server ... IMAGE sha256:5b4513a1d8cc9cec26539c85ec33a84fb888a88192180d38ebf12b83767e8875 sha256:1ac92c08fe32224683a82556b96ae94ee5f67381bd53ef671b3e9b7869959746 sha256:9a0e7809e9b5ad1fd706e7e66db51fdc5d6604123ffd28465e66445f8f2fad5d sha256:9fca5cd08b22948edb51b5238a2ffc5e63390809aa325b58a85bf4b525d19603 sha256:077bf3c64737df592307720f9e342b76a46830fed0885cab58c229273c93ae30 sha256:989d62758d6a0b8a0c326d2ed84e5b9eaa4975f8e9a7880f7b1b55a5f77930d1 to contain substring <string>: sha256:49cbcaeec119cd52cc777fd0b62cb5ad44103eae1cba41d6c10be8c822fe3650 /data/src/github.com/openshift/origin/test/extended/images/prune.go:136 ```
1.0
prune images of schema 2 - should prune old image with config - I'm seeing this extended test failure both in current master and in the current rebase: ``` • Failure [56.993 seconds] [images] prune images /data/src/github.com/openshift/origin/test/extended/images/prune.go:85 of schema 2 /data/src/github.com/openshift/origin/test/extended/images/prune.go:84 should prune old image with config [It] /data/src/github.com/openshift/origin/test/extended/images/prune.go:83 Expected <string>: Dry run enabled - no modifications will be made. Add --confirm to remove images Deleting references from image streams to images ... STREAM IMAGE TAGS extended-test-prune-images-3lzp8-dvg65/prune sha256:989d62758d6a0b8a0c326d2ed84e5b9eaa4975f8e9a7880f7b1b55a5f77930d1 latest Deleting registry repository layer links ... REPO LAYER LINK extended-test-prune-images-3lzp8-dvg65/prune sha256:f1041d65727d2d7ee29d8f3f8975d2782b34246b100579b978ffaa9338816b7a extended-test-prune-images-3lzp8-dvg65/prune sha256:09ec8a0721608045992660137db1cc817b0fd945214b2452adad00bc5e12e7e8 Deleting registry layer blobs ... BLOB sha256:f1041d65727d2d7ee29d8f3f8975d2782b34246b100579b978ffaa9338816b7a sha256:e7e405bfcd8e5ae1e1b162c1f0a8f51ccfd76e5b04f55f4cc0b4ade19ea920e8 sha256:09ec8a0721608045992660137db1cc817b0fd945214b2452adad00bc5e12e7e8 sha256:cbd820c371f7c529183be0b318fc8763769817b5a229f911d9aa5064ee80f59a sha256:729b1393f8be141b88c8d7e220f12fb24121feabe6dbe5d291b6a6cc02bbff7c sha256:6d5981658b9e8728da196504c0818138a86d96e8ccafefa5e01672800d1a0f1f sha256:01875e248676e4a572ebc1659cef0effa5939d3e11130fa514d6220ec2bada20 sha256:1606620ef7ebe33d68dc819980b2ae7e0ad43a0c33ece9833a9fc468307286c3 sha256:2a88dc48f0522f21ec6889dd6177ef1b6132cedf10cad95b461759cc66343a1c sha256:d6d897d19b25daf0149bb3b1f70f9eb3c0b35df82db95e8999a8802e7136630b sha256:a3ed95caeb02ffe68cdd9fd84406680ae93d633cb16422d00e8a7c22955b46d4 sha256:45a2e645736c4c66ef34acce2407ded21f7a9b231199d3b92d6c9776df264729 sha256:b74941a5e64da5fe7f91e8e3e40529661b18b3e54ecacabf7b82a920a606921e sha256:23f0135afcd024032f3bba55e857222dd5a4932758ca22e93caa3f10945e18eb sha256:f7d81f468b937c89cd7803f82eb0441988fd31ae90429f33a731d0b656cfe26a sha256:50d0864be036eb14696d1c889d7884b863e4bc65bb79ddfb3160f3727cd1845d sha256:3f92963ff97b7ecef976010b5e1d257ca91ecf8a6bc70542c6dc4540b587ed7b Deleting registry repository manifest data ... REPO IMAGE extended-test-prune-images-3lzp8-dvg65/prune sha256:989d62758d6a0b8a0c326d2ed84e5b9eaa4975f8e9a7880f7b1b55a5f77930d1 Deleting images from server ... IMAGE sha256:5b4513a1d8cc9cec26539c85ec33a84fb888a88192180d38ebf12b83767e8875 sha256:1ac92c08fe32224683a82556b96ae94ee5f67381bd53ef671b3e9b7869959746 sha256:9a0e7809e9b5ad1fd706e7e66db51fdc5d6604123ffd28465e66445f8f2fad5d sha256:9fca5cd08b22948edb51b5238a2ffc5e63390809aa325b58a85bf4b525d19603 sha256:077bf3c64737df592307720f9e342b76a46830fed0885cab58c229273c93ae30 sha256:989d62758d6a0b8a0c326d2ed84e5b9eaa4975f8e9a7880f7b1b55a5f77930d1 to contain substring <string>: sha256:49cbcaeec119cd52cc777fd0b62cb5ad44103eae1cba41d6c10be8c822fe3650 /data/src/github.com/openshift/origin/test/extended/images/prune.go:136 ```
non_process
prune images of schema should prune old image with config i m seeing this extended test failure both in current master and in the current rebase • failure prune images data src github com openshift origin test extended images prune go of schema data src github com openshift origin test extended images prune go should prune old image with config data src github com openshift origin test extended images prune go expected dry run enabled no modifications will be made add confirm to remove images deleting references from image streams to images stream image tags extended test prune images prune latest deleting registry repository layer links repo layer link extended test prune images prune extended test prune images prune deleting registry layer blobs blob deleting registry repository manifest data repo image extended test prune images prune deleting images from server image to contain substring data src github com openshift origin test extended images prune go
0
707,433
24,306,671,254
IssuesEvent
2022-09-29 18:04:06
opendp/dpcreator
https://api.github.com/repos/opendp/dpcreator
closed
Profiler on Boolean Type
DP Creator Stats-enhancement near-term-priority
I noticed something about the profiler in testing the demo DP Creator today. Out of the three boolean columns in the test data file I used (http://ec2-35-171-244-178.compute-1.amazonaws.com/dataset.xhtml?persistentId=doi:10.5072/FK2/V6YFKL&version=1.0), the first one (Boolean1) didn’t have any missing value, while the other two (Boolean2 and Boolean3) had missing values. The profiler recognized the first one as boolean and the other two as categorical. Is this expected behavior?
1.0
Profiler on Boolean Type - I noticed something about the profiler in testing the demo DP Creator today. Out of the three boolean columns in the test data file I used (http://ec2-35-171-244-178.compute-1.amazonaws.com/dataset.xhtml?persistentId=doi:10.5072/FK2/V6YFKL&version=1.0), the first one (Boolean1) didn’t have any missing value, while the other two (Boolean2 and Boolean3) had missing values. The profiler recognized the first one as boolean and the other two as categorical. Is this expected behavior?
non_process
profiler on boolean type i noticed something about the profiler in testing the demo dp creator today out of the three boolean columns in the test data file i used the first one didn’t have any missing value while the other two and had missing values the profiler recognized the first one as boolean and the other two as categorical is this expected behavior
0
141,850
12,980,301,593
IssuesEvent
2020-07-22 04:51:13
dotnet/SqlClient
https://api.github.com/repos/dotnet/SqlClient
closed
Documentation on how to use SqlAuthenticationProvider
:page_with_curl: Documentation
How do I register my own AuthenticationProvider? What changes do I need to make to the connection string to call my custom AcquireTokenAsync method?
1.0
Documentation on how to use SqlAuthenticationProvider - How do I register my own AuthenticationProvider? What changes do I need to make to the connection string to call my custom AcquireTokenAsync method?
non_process
documentation on how to use sqlauthenticationprovider how do i register my own authenticationprovider what changes do i need to make to the connection string to call my custom acquiretokenasync method
0
29,244
4,479,859,450
IssuesEvent
2016-08-27 21:57:26
artiator/artiator-frontend
https://api.github.com/repos/artiator/artiator-frontend
opened
Split components and containers
refactoring / automation / testing
Smart components in `src/containers`. Dumb components in `src/components`.
1.0
Split components and containers - Smart components in `src/containers`. Dumb components in `src/components`.
non_process
split components and containers smart components in src containers dumb components in src components
0
11,252
14,019,016,281
IssuesEvent
2020-10-29 17:35:41
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
How to consume pipeline resource predefined variable `resources.pipeline.<alias>.runName` in `name`
Pri2 devops-cicd-process/tech devops/prod doc-bug
I would like to know how I can reference the previous execution of a pipeline resource as part of my pipeline build number format, controlled by the attribute `name`. It is known that these [predefined pipeline resource variables are available (documentation)](https://docs.microsoft.com/en-us/azure/devops/pipelines/yaml-schema?view=azure-devops&tabs=schema%2Cparameter-schema#the-pipeline-resource-metadata-as-predefined-variables) Assuming my pipeline resource reference is such: ``` resources: pipelines: - pipeline: 'guide' source: 'guide (yaml)' trigger: tags: - example ``` And the referenced build completes and my referencing pipeline is correctly triggered, how do I express these predefined variables in the `name:` attribute? I've tried just a macro style expression (conscious that this is not a "task") ... ``` name: $(resources.pipeline.guide.runName) ``` ... and with quotes ... ``` name: '$(resources.pipeline.guide.runName)' ``` ... compile time expression syntax yielding a macro literal in quotes ... ``` name: ${{ '$(resources.pipeline.guide.runName)' }} ``` ... runtime expression variable index syntax (note the RHS completeness as per requirement) ... ``` name: $[ variables['resources.pipeline.guide.runName'] ] ``` ... runtime expression variable dereference syntax ... ``` name: $[ variables.resources.pipeline.guide.runName ] ``` ... even some weird combinations like compile expression with variable index syntax _just in case I was misunderstanding something_ ... ``` name: ${{ variables['resources.pipeline.guide.runName'] }} ``` ... even doing all the above into a variable first before sending it name, e.g. ... ``` variables: bn: $[ variables['resources.pipeline.guide.runName'] ] name: $(bn) ``` Nope! Nothing. Just defaults back to the weird default name, or in some cases, complains the build number format contained invalid characters: 1. It definitely doesn't, e.g. `0.1.0.80+2020q4` 2. If I later in the pipeline in a task call `Write-Host "##vso[build.updatebuildnumber]${env:RESOURCES_PIPELINE_GUIDE_RUNNAME}"` then it gets updated correctly. _The problem with (2), however, is that when you view the deployment history of an environment, the build number is not the one that is set in (2) - it instead lists whatever the build number was defined for `name:`._ So something is either bugged or the docs are not making any sense. Please help! _As a user, I want to give my pipeline runs much more useful names that are meaningful to my team._ --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93 * Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7 * Content: [Run (build) number - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/run-number.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
How to consume pipeline resource predefined variable `resources.pipeline.<alias>.runName` in `name` - I would like to know how I can reference the previous execution of a pipeline resource as part of my pipeline build number format, controlled by the attribute `name`. It is known that these [predefined pipeline resource variables are available (documentation)](https://docs.microsoft.com/en-us/azure/devops/pipelines/yaml-schema?view=azure-devops&tabs=schema%2Cparameter-schema#the-pipeline-resource-metadata-as-predefined-variables) Assuming my pipeline resource reference is such: ``` resources: pipelines: - pipeline: 'guide' source: 'guide (yaml)' trigger: tags: - example ``` And the referenced build completes and my referencing pipeline is correctly triggered, how do I express these predefined variables in the `name:` attribute? I've tried just a macro style expression (conscious that this is not a "task") ... ``` name: $(resources.pipeline.guide.runName) ``` ... and with quotes ... ``` name: '$(resources.pipeline.guide.runName)' ``` ... compile time expression syntax yielding a macro literal in quotes ... ``` name: ${{ '$(resources.pipeline.guide.runName)' }} ``` ... runtime expression variable index syntax (note the RHS completeness as per requirement) ... ``` name: $[ variables['resources.pipeline.guide.runName'] ] ``` ... runtime expression variable dereference syntax ... ``` name: $[ variables.resources.pipeline.guide.runName ] ``` ... even some weird combinations like compile expression with variable index syntax _just in case I was misunderstanding something_ ... ``` name: ${{ variables['resources.pipeline.guide.runName'] }} ``` ... even doing all the above into a variable first before sending it name, e.g. ... ``` variables: bn: $[ variables['resources.pipeline.guide.runName'] ] name: $(bn) ``` Nope! Nothing. Just defaults back to the weird default name, or in some cases, complains the build number format contained invalid characters: 1. It definitely doesn't, e.g. `0.1.0.80+2020q4` 2. If I later in the pipeline in a task call `Write-Host "##vso[build.updatebuildnumber]${env:RESOURCES_PIPELINE_GUIDE_RUNNAME}"` then it gets updated correctly. _The problem with (2), however, is that when you view the deployment history of an environment, the build number is not the one that is set in (2) - it instead lists whatever the build number was defined for `name:`._ So something is either bugged or the docs are not making any sense. Please help! _As a user, I want to give my pipeline runs much more useful names that are meaningful to my team._ --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93 * Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7 * Content: [Run (build) number - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/run-number.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
how to consume pipeline resource predefined variable resources pipeline runname in name i would like to know how i can reference the previous execution of a pipeline resource as part of my pipeline build number format controlled by the attribute name it is known that these assuming my pipeline resource reference is such resources pipelines pipeline guide source guide yaml trigger tags example and the referenced build completes and my referencing pipeline is correctly triggered how do i express these predefined variables in the name attribute i ve tried just a macro style expression conscious that this is not a task name resources pipeline guide runname and with quotes name resources pipeline guide runname compile time expression syntax yielding a macro literal in quotes name resources pipeline guide runname runtime expression variable index syntax note the rhs completeness as per requirement name runtime expression variable dereference syntax name even some weird combinations like compile expression with variable index syntax just in case i was misunderstanding something name variables even doing all the above into a variable first before sending it name e g variables bn name bn nope nothing just defaults back to the weird default name or in some cases complains the build number format contained invalid characters it definitely doesn t e g if i later in the pipeline in a task call write host vso env resources pipeline guide runname then it gets updated correctly the problem with however is that when you view the deployment history of an environment the build number is not the one that is set in it instead lists whatever the build number was defined for name so something is either bugged or the docs are not making any sense please help as a user i want to give my pipeline runs much more useful names that are meaningful to my team document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
129,939
12,421,015,136
IssuesEvent
2020-05-23 14:52:19
Samograje/tasks
https://api.github.com/repos/Samograje/tasks
closed
Dokumentacja
documentation
W celu zaliczenia efektu 4 („zna metody testowania aplikacji z uwzględnieniem specyfiki urządzeń mobilnych i ich wyposażenia, potrafi udokumentować projekt w celu umożliwienia testów innym programistom”) należy przygotować dokumentację zawierającą: a. opis systematycznych testów automatycznych i/lub manualnych (dotyczy to zarówno testów jednostkowych poszczególnych modułów kodu, jak i testów UX). Z dokumentacji powinno wynikać rozwiązanie co najmniej jednego problemu zidentyfikowanego w trakcie testowania b. schemat struktury wewnętrznej aplikacji: projekt wykorzystanych klas/struktur danych, pozwalający zrozumieć, interpretować i modyfikować działanie aplikacji. c. zrzut tablicy Trello/innego narzędzia wspomagającego zarządzanie projektem. Całość dokumentacji może mieć formę prezentacji lub dokumentu pdf. **Sugerowany termin przesłania tej części dokumentacji: 29.05.2020.**
1.0
Dokumentacja - W celu zaliczenia efektu 4 („zna metody testowania aplikacji z uwzględnieniem specyfiki urządzeń mobilnych i ich wyposażenia, potrafi udokumentować projekt w celu umożliwienia testów innym programistom”) należy przygotować dokumentację zawierającą: a. opis systematycznych testów automatycznych i/lub manualnych (dotyczy to zarówno testów jednostkowych poszczególnych modułów kodu, jak i testów UX). Z dokumentacji powinno wynikać rozwiązanie co najmniej jednego problemu zidentyfikowanego w trakcie testowania b. schemat struktury wewnętrznej aplikacji: projekt wykorzystanych klas/struktur danych, pozwalający zrozumieć, interpretować i modyfikować działanie aplikacji. c. zrzut tablicy Trello/innego narzędzia wspomagającego zarządzanie projektem. Całość dokumentacji może mieć formę prezentacji lub dokumentu pdf. **Sugerowany termin przesłania tej części dokumentacji: 29.05.2020.**
non_process
dokumentacja w celu zaliczenia efektu „zna metody testowania aplikacji z uwzględnieniem specyfiki urządzeń mobilnych i ich wyposażenia potrafi udokumentować projekt w celu umożliwienia testów innym programistom” należy przygotować dokumentację zawierającą a opis systematycznych testów automatycznych i lub manualnych dotyczy to zarówno testów jednostkowych poszczególnych modułów kodu jak i testów ux z dokumentacji powinno wynikać rozwiązanie co najmniej jednego problemu zidentyfikowanego w trakcie testowania b schemat struktury wewnętrznej aplikacji projekt wykorzystanych klas struktur danych pozwalający zrozumieć interpretować i modyfikować działanie aplikacji c zrzut tablicy trello innego narzędzia wspomagającego zarządzanie projektem całość dokumentacji może mieć formę prezentacji lub dokumentu pdf sugerowany termin przesłania tej części dokumentacji
0
15,999
3,491,460,974
IssuesEvent
2016-01-04 15:48:42
department-of-veterans-affairs/caseflow
https://api.github.com/repos/department-of-veterans-affairs/caseflow
closed
Run basic accessibility tests in travis with pa11y
frontend in progress new-feature testing
starting with even just hitting the login page would cover the base styles which would be a start
1.0
Run basic accessibility tests in travis with pa11y - starting with even just hitting the login page would cover the base styles which would be a start
non_process
run basic accessibility tests in travis with starting with even just hitting the login page would cover the base styles which would be a start
0
4,286
7,190,622,972
IssuesEvent
2018-02-02 17:55:15
Great-Hill-Corporation/quickBlocks
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
closed
Usage of ethprice should be improved, many doubts here as user
apps-ethPrice status-inprocess type-enhancement
I detail below the options that I do not understand as final user: -a (or --at) Report the price since nearest five minutes to :timestamp DOUBTS: What is the format of this option? What is the format of the timestamp passed? -p (or --period) Time increment for display. Default 120 mins or :t where t is in [5|15|30|120|240|1440] DOUBT: Again the same doubt, what is the expected format and how we pass a value here? -w (or --when) Time of day to start the display. :h determines the hour to start DOUBT: How to pass the hour and its supported format? -o (or --output) Output level (default 1). :2 exposes more data per record DOUBT: How to pass the level? and its supported format? Range of values (only 1 and 2)?
1.0
Usage of ethprice should be improved, many doubts here as user - I detail below the options that I do not understand as final user: -a (or --at) Report the price since nearest five minutes to :timestamp DOUBTS: What is the format of this option? What is the format of the timestamp passed? -p (or --period) Time increment for display. Default 120 mins or :t where t is in [5|15|30|120|240|1440] DOUBT: Again the same doubt, what is the expected format and how we pass a value here? -w (or --when) Time of day to start the display. :h determines the hour to start DOUBT: How to pass the hour and its supported format? -o (or --output) Output level (default 1). :2 exposes more data per record DOUBT: How to pass the level? and its supported format? Range of values (only 1 and 2)?
process
usage of ethprice should be improved many doubts here as user i detail below the options that i do not understand as final user a or at report the price since nearest five minutes to timestamp doubts what is the format of this option what is the format of the timestamp passed p or period time increment for display default mins or t where t is in doubt again the same doubt what is the expected format and how we pass a value here w or when time of day to start the display h determines the hour to start doubt how to pass the hour and its supported format o or output output level default exposes more data per record doubt how to pass the level and its supported format range of values only and
1
14,721
17,929,674,503
IssuesEvent
2021-09-10 07:31:10
googleapis/teeny-request
https://api.github.com/repos/googleapis/teeny-request
closed
Dependency Dashboard
type: process
This issue provides visibility into Renovate updates and their statuses. [Learn more](https://docs.renovatebot.com/key-concepts/dashboard/) ## Awaiting Schedule These updates are awaiting their schedule. Click on a checkbox to get an update now. - [ ] <!-- unschedule-branch=renovate/actions-setup-node-2.x -->chore(deps): update actions/setup-node action to v2 ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/mocha-9.x -->[chore(deps): update dependency mocha to v9](../pull/230) (`mocha`, `@types/mocha`) - [ ] <!-- recreate-branch=renovate/sinon-11.x -->[chore(deps): update dependency sinon to v11](../pull/228) - [ ] <!-- recreate-branch=renovate/node-fetch-3.x -->[fix(deps): update dependency node-fetch to v3](../pull/246) (`node-fetch`, `@types/node-fetch`) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue provides visibility into Renovate updates and their statuses. [Learn more](https://docs.renovatebot.com/key-concepts/dashboard/) ## Awaiting Schedule These updates are awaiting their schedule. Click on a checkbox to get an update now. - [ ] <!-- unschedule-branch=renovate/actions-setup-node-2.x -->chore(deps): update actions/setup-node action to v2 ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/mocha-9.x -->[chore(deps): update dependency mocha to v9](../pull/230) (`mocha`, `@types/mocha`) - [ ] <!-- recreate-branch=renovate/sinon-11.x -->[chore(deps): update dependency sinon to v11](../pull/228) - [ ] <!-- recreate-branch=renovate/node-fetch-3.x -->[fix(deps): update dependency node-fetch to v3](../pull/246) (`node-fetch`, `@types/node-fetch`) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue provides visibility into renovate updates and their statuses awaiting schedule these updates are awaiting their schedule click on a checkbox to get an update now chore deps update actions setup node action to ignored or blocked these are blocked by an existing closed pr and will not be recreated unless you click a checkbox below pull mocha types mocha pull pull node fetch types node fetch check this box to trigger a request for renovate to run again on this repository
1
76,182
26,279,928,498
IssuesEvent
2023-01-07 07:03:58
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
BUG: RegularGridInterpolator.__call__ fails when grid points arrays aren't C-contiguous in v1.10
defect
### Describe your issue. `scipy.interpolate.RegularGridInterpolator.__call__()` raises an exception when one or more arrays of grid coordinates used to construct the interpolator are non-C-contiguous. This appears to be new behavior in v1.10.0. In previous versions, there was no such restriction that these arrays must be C-contiguous -- the interpolator would produce valid results instead of raising an exception. ### Reproducing Code Example ```python from scipy.interpolate import RegularGridInterpolator import numpy as np def f(x, y, z): return 2 * x**3 + 3 * y**2 - z x = np.linspace(1, 4, 11) y = np.linspace(4, 7, 22) z = np.linspace(7, 9, 33) xg, yg ,zg = np.meshgrid(x, y[::2], z[::3], indexing='ij', sparse=True) data = f(xg, yg, zg) # Take a strided slice of `y` and `z` so that these input arrays are not # C-contiguous. interp = RegularGridInterpolator((x, y[::2], z[::3]), data) pts = np.array([[2.1, 6.2, 8.3], [3.3, 5.2, 7.1]]) interp(pts) ``` ### Error message ```shell Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/scratch/gunter/miniconda3/envs/tmp/lib/python3.9/site-packages/scipy/interpolate/_rgi.py", line 331, in __call__ indices, norm_distances = self._find_indices(xi.T) File "/scratch/gunter/miniconda3/envs/tmp/lib/python3.9/site-packages/scipy/interpolate/_rgi.py", line 499, in _find_indices return find_indices(self.grid, xi) File "_rgi_cython.pyx", line 95, in scipy.interpolate._rgi_cython.find_indices File "stringsource", line 660, in View.MemoryView.memoryview_cwrapper File "stringsource", line 350, in View.MemoryView.memoryview.__cinit__ ValueError: ndarray is not C-contiguous ``` ### SciPy/NumPy/Python version information 1.10.0 1.24.1 sys.version_info(major=3, minor=9, micro=15, releaselevel='final', serial=0)
1.0
BUG: RegularGridInterpolator.__call__ fails when grid points arrays aren't C-contiguous in v1.10 - ### Describe your issue. `scipy.interpolate.RegularGridInterpolator.__call__()` raises an exception when one or more arrays of grid coordinates used to construct the interpolator are non-C-contiguous. This appears to be new behavior in v1.10.0. In previous versions, there was no such restriction that these arrays must be C-contiguous -- the interpolator would produce valid results instead of raising an exception. ### Reproducing Code Example ```python from scipy.interpolate import RegularGridInterpolator import numpy as np def f(x, y, z): return 2 * x**3 + 3 * y**2 - z x = np.linspace(1, 4, 11) y = np.linspace(4, 7, 22) z = np.linspace(7, 9, 33) xg, yg ,zg = np.meshgrid(x, y[::2], z[::3], indexing='ij', sparse=True) data = f(xg, yg, zg) # Take a strided slice of `y` and `z` so that these input arrays are not # C-contiguous. interp = RegularGridInterpolator((x, y[::2], z[::3]), data) pts = np.array([[2.1, 6.2, 8.3], [3.3, 5.2, 7.1]]) interp(pts) ``` ### Error message ```shell Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/scratch/gunter/miniconda3/envs/tmp/lib/python3.9/site-packages/scipy/interpolate/_rgi.py", line 331, in __call__ indices, norm_distances = self._find_indices(xi.T) File "/scratch/gunter/miniconda3/envs/tmp/lib/python3.9/site-packages/scipy/interpolate/_rgi.py", line 499, in _find_indices return find_indices(self.grid, xi) File "_rgi_cython.pyx", line 95, in scipy.interpolate._rgi_cython.find_indices File "stringsource", line 660, in View.MemoryView.memoryview_cwrapper File "stringsource", line 350, in View.MemoryView.memoryview.__cinit__ ValueError: ndarray is not C-contiguous ``` ### SciPy/NumPy/Python version information 1.10.0 1.24.1 sys.version_info(major=3, minor=9, micro=15, releaselevel='final', serial=0)
non_process
bug regulargridinterpolator call fails when grid points arrays aren t c contiguous in describe your issue scipy interpolate regulargridinterpolator call raises an exception when one or more arrays of grid coordinates used to construct the interpolator are non c contiguous this appears to be new behavior in in previous versions there was no such restriction that these arrays must be c contiguous the interpolator would produce valid results instead of raising an exception reproducing code example python from scipy interpolate import regulargridinterpolator import numpy as np def f x y z return x y z x np linspace y np linspace z np linspace xg yg zg np meshgrid x y z indexing ij sparse true data f xg yg zg take a strided slice of y and z so that these input arrays are not c contiguous interp regulargridinterpolator x y z data pts np array interp pts error message shell traceback most recent call last file line in file scratch gunter envs tmp lib site packages scipy interpolate rgi py line in call indices norm distances self find indices xi t file scratch gunter envs tmp lib site packages scipy interpolate rgi py line in find indices return find indices self grid xi file rgi cython pyx line in scipy interpolate rgi cython find indices file stringsource line in view memoryview memoryview cwrapper file stringsource line in view memoryview memoryview cinit valueerror ndarray is not c contiguous scipy numpy python version information sys version info major minor micro releaselevel final serial
0
16,518
21,527,816,316
IssuesEvent
2022-04-28 20:22:34
googleapis/google-cloud-php-common-protos
https://api.github.com/repos/googleapis/google-cloud-php-common-protos
closed
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * client_documentation must match pattern "^https://.*" in .repo-metadata.json * release_level must be equal to one of the allowed values in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * client_documentation must match pattern "^https://.*" in .repo-metadata.json * release_level must be equal to one of the allowed values in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 client documentation must match pattern in repo metadata json release level must be equal to one of the allowed values in repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
1
10,452
13,233,328,135
IssuesEvent
2020-08-18 14:38:56
googleapis/google-api-dotnet-client
https://api.github.com/repos/googleapis/google-api-dotnet-client
closed
Remove odd comment at the start of each API file
type: process
Our generator adds comments like this at the start of each file: ```csharp /** * \brief * Access Approval API Version v1 * * \section ApiInfo API Version Information * <table> * <tr><th>API * <td><a href='https://cloud.google.com/access-approval/docs'>Access Approval API</a> * <tr><th>API Version<td>v1 * <tr><th>API Rev<td>20200726 (2033) * <tr><th>API Docs * <td><a href='https://cloud.google.com/access-approval/docs'> * https://cloud.google.com/access-approval/docs</a> * <tr><th>Discovery Name<td>accessapproval * </table> * * \section ForMoreInfo For More Information * * The complete API documentation for using Access Approval API can be found at * <a href='https://cloud.google.com/access-approval/docs'>https://cloud.google.com/access-approval/docs</a>. * * For more information about the Google APIs Client Library for .NET, see * <a href='https://developers.google.com/api-client-library/dotnet/get_started'> * https://developers.google.com/api-client-library/dotnet/get_started</a> */ ``` I'm sure there was a good reason for this at one point, but I don't think it's needed now, and I'd rather not generate it in the new generator. There's another comment like this: ```csharp //------------------------------------------------------------------------------ // <auto-generated> // This code was generated by google-apis-code-generator 1.5.1 // C# generator version: 1.48.0 // // Changes to this file may cause incorrect behavior and will be lost if // the code is regenerated. // </auto-generated> //------------------------------------------------------------------------------ ``` We could remove that too, probably just replacing it with the same thing we use in GAPIC: ```csharp // Generated code. DO NOT EDIT! ``` It's possible that including the version information could be useful, but I'm not immediately convinced. We could always bring it back later, after the transition.
1.0
Remove odd comment at the start of each API file - Our generator adds comments like this at the start of each file: ```csharp /** * \brief * Access Approval API Version v1 * * \section ApiInfo API Version Information * <table> * <tr><th>API * <td><a href='https://cloud.google.com/access-approval/docs'>Access Approval API</a> * <tr><th>API Version<td>v1 * <tr><th>API Rev<td>20200726 (2033) * <tr><th>API Docs * <td><a href='https://cloud.google.com/access-approval/docs'> * https://cloud.google.com/access-approval/docs</a> * <tr><th>Discovery Name<td>accessapproval * </table> * * \section ForMoreInfo For More Information * * The complete API documentation for using Access Approval API can be found at * <a href='https://cloud.google.com/access-approval/docs'>https://cloud.google.com/access-approval/docs</a>. * * For more information about the Google APIs Client Library for .NET, see * <a href='https://developers.google.com/api-client-library/dotnet/get_started'> * https://developers.google.com/api-client-library/dotnet/get_started</a> */ ``` I'm sure there was a good reason for this at one point, but I don't think it's needed now, and I'd rather not generate it in the new generator. There's another comment like this: ```csharp //------------------------------------------------------------------------------ // <auto-generated> // This code was generated by google-apis-code-generator 1.5.1 // C# generator version: 1.48.0 // // Changes to this file may cause incorrect behavior and will be lost if // the code is regenerated. // </auto-generated> //------------------------------------------------------------------------------ ``` We could remove that too, probably just replacing it with the same thing we use in GAPIC: ```csharp // Generated code. DO NOT EDIT! ``` It's possible that including the version information could be useful, but I'm not immediately convinced. We could always bring it back later, after the transition.
process
remove odd comment at the start of each api file our generator adds comments like this at the start of each file csharp brief access approval api version section apiinfo api version information api api version api rev api docs a href discovery name accessapproval section formoreinfo for more information the complete api documentation for using access approval api can be found at a href for more information about the google apis client library for net see a href i m sure there was a good reason for this at one point but i don t think it s needed now and i d rather not generate it in the new generator there s another comment like this csharp this code was generated by google apis code generator c generator version changes to this file may cause incorrect behavior and will be lost if the code is regenerated we could remove that too probably just replacing it with the same thing we use in gapic csharp generated code do not edit it s possible that including the version information could be useful but i m not immediately convinced we could always bring it back later after the transition
1
7,761
10,883,200,015
IssuesEvent
2019-11-18 03:50:07
zchenry/papers
https://api.github.com/repos/zchenry/papers
opened
Active Learning for Level Set Estimation
Active Search Gaussian Process
Alkis Gotovos, Nathalie Casati, Gregory Hitz, Andreas Krause IJCAI 2013 https://drive.google.com/open?id=1eU_9lInTqUxByVmnDPKtANdNxfcZpt0p Gaussian process based method to solve a general active search problem.
1.0
Active Learning for Level Set Estimation - Alkis Gotovos, Nathalie Casati, Gregory Hitz, Andreas Krause IJCAI 2013 https://drive.google.com/open?id=1eU_9lInTqUxByVmnDPKtANdNxfcZpt0p Gaussian process based method to solve a general active search problem.
process
active learning for level set estimation alkis gotovos nathalie casati gregory hitz andreas krause ijcai gaussian process based method to solve a general active search problem
1
593,556
18,011,079,725
IssuesEvent
2021-09-16 08:40:57
ntop/ntopng
https://api.github.com/repos/ntop/ntopng
opened
Color inconsistency
low-priority bug
In this example, there is only one alert with score 50 (yellow) and the icon in the menubar is red instead of yellow ![Screenshot 2021-09-16 at 10 38 18](https://user-images.githubusercontent.com/4493366/133579772-de6042d1-754e-4bb1-a2c6-9ec1a2178959.png)
1.0
Color inconsistency - In this example, there is only one alert with score 50 (yellow) and the icon in the menubar is red instead of yellow ![Screenshot 2021-09-16 at 10 38 18](https://user-images.githubusercontent.com/4493366/133579772-de6042d1-754e-4bb1-a2c6-9ec1a2178959.png)
non_process
color inconsistency in this example there is only one alert with score yellow and the icon in the menubar is red instead of yellow
0
65,698
8,827,047,210
IssuesEvent
2019-01-03 06:54:58
PegaSysEng/pantheon
https://api.github.com/repos/PegaSysEng/pantheon
opened
Document ibft_* JSON-RPC API methods
documentation
### Description As a user, I want the ibft_* JSON-RPC API methods documented so that I know when and how to use them. ### Acceptance Criteria * Methods documented in JSON-RPC API reference: - `ibft_getPendingVotes` - `ibft_getValidatorsByBlockHash` - `ibft_getValidatorsByBlockNumber` - `ibft_proposeValidatorVote` - `ibft_proposeValidatorVote` * Conceptual content about when and how to use added * Documentation tested
1.0
Document ibft_* JSON-RPC API methods - ### Description As a user, I want the ibft_* JSON-RPC API methods documented so that I know when and how to use them. ### Acceptance Criteria * Methods documented in JSON-RPC API reference: - `ibft_getPendingVotes` - `ibft_getValidatorsByBlockHash` - `ibft_getValidatorsByBlockNumber` - `ibft_proposeValidatorVote` - `ibft_proposeValidatorVote` * Conceptual content about when and how to use added * Documentation tested
non_process
document ibft json rpc api methods description as a user i want the ibft json rpc api methods documented so that i know when and how to use them acceptance criteria methods documented in json rpc api reference ibft getpendingvotes ibft getvalidatorsbyblockhash ibft getvalidatorsbyblocknumber ibft proposevalidatorvote ibft proposevalidatorvote conceptual content about when and how to use added documentation tested
0
418,772
28,127,072,603
IssuesEvent
2023-03-31 18:44:25
liveview-native/liveview-client-swiftui
https://api.github.com/repos/liveview-native/liveview-client-swiftui
closed
Documentation: SwiftUI -> User Interface Element -. Controls and indicator -> Buttons: EditButton
documentation
As a PO I'd like to have documentation for the [EditButton](https://github.com/liveviewnative/liveview-client-swiftui/issues/61)
1.0
Documentation: SwiftUI -> User Interface Element -. Controls and indicator -> Buttons: EditButton - As a PO I'd like to have documentation for the [EditButton](https://github.com/liveviewnative/liveview-client-swiftui/issues/61)
non_process
documentation swiftui user interface element controls and indicator buttons editbutton as a po i d like to have documentation for the
0
22,557
31,777,733,001
IssuesEvent
2023-09-12 15:20:51
IMAP-Science-Operations-Center/imap_processing
https://api.github.com/repos/IMAP-Science-Operations-Center/imap_processing
opened
[L5] Produce L1B CDF formatted data products for CoDICE
CoDICE L1 Requirement: Level 5 Untested Parent Req:SDC Execution of L1 Processing Software
### Summary of the L5 requirement The SDC needs to produce L1B CDF formatted data products for the CoDICEi instrument. The data products for L1B can be found in the CoDICE algorithm document here: https://lasp.colorado.edu/galaxy/display/IMAP/IMAP+Algorithm+Document+from+Instrument+Teams ### Parent requirement SOC-SDC-L4-29 | SDC Execution of L1 Processing Software | The SDC shall be capable of executing L1 processing software to produce L1 data products within seven (7) days of receipt of all requisite data inputs. | Requirement | The SDC produces L1 data products via Instrument Team supplied algorithms and SDC produced processing code. | SOC-L3-57 | Test | SIT-3: L0 -> L1 processing | SIT-3 -- | -- | -- | -- | -- | -- | -- | -- | -- ### Tasks Create the following L1B data products: - [ ] TBD
1.0
[L5] Produce L1B CDF formatted data products for CoDICE - ### Summary of the L5 requirement The SDC needs to produce L1B CDF formatted data products for the CoDICEi instrument. The data products for L1B can be found in the CoDICE algorithm document here: https://lasp.colorado.edu/galaxy/display/IMAP/IMAP+Algorithm+Document+from+Instrument+Teams ### Parent requirement SOC-SDC-L4-29 | SDC Execution of L1 Processing Software | The SDC shall be capable of executing L1 processing software to produce L1 data products within seven (7) days of receipt of all requisite data inputs. | Requirement | The SDC produces L1 data products via Instrument Team supplied algorithms and SDC produced processing code. | SOC-L3-57 | Test | SIT-3: L0 -> L1 processing | SIT-3 -- | -- | -- | -- | -- | -- | -- | -- | -- ### Tasks Create the following L1B data products: - [ ] TBD
process
produce cdf formatted data products for codice summary of the requirement the sdc needs to produce cdf formatted data products for the codicei instrument the data products for can be found in the codice algorithm document here parent requirement soc sdc sdc execution of processing software the sdc shall be capable of executing processing software to produce data products within seven days of receipt of all requisite data inputs requirement the sdc produces data products via instrument team supplied algorithms and sdc produced processing code soc test sit processing sit tasks create the following data products tbd
1
827,079
31,724,023,284
IssuesEvent
2023-09-10 18:36:43
easystats/performance
https://api.github.com/repos/easystats/performance
closed
Errors in CRAN checks
bug :bug: high priority :runner:
@rempsyc Not sure why these tests suddenly fail: https://cran.r-project.org/web/checks/check_results_performance.html This commit should fix it, i.e. it should no longer error when `NULL` is returned: https://github.com/easystats/performance/commit/7b181e128215e395d09976053feb9bb8753fb137 Need to fix tests though, because these tests now would return `NULL` on CRAN: (https://github.com/easystats/performance/blob/7b181e128215e395d09976053feb9bb8753fb137/tests/testthat/test-check_outliers.R#L100-L112) Maybe we just skip them on CRAN, and I wonder why we need `rstan` here?
1.0
Errors in CRAN checks - @rempsyc Not sure why these tests suddenly fail: https://cran.r-project.org/web/checks/check_results_performance.html This commit should fix it, i.e. it should no longer error when `NULL` is returned: https://github.com/easystats/performance/commit/7b181e128215e395d09976053feb9bb8753fb137 Need to fix tests though, because these tests now would return `NULL` on CRAN: (https://github.com/easystats/performance/blob/7b181e128215e395d09976053feb9bb8753fb137/tests/testthat/test-check_outliers.R#L100-L112) Maybe we just skip them on CRAN, and I wonder why we need `rstan` here?
non_process
errors in cran checks rempsyc not sure why these tests suddenly fail this commit should fix it i e it should no longer error when null is returned need to fix tests though because these tests now would return null on cran maybe we just skip them on cran and i wonder why we need rstan here
0