Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
5,816
| 8,651,268,682
|
IssuesEvent
|
2018-11-27 02:18:38
|
google/codeworld
|
https://api.github.com/repos/google/codeworld
|
closed
|
Automatic requirement checker for exercises
|
enhancement in process
|
Idea from class today, that @marcianx and I plan to work on this evening.
When students are completing an exercise in class, there are usually some requirements for success. For example, in class today, we wanted students to redefine three variables to be applications of the same function. Enforcing this by hand is work-intensive, and students are likely to ignore specific requirements if they are not checked. So, we have the following idea.
First, define a language for requirements that can be embedded into comments in a program. When such a comment is found, the server will parse it, check each of these requirements against the syntax tree for the code, and emit an extra diagnostic that summarizes the status of each item on the checklist. The client will look for that diagnostic, and if it's found, will display a summary of the requirements in a popup window when the program runs, with checkmarks for those that are satisfied, and an X for anything that's not satisfied.
We're starting with the ideas we need for a specific exercise. The specific requirements:
1. "This top-level definition is not modified." Requires a name for the top-level definition, and a hash of the definition. Verifies that the parsed definition in the submitted code (ignoring `SrcLoc`s) matches the one provided.
2. "This variable is defined by applying that function." Requires a variable name, and a function name. The definition must be `var = func(...)`
3. "This function has no incomplete pattern matches." This is a built-in warning in GHC, but I believe it's disabled by default. We can give a conservative approximation by just requiring that all function arguments are variables.
The language is to-be-decided, but some nice features are: (1) groups of specific requirements can be given a human-readable name, to make the diagnostic more readable, and (2) there can be multiple named groups, to be shown as a checklist.
|
1.0
|
Automatic requirement checker for exercises - Idea from class today, that @marcianx and I plan to work on this evening.
When students are completing an exercise in class, there are usually some requirements for success. For example, in class today, we wanted students to redefine three variables to be applications of the same function. Enforcing this by hand is work-intensive, and students are likely to ignore specific requirements if they are not checked. So, we have the following idea.
First, define a language for requirements that can be embedded into comments in a program. When such a comment is found, the server will parse it, check each of these requirements against the syntax tree for the code, and emit an extra diagnostic that summarizes the status of each item on the checklist. The client will look for that diagnostic, and if it's found, will display a summary of the requirements in a popup window when the program runs, with checkmarks for those that are satisfied, and an X for anything that's not satisfied.
We're starting with the ideas we need for a specific exercise. The specific requirements:
1. "This top-level definition is not modified." Requires a name for the top-level definition, and a hash of the definition. Verifies that the parsed definition in the submitted code (ignoring `SrcLoc`s) matches the one provided.
2. "This variable is defined by applying that function." Requires a variable name, and a function name. The definition must be `var = func(...)`
3. "This function has no incomplete pattern matches." This is a built-in warning in GHC, but I believe it's disabled by default. We can give a conservative approximation by just requiring that all function arguments are variables.
The language is to-be-decided, but some nice features are: (1) groups of specific requirements can be given a human-readable name, to make the diagnostic more readable, and (2) there can be multiple named groups, to be shown as a checklist.
|
process
|
automatic requirement checker for exercises idea from class today that marcianx and i plan to work on this evening when students are completing an exercise in class there are usually some requirements for success for example in class today we wanted students to redefine three variables to be applications of the same function enforcing this by hand is work intensive and students are likely to ignore specific requirements if they are not checked so we have the following idea first define a language for requirements that can be embedded into comments in a program when such a comment is found the server will parse it check each of these requirements against the syntax tree for the code and emit an extra diagnostic that summarizes the status of each item on the checklist the client will look for that diagnostic and if it s found will display a summary of the requirements in a popup window when the program runs with checkmarks for those that are satisfied and an x for anything that s not satisfied we re starting with the ideas we need for a specific exercise the specific requirements this top level definition is not modified requires a name for the top level definition and a hash of the definition verifies that the parsed definition in the submitted code ignoring srcloc s matches the one provided this variable is defined by applying that function requires a variable name and a function name the definition must be var func this function has no incomplete pattern matches this is a built in warning in ghc but i believe it s disabled by default we can give a conservative approximation by just requiring that all function arguments are variables the language is to be decided but some nice features are groups of specific requirements can be given a human readable name to make the diagnostic more readable and there can be multiple named groups to be shown as a checklist
| 1
|
12,855
| 15,240,065,186
|
IssuesEvent
|
2021-02-19 05:56:41
|
gfx-rs/naga
|
https://api.github.com/repos/gfx-rs/naga
|
closed
|
Analysis doesn't process Call statement correctly
|
area: processing kind: bug
|
It should follow the same logic as `Expression::Call`, but currently it's doing less than that.
|
1.0
|
Analysis doesn't process Call statement correctly - It should follow the same logic as `Expression::Call`, but currently it's doing less than that.
|
process
|
analysis doesn t process call statement correctly it should follow the same logic as expression call but currently it s doing less than that
| 1
|
171,838
| 27,186,110,280
|
IssuesEvent
|
2023-02-19 08:17:32
|
WordPress/wporg-developer
|
https://api.github.com/repos/WordPress/wporg-developer
|
closed
|
Implement WP version archive view.
|
Redesign
|
Users should be able to view the code changes between wp versions. This page also contains filters and a unique update to the breadcrumb.
## Screenshot

[Figma](https://www.figma.com/file/2WxlJFzMJvqPfZL1EkAOVp/Developer-Resources?node-id=1382%3A14258&t=w09GjUs65FuEPvwO-0)
|
1.0
|
Implement WP version archive view. - Users should be able to view the code changes between wp versions. This page also contains filters and a unique update to the breadcrumb.
## Screenshot

[Figma](https://www.figma.com/file/2WxlJFzMJvqPfZL1EkAOVp/Developer-Resources?node-id=1382%3A14258&t=w09GjUs65FuEPvwO-0)
|
non_process
|
implement wp version archive view users should be able to view the code changes between wp versions this page also contains filters and a unique update to the breadcrumb screenshot
| 0
|
21,648
| 30,083,376,504
|
IssuesEvent
|
2023-06-29 06:35:47
|
inmanta/web-console
|
https://api.github.com/repos/inmanta/web-console
|
opened
|
Dump server and agent log when e2e tests fail
|
process
|
The e2e tests use an Inmanta server created by the local-setup tool. When a test case fails, it's very useful to have access to the server and agent logs to investigate what went wrong. This information goes lost when `yarn kill-server` is executed at the end of the Jenkins pipeline. It would be good to archive this information when a test case failed.
Tasks:
- [ ] Adjust the Jenkins pipeline to archive the server and agent logs when an e2e test fails.
- [ ] Log a timestamp at the start of each e2e test case. This will allow us to correlate the logs with a specific test case.
|
1.0
|
Dump server and agent log when e2e tests fail - The e2e tests use an Inmanta server created by the local-setup tool. When a test case fails, it's very useful to have access to the server and agent logs to investigate what went wrong. This information goes lost when `yarn kill-server` is executed at the end of the Jenkins pipeline. It would be good to archive this information when a test case failed.
Tasks:
- [ ] Adjust the Jenkins pipeline to archive the server and agent logs when an e2e test fails.
- [ ] Log a timestamp at the start of each e2e test case. This will allow us to correlate the logs with a specific test case.
|
process
|
dump server and agent log when tests fail the tests use an inmanta server created by the local setup tool when a test case fails it s very useful to have access to the server and agent logs to investigate what went wrong this information goes lost when yarn kill server is executed at the end of the jenkins pipeline it would be good to archive this information when a test case failed tasks adjust the jenkins pipeline to archive the server and agent logs when an test fails log a timestamp at the start of each test case this will allow us to correlate the logs with a specific test case
| 1
|
6,942
| 9,219,359,200
|
IssuesEvent
|
2019-03-11 15:16:08
|
LLK/scratch-blocks
|
https://api.github.com/repos/LLK/scratch-blocks
|
closed
|
Blocks in "Echo of the Light - prototype" do not show up
|
bug compatibility
|
### Expected Behavior
The blocks show up (when you have the "Pen Text Engine" sprite selected), looking something like this:

### Actual Behavior
The blocks don't show up at all! Note that the the scripts *run*, though - if you click the green flag, the project will function correctly. So I think this is a problem in scratch-blocks, preventing the blocks from blocks from showing up, but they exist in the VM properly.
### Steps to Reproduce
Load the project in Scratch 3.0: https://llk.github.io/scratch-gui/develop/#269991846
### Operating System and Browser
I can repro this in Firefox (65.0b2). The original post pointing out this compatibility issue was from done while Chrome 70.0.3538.110.
---
Copied from [a post on the Scratch forums](https://scratch.mit.edu/discuss/topic/324629/), quoted below:
>When I try to load this project in the beta editor, the editor comes up completely blank. It's like all the scripts in the project have been completely deleted by something. Miraculously, the costumes and variables are untouched - the single costume is still there and unchanged, and none of the variables in the project have been deleted. It's only the scripts themselves that have been deleted. I've tried duplicating the project; the duplicate has the same problem. I've also tried backpacking the scripts over to another project - initially, that seems to solve the problem, but after backpacking several scripts over the new project ends up just like the original project. Strangely, I've noticed that if you interact with the project as if it were running and then click the stop button, it displays a snapshot of the project as if it were actually working in the editor.
OP [also said](https://scratch.mit.edu/discuss/post/3347932/) they don't think they used any hacked blocks.
>I'm fairly sure I didn't use any hacked blocks, and I doubt that there are too many scripts given that another project with 31 scripts works fine… (the troublesome one has 18…)
|
True
|
Blocks in "Echo of the Light - prototype" do not show up - ### Expected Behavior
The blocks show up (when you have the "Pen Text Engine" sprite selected), looking something like this:

### Actual Behavior
The blocks don't show up at all! Note that the the scripts *run*, though - if you click the green flag, the project will function correctly. So I think this is a problem in scratch-blocks, preventing the blocks from blocks from showing up, but they exist in the VM properly.
### Steps to Reproduce
Load the project in Scratch 3.0: https://llk.github.io/scratch-gui/develop/#269991846
### Operating System and Browser
I can repro this in Firefox (65.0b2). The original post pointing out this compatibility issue was from done while Chrome 70.0.3538.110.
---
Copied from [a post on the Scratch forums](https://scratch.mit.edu/discuss/topic/324629/), quoted below:
>When I try to load this project in the beta editor, the editor comes up completely blank. It's like all the scripts in the project have been completely deleted by something. Miraculously, the costumes and variables are untouched - the single costume is still there and unchanged, and none of the variables in the project have been deleted. It's only the scripts themselves that have been deleted. I've tried duplicating the project; the duplicate has the same problem. I've also tried backpacking the scripts over to another project - initially, that seems to solve the problem, but after backpacking several scripts over the new project ends up just like the original project. Strangely, I've noticed that if you interact with the project as if it were running and then click the stop button, it displays a snapshot of the project as if it were actually working in the editor.
OP [also said](https://scratch.mit.edu/discuss/post/3347932/) they don't think they used any hacked blocks.
>I'm fairly sure I didn't use any hacked blocks, and I doubt that there are too many scripts given that another project with 31 scripts works fine… (the troublesome one has 18…)
|
non_process
|
blocks in echo of the light prototype do not show up expected behavior the blocks show up when you have the pen text engine sprite selected looking something like this actual behavior the blocks don t show up at all note that the the scripts run though if you click the green flag the project will function correctly so i think this is a problem in scratch blocks preventing the blocks from blocks from showing up but they exist in the vm properly steps to reproduce load the project in scratch operating system and browser i can repro this in firefox the original post pointing out this compatibility issue was from done while chrome copied from quoted below when i try to load this project in the beta editor the editor comes up completely blank it s like all the scripts in the project have been completely deleted by something miraculously the costumes and variables are untouched the single costume is still there and unchanged and none of the variables in the project have been deleted it s only the scripts themselves that have been deleted i ve tried duplicating the project the duplicate has the same problem i ve also tried backpacking the scripts over to another project initially that seems to solve the problem but after backpacking several scripts over the new project ends up just like the original project strangely i ve noticed that if you interact with the project as if it were running and then click the stop button it displays a snapshot of the project as if it were actually working in the editor op they don t think they used any hacked blocks i m fairly sure i didn t use any hacked blocks and i doubt that there are too many scripts given that another project with scripts works fine… the troublesome one has …
| 0
|
21,959
| 30,454,236,407
|
IssuesEvent
|
2023-07-16 17:28:05
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
Typo/misquote for Invitation to Love
|
bug in process
|
**Describe the bug**
The quote in Invitation to Love is incorrect.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to the listing for Invitation to Love (as seen in Twin Peaks)
2. See the quote at the bottom listed as "Chetster, you little fruit loop. You’re done!"
**Expected behavior**
The correct quote is "Chester, you little Froot Loop, you’re done!"
**Additional context**
The full quote is "Chester, you little Froot Loop, you’re done – done!" with an extra "done" at the end, but just the first "done" is probably enough to get people jazzed about this soap.
|
1.0
|
Typo/misquote for Invitation to Love - **Describe the bug**
The quote in Invitation to Love is incorrect.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to the listing for Invitation to Love (as seen in Twin Peaks)
2. See the quote at the bottom listed as "Chetster, you little fruit loop. You’re done!"
**Expected behavior**
The correct quote is "Chester, you little Froot Loop, you’re done!"
**Additional context**
The full quote is "Chester, you little Froot Loop, you’re done – done!" with an extra "done" at the end, but just the first "done" is probably enough to get people jazzed about this soap.
|
process
|
typo misquote for invitation to love describe the bug the quote in invitation to love is incorrect to reproduce steps to reproduce the behavior go to the listing for invitation to love as seen in twin peaks see the quote at the bottom listed as chetster you little fruit loop you’re done expected behavior the correct quote is chester you little froot loop you’re done additional context the full quote is chester you little froot loop you’re done – done with an extra done at the end but just the first done is probably enough to get people jazzed about this soap
| 1
|
190,194
| 14,537,283,334
|
IssuesEvent
|
2020-12-15 08:57:04
|
microsoft/azure-pipelines-tasks
|
https://api.github.com/repos/microsoft/azure-pipelines-tasks
|
closed
|
Visual Studio Test (V 2.*) Task: How to retrieve TestRunId when Test Selection uses Test Plan.
|
Area: Test question
|
**Question, Bug, or Feature?**
*Type*: Question
**Enter Task Name**: Visual Studio Test V2
## Environment
Azure Pipelines
## Issue Description
Hi Team,
We are using Visual Studio Test (V 2.*) Task to execute Automated Tests under given Test plan. In Automated Test methods , we have custom logic to add more detail to Test Results using TestRunID. When we are executing from TestHub , we can see that $(test.RunId) is available. But we would like to get this when Test Selection uses Test Plan.
RunId Available upfront during Select tests using as Test Run

**Question:**
When Select tests using as TestPlan like below, TestRunId is getting created during this task execution internally and we don’t have any control. Actually we want to use that TestRunID in our Automated Test method implementations.
**_Is there any way to pass on this TestRunID to associated TestMethods ducting execution._**

|
1.0
|
Visual Studio Test (V 2.*) Task: How to retrieve TestRunId when Test Selection uses Test Plan. - **Question, Bug, or Feature?**
*Type*: Question
**Enter Task Name**: Visual Studio Test V2
## Environment
Azure Pipelines
## Issue Description
Hi Team,
We are using Visual Studio Test (V 2.*) Task to execute Automated Tests under given Test plan. In Automated Test methods , we have custom logic to add more detail to Test Results using TestRunID. When we are executing from TestHub , we can see that $(test.RunId) is available. But we would like to get this when Test Selection uses Test Plan.
RunId Available upfront during Select tests using as Test Run

**Question:**
When Select tests using as TestPlan like below, TestRunId is getting created during this task execution internally and we don’t have any control. Actually we want to use that TestRunID in our Automated Test method implementations.
**_Is there any way to pass on this TestRunID to associated TestMethods ducting execution._**

|
non_process
|
visual studio test v task how to retrieve testrunid when test selection uses test plan question bug or feature type question enter task name visual studio test environment azure pipelines issue description hi team we are using visual studio test v task to execute automated tests under given test plan in automated test methods we have custom logic to add more detail to test results using testrunid when we are executing from testhub we can see that test runid is available but we would like to get this when test selection uses test plan runid available upfront during select tests using as test run question when select tests using as testplan like below testrunid is getting created during this task execution internally and we don’t have any control actually we want to use that testrunid in our automated test method implementations is there any way to pass on this testrunid to associated testmethods ducting execution
| 0
|
620,950
| 19,573,887,893
|
IssuesEvent
|
2022-01-04 13:20:31
|
literakl/mezinamiridici
|
https://api.github.com/repos/literakl/mezinamiridici
|
opened
|
Service worker does not update obsolete index file
|
type: bug priority: P1
|
Index page has the correct expiration metadata though it is serves cached content even on new page view. Manual refresh was neccessary.
|
1.0
|
Service worker does not update obsolete index file - Index page has the correct expiration metadata though it is serves cached content even on new page view. Manual refresh was neccessary.
|
non_process
|
service worker does not update obsolete index file index page has the correct expiration metadata though it is serves cached content even on new page view manual refresh was neccessary
| 0
|
367,387
| 10,853,158,545
|
IssuesEvent
|
2019-11-13 14:13:39
|
WordPress/gutenberg
|
https://api.github.com/repos/WordPress/gutenberg
|
opened
|
Motion again super slow in Safari
|
Browser Issues [Priority] High
|
**Describe the bug**
Previously: #17383.
For me this issue makes the editor **very** hard to use. If I'm editing a post quick block movement is perpetually delayed. :(
**To reproduce**
* Undo or redo blocks OR
* Paste blocks OR
* Delete a block
I think it becomes more obvious as there are more blocks present.
**Expected behavior**
I expect the blocks to move nearly instantly. Currently it takes a few seconds to move.
**Screenshots**

**Desktop (please complete the following information):**
- OS: Mac
- Browser: Safari
- Version: 13.0.3
WordPress version 5.3
|
1.0
|
Motion again super slow in Safari - **Describe the bug**
Previously: #17383.
For me this issue makes the editor **very** hard to use. If I'm editing a post quick block movement is perpetually delayed. :(
**To reproduce**
* Undo or redo blocks OR
* Paste blocks OR
* Delete a block
I think it becomes more obvious as there are more blocks present.
**Expected behavior**
I expect the blocks to move nearly instantly. Currently it takes a few seconds to move.
**Screenshots**

**Desktop (please complete the following information):**
- OS: Mac
- Browser: Safari
- Version: 13.0.3
WordPress version 5.3
|
non_process
|
motion again super slow in safari describe the bug previously for me this issue makes the editor very hard to use if i m editing a post quick block movement is perpetually delayed to reproduce undo or redo blocks or paste blocks or delete a block i think it becomes more obvious as there are more blocks present expected behavior i expect the blocks to move nearly instantly currently it takes a few seconds to move screenshots desktop please complete the following information os mac browser safari version wordpress version
| 0
|
424,535
| 29,144,612,086
|
IssuesEvent
|
2023-05-18 00:57:38
|
jrsteensen/OpenHornet
|
https://api.github.com/repos/jrsteensen/OpenHornet
|
opened
|
Generate MFG Files: OH1A8A1-1 - ASSY, FWD UFC
|
Type: Documentation "Category: MCAD Priority: Normal"
|
Generate the manufacturing files for Generate MFG Files: OH1A8A1-1 - ASSY, FWD UFC.
__Check off each item in issue as you complete it.__
### File generation
- [OH Wiki HOWTO Link](https://github.com/jrsteensen/OpenHornet/wiki/HOWTO:-Generating-Fusion360-Manufacturing-Files)
- [ ] Generate SVG files (if required.)
- [ ] Generate 3MF files (if required.)
- [ ] Generate STEP files (if required.)
- [ ] Copy the relevant decal PDFs from the art folder to the relevant manufacturing folder (if required.)
### Review your files
- [ ] Verify against drawing parts list that all the relevant manufacturing files have been created.
- [ ] Open each SVG in your browser and compare against part to ensure it appears the same and its filename is correct.
- [ ] Open each 3MF in a slicer of your choice and verify geometry matches F360 model and its filename is correct.
- [ ] Open each STEP in a STEP file viewer of your choice and verify geometry matches F360 model and its filename is correct.
### Submit your files
- [ ] Create a github PR against the beta 1 branch with the manufacturing files located in correct location of the release folder.
- [ ] Request a review of the PR.
#### Why a PR?
It gives you credit when I generate the changelog in the release, and (more importantly) adds traceability to the history of the issues.
|
1.0
|
Generate MFG Files: OH1A8A1-1 - ASSY, FWD UFC - Generate the manufacturing files for Generate MFG Files: OH1A8A1-1 - ASSY, FWD UFC.
__Check off each item in issue as you complete it.__
### File generation
- [OH Wiki HOWTO Link](https://github.com/jrsteensen/OpenHornet/wiki/HOWTO:-Generating-Fusion360-Manufacturing-Files)
- [ ] Generate SVG files (if required.)
- [ ] Generate 3MF files (if required.)
- [ ] Generate STEP files (if required.)
- [ ] Copy the relevant decal PDFs from the art folder to the relevant manufacturing folder (if required.)
### Review your files
- [ ] Verify against drawing parts list that all the relevant manufacturing files have been created.
- [ ] Open each SVG in your browser and compare against part to ensure it appears the same and its filename is correct.
- [ ] Open each 3MF in a slicer of your choice and verify geometry matches F360 model and its filename is correct.
- [ ] Open each STEP in a STEP file viewer of your choice and verify geometry matches F360 model and its filename is correct.
### Submit your files
- [ ] Create a github PR against the beta 1 branch with the manufacturing files located in correct location of the release folder.
- [ ] Request a review of the PR.
#### Why a PR?
It gives you credit when I generate the changelog in the release, and (more importantly) adds traceability to the history of the issues.
|
non_process
|
generate mfg files assy fwd ufc generate the manufacturing files for generate mfg files assy fwd ufc check off each item in issue as you complete it file generation generate svg files if required generate files if required generate step files if required copy the relevant decal pdfs from the art folder to the relevant manufacturing folder if required review your files verify against drawing parts list that all the relevant manufacturing files have been created open each svg in your browser and compare against part to ensure it appears the same and its filename is correct open each in a slicer of your choice and verify geometry matches model and its filename is correct open each step in a step file viewer of your choice and verify geometry matches model and its filename is correct submit your files create a github pr against the beta branch with the manufacturing files located in correct location of the release folder request a review of the pr why a pr it gives you credit when i generate the changelog in the release and more importantly adds traceability to the history of the issues
| 0
|
95,267
| 3,941,413,440
|
IssuesEvent
|
2016-04-27 07:36:17
|
raml-org/raml-js-parser-2
|
https://api.github.com/repos/raml-org/raml-js-parser-2
|
closed
|
types are getting parsed as a whole, not individually anymore
|
bug priority:normal
|
Using the following RAML:
```yaml
#%RAML 1.0
title: API
types:
TypeA:
type: object
properties:
a: string[]
TypeB:
type: object
properties:
b:
type: string[]
TypeC:
type: object
properties:
d:
type: array
items:
type: string
TypeE:
type: string[]
TypeF:
type: array
items:
type: string
```
everything seems to be fine when you validate the complete RAML file. As soon as you look at each node and calling `node.errors()` I am getting error messages that `string[]` is not recognized. It seems that the behaviour of the parser validating the RAML as a whole interferes with `errors()` method which validates each node separately.
|
1.0
|
types are getting parsed as a whole, not individually anymore - Using the following RAML:
```yaml
#%RAML 1.0
title: API
types:
TypeA:
type: object
properties:
a: string[]
TypeB:
type: object
properties:
b:
type: string[]
TypeC:
type: object
properties:
d:
type: array
items:
type: string
TypeE:
type: string[]
TypeF:
type: array
items:
type: string
```
everything seems to be fine when you validate the complete RAML file. As soon as you look at each node and calling `node.errors()` I am getting error messages that `string[]` is not recognized. It seems that the behaviour of the parser validating the RAML as a whole interferes with `errors()` method which validates each node separately.
|
non_process
|
types are getting parsed as a whole not individually anymore using the following raml yaml raml title api types typea type object properties a string typeb type object properties b type string typec type object properties d type array items type string typee type string typef type array items type string everything seems to be fine when you validate the complete raml file as soon as you look at each node and calling node errors i am getting error messages that string is not recognized it seems that the behaviour of the parser validating the raml as a whole interferes with errors method which validates each node separately
| 0
|
7,465
| 10,563,144,519
|
IssuesEvent
|
2019-10-04 20:10:10
|
nodejs/security-wg
|
https://api.github.com/repos/nodejs/security-wg
|
closed
|
HackerOne Support team ideas
|
process
|
Some ideas on further improving the triage process with the help of the HackerOne team which has been valuable from the recent reports I've seen (thanks a lot Alek, Megan and Reed).
1. We currently have a `Node.js WG triage team` bucket for everything that has been triaged. Quite a bit of the reports are still for low downloads modules such as 5 or 30 weekly and I could get busy triaging those while there are others that are more important but I will only get to later (because I'm just doing FIFO from the bucket). My suggestion is then to have Low and High priority buckets which we'd push triaged reports into based on module popularity/criticality or some other criteria that works..
2. The HackerOne team summary is helpful. Is it possible to add to it the vulnerable versions as well?
cc @reedloden
|
1.0
|
HackerOne Support team ideas - Some ideas on further improving the triage process with the help of the HackerOne team which has been valuable from the recent reports I've seen (thanks a lot Alek, Megan and Reed).
1. We currently have a `Node.js WG triage team` bucket for everything that has been triaged. Quite a bit of the reports are still for low downloads modules such as 5 or 30 weekly and I could get busy triaging those while there are others that are more important but I will only get to later (because I'm just doing FIFO from the bucket). My suggestion is then to have Low and High priority buckets which we'd push triaged reports into based on module popularity/criticality or some other criteria that works..
2. The HackerOne team summary is helpful. Is it possible to add to it the vulnerable versions as well?
cc @reedloden
|
process
|
hackerone support team ideas some ideas on further improving the triage process with the help of the hackerone team which has been valuable from the recent reports i ve seen thanks a lot alek megan and reed we currently have a node js wg triage team bucket for everything that has been triaged quite a bit of the reports are still for low downloads modules such as or weekly and i could get busy triaging those while there are others that are more important but i will only get to later because i m just doing fifo from the bucket my suggestion is then to have low and high priority buckets which we d push triaged reports into based on module popularity criticality or some other criteria that works the hackerone team summary is helpful is it possible to add to it the vulnerable versions as well cc reedloden
| 1
|
19,783
| 26,163,381,744
|
IssuesEvent
|
2022-12-31 23:47:56
|
RobertCraigie/prisma-client-py
|
https://api.github.com/repos/RobertCraigie/prisma-client-py
|
closed
|
Required array fields cannot be used in type safe raw queries
|
bug/2-confirmed topic: types kind/bug process/candidate level/advanced priority/high topic: crash topic: raw queries
|
<!--
Thanks for helping us improve Prisma Client Python! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by enabling additional logging output.
See https://prisma-client-py.readthedocs.io/en/stable/reference/logging/ for how to enable additional logging output.
-->
## Bug description
<!-- A clear and concise description of what the bug is. -->
When using `[]` fields that are required, e.g.
```prisma
model Lists {
id String @id @default(cuid())
strings String[]
}
```
Trying to query against this model using Pydantic will cause an error, e.g.
```py
model = await client.lists.create({})
# this will cause an error because `strings` is not allowed to be `None`
found = await client.query_first('SELECT * FROM Lists WHERE id = $1', model.id, model=Lists)
```
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
An empty list should be set instead.
|
1.0
|
Required array fields cannot be used in type safe raw queries - <!--
Thanks for helping us improve Prisma Client Python! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by enabling additional logging output.
See https://prisma-client-py.readthedocs.io/en/stable/reference/logging/ for how to enable additional logging output.
-->
## Bug description
<!-- A clear and concise description of what the bug is. -->
When using `[]` fields that are required, e.g.
```prisma
model Lists {
id String @id @default(cuid())
strings String[]
}
```
Trying to query against this model using Pydantic will cause an error, e.g.
```py
model = await client.lists.create({})
# this will cause an error because `strings` is not allowed to be `None`
found = await client.query_first('SELECT * FROM Lists WHERE id = $1', model.id, model=Lists)
```
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
An empty list should be set instead.
|
process
|
required array fields cannot be used in type safe raw queries thanks for helping us improve prisma client python 🙏 please follow the sections in the template and provide as much information as possible about your problem e g by enabling additional logging output see for how to enable additional logging output bug description when using fields that are required e g prisma model lists id string id default cuid strings string trying to query against this model using pydantic will cause an error e g py model await client lists create this will cause an error because strings is not allowed to be none found await client query first select from lists where id model id model lists expected behavior an empty list should be set instead
| 1
|
146,812
| 11,757,819,698
|
IssuesEvent
|
2020-03-13 14:21:50
|
CIMDBORG/CIMMigrationProject
|
https://api.github.com/repos/CIMDBORG/CIMMigrationProject
|
opened
|
Code Clean-Up
|
Testing enhancement
|
[codechanges.txt](https://github.com/CIMDBORG/CIMMigrationProject/files/4330006/codechanges.txt)
The attached document is a very informal list of what needs to be looked at in the database. These small clean-ups will make the application look and run smoother.
|
1.0
|
Code Clean-Up - [codechanges.txt](https://github.com/CIMDBORG/CIMMigrationProject/files/4330006/codechanges.txt)
The attached document is a very informal list of what needs to be looked at in the database. These small clean-ups will make the application look and run smoother.
|
non_process
|
code clean up the attached document is a very informal list of what needs to be looked at in the database these small clean ups will make the application look and run smoother
| 0
|
12,036
| 14,738,645,232
|
IssuesEvent
|
2021-01-07 05:20:58
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Invoice summary email
|
anc-process anp-2.5 ant-child/secondary ant-enhancement grt-billing
|
In GitLab by @kdjstudios on Jul 5, 2018, 09:44
Hello team,
While working on #937; it became apparent that not always is the user who is running the billing cycle have access to the email address used at the site level. Which is where the email/fax summaries will be sent. how difficult would it be to have this email be sent to the site email address and the user's email address when they choose either of those fax/email buttons on the billing cycle page?
|
1.0
|
Invoice summary email - In GitLab by @kdjstudios on Jul 5, 2018, 09:44
Hello team,
While working on #937; it became apparent that not always is the user who is running the billing cycle have access to the email address used at the site level. Which is where the email/fax summaries will be sent. how difficult would it be to have this email be sent to the site email address and the user's email address when they choose either of those fax/email buttons on the billing cycle page?
|
process
|
invoice summary email in gitlab by kdjstudios on jul hello team while working on it became apparent that not always is the user who is running the billing cycle have access to the email address used at the site level which is where the email fax summaries will be sent how difficult would it be to have this email be sent to the site email address and the user s email address when they choose either of those fax email buttons on the billing cycle page
| 1
|
13,255
| 15,725,719,140
|
IssuesEvent
|
2021-03-29 10:20:12
|
hashicorp/packer-plugin-amazon
|
https://api.github.com/repos/hashicorp/packer-plugin-amazon
|
opened
|
amazon-import - improvement
|
enhancement post-processor/amazon-import
|
_This issue was originally opened by @Roxyrob as hashicorp/packer#9582. It was migrated here as a result of the [Packer plugin split](https://github.com/hashicorp/packer/issues/8610#issuecomment-770034737). The original body of the issue is below._
<hr>
#### Feature Description
amazon-import supporting "import-snapshot" (ami import passing through snapshot.).
#### Use Case(s)
I know support from vendor is important, this case can be an exception. AWS is not declaring support for kernel version present in many new OS (CentOS 8, Fedora 31, ...) and stop vmimport task on ami import stage with error: " Unable to determine kernel version" (see hashicorp/packer#8302 from where we start to search for a solution/workaround).
AWS does not stop however import task as snapshot (instead ami directly). We succesfully reach a working CentOS8 custom AMI manually starting from an virtualbox ova e.g created with packer virtualbox-iso
* extracting VMDK inside ova (tar -xvf packer-centos8min-x86_64.ova)
* uploading vmdk (aws s3 cp ...)
* importing VMDK to aws as "snapshot" (aws ec2 import-snapshot...)
* Creating AMI from snapshot (aws ec2 register-image...)
ova was made using packer provisioned to set right iniramfs with needed modules to allow root device discovery on boot stages for xen-base es2 instances (xen-blkfront), nitro-based ec2 instances (nvme, ena).
It would be nice if amazon-import can support import-snapshot (using VMDK in ova) automating all image creation steps useful to make custom image of new OS release (when kernel version is not yet supported by AWS as often it is a long time lack).
|
1.0
|
amazon-import - improvement - _This issue was originally opened by @Roxyrob as hashicorp/packer#9582. It was migrated here as a result of the [Packer plugin split](https://github.com/hashicorp/packer/issues/8610#issuecomment-770034737). The original body of the issue is below._
<hr>
#### Feature Description
amazon-import supporting "import-snapshot" (ami import passing through snapshot.).
#### Use Case(s)
I know support from vendor is important, this case can be an exception. AWS is not declaring support for kernel version present in many new OS (CentOS 8, Fedora 31, ...) and stop vmimport task on ami import stage with error: " Unable to determine kernel version" (see hashicorp/packer#8302 from where we start to search for a solution/workaround).
AWS does not stop however import task as snapshot (instead ami directly). We succesfully reach a working CentOS8 custom AMI manually starting from an virtualbox ova e.g created with packer virtualbox-iso
* extracting VMDK inside ova (tar -xvf packer-centos8min-x86_64.ova)
* uploading vmdk (aws s3 cp ...)
* importing VMDK to aws as "snapshot" (aws ec2 import-snapshot...)
* Creating AMI from snapshot (aws ec2 register-image...)
ova was made using packer provisioned to set right iniramfs with needed modules to allow root device discovery on boot stages for xen-base es2 instances (xen-blkfront), nitro-based ec2 instances (nvme, ena).
It would be nice if amazon-import can support import-snapshot (using VMDK in ova) automating all image creation steps useful to make custom image of new OS release (when kernel version is not yet supported by AWS as often it is a long time lack).
|
process
|
amazon import improvement this issue was originally opened by roxyrob as hashicorp packer it was migrated here as a result of the the original body of the issue is below feature description amazon import supporting import snapshot ami import passing through snapshot use case s i know support from vendor is important this case can be an exception aws is not declaring support for kernel version present in many new os centos fedora and stop vmimport task on ami import stage with error unable to determine kernel version see hashicorp packer from where we start to search for a solution workaround aws does not stop however import task as snapshot instead ami directly we succesfully reach a working custom ami manually starting from an virtualbox ova e g created with packer virtualbox iso extracting vmdk inside ova tar xvf packer ova uploading vmdk aws cp importing vmdk to aws as snapshot aws import snapshot creating ami from snapshot aws register image ova was made using packer provisioned to set right iniramfs with needed modules to allow root device discovery on boot stages for xen base instances xen blkfront nitro based instances nvme ena it would be nice if amazon import can support import snapshot using vmdk in ova automating all image creation steps useful to make custom image of new os release when kernel version is not yet supported by aws as often it is a long time lack
| 1
|
22,167
| 3,940,523,602
|
IssuesEvent
|
2016-04-27 01:26:42
|
extnet/Ext.NET
|
https://api.github.com/repos/extnet/Ext.NET
|
closed
|
Support multiple axes per position in a chart
|
2.x 3.x 4.x feature fixed-in-latest-extjs sencha
|
http://forums.ext.net/showthread.php?26496
http://www.sencha.com/forum/showthread.php?275032
**Update:** Allegedly fixed in a _pending_ release at the time of 6.0.1 release.
|
1.0
|
Support multiple axes per position in a chart - http://forums.ext.net/showthread.php?26496
http://www.sencha.com/forum/showthread.php?275032
**Update:** Allegedly fixed in a _pending_ release at the time of 6.0.1 release.
|
non_process
|
support multiple axes per position in a chart update allegedly fixed in a pending release at the time of release
| 0
|
6,380
| 9,429,859,405
|
IssuesEvent
|
2019-04-12 07:30:11
|
googleapis/google-cloud-node
|
https://api.github.com/repos/googleapis/google-cloud-node
|
closed
|
use `process.execSync` in favor of `execa` in tests
|
type: process
|
`process.execSync` is great for writing unit tests, it's what we use in Node.js itself.
I think that `execa`, `cross-spawn`, etc., address some historic inconsistencies in the Node.js API that we shouldn't be bumping into in simple assertions (and I'm not sure what, if any of the issues, continue to exist on Node.js > 6).
|
1.0
|
use `process.execSync` in favor of `execa` in tests - `process.execSync` is great for writing unit tests, it's what we use in Node.js itself.
I think that `execa`, `cross-spawn`, etc., address some historic inconsistencies in the Node.js API that we shouldn't be bumping into in simple assertions (and I'm not sure what, if any of the issues, continue to exist on Node.js > 6).
|
process
|
use process execsync in favor of execa in tests process execsync is great for writing unit tests it s what we use in node js itself i think that execa cross spawn etc address some historic inconsistencies in the node js api that we shouldn t be bumping into in simple assertions and i m not sure what if any of the issues continue to exist on node js
| 1
|
175,155
| 13,536,866,458
|
IssuesEvent
|
2020-09-16 09:39:06
|
Muhammad-Mamduh/Muhammad-Mamduh
|
https://api.github.com/repos/Muhammad-Mamduh/Muhammad-Mamduh
|
opened
|
ali
|
Taggg (a) Taggggggg (b) Taggggggg (c) test tag0 test tag1 test tag2 test tag3 test tag4 test tag5 test tag6 test tag7 test tag8 test tag9
|
# :clipboard: Bug Details
>ali
key | value
--|--
Reported At | 2020-09-16 09:38:45 UTC
Email | test@example.com
Categories | Report a bug
Tags | test tag0, test tag1, test tag2, test tag3, test tag4, test tag5, test tag6, test tag7, test tag8, test tag9, Taggg (a), Taggggggg (b), Taggggggg (c)
App Version | 1.0-kotlin (1)
Session Duration | 27
Device | Google Android SDK built for x86, OS Level 29
Display | 1080x1920 (xhdpi)
Location | Tanta, Egypt (en)
## :point_right: [View Full Bug Report on Instabug](https://dashboard.instabug.com/applications/test/beta/bugs/20?utm_source=github&utm_medium=integrations) :point_left:
___
# :iphone: View Hierarchy
This bug was reported from **com.example.instabug.developerOption.DeveloperOptionFragment**
Find its interactive view hierarchy with all its subviews here: :point_right: **[Check View Hierarchy](https://dashboard.instabug.com/applications/test/beta/bugs/20?show-hierarchy-view=true&utm_source=github&utm_medium=integrations)** :point_left:
___
# :chart_with_downwards_trend: Session Profiler
Here is what the app was doing right before the bug was reported:
Key | Value
--|--
Used Memory | 48.9% - 0.95/1.95 GB
Used Storage | 24.9% - 2.89/11.58 GB
Connectivity | WiFi
Battery | 100% - unplugged
Orientation | portrait
Find all the changes that happened in the parameters mentioned above during the last 60 seconds before the bug was reported here: :point_right: **[View Full Session Profiler](https://dashboard.instabug.com/applications/test/beta/bugs/20?show-session-profiler=true&utm_source=github&utm_medium=integrations)** :point_left:
___
# :bust_in_silhouette: User Info
### User Attributes
```
user1: user1:value
user2: user2:value
user0: user0:value
carteron: meshi
zamalke: second
user9: user9:value
user7: user7:value
user8: user8:value
LoggedIn: True
user5: user5:value
user6: user6:value
user3: user3:value
Ahlu: fu2 l gami3
Name: True
user4: user4:value
ismaili: darawish
School: rowad
Length: 16
Club: ahil
Hobby: foot
Width: 10
Age: 18
```
___
# :mag_right: Logs
### Network Log
This bug report has **1 failed** (4xx, 5xx or no response) requests. Here are some of them:
```
09:35:09 0 GET http://10.0.2.2:8080/resp
```
Find all the network requests with more details about each one, its duration, method, URL, headers, and response here: :point_right: **[View Detailed Network Logs](https://dashboard.instabug.com/applications/test/beta/bugs/20?show-logs=network_log&utm_source=github&utm_medium=integrations)** :point_left:
### User Steps
Here are the last 10 steps done by the user right before the bug was reported:
```
09:38:38 Long press in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.instabug.main.MainActivity"
09:38:38 Long press in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.instabug.main.MainActivity"
09:38:39 Long press in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.instabug.main.MainActivity"
09:38:40 Long press in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.instabug.main.MainActivity"
09:38:40 Long press in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.instabug.main.MainActivity"
09:38:41 Long press in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.instabug.main.MainActivity"
09:38:42 com.example.instabug.main.MainActivity was paused.
09:38:42 In activity com.example.instabug.main.MainActivity: fragment com.example.instabug.developerOption.DeveloperOptionFragment was paused.
09:38:43 Long press in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.instabug.main.MainActivity"
09:38:45 Long press in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.instabug.main.MainActivity"
```
Find all the user steps done by the user throughout the session here: :point_right: **[View All User Steps](https://dashboard.instabug.com/applications/test/beta/bugs/20?show-logs=user_steps&utm_source=github&utm_medium=integrations)** :point_left:
### User Events
Here are the last 10 user events logged right before the bug was reported:
```
09:38:41 Logged proud
09:38:41 Logged loud
09:38:41 Logged profound
09:38:41 Logged less
09:38:41 Logged mess
09:38:41 Logged dress
09:38:41 Logged weak
09:38:41 Logged freak
09:38:41 Logged car
09:38:41 Logged tar
```
Find all the logged user events throughout the session here: :point_right: **[View All User Events](https://dashboard.instabug.com/applications/test/beta/bugs/20?show-logs=user_events&utm_source=github&utm_medium=integrations)** :point_left:
### Instabug Log
Here are the last 10 instabug logs logged right before the bug was reported:
```
09:10:47 log #83
09:10:47 log #84
09:10:47 log #85
09:10:47 log #80
09:10:47 log #81
09:10:47 log #82
09:10:47 log #76
09:10:47 log #77
09:10:47 log #78
09:10:47 log #79
```
Find all the logged Instabug logs throughout the session here: :point_right: **[View All Instabug Log](https://dashboard.instabug.com/applications/test/beta/bugs/20?show-logs=instabug_log&utm_source=github&utm_medium=integrations)** :point_left:
### Console Log
Here are the last 10 console logs logged right before the bug was reported:
```
09:38:46 D/EGL_emulation(31754): eglMakeCurrent: 0xdc01a960: ver 3 0 (tinfo 0xdc00fa60)
09:38:47 D/IB-BaseReportingPresenter(31754): checkUserEmailValid :test@example.com
09:38:47 V/IB-PoolProvider(31754): postTask
09:38:47 D/IB-ActionsOrchestrator(31754): runAction
09:38:47 D/IB-AttachmentsUtility(31754): encryptAttachments
09:38:47 D/EGL_emulation(31754): eglMakeCurrent: 0xdc01a960: ver 3 0 (tinfo 0xdc00fa60)
09:38:47 D/EGL_emulation(31754): eglMakeCurrent: 0xdc01a960: ver 3 0 (tinfo 0xdc00fa60)
09:38:47 I/chatty (31754): uid=10143(com.example.instabug.kotlin) RenderThread identical 1 line
09:38:47 D/EGL_emulation(31754): eglMakeCurrent: 0xdc01a960: ver 3 0 (tinfo 0xdc00fa60)
09:38:47 W/System (31754): A resource failed to call close.
```
Find all the logged console logs throughout the session here: :point_right: **[View All Console Log](https://dashboard.instabug.com/applications/test/beta/bugs/20?show-logs=console_log&utm_source=github&utm_medium=integrations)** :point_left:
___
# :camera: Images
[](https://s3.us-east-1.amazonaws.com/instabug-backend-4540/attachments/bugs/8631640/4f8b7d25c72fc5ead65fa9ad0a861d99_original/7547411/bug_1600249122261_.jpg?1600249137)
|
10.0
|
ali - # :clipboard: Bug Details
>ali
key | value
--|--
Reported At | 2020-09-16 09:38:45 UTC
Email | test@example.com
Categories | Report a bug
Tags | test tag0, test tag1, test tag2, test tag3, test tag4, test tag5, test tag6, test tag7, test tag8, test tag9, Taggg (a), Taggggggg (b), Taggggggg (c)
App Version | 1.0-kotlin (1)
Session Duration | 27
Device | Google Android SDK built for x86, OS Level 29
Display | 1080x1920 (xhdpi)
Location | Tanta, Egypt (en)
## :point_right: [View Full Bug Report on Instabug](https://dashboard.instabug.com/applications/test/beta/bugs/20?utm_source=github&utm_medium=integrations) :point_left:
___
# :iphone: View Hierarchy
This bug was reported from **com.example.instabug.developerOption.DeveloperOptionFragment**
Find its interactive view hierarchy with all its subviews here: :point_right: **[Check View Hierarchy](https://dashboard.instabug.com/applications/test/beta/bugs/20?show-hierarchy-view=true&utm_source=github&utm_medium=integrations)** :point_left:
___
# :chart_with_downwards_trend: Session Profiler
Here is what the app was doing right before the bug was reported:
Key | Value
--|--
Used Memory | 48.9% - 0.95/1.95 GB
Used Storage | 24.9% - 2.89/11.58 GB
Connectivity | WiFi
Battery | 100% - unplugged
Orientation | portrait
Find all the changes that happened in the parameters mentioned above during the last 60 seconds before the bug was reported here: :point_right: **[View Full Session Profiler](https://dashboard.instabug.com/applications/test/beta/bugs/20?show-session-profiler=true&utm_source=github&utm_medium=integrations)** :point_left:
___
# :bust_in_silhouette: User Info
### User Attributes
```
user1: user1:value
user2: user2:value
user0: user0:value
carteron: meshi
zamalke: second
user9: user9:value
user7: user7:value
user8: user8:value
LoggedIn: True
user5: user5:value
user6: user6:value
user3: user3:value
Ahlu: fu2 l gami3
Name: True
user4: user4:value
ismaili: darawish
School: rowad
Length: 16
Club: ahil
Hobby: foot
Width: 10
Age: 18
```
___
# :mag_right: Logs
### Network Log
This bug report has **1 failed** (4xx, 5xx or no response) requests. Here are some of them:
```
09:35:09 0 GET http://10.0.2.2:8080/resp
```
Find all the network requests with more details about each one, its duration, method, URL, headers, and response here: :point_right: **[View Detailed Network Logs](https://dashboard.instabug.com/applications/test/beta/bugs/20?show-logs=network_log&utm_source=github&utm_medium=integrations)** :point_left:
### User Steps
Here are the last 10 steps done by the user right before the bug was reported:
```
09:38:38 Long press in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.instabug.main.MainActivity"
09:38:38 Long press in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.instabug.main.MainActivity"
09:38:39 Long press in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.instabug.main.MainActivity"
09:38:40 Long press in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.instabug.main.MainActivity"
09:38:40 Long press in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.instabug.main.MainActivity"
09:38:41 Long press in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.instabug.main.MainActivity"
09:38:42 com.example.instabug.main.MainActivity was paused.
09:38:42 In activity com.example.instabug.main.MainActivity: fragment com.example.instabug.developerOption.DeveloperOptionFragment was paused.
09:38:43 Long press in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.instabug.main.MainActivity"
09:38:45 Long press in "androidx.constraintlayout.widget.ConstraintLayout" in "com.example.instabug.main.MainActivity"
```
Find all the user steps done by the user throughout the session here: :point_right: **[View All User Steps](https://dashboard.instabug.com/applications/test/beta/bugs/20?show-logs=user_steps&utm_source=github&utm_medium=integrations)** :point_left:
### User Events
Here are the last 10 user events logged right before the bug was reported:
```
09:38:41 Logged proud
09:38:41 Logged loud
09:38:41 Logged profound
09:38:41 Logged less
09:38:41 Logged mess
09:38:41 Logged dress
09:38:41 Logged weak
09:38:41 Logged freak
09:38:41 Logged car
09:38:41 Logged tar
```
Find all the logged user events throughout the session here: :point_right: **[View All User Events](https://dashboard.instabug.com/applications/test/beta/bugs/20?show-logs=user_events&utm_source=github&utm_medium=integrations)** :point_left:
### Instabug Log
Here are the last 10 instabug logs logged right before the bug was reported:
```
09:10:47 log #83
09:10:47 log #84
09:10:47 log #85
09:10:47 log #80
09:10:47 log #81
09:10:47 log #82
09:10:47 log #76
09:10:47 log #77
09:10:47 log #78
09:10:47 log #79
```
Find all the logged Instabug logs throughout the session here: :point_right: **[View All Instabug Log](https://dashboard.instabug.com/applications/test/beta/bugs/20?show-logs=instabug_log&utm_source=github&utm_medium=integrations)** :point_left:
### Console Log
Here are the last 10 console logs logged right before the bug was reported:
```
09:38:46 D/EGL_emulation(31754): eglMakeCurrent: 0xdc01a960: ver 3 0 (tinfo 0xdc00fa60)
09:38:47 D/IB-BaseReportingPresenter(31754): checkUserEmailValid :test@example.com
09:38:47 V/IB-PoolProvider(31754): postTask
09:38:47 D/IB-ActionsOrchestrator(31754): runAction
09:38:47 D/IB-AttachmentsUtility(31754): encryptAttachments
09:38:47 D/EGL_emulation(31754): eglMakeCurrent: 0xdc01a960: ver 3 0 (tinfo 0xdc00fa60)
09:38:47 D/EGL_emulation(31754): eglMakeCurrent: 0xdc01a960: ver 3 0 (tinfo 0xdc00fa60)
09:38:47 I/chatty (31754): uid=10143(com.example.instabug.kotlin) RenderThread identical 1 line
09:38:47 D/EGL_emulation(31754): eglMakeCurrent: 0xdc01a960: ver 3 0 (tinfo 0xdc00fa60)
09:38:47 W/System (31754): A resource failed to call close.
```
Find all the logged console logs throughout the session here: :point_right: **[View All Console Log](https://dashboard.instabug.com/applications/test/beta/bugs/20?show-logs=console_log&utm_source=github&utm_medium=integrations)** :point_left:
___
# :camera: Images
[](https://s3.us-east-1.amazonaws.com/instabug-backend-4540/attachments/bugs/8631640/4f8b7d25c72fc5ead65fa9ad0a861d99_original/7547411/bug_1600249122261_.jpg?1600249137)
|
non_process
|
ali clipboard bug details ali key value reported at utc email test example com categories report a bug tags test test test test test test test test test test taggg a taggggggg b taggggggg c app version kotlin session duration device google android sdk built for os level display xhdpi location tanta egypt en point right point left iphone view hierarchy this bug was reported from com example instabug developeroption developeroptionfragment find its interactive view hierarchy with all its subviews here point right point left chart with downwards trend session profiler here is what the app was doing right before the bug was reported key value used memory gb used storage gb connectivity wifi battery unplugged orientation portrait find all the changes that happened in the parameters mentioned above during the last seconds before the bug was reported here point right point left bust in silhouette user info user attributes value value value carteron meshi zamalke second value value value loggedin true value value value ahlu l name true value ismaili darawish school rowad length club ahil hobby foot width age mag right logs network log this bug report has failed or no response requests here are some of them get find all the network requests with more details about each one its duration method url headers and response here point right point left user steps here are the last steps done by the user right before the bug was reported long press in androidx constraintlayout widget constraintlayout in com example instabug main mainactivity long press in androidx constraintlayout widget constraintlayout in com example instabug main mainactivity long press in androidx constraintlayout widget constraintlayout in com example instabug main mainactivity long press in androidx constraintlayout widget constraintlayout in com example instabug main mainactivity long press in androidx constraintlayout widget constraintlayout in com example instabug main mainactivity long press in androidx constraintlayout widget constraintlayout in com example instabug main mainactivity com example instabug main mainactivity was paused in activity com example instabug main mainactivity fragment com example instabug developeroption developeroptionfragment was paused long press in androidx constraintlayout widget constraintlayout in com example instabug main mainactivity long press in androidx constraintlayout widget constraintlayout in com example instabug main mainactivity find all the user steps done by the user throughout the session here point right point left user events here are the last user events logged right before the bug was reported logged proud logged loud logged profound logged less logged mess logged dress logged weak logged freak logged car logged tar find all the logged user events throughout the session here point right point left instabug log here are the last instabug logs logged right before the bug was reported log log log log log log log log log log find all the logged instabug logs throughout the session here point right point left console log here are the last console logs logged right before the bug was reported d egl emulation eglmakecurrent ver tinfo d ib basereportingpresenter checkuseremailvalid test example com v ib poolprovider posttask d ib actionsorchestrator runaction d ib attachmentsutility encryptattachments d egl emulation eglmakecurrent ver tinfo d egl emulation eglmakecurrent ver tinfo i chatty uid com example instabug kotlin renderthread identical line d egl emulation eglmakecurrent ver tinfo w system a resource failed to call close find all the logged console logs throughout the session here point right point left camera images
| 0
|
60,764
| 8,461,218,227
|
IssuesEvent
|
2018-10-22 21:07:31
|
droidkfx/Yet-Another-Productivity-App
|
https://api.github.com/repos/droidkfx/Yet-Another-Productivity-App
|
closed
|
document new features
|
documentation
|
- [ ] Logout needs to be documented
- [ ] Document update to yapa-api - added authentication, bugfix multithread access to task
|
1.0
|
document new features - - [ ] Logout needs to be documented
- [ ] Document update to yapa-api - added authentication, bugfix multithread access to task
|
non_process
|
document new features logout needs to be documented document update to yapa api added authentication bugfix multithread access to task
| 0
|
16,545
| 21,568,598,940
|
IssuesEvent
|
2022-05-02 04:17:56
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
Add The Cooper Clan
|
suggested title in process
|
Title: The Cooper Clan
Type (film/tv show): TV show
Show in which it appears: Monk
Is the parent show streaming anywhere? Amazon Prime and Peacock
About when in the parent show does it appear? Season 8, Episode 1, "Mr. Monk's Favorite Show"
Actual footage of the show can be seen (yes/no)? Yes
Cast:
Christine Rapp as Kathy Cooper
Brendan Collins as Mr. Cooper
Gail Rodono as Mrs. Cooper
Lauren Perkins as Janey Cooper
Steven Dorn as William "Billy" Cooper
Show most likely ended in 1974 (the show is mentioned to have ended "35 years ago," and if the Monk episode occurred the same year it aired, 2009, that would mean the show was cancelled in 1974. That happens to be the same year The Brady Bunch was cancelled.)
Link to Monk wiki: https://monk.fandom.com/wiki/Mr._Monk%27s_Favorite_Show
|
1.0
|
Add The Cooper Clan - Title: The Cooper Clan
Type (film/tv show): TV show
Show in which it appears: Monk
Is the parent show streaming anywhere? Amazon Prime and Peacock
About when in the parent show does it appear? Season 8, Episode 1, "Mr. Monk's Favorite Show"
Actual footage of the show can be seen (yes/no)? Yes
Cast:
Christine Rapp as Kathy Cooper
Brendan Collins as Mr. Cooper
Gail Rodono as Mrs. Cooper
Lauren Perkins as Janey Cooper
Steven Dorn as William "Billy" Cooper
Show most likely ended in 1974 (the show is mentioned to have ended "35 years ago," and if the Monk episode occurred the same year it aired, 2009, that would mean the show was cancelled in 1974. That happens to be the same year The Brady Bunch was cancelled.)
Link to Monk wiki: https://monk.fandom.com/wiki/Mr._Monk%27s_Favorite_Show
|
process
|
add the cooper clan title the cooper clan type film tv show tv show show in which it appears monk is the parent show streaming anywhere amazon prime and peacock about when in the parent show does it appear season episode mr monk s favorite show actual footage of the show can be seen yes no yes cast christine rapp as kathy cooper brendan collins as mr cooper gail rodono as mrs cooper lauren perkins as janey cooper steven dorn as william billy cooper show most likely ended in the show is mentioned to have ended years ago and if the monk episode occurred the same year it aired that would mean the show was cancelled in that happens to be the same year the brady bunch was cancelled link to monk wiki
| 1
|
184,476
| 14,981,394,459
|
IssuesEvent
|
2021-01-28 14:47:44
|
google/jsonnet
|
https://api.github.com/repos/google/jsonnet
|
reopened
|
Document -S (--string) in language reference (and maybe tutorial, too).
|
documentation
|
Not much else to say.
```sh
$ jsonnet --version && cat test.jsonnet && echo && jsonnet test.jsonnet
Jsonnet commandline interpreter v0.16.0
local wrapper = [
{
some_prop: true
}
];
std.manifestYamlDoc(wrapper)
"- \"some_prop\": true"
```
|
1.0
|
Document -S (--string) in language reference (and maybe tutorial, too). - Not much else to say.
```sh
$ jsonnet --version && cat test.jsonnet && echo && jsonnet test.jsonnet
Jsonnet commandline interpreter v0.16.0
local wrapper = [
{
some_prop: true
}
];
std.manifestYamlDoc(wrapper)
"- \"some_prop\": true"
```
|
non_process
|
document s string in language reference and maybe tutorial too not much else to say sh jsonnet version cat test jsonnet echo jsonnet test jsonnet jsonnet commandline interpreter local wrapper some prop true std manifestyamldoc wrapper some prop true
| 0
|
20,367
| 27,024,939,008
|
IssuesEvent
|
2023-02-11 13:28:47
|
firebase/firebase-cpp-sdk
|
https://api.github.com/repos/firebase/firebase-cpp-sdk
|
closed
|
[C++] Nightly Integration Testing Report for Firestore
|
type: process nightly-testing
|
<hidden value="integration-test-status-comment"></hidden>
### ✅ [build against repo] Integration test succeeded!
Requested by @sunmou99 on commit b1a5444b10c751c4c039a29bb9fe531b39d1b9ad
Last updated: Sat Feb 11 04:02 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4150834848)**
<hidden value="integration-test-status-comment"></hidden>
***
### ❌ [build against SDK] Integration test FAILED
Requested by @firebase-workflow-trigger[bot] on commit b1a5444b10c751c4c039a29bb9fe531b39d1b9ad
Last updated: Fri Feb 10 06:02 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4143880709)**
| Failures | Configs |
|----------|---------|
| missing_log | [BUILD] [ERROR] [Android] [1/3 os: ubuntu]<br/> |
Add flaky tests to **[go/fpl-cpp-flake-tracker](http://go/fpl-cpp-flake-tracker)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against tip] Integration test succeeded!
Requested by @sunmou99 on commit b1a5444b10c751c4c039a29bb9fe531b39d1b9ad
Last updated: Sat Feb 11 03:41 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4151028457)**
|
1.0
|
[C++] Nightly Integration Testing Report for Firestore -
<hidden value="integration-test-status-comment"></hidden>
### ✅ [build against repo] Integration test succeeded!
Requested by @sunmou99 on commit b1a5444b10c751c4c039a29bb9fe531b39d1b9ad
Last updated: Sat Feb 11 04:02 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4150834848)**
<hidden value="integration-test-status-comment"></hidden>
***
### ❌ [build against SDK] Integration test FAILED
Requested by @firebase-workflow-trigger[bot] on commit b1a5444b10c751c4c039a29bb9fe531b39d1b9ad
Last updated: Fri Feb 10 06:02 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4143880709)**
| Failures | Configs |
|----------|---------|
| missing_log | [BUILD] [ERROR] [Android] [1/3 os: ubuntu]<br/> |
Add flaky tests to **[go/fpl-cpp-flake-tracker](http://go/fpl-cpp-flake-tracker)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against tip] Integration test succeeded!
Requested by @sunmou99 on commit b1a5444b10c751c4c039a29bb9fe531b39d1b9ad
Last updated: Sat Feb 11 03:41 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4151028457)**
|
process
|
nightly integration testing report for firestore ✅ nbsp integration test succeeded requested by on commit last updated sat feb pst ❌ nbsp integration test failed requested by firebase workflow trigger on commit last updated fri feb pst failures configs missing log add flaky tests to ✅ nbsp integration test succeeded requested by on commit last updated sat feb pst
| 1
|
54,299
| 6,378,250,202
|
IssuesEvent
|
2017-08-02 12:15:57
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
TestClientGoCustomResourceExample flakes
|
kind/flake priority/failing-test sig/api-machinery
|
flaking on HEAD:
https://storage.googleapis.com/k8s-gubernator/triage/index.html?pr=1&job=ci-kubernetes-test-go&test=TestClientGoCustomResourceExample
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-test-go/8135#k8siokubernetesvendork8sioapiextensions-apiservertestintegration-testclientgocustomresourceexample
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-test-go/8118#k8siokubernetesvendork8sioapiextensions-apiservertestintegration-testclientgocustomresourceexample
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-test-go/8105#k8siokubernetesvendork8sioapiextensions-apiservertestintegration-testclientgocustomresourceexample
@kubernetes/sig-api-machinery-test-failures
|
1.0
|
TestClientGoCustomResourceExample flakes - flaking on HEAD:
https://storage.googleapis.com/k8s-gubernator/triage/index.html?pr=1&job=ci-kubernetes-test-go&test=TestClientGoCustomResourceExample
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-test-go/8135#k8siokubernetesvendork8sioapiextensions-apiservertestintegration-testclientgocustomresourceexample
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-test-go/8118#k8siokubernetesvendork8sioapiextensions-apiservertestintegration-testclientgocustomresourceexample
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-test-go/8105#k8siokubernetesvendork8sioapiextensions-apiservertestintegration-testclientgocustomresourceexample
@kubernetes/sig-api-machinery-test-failures
|
non_process
|
testclientgocustomresourceexample flakes flaking on head kubernetes sig api machinery test failures
| 0
|
22,745
| 32,062,153,572
|
IssuesEvent
|
2023-09-24 19:33:11
|
bitfocus/companion-module-requests
|
https://api.github.com/repos/bitfocus/companion-module-requests
|
opened
|
ENTTEC S-Play
|
NOT YET PROCESSED
|
- [x] **I have researched the list of existing Companion modules and requests and have determined this has not yet been requested**
The name of the device, hardware, or software you would like to control:
ENTTEC S-PLAY via the http api
https://www.enttec.com/product/lighting-show-recorder/smart-light-show-controller/
S-Play has OSC support but it's very limited.
What you would like to be able to make it do from Companion:
- Record and playback of dmx.
- Set dmx cues
- Trigger events
- Return feedback varibles on states and status
- Hopefully in the future all supported features in the api
Direct links or attachments to the ethernet control protocol or API:
Official http api specification: https://github.com/ENTTEC/SplayApi
|
1.0
|
ENTTEC S-Play - - [x] **I have researched the list of existing Companion modules and requests and have determined this has not yet been requested**
The name of the device, hardware, or software you would like to control:
ENTTEC S-PLAY via the http api
https://www.enttec.com/product/lighting-show-recorder/smart-light-show-controller/
S-Play has OSC support but it's very limited.
What you would like to be able to make it do from Companion:
- Record and playback of dmx.
- Set dmx cues
- Trigger events
- Return feedback varibles on states and status
- Hopefully in the future all supported features in the api
Direct links or attachments to the ethernet control protocol or API:
Official http api specification: https://github.com/ENTTEC/SplayApi
|
process
|
enttec s play i have researched the list of existing companion modules and requests and have determined this has not yet been requested the name of the device hardware or software you would like to control enttec s play via the http api s play has osc support but it s very limited what you would like to be able to make it do from companion record and playback of dmx set dmx cues trigger events return feedback varibles on states and status hopefully in the future all supported features in the api direct links or attachments to the ethernet control protocol or api official http api specification
| 1
|
120,156
| 12,060,360,654
|
IssuesEvent
|
2020-04-15 21:03:12
|
SAP/openui5
|
https://api.github.com/repos/SAP/openui5
|
closed
|
Tooltips still not satisfactory
|
documentation
|
OpenUI5 version: 1.68.1
I have found this issue https://github.com/SAP/openui5/issues/1612 and I have to say I agree and conclude it is not fixed still.
1. All Controls show a tooltip aggregation and the documentation keeps sending you in circles. For example, the https://openui5.hana.ondemand.com/#/api/sap.f.Card has a tooltip aggregation, it tells you the option to use the two TooltipBase objects but they are deprecated. Shouldn't you simply change that to a tooltip=string property in the docs then? As this is the only supported option at the moment.
1. For most controls the tooltip=string simply does not work. Buttons are fine, but above Card for example does not show anything.
1. Tooltips are so common, they should be supported out of the box. Not with mouseover, delegate and popover.
Agree?
|
1.0
|
Tooltips still not satisfactory - OpenUI5 version: 1.68.1
I have found this issue https://github.com/SAP/openui5/issues/1612 and I have to say I agree and conclude it is not fixed still.
1. All Controls show a tooltip aggregation and the documentation keeps sending you in circles. For example, the https://openui5.hana.ondemand.com/#/api/sap.f.Card has a tooltip aggregation, it tells you the option to use the two TooltipBase objects but they are deprecated. Shouldn't you simply change that to a tooltip=string property in the docs then? As this is the only supported option at the moment.
1. For most controls the tooltip=string simply does not work. Buttons are fine, but above Card for example does not show anything.
1. Tooltips are so common, they should be supported out of the box. Not with mouseover, delegate and popover.
Agree?
|
non_process
|
tooltips still not satisfactory version i have found this issue and i have to say i agree and conclude it is not fixed still all controls show a tooltip aggregation and the documentation keeps sending you in circles for example the has a tooltip aggregation it tells you the option to use the two tooltipbase objects but they are deprecated shouldn t you simply change that to a tooltip string property in the docs then as this is the only supported option at the moment for most controls the tooltip string simply does not work buttons are fine but above card for example does not show anything tooltips are so common they should be supported out of the box not with mouseover delegate and popover agree
| 0
|
381,726
| 11,286,759,318
|
IssuesEvent
|
2020-01-16 01:49:07
|
Radarr/Radarr
|
https://api.github.com/repos/Radarr/Radarr
|
closed
|
Update cleanReleaseGroupRegEx parsing
|
confirmed done aphrodite enhancement parser priority:low sonarr upstream
|
Scene name gets messed up on a repost/xpost/scrambled post-fix.
The Sonarr code was updated, so should be easy fix.
https://github.com/Radarr/Radarr/blob/develop/src/NzbDrone.Core/Parser/Parser.cs#L113
Added strings for movies would be xpost/Scrambled, but for usability ideally make it user configurable.
|
1.0
|
Update cleanReleaseGroupRegEx parsing - Scene name gets messed up on a repost/xpost/scrambled post-fix.
The Sonarr code was updated, so should be easy fix.
https://github.com/Radarr/Radarr/blob/develop/src/NzbDrone.Core/Parser/Parser.cs#L113
Added strings for movies would be xpost/Scrambled, but for usability ideally make it user configurable.
|
non_process
|
update cleanreleasegroupregex parsing scene name gets messed up on a repost xpost scrambled post fix the sonarr code was updated so should be easy fix added strings for movies would be xpost scrambled but for usability ideally make it user configurable
| 0
|
9,421
| 12,416,849,919
|
IssuesEvent
|
2020-05-22 19:09:36
|
NationalSecurityAgency/ghidra
|
https://api.github.com/repos/NationalSecurityAgency/ghidra
|
closed
|
M68000 decompiler: Read of volatile region ignored if followed by one or more writes
|
Feature: Processor/68000 Type: Bug
|
**Describe the bug**
When a read of a volatile memory region is followed by a sequence of zero-writes, the read is silently dropped from the decompiled output.
**To Reproduce**
Steps to reproduce the behavior:
1. Load the code segment included below as M68020 Big Endian code.
2. Define a volatile read-write region in the memory map -- starting address 0x210000, size 32 bytes. Define the whole region as a 32-byte array labelled "NVRAMCON".
3. Disassemble the code.
4. Observe that there is a read from the volatile region immediately prior to the writes which is not included in the disassembly (but should be).
**Expected behavior**
Decompiler shows a read from volatile space (with the result discarded), followed by a sequence of volatile writes of value zero.
**Environment (please complete the following information):**
- OS: Ubuntu 18.04
- Java Version: 11
- Ghidra Version: 9.0.2
**Additional context**
Disassembly extract:
```
0000040a 10 28 00 1e move.b (offset NVRAMCON[30],A0),D0b = null
0000040e 42 00 clr.b D0b
00000410 11 40 00 14 move.b D0b,(offset NVRAMCON[20],A0) = null
00000414 11 40 00 0a move.b D0b,(offset NVRAMCON[10],A0) = null
00000418 11 40 00 14 move.b D0b,(offset NVRAMCON[20],A0) = null
0000041c 11 40 00 0a move.b D0b,(offset NVRAMCON[10],A0) = null
00000420 11 40 00 0a move.b D0b,(offset NVRAMCON[10],A0) = null
```
Decompiler output:
```
write_volatile_1(NVRAMCON[20],0);
write_volatile_1(NVRAMCON[10],0);
write_volatile_1(NVRAMCON[20],0);
write_volatile_1(NVRAMCON[10],0);
write_volatile_1(NVRAMCON[10],0);
```
Note that the byte read from NVRAMCON[30] into D0 is ignored, but the writes are accepted.
|
1.0
|
M68000 decompiler: Read of volatile region ignored if followed by one or more writes - **Describe the bug**
When a read of a volatile memory region is followed by a sequence of zero-writes, the read is silently dropped from the decompiled output.
**To Reproduce**
Steps to reproduce the behavior:
1. Load the code segment included below as M68020 Big Endian code.
2. Define a volatile read-write region in the memory map -- starting address 0x210000, size 32 bytes. Define the whole region as a 32-byte array labelled "NVRAMCON".
3. Disassemble the code.
4. Observe that there is a read from the volatile region immediately prior to the writes which is not included in the disassembly (but should be).
**Expected behavior**
Decompiler shows a read from volatile space (with the result discarded), followed by a sequence of volatile writes of value zero.
**Environment (please complete the following information):**
- OS: Ubuntu 18.04
- Java Version: 11
- Ghidra Version: 9.0.2
**Additional context**
Disassembly extract:
```
0000040a 10 28 00 1e move.b (offset NVRAMCON[30],A0),D0b = null
0000040e 42 00 clr.b D0b
00000410 11 40 00 14 move.b D0b,(offset NVRAMCON[20],A0) = null
00000414 11 40 00 0a move.b D0b,(offset NVRAMCON[10],A0) = null
00000418 11 40 00 14 move.b D0b,(offset NVRAMCON[20],A0) = null
0000041c 11 40 00 0a move.b D0b,(offset NVRAMCON[10],A0) = null
00000420 11 40 00 0a move.b D0b,(offset NVRAMCON[10],A0) = null
```
Decompiler output:
```
write_volatile_1(NVRAMCON[20],0);
write_volatile_1(NVRAMCON[10],0);
write_volatile_1(NVRAMCON[20],0);
write_volatile_1(NVRAMCON[10],0);
write_volatile_1(NVRAMCON[10],0);
```
Note that the byte read from NVRAMCON[30] into D0 is ignored, but the writes are accepted.
|
process
|
decompiler read of volatile region ignored if followed by one or more writes describe the bug when a read of a volatile memory region is followed by a sequence of zero writes the read is silently dropped from the decompiled output to reproduce steps to reproduce the behavior load the code segment included below as big endian code define a volatile read write region in the memory map starting address size bytes define the whole region as a byte array labelled nvramcon disassemble the code observe that there is a read from the volatile region immediately prior to the writes which is not included in the disassembly but should be expected behavior decompiler shows a read from volatile space with the result discarded followed by a sequence of volatile writes of value zero environment please complete the following information os ubuntu java version ghidra version additional context disassembly extract move b offset nvramcon null clr b move b offset nvramcon null move b offset nvramcon null move b offset nvramcon null move b offset nvramcon null move b offset nvramcon null decompiler output write volatile nvramcon write volatile nvramcon write volatile nvramcon write volatile nvramcon write volatile nvramcon note that the byte read from nvramcon into is ignored but the writes are accepted
| 1
|
68,657
| 14,941,691,911
|
IssuesEvent
|
2021-01-25 20:08:23
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
"SQL Information Protection" is not available under "Security Center --> Pricing & Settings"
|
Pri2 cxp doc-enhancement security-center/svc triaged
|
The option "SQL Information Protection" is not available under "Security Center --> Pricing & Settings"
[Enter feedback here]
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 4ecc186b-cbfd-1c4e-fc2e-165883dc0a02
* Version Independent ID: 207f8456-0e71-c5c6-ddbd-88c0762dfc1e
* Content: [SQL information protection policy in Azure Security Center](https://docs.microsoft.com/en-us/azure/security-center/security-center-info-protection-policy?tabs=sqlip-tenant)
* Content Source: [articles/security-center/security-center-info-protection-policy.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/security-center/security-center-info-protection-policy.md)
* Service: **security-center**
* GitHub Login: @memildin
* Microsoft Alias: **memildin**
|
True
|
"SQL Information Protection" is not available under "Security Center --> Pricing & Settings" - The option "SQL Information Protection" is not available under "Security Center --> Pricing & Settings"
[Enter feedback here]
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 4ecc186b-cbfd-1c4e-fc2e-165883dc0a02
* Version Independent ID: 207f8456-0e71-c5c6-ddbd-88c0762dfc1e
* Content: [SQL information protection policy in Azure Security Center](https://docs.microsoft.com/en-us/azure/security-center/security-center-info-protection-policy?tabs=sqlip-tenant)
* Content Source: [articles/security-center/security-center-info-protection-policy.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/security-center/security-center-info-protection-policy.md)
* Service: **security-center**
* GitHub Login: @memildin
* Microsoft Alias: **memildin**
|
non_process
|
sql information protection is not available under security center pricing settings the option sql information protection is not available under security center pricing settings document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id cbfd version independent id ddbd content content source service security center github login memildin microsoft alias memildin
| 0
|
19,486
| 25,795,378,490
|
IssuesEvent
|
2022-12-10 13:53:47
|
vesoft-inc/nebula-studio
|
https://api.github.com/repos/vesoft-inc/nebula-studio
|
closed
|
connect a non-existent address timeout and get 500 code
|
process/fixed severity/major affects/v3.2 type/bug
|
**Describe the bug (__must be provided__)**
A clear and concise description of what the bug is.
**Your Environments (__must be provided__)**
* OS: `Mac OS 12.3.1`
* Node-version: `18.7.0`
* Studio-version: `all`
**How To Reproduce(__must be provided__)**
Steps to reproduce the behavior:
1. input address filed with '192.168.1.1:9669'
2. connecting for a long time until timeout
3. all connections failed
**Expected behavior**
A clear and concise description of what you expected to happen.
**Additional context**
Provide logs and configs, or any other context to trace the problem.
|
1.0
|
connect a non-existent address timeout and get 500 code - **Describe the bug (__must be provided__)**
A clear and concise description of what the bug is.
**Your Environments (__must be provided__)**
* OS: `Mac OS 12.3.1`
* Node-version: `18.7.0`
* Studio-version: `all`
**How To Reproduce(__must be provided__)**
Steps to reproduce the behavior:
1. input address filed with '192.168.1.1:9669'
2. connecting for a long time until timeout
3. all connections failed
**Expected behavior**
A clear and concise description of what you expected to happen.
**Additional context**
Provide logs and configs, or any other context to trace the problem.
|
process
|
connect a non existent address timeout and get code describe the bug must be provided a clear and concise description of what the bug is your environments must be provided os mac os node version studio version all how to reproduce must be provided steps to reproduce the behavior input address filed with connecting for a long time until timeout all connections failed expected behavior a clear and concise description of what you expected to happen additional context provide logs and configs or any other context to trace the problem
| 1
|
304,995
| 26,354,230,353
|
IssuesEvent
|
2023-01-11 08:26:15
|
OffchainLabs/arb-token-lists
|
https://api.github.com/repos/OffchainLabs/arb-token-lists
|
closed
|
CI integration tests
|
Type: Tests
|
Test "arbify":
- Fetch public arbified list (i.e., CMC); we'll call it L
- aribify it locally; i.e., check that arbifying its corresponding layer 1 list with L as the previous list still (deep)-equals L
Test "update":
- Fetch public arbified list (i.e., CMC); we'll call it L
- Updating L should still equal L
Test "fullList"
- Generate a full list locally for some network, check that corresponds to the public one
|
1.0
|
CI integration tests - Test "arbify":
- Fetch public arbified list (i.e., CMC); we'll call it L
- aribify it locally; i.e., check that arbifying its corresponding layer 1 list with L as the previous list still (deep)-equals L
Test "update":
- Fetch public arbified list (i.e., CMC); we'll call it L
- Updating L should still equal L
Test "fullList"
- Generate a full list locally for some network, check that corresponds to the public one
|
non_process
|
ci integration tests test arbify fetch public arbified list i e cmc we ll call it l aribify it locally i e check that arbifying its corresponding layer list with l as the previous list still deep equals l test update fetch public arbified list i e cmc we ll call it l updating l should still equal l test fulllist generate a full list locally for some network check that corresponds to the public one
| 0
|
13,313
| 15,783,613,299
|
IssuesEvent
|
2021-04-01 14:11:55
|
ooi-data/CE04OSPS-SF01B-3C-PARADA102-streamed-parad_sa_sample
|
https://api.github.com/repos/ooi-data/CE04OSPS-SF01B-3C-PARADA102-streamed-parad_sa_sample
|
opened
|
🛑 Processing failed: OSError
|
process
|
## Overview
`OSError` found in `processing_task` task during run ended on 2021-04-01T14:11:54.836710.
## Details
Flow name: `CE04OSPS-SF01B-3C-PARADA102-streamed-parad_sa_sample`
Task name: `processing_task`
Error type: `OSError`
Error message: [Errno 16] Please reduce your request rate.
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 234, in _call_s3
return await method(**additional_kwargs)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 154, in _make_api_call
raise error_class(parsed_response, operation_name)
botocore.exceptions.ClientError: An error occurred (SlowDown) when calling the DeleteObjects operation (reached max retries: 4): Please reduce your request rate.
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/share/miniconda/envs/harvester/lib/python3.8/site-packages/ooi_harvester/processor/pipeline.py", line 71, in processing_task
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/ooi_harvester/processor/__init__.py", line 311, in finalize_zarr
source_store.fs.delete(source_store.root, recursive=True)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/spec.py", line 1146, in delete
return self.rm(path, recursive=recursive, maxdepth=maxdepth)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1445, in rm
super().rm(path, recursive=recursive, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 196, in rm
maybe_sync(self._rm, self, path, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 100, in maybe_sync
return sync(loop, func, *args, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 71, in sync
raise exc.with_traceback(tb)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 55, in f
result[0] = await future
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1404, in _rm
await asyncio.gather(
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1396, in _bulk_delete
await self._call_s3(
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 252, in _call_s3
raise translate_boto_error(err) from err
OSError: [Errno 16] Please reduce your request rate.
```
</details>
|
1.0
|
🛑 Processing failed: OSError - ## Overview
`OSError` found in `processing_task` task during run ended on 2021-04-01T14:11:54.836710.
## Details
Flow name: `CE04OSPS-SF01B-3C-PARADA102-streamed-parad_sa_sample`
Task name: `processing_task`
Error type: `OSError`
Error message: [Errno 16] Please reduce your request rate.
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 234, in _call_s3
return await method(**additional_kwargs)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 154, in _make_api_call
raise error_class(parsed_response, operation_name)
botocore.exceptions.ClientError: An error occurred (SlowDown) when calling the DeleteObjects operation (reached max retries: 4): Please reduce your request rate.
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/share/miniconda/envs/harvester/lib/python3.8/site-packages/ooi_harvester/processor/pipeline.py", line 71, in processing_task
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/ooi_harvester/processor/__init__.py", line 311, in finalize_zarr
source_store.fs.delete(source_store.root, recursive=True)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/spec.py", line 1146, in delete
return self.rm(path, recursive=recursive, maxdepth=maxdepth)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1445, in rm
super().rm(path, recursive=recursive, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 196, in rm
maybe_sync(self._rm, self, path, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 100, in maybe_sync
return sync(loop, func, *args, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 71, in sync
raise exc.with_traceback(tb)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 55, in f
result[0] = await future
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1404, in _rm
await asyncio.gather(
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1396, in _bulk_delete
await self._call_s3(
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 252, in _call_s3
raise translate_boto_error(err) from err
OSError: [Errno 16] Please reduce your request rate.
```
</details>
|
process
|
🛑 processing failed oserror overview oserror found in processing task task during run ended on details flow name streamed parad sa sample task name processing task error type oserror error message please reduce your request rate traceback traceback most recent call last file srv conda envs notebook lib site packages core py line in call return await method additional kwargs file srv conda envs notebook lib site packages aiobotocore client py line in make api call raise error class parsed response operation name botocore exceptions clienterror an error occurred slowdown when calling the deleteobjects operation reached max retries please reduce your request rate the above exception was the direct cause of the following exception traceback most recent call last file usr share miniconda envs harvester lib site packages ooi harvester processor pipeline py line in processing task file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize zarr source store fs delete source store root recursive true file srv conda envs notebook lib site packages fsspec spec py line in delete return self rm path recursive recursive maxdepth maxdepth file srv conda envs notebook lib site packages core py line in rm super rm path recursive recursive kwargs file srv conda envs notebook lib site packages fsspec asyn py line in rm maybe sync self rm self path kwargs file srv conda envs notebook lib site packages fsspec asyn py line in maybe sync return sync loop func args kwargs file srv conda envs notebook lib site packages fsspec asyn py line in sync raise exc with traceback tb file srv conda envs notebook lib site packages fsspec asyn py line in f result await future file srv conda envs notebook lib site packages core py line in rm await asyncio gather file srv conda envs notebook lib site packages core py line in bulk delete await self call file srv conda envs notebook lib site packages core py line in call raise translate boto error err from err oserror please reduce your request rate
| 1
|
75,804
| 21,001,817,879
|
IssuesEvent
|
2022-03-29 18:13:37
|
cse-sim/cse
|
https://api.github.com/repos/cse-sim/cse
|
opened
|
Add economizer to RSYS
|
enhancement building science 2 - medium priority
|
- Add input for minimum OA fraction with hourly/subhourly variability
- Add inputs for economizer controls (see [EnergyPlus documentation](https://bigladdersoftware.com/epx/docs/9-6/input-output-reference/group-controllers.html#controlleroutdoorair))
- Maximum OA flow/fraction?
- Control type (Fixed/differential, drybulb/enthalpy)
- Lockout?
Some controls can probably be set based on expressions, others may need current timestep information that isn't available via probes.
|
1.0
|
Add economizer to RSYS - - Add input for minimum OA fraction with hourly/subhourly variability
- Add inputs for economizer controls (see [EnergyPlus documentation](https://bigladdersoftware.com/epx/docs/9-6/input-output-reference/group-controllers.html#controlleroutdoorair))
- Maximum OA flow/fraction?
- Control type (Fixed/differential, drybulb/enthalpy)
- Lockout?
Some controls can probably be set based on expressions, others may need current timestep information that isn't available via probes.
|
non_process
|
add economizer to rsys add input for minimum oa fraction with hourly subhourly variability add inputs for economizer controls see maximum oa flow fraction control type fixed differential drybulb enthalpy lockout some controls can probably be set based on expressions others may need current timestep information that isn t available via probes
| 0
|
164,114
| 25,921,480,784
|
IssuesEvent
|
2022-12-15 22:28:02
|
coder/coder
|
https://api.github.com/repos/coder/coder
|
closed
|
Long metadata strings get truncated
|
site design waiting-for-info
|
Can we have a way to specify that a particular metadata string not be truncated even if its too long?
<img width="826" alt="Screen Shot 2022-10-25 at 9 41 29 pm" src="https://user-images.githubusercontent.com/21701128/197753151-e89e7778-8f67-4470-a77e-6fd3b74acb9d.png">
|
1.0
|
Long metadata strings get truncated - Can we have a way to specify that a particular metadata string not be truncated even if its too long?
<img width="826" alt="Screen Shot 2022-10-25 at 9 41 29 pm" src="https://user-images.githubusercontent.com/21701128/197753151-e89e7778-8f67-4470-a77e-6fd3b74acb9d.png">
|
non_process
|
long metadata strings get truncated can we have a way to specify that a particular metadata string not be truncated even if its too long img width alt screen shot at pm src
| 0
|
4,064
| 6,995,525,398
|
IssuesEvent
|
2017-12-15 19:40:42
|
syndesisio/syndesis
|
https://api.github.com/repos/syndesisio/syndesis
|
opened
|
Improve Development Process
|
cat/discussion cat/process cat/retro
|
I personally think this should be an Epic, as it's so broad and there are a lot of areas we can improve on here. This was a card in the retrospective that drew my attention. The general comment on it was that we need time to accommodate for QE, UXD, and Docs review after implementing specific features, and that this was a bit of an issue with our current development process.
|
1.0
|
Improve Development Process - I personally think this should be an Epic, as it's so broad and there are a lot of areas we can improve on here. This was a card in the retrospective that drew my attention. The general comment on it was that we need time to accommodate for QE, UXD, and Docs review after implementing specific features, and that this was a bit of an issue with our current development process.
|
process
|
improve development process i personally think this should be an epic as it s so broad and there are a lot of areas we can improve on here this was a card in the retrospective that drew my attention the general comment on it was that we need time to accommodate for qe uxd and docs review after implementing specific features and that this was a bit of an issue with our current development process
| 1
|
265,546
| 28,297,918,719
|
IssuesEvent
|
2023-04-10 01:16:19
|
hshivhare67/platform_external_tcpdump_AOSP10_r33_4.9.2
|
https://api.github.com/repos/hshivhare67/platform_external_tcpdump_AOSP10_r33_4.9.2
|
closed
|
CVE-2020-8037 (High) detected in platform_external_tcpdumpandroid-mainline-12.0.0_r17 - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2020-8037 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>platform_external_tcpdumpandroid-mainline-12.0.0_r17</b></p></summary>
<p>
<p>Library home page: <a href=https://github.com/aosp-mirror/platform_external_tcpdump.git>https://github.com/aosp-mirror/platform_external_tcpdump.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/hshivhare67/platform_external_tcpdump_AOSP10_r33_4.9.2/commit/8bc69f757d775a894d925b18a07be1dcc294773c">8bc69f757d775a894d925b18a07be1dcc294773c</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/print-ppp.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The ppp decapsulator in tcpdump 4.9.3 can be convinced to allocate a large amount of memory.
<p>Publish Date: 2020-11-04
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-8037>CVE-2020-8037</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8037">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8037</a></p>
<p>Release Date: 2020-11-04</p>
<p>Fix Resolution: 4.10.0-bp</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-8037 (High) detected in platform_external_tcpdumpandroid-mainline-12.0.0_r17 - autoclosed - ## CVE-2020-8037 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>platform_external_tcpdumpandroid-mainline-12.0.0_r17</b></p></summary>
<p>
<p>Library home page: <a href=https://github.com/aosp-mirror/platform_external_tcpdump.git>https://github.com/aosp-mirror/platform_external_tcpdump.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/hshivhare67/platform_external_tcpdump_AOSP10_r33_4.9.2/commit/8bc69f757d775a894d925b18a07be1dcc294773c">8bc69f757d775a894d925b18a07be1dcc294773c</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/print-ppp.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The ppp decapsulator in tcpdump 4.9.3 can be convinced to allocate a large amount of memory.
<p>Publish Date: 2020-11-04
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-8037>CVE-2020-8037</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8037">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8037</a></p>
<p>Release Date: 2020-11-04</p>
<p>Fix Resolution: 4.10.0-bp</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in platform external tcpdumpandroid mainline autoclosed cve high severity vulnerability vulnerable library platform external tcpdumpandroid mainline library home page a href found in head commit a href found in base branch master vulnerable source files print ppp c vulnerability details the ppp decapsulator in tcpdump can be convinced to allocate a large amount of memory publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bp step up your open source security game with mend
| 0
|
148,952
| 19,560,750,033
|
IssuesEvent
|
2022-01-03 15:54:15
|
shaimael/Webgoat
|
https://api.github.com/repos/shaimael/Webgoat
|
opened
|
CVE-2017-7957 (High) detected in xstream-1.4.5.jar
|
security vulnerability
|
## CVE-2017-7957 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.5.jar</b></p></summary>
<p>XStream is a serialization library from Java objects to XML and back.</p>
<p>Path to dependency file: /webgoat-lessons/vulnerable-components/pom.xml</p>
<p>Path to vulnerable library: /m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **xstream-1.4.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/shaimael/Webgoat/commit/06d8d0b5bb8a459ba3d47f61a64fe00c62662d81">06d8d0b5bb8a459ba3d47f61a64fe00c62662d81</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XStream through 1.4.9, when a certain denyTypes workaround is not used, mishandles attempts to create an instance of the primitive type 'void' during unmarshalling, leading to a remote application crash, as demonstrated by an xstream.fromXML("<void/>") call.
<p>Publish Date: 2017-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-7957>CVE-2017-7957</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://x-stream.github.io/CVE-2017-7957.html">http://x-stream.github.io/CVE-2017-7957.html</a></p>
<p>Release Date: 2017-04-29</p>
<p>Fix Resolution: 1.4.10</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.5","packageFilePaths":["/webgoat-lessons/vulnerable-components/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.thoughtworks.xstream:xstream:1.4.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.10","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-7957","vulnerabilityDetails":"XStream through 1.4.9, when a certain denyTypes workaround is not used, mishandles attempts to create an instance of the primitive type \u0027void\u0027 during unmarshalling, leading to a remote application crash, as demonstrated by an xstream.fromXML(\"\u003cvoid/\u003e\") call.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-7957","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2017-7957 (High) detected in xstream-1.4.5.jar - ## CVE-2017-7957 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.5.jar</b></p></summary>
<p>XStream is a serialization library from Java objects to XML and back.</p>
<p>Path to dependency file: /webgoat-lessons/vulnerable-components/pom.xml</p>
<p>Path to vulnerable library: /m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.5/xstream-1.4.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **xstream-1.4.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/shaimael/Webgoat/commit/06d8d0b5bb8a459ba3d47f61a64fe00c62662d81">06d8d0b5bb8a459ba3d47f61a64fe00c62662d81</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XStream through 1.4.9, when a certain denyTypes workaround is not used, mishandles attempts to create an instance of the primitive type 'void' during unmarshalling, leading to a remote application crash, as demonstrated by an xstream.fromXML("<void/>") call.
<p>Publish Date: 2017-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-7957>CVE-2017-7957</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://x-stream.github.io/CVE-2017-7957.html">http://x-stream.github.io/CVE-2017-7957.html</a></p>
<p>Release Date: 2017-04-29</p>
<p>Fix Resolution: 1.4.10</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.4.5","packageFilePaths":["/webgoat-lessons/vulnerable-components/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.thoughtworks.xstream:xstream:1.4.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.10","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-7957","vulnerabilityDetails":"XStream through 1.4.9, when a certain denyTypes workaround is not used, mishandles attempts to create an instance of the primitive type \u0027void\u0027 during unmarshalling, leading to a remote application crash, as demonstrated by an xstream.fromXML(\"\u003cvoid/\u003e\") call.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-7957","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in xstream jar cve high severity vulnerability vulnerable library xstream jar xstream is a serialization library from java objects to xml and back path to dependency file webgoat lessons vulnerable components pom xml path to vulnerable library repository com thoughtworks xstream xstream xstream jar home wss scanner repository com thoughtworks xstream xstream xstream jar home wss scanner repository com thoughtworks xstream xstream xstream jar dependency hierarchy x xstream jar vulnerable library found in head commit a href found in base branch main vulnerability details xstream through when a certain denytypes workaround is not used mishandles attempts to create an instance of the primitive type void during unmarshalling leading to a remote application crash as demonstrated by an xstream fromxml call publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com thoughtworks xstream xstream isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails xstream through when a certain denytypes workaround is not used mishandles attempts to create an instance of the primitive type during unmarshalling leading to a remote application crash as demonstrated by an xstream fromxml call vulnerabilityurl
| 0
|
258,894
| 8,180,602,948
|
IssuesEvent
|
2018-08-28 20:00:29
|
MultiPoolMiner/MultiPoolMiner
|
https://api.github.com/repos/MultiPoolMiner/MultiPoolMiner
|
closed
|
WIP: Algorithms.txt with all new Equihash and CryptoNight algorithms
|
priority workaround
|
### Important!!!
**Algorithms.txt in the download zip file is broken - please use the file in this thread!**
Due to the quick changes in the CryptoNight and Equihash ecosystem we are struggling to keep Algorithms.txt up-to-date.
Should you encounter issues with CryptoNight or Equihash miners then make sure you are replacing
[MPM-Directory]\Algorithms.txt with the version posted here.
Version 11/08/2018
[Algorithms.txt](https://github.com/MultiPoolMiner/MultiPoolMiner/files/2280045/Algorithms.txt)
**Note: All CryptoNight & Equihash miners require always the latest version of Algorithms.txt**
|
1.0
|
WIP: Algorithms.txt with all new Equihash and CryptoNight algorithms - ### Important!!!
**Algorithms.txt in the download zip file is broken - please use the file in this thread!**
Due to the quick changes in the CryptoNight and Equihash ecosystem we are struggling to keep Algorithms.txt up-to-date.
Should you encounter issues with CryptoNight or Equihash miners then make sure you are replacing
[MPM-Directory]\Algorithms.txt with the version posted here.
Version 11/08/2018
[Algorithms.txt](https://github.com/MultiPoolMiner/MultiPoolMiner/files/2280045/Algorithms.txt)
**Note: All CryptoNight & Equihash miners require always the latest version of Algorithms.txt**
|
non_process
|
wip algorithms txt with all new equihash and cryptonight algorithms important algorithms txt in the download zip file is broken please use the file in this thread due to the quick changes in the cryptonight and equihash ecosystem we are struggling to keep algorithms txt up to date should you encounter issues with cryptonight or equihash miners then make sure you are replacing algorithms txt with the version posted here version note all cryptonight equihash miners require always the latest version of algorithms txt
| 0
|
254
| 2,677,662,878
|
IssuesEvent
|
2015-03-26 02:10:45
|
meteor/meteor
|
https://api.github.com/repos/meteor/meteor
|
closed
|
Release 1.0.5
|
Project:Release Process
|
This issue tracks the release of Meteor 1.0.5. Have any concerns? Mention them here. Finding that the RCs are working great? Mention that here. We will keep this top-level description updated with known issues.
The current RC is `1.0.5-rc.0`. Update to the release candidate with: `meteor update --release 1.0.5-rc.0`
Read [the release notes](https://github.com/meteor/meteor/blob/release-1.0.5/History.md) to learn what's new in this release. This release contains only one change not in 1.0.4.2: an upgrade to the `facebook` package to use a version of their Graph API that will not be deprecated at the end of April. If you do not use the `facebook` package (perhaps via `accounts-facebook`), this release will not affect you. If you only use Facebook for login (and don't use the Facebook API directly), you will not have to change any code after taking this upgrade. If you use the Facebook API directly, you may need to ask for different permissions. More information is at https://github.com/meteor/meteor/wiki/Facebook-Graph-API-Upgrade
- [x] History.md initial pass
- [x] History.md final draft
- [x] mailmap update
- [x] license check
- [x] full QA
- [ ] blog post text
- [x] banner text
RC history:
* `1.0.5-rc.0`: initial release candidate
|
1.0
|
Release 1.0.5 - This issue tracks the release of Meteor 1.0.5. Have any concerns? Mention them here. Finding that the RCs are working great? Mention that here. We will keep this top-level description updated with known issues.
The current RC is `1.0.5-rc.0`. Update to the release candidate with: `meteor update --release 1.0.5-rc.0`
Read [the release notes](https://github.com/meteor/meteor/blob/release-1.0.5/History.md) to learn what's new in this release. This release contains only one change not in 1.0.4.2: an upgrade to the `facebook` package to use a version of their Graph API that will not be deprecated at the end of April. If you do not use the `facebook` package (perhaps via `accounts-facebook`), this release will not affect you. If you only use Facebook for login (and don't use the Facebook API directly), you will not have to change any code after taking this upgrade. If you use the Facebook API directly, you may need to ask for different permissions. More information is at https://github.com/meteor/meteor/wiki/Facebook-Graph-API-Upgrade
- [x] History.md initial pass
- [x] History.md final draft
- [x] mailmap update
- [x] license check
- [x] full QA
- [ ] blog post text
- [x] banner text
RC history:
* `1.0.5-rc.0`: initial release candidate
|
process
|
release this issue tracks the release of meteor have any concerns mention them here finding that the rcs are working great mention that here we will keep this top level description updated with known issues the current rc is rc update to the release candidate with meteor update release rc read to learn what s new in this release this release contains only one change not in an upgrade to the facebook package to use a version of their graph api that will not be deprecated at the end of april if you do not use the facebook package perhaps via accounts facebook this release will not affect you if you only use facebook for login and don t use the facebook api directly you will not have to change any code after taking this upgrade if you use the facebook api directly you may need to ask for different permissions more information is at history md initial pass history md final draft mailmap update license check full qa blog post text banner text rc history rc initial release candidate
| 1
|
54,447
| 13,348,768,024
|
IssuesEvent
|
2020-08-29 20:20:28
|
cosmos/cosmos-sdk
|
https://api.github.com/repos/cosmos/cosmos-sdk
|
closed
|
go mod tidy fails with crypto/secp256k1/internal/secp256k1: no matching versions for query "latest"
|
build
|
When you run `go mod tidy`, the process exits with error:
```bash
alessio@phoenix:~/work/cosmos-sdk$ go mod tidy
go: finding module for package github.com/cosmos/cosmos-sdk/crypto/secp256k1/internal/secp256k1
github.com/cosmos/cosmos-sdk/crypto/keys/secp256k1 imports
github.com/cosmos/cosmos-sdk/crypto/secp256k1/internal/secp256k1: no matching versions for query "latest"
alessio@phoenix:~/work/cosmos-sdk$ echo $?
1
```
This seems to have been introduced by https://github.com/cosmos/cosmos-sdk/commit/443e0c1f89bd3730a731aea30453bd732f7efa35.
It's reproducible on master's current latest commit: 018915b1a833717c0e5cc90db66b9540c90ea37e
cc'ing @sahith-narahari @amaurymartiny @aaronc
____
## For Admin Use
- [ ] Not duplicate issue
- [ ] Appropriate labels applied
- [ ] Appropriate contributors tagged
- [ ] Contributor assigned/self-assigned
|
1.0
|
go mod tidy fails with crypto/secp256k1/internal/secp256k1: no matching versions for query "latest" - When you run `go mod tidy`, the process exits with error:
```bash
alessio@phoenix:~/work/cosmos-sdk$ go mod tidy
go: finding module for package github.com/cosmos/cosmos-sdk/crypto/secp256k1/internal/secp256k1
github.com/cosmos/cosmos-sdk/crypto/keys/secp256k1 imports
github.com/cosmos/cosmos-sdk/crypto/secp256k1/internal/secp256k1: no matching versions for query "latest"
alessio@phoenix:~/work/cosmos-sdk$ echo $?
1
```
This seems to have been introduced by https://github.com/cosmos/cosmos-sdk/commit/443e0c1f89bd3730a731aea30453bd732f7efa35.
It's reproducible on master's current latest commit: 018915b1a833717c0e5cc90db66b9540c90ea37e
cc'ing @sahith-narahari @amaurymartiny @aaronc
____
## For Admin Use
- [ ] Not duplicate issue
- [ ] Appropriate labels applied
- [ ] Appropriate contributors tagged
- [ ] Contributor assigned/self-assigned
|
non_process
|
go mod tidy fails with crypto internal no matching versions for query latest when you run go mod tidy the process exits with error bash alessio phoenix work cosmos sdk go mod tidy go finding module for package github com cosmos cosmos sdk crypto internal github com cosmos cosmos sdk crypto keys imports github com cosmos cosmos sdk crypto internal no matching versions for query latest alessio phoenix work cosmos sdk echo this seems to have been introduced by it s reproducible on master s current latest commit cc ing sahith narahari amaurymartiny aaronc for admin use not duplicate issue appropriate labels applied appropriate contributors tagged contributor assigned self assigned
| 0
|
180,415
| 6,649,520,876
|
IssuesEvent
|
2017-09-28 13:31:33
|
tootsuite/mastodon
|
https://api.github.com/repos/tootsuite/mastodon
|
closed
|
Ability to add alt text to images/media
|
a11y enhancement new user experience priority - high
|
Images must have text alternatives that describe the information or function represented by the images. This ensures that images can be used by people with various disabilities. The main technique to add alt text is via the `alt` attribute.
Twitter added the ability to add alt text to images earlier this year. Something in Mastodon would be great!
Some more information on alt text:
- [Short note on coding alt text](https://www.paciellogroup.com/blog/2015/09/short-note-on-coding-alt-text/)
- [HTML 5.1 §4.7.5.1. Requirements for providing text to act as an alternative for images](https://www.w3.org/TR/html51/semantics-embedded-content.html#alt-text)
- [Web Content Accessibility Guideline 1.1 – Text Alternatives](https://www.w3.org/WAI/WCAG20/quickref/#qr-text-equiv-all)
- [Web Accessibility Tutorials Images Concepts](https://www.w3.org/WAI/tutorials/images/)
|
1.0
|
Ability to add alt text to images/media - Images must have text alternatives that describe the information or function represented by the images. This ensures that images can be used by people with various disabilities. The main technique to add alt text is via the `alt` attribute.
Twitter added the ability to add alt text to images earlier this year. Something in Mastodon would be great!
Some more information on alt text:
- [Short note on coding alt text](https://www.paciellogroup.com/blog/2015/09/short-note-on-coding-alt-text/)
- [HTML 5.1 §4.7.5.1. Requirements for providing text to act as an alternative for images](https://www.w3.org/TR/html51/semantics-embedded-content.html#alt-text)
- [Web Content Accessibility Guideline 1.1 – Text Alternatives](https://www.w3.org/WAI/WCAG20/quickref/#qr-text-equiv-all)
- [Web Accessibility Tutorials Images Concepts](https://www.w3.org/WAI/tutorials/images/)
|
non_process
|
ability to add alt text to images media images must have text alternatives that describe the information or function represented by the images this ensures that images can be used by people with various disabilities the main technique to add alt text is via the alt attribute twitter added the ability to add alt text to images earlier this year something in mastodon would be great some more information on alt text
| 0
|
4,450
| 7,315,757,361
|
IssuesEvent
|
2018-03-01 12:16:26
|
rubberduck-vba/Rubberduck
|
https://api.github.com/repos/rubberduck-vba/Rubberduck
|
closed
|
comment seen as code
|
antlr bug difficulty-02-ducky feature-inspections parse-tree-processing up-for-grabs
|
Error: Local variable 'testbatch' is not declared - (XLTest163a.xlsm) XLTest.Informations, line 227
```
Sub testAudit() '@notrace testautotest testbatch
```
|
1.0
|
comment seen as code - Error: Local variable 'testbatch' is not declared - (XLTest163a.xlsm) XLTest.Informations, line 227
```
Sub testAudit() '@notrace testautotest testbatch
```
|
process
|
comment seen as code error local variable testbatch is not declared xlsm xltest informations line sub testaudit notrace testautotest testbatch
| 1
|
13,058
| 15,394,228,219
|
IssuesEvent
|
2021-03-03 17:38:45
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Obsolete 'detection of stimulus involved in cell cycle checkpoint' branch
|
cell cycle and DNA processes obsoletion
|
Dear all,
The proposal has been made to obsolete the 'detection of stimulus involved in cell cycle checkpoint' branch. The reason for obsoletion is that these terms correspond to molecular functions. The terms impacted are:
GO:0090429 detection of endogenous biotic stimulus
GO:0072394 detection of stimulus involved in cell cycle checkpoint
GO:0072409 detection of stimulus involved in meiotic cell cycle checkpoint
GO:0072412 detection of stimulus involved in mitotic cell cycle checkpoint
GO:0072415 detection of stimulus involved in spindle checkpoint
GO:0072460 detection of stimulus involved in meiotic recombination checkpoint
GO:0072463 detection of stimulus involved in meiotic spindle assembly checkpoint
GO:0072475 detection of stimulus involved in mitotic spindle checkpoint
GO:0072484 detection of stimulus involved in spindle assembly checkpoint
GO:0072481 detection of stimulus involved in mitotic spindle orientation checkpoint
GO:0072469 detection of stimulus involved in cell size control checkpoint
GO:0072448 detection of stimulus involved in G1 cell size control checkpoint
GO:0072451 detection of stimulus involved in G2 cell size control checkpoint
GO:0072397 detection of stimulus involved in cytokinesis checkpoint
GO:1903819 detection of stimulus involved in mitotic cytokinesis checkpoint
GO:0072400 detection of stimulus involved in mitotic DNA integrity checkpoint
GO:1902401 detection of stimulus involved in mitotic DNA damage checkpoint
GO:0072442 detection of stimulus involved in mitotic DNA replication checkpoint
GO:0072427 detection of DNA damage stimulus involved in intra-S DNA damage checkpoint
GO:0072430 detection of DNA damage stimulus involved in mitotic G1 DNA damage checkpoint
GO:0072433 detection of DNA damage stimulus involved in mitotic G2 DNA damage checkpoint
GO:1903821 detection of stimulus involved in morphogenesis checkpoint
There is a single annotation to those terms, by MGI (MGI:1328317). There are no mappings, those terms are not present in any subsets.
You can comment on the ticket: https://github.com/geneontology/go-ontology/issues/20976
Thanks, Pascale
|
1.0
|
Obsolete 'detection of stimulus involved in cell cycle checkpoint' branch - Dear all,
The proposal has been made to obsolete the 'detection of stimulus involved in cell cycle checkpoint' branch. The reason for obsoletion is that these terms correspond to molecular functions. The terms impacted are:
GO:0090429 detection of endogenous biotic stimulus
GO:0072394 detection of stimulus involved in cell cycle checkpoint
GO:0072409 detection of stimulus involved in meiotic cell cycle checkpoint
GO:0072412 detection of stimulus involved in mitotic cell cycle checkpoint
GO:0072415 detection of stimulus involved in spindle checkpoint
GO:0072460 detection of stimulus involved in meiotic recombination checkpoint
GO:0072463 detection of stimulus involved in meiotic spindle assembly checkpoint
GO:0072475 detection of stimulus involved in mitotic spindle checkpoint
GO:0072484 detection of stimulus involved in spindle assembly checkpoint
GO:0072481 detection of stimulus involved in mitotic spindle orientation checkpoint
GO:0072469 detection of stimulus involved in cell size control checkpoint
GO:0072448 detection of stimulus involved in G1 cell size control checkpoint
GO:0072451 detection of stimulus involved in G2 cell size control checkpoint
GO:0072397 detection of stimulus involved in cytokinesis checkpoint
GO:1903819 detection of stimulus involved in mitotic cytokinesis checkpoint
GO:0072400 detection of stimulus involved in mitotic DNA integrity checkpoint
GO:1902401 detection of stimulus involved in mitotic DNA damage checkpoint
GO:0072442 detection of stimulus involved in mitotic DNA replication checkpoint
GO:0072427 detection of DNA damage stimulus involved in intra-S DNA damage checkpoint
GO:0072430 detection of DNA damage stimulus involved in mitotic G1 DNA damage checkpoint
GO:0072433 detection of DNA damage stimulus involved in mitotic G2 DNA damage checkpoint
GO:1903821 detection of stimulus involved in morphogenesis checkpoint
There is a single annotation to those terms, by MGI (MGI:1328317). There are no mappings, those terms are not present in any subsets.
You can comment on the ticket: https://github.com/geneontology/go-ontology/issues/20976
Thanks, Pascale
|
process
|
obsolete detection of stimulus involved in cell cycle checkpoint branch dear all the proposal has been made to obsolete the detection of stimulus involved in cell cycle checkpoint branch the reason for obsoletion is that these terms correspond to molecular functions the terms impacted are go detection of endogenous biotic stimulus go detection of stimulus involved in cell cycle checkpoint go detection of stimulus involved in meiotic cell cycle checkpoint go detection of stimulus involved in mitotic cell cycle checkpoint go detection of stimulus involved in spindle checkpoint go detection of stimulus involved in meiotic recombination checkpoint go detection of stimulus involved in meiotic spindle assembly checkpoint go detection of stimulus involved in mitotic spindle checkpoint go detection of stimulus involved in spindle assembly checkpoint go detection of stimulus involved in mitotic spindle orientation checkpoint go detection of stimulus involved in cell size control checkpoint go detection of stimulus involved in cell size control checkpoint go detection of stimulus involved in cell size control checkpoint go detection of stimulus involved in cytokinesis checkpoint go detection of stimulus involved in mitotic cytokinesis checkpoint go detection of stimulus involved in mitotic dna integrity checkpoint go detection of stimulus involved in mitotic dna damage checkpoint go detection of stimulus involved in mitotic dna replication checkpoint go detection of dna damage stimulus involved in intra s dna damage checkpoint go detection of dna damage stimulus involved in mitotic dna damage checkpoint go detection of dna damage stimulus involved in mitotic dna damage checkpoint go detection of stimulus involved in morphogenesis checkpoint there is a single annotation to those terms by mgi mgi there are no mappings those terms are not present in any subsets you can comment on the ticket thanks pascale
| 1
|
13,814
| 5,467,265,993
|
IssuesEvent
|
2017-03-10 00:31:53
|
mitchellh/packer
|
https://api.github.com/repos/mitchellh/packer
|
closed
|
build -force should remove existing AWS AMI
|
bug builder/amazon
|
According to the docs, -force should remove existing artifacts before building, however this seems to work for QEMU but not for amazon-ebs builder:
# packer build -force template.json
[...]
==> amazon-ebs: Error: name conflicts with an existing AMI: ami-0f160a63
Build 'amazon-ebs' errored: Error: name conflicts with an existing AMI: ami-0f160a63
|
1.0
|
build -force should remove existing AWS AMI - According to the docs, -force should remove existing artifacts before building, however this seems to work for QEMU but not for amazon-ebs builder:
# packer build -force template.json
[...]
==> amazon-ebs: Error: name conflicts with an existing AMI: ami-0f160a63
Build 'amazon-ebs' errored: Error: name conflicts with an existing AMI: ami-0f160a63
|
non_process
|
build force should remove existing aws ami according to the docs force should remove existing artifacts before building however this seems to work for qemu but not for amazon ebs builder packer build force template json amazon ebs error name conflicts with an existing ami ami build amazon ebs errored error name conflicts with an existing ami ami
| 0
|
135,309
| 18,678,892,678
|
IssuesEvent
|
2021-11-01 01:02:30
|
benchmarkdebricked/kubernetes
|
https://api.github.com/repos/benchmarkdebricked/kubernetes
|
opened
|
CVE-2020-15113 (High) detected in mobyv17.03.2-ce, kubernetesv1.16.0-alpha.0
|
security vulnerability
|
## CVE-2020-15113 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>mobyv17.03.2-ce</b>, <b>kubernetesv1.16.0-alpha.0</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In etcd before versions 3.3.23 and 3.4.10, certain directory paths are created (etcd data directory and the directory path when provided to automatically generate self-signed certificates for TLS connections with clients) with restricted access permissions (700) by using the os.MkdirAll. This function does not perform any permission checks when a given directory path exists already. A possible workaround is to ensure the directories have the desired permission (700).
<p>Publish Date: 2020-08-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15113>CVE-2020-15113</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/etcd-io/etcd/compare/v3.4.9...v3.4.10">https://github.com/etcd-io/etcd/compare/v3.4.9...v3.4.10</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 3.4.10, 3.3.23</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-15113 (High) detected in mobyv17.03.2-ce, kubernetesv1.16.0-alpha.0 - ## CVE-2020-15113 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>mobyv17.03.2-ce</b>, <b>kubernetesv1.16.0-alpha.0</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In etcd before versions 3.3.23 and 3.4.10, certain directory paths are created (etcd data directory and the directory path when provided to automatically generate self-signed certificates for TLS connections with clients) with restricted access permissions (700) by using the os.MkdirAll. This function does not perform any permission checks when a given directory path exists already. A possible workaround is to ensure the directories have the desired permission (700).
<p>Publish Date: 2020-08-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15113>CVE-2020-15113</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/etcd-io/etcd/compare/v3.4.9...v3.4.10">https://github.com/etcd-io/etcd/compare/v3.4.9...v3.4.10</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 3.4.10, 3.3.23</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in ce alpha cve high severity vulnerability vulnerable libraries ce alpha vulnerability details in etcd before versions and certain directory paths are created etcd data directory and the directory path when provided to automatically generate self signed certificates for tls connections with clients with restricted access permissions by using the os mkdirall this function does not perform any permission checks when a given directory path exists already a possible workaround is to ensure the directories have the desired permission publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
668,369
| 22,580,879,121
|
IssuesEvent
|
2022-06-28 11:31:36
|
Mezzanine-UI/mezzanine
|
https://api.github.com/repos/Mezzanine-UI/mezzanine
|
closed
|
Add support for React v18 and server side render
|
feature Priority: HIGH react
|
### Is your feature request related to a problem? Please describe.
1. mezzanine `createNotifier` use `render` and `unmountComponentAtNode` which comes from `react-dom` package and is already marked as DEPRECATED in react v18.
2. In SSR mode, document will be undefined.
### Describe the solution you'd like
See https://github.com/facebook/react/releases/tag/v18.0.0 for more information
### Additional context or references
none
|
1.0
|
Add support for React v18 and server side render - ### Is your feature request related to a problem? Please describe.
1. mezzanine `createNotifier` use `render` and `unmountComponentAtNode` which comes from `react-dom` package and is already marked as DEPRECATED in react v18.
2. In SSR mode, document will be undefined.
### Describe the solution you'd like
See https://github.com/facebook/react/releases/tag/v18.0.0 for more information
### Additional context or references
none
|
non_process
|
add support for react and server side render is your feature request related to a problem please describe mezzanine createnotifier use render and unmountcomponentatnode which comes from react dom package and is already marked as deprecated in react in ssr mode document will be undefined describe the solution you d like see for more information additional context or references none
| 0
|
12,004
| 14,738,160,194
|
IssuesEvent
|
2021-01-07 03:56:16
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
001 NCSM - NCSM 1st cycle accounts
|
anc-ops anc-process anp-important ant-bug ant-support
|
In GitLab by @kdjstudios on May 9, 2018, 12:17
**Submitted by:** "Jesus Corchado" <jesus.corchado@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-05-09-47848/conversation
**Server:** Internal
**Client/Site:** Multiple
**Account:** NA
**Issue:**
Sorry, we found another item that we need fixed. I processed the billing on the outbound sites for the NCSM 1st cycle accounts. However, it is showing that the last billing date was 5/29/18 and when you go to their billing history, it shows invoice dates of 6/1/18. I checked to see if I could make the change myself but it’s showing last billing date of 4/29/18. I need to run the billing for the other sites but need these fixed since it may confuse accounting once they go in to the accounts.
|
1.0
|
001 NCSM - NCSM 1st cycle accounts - In GitLab by @kdjstudios on May 9, 2018, 12:17
**Submitted by:** "Jesus Corchado" <jesus.corchado@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-05-09-47848/conversation
**Server:** Internal
**Client/Site:** Multiple
**Account:** NA
**Issue:**
Sorry, we found another item that we need fixed. I processed the billing on the outbound sites for the NCSM 1st cycle accounts. However, it is showing that the last billing date was 5/29/18 and when you go to their billing history, it shows invoice dates of 6/1/18. I checked to see if I could make the change myself but it’s showing last billing date of 4/29/18. I need to run the billing for the other sites but need these fixed since it may confuse accounting once they go in to the accounts.
|
process
|
ncsm ncsm cycle accounts in gitlab by kdjstudios on may submitted by jesus corchado helpdesk server internal client site multiple account na issue sorry we found another item that we need fixed i processed the billing on the outbound sites for the ncsm cycle accounts however it is showing that the last billing date was and when you go to their billing history it shows invoice dates of i checked to see if i could make the change myself but it’s showing last billing date of i need to run the billing for the other sites but need these fixed since it may confuse accounting once they go in to the accounts
| 1
|
174,454
| 14,483,172,461
|
IssuesEvent
|
2020-12-10 14:51:51
|
nilearn/nilearn
|
https://api.github.com/repos/nilearn/nilearn
|
closed
|
Improve Documentation of masker family
|
Documentation
|
Perhaps the docstrings for the following classes could be improved:
```
- NiftiMasker
- NiftiLabelsMasker
- NiftiMapsMasker
- MultiNiftiMasker
```
More specifically, at the top of the respective class docstrings, one might add
- a more generous explanation of the purpose of the class
- its commonalities and differences comparing to the respective other masker classes
- provide one simple use case when this is typically useful
Apart from that, a new nilearn example might be useful that uses the various masker and, thus, gives an overview of this part of the API.
Edited by @AlexandreAbraham:
Here is the list of remarks made on _Nifti_Masker doc across issues:
- [ ] Specify which data is detrended (#418)
- [ ] Add intersect_masks to NiftiMasker's "related functions" (#436)
- [ ] Specify that only `mask_args` and `mask_strategy` are used for fitting. It's specified in the class docstring, but not on the fit function docstring (#436). Gael suggested to add it in the Notes section.
- [ ] Add in transform (#436):
- Parameters (detrend, standardize) are computed from the images passed to transform (i.e. not computed and stored from fit)
- Whether the inverse of these parameters is applied when inverse_transform is called (not sure if they are, for detrend, standardize, and confounds).
- [ ] NiftiMapsMasker and NiftiLabelsMasker don't have docstring for fit_transforms (#603).
|
1.0
|
Improve Documentation of masker family - Perhaps the docstrings for the following classes could be improved:
```
- NiftiMasker
- NiftiLabelsMasker
- NiftiMapsMasker
- MultiNiftiMasker
```
More specifically, at the top of the respective class docstrings, one might add
- a more generous explanation of the purpose of the class
- its commonalities and differences comparing to the respective other masker classes
- provide one simple use case when this is typically useful
Apart from that, a new nilearn example might be useful that uses the various masker and, thus, gives an overview of this part of the API.
Edited by @AlexandreAbraham:
Here is the list of remarks made on _Nifti_Masker doc across issues:
- [ ] Specify which data is detrended (#418)
- [ ] Add intersect_masks to NiftiMasker's "related functions" (#436)
- [ ] Specify that only `mask_args` and `mask_strategy` are used for fitting. It's specified in the class docstring, but not on the fit function docstring (#436). Gael suggested to add it in the Notes section.
- [ ] Add in transform (#436):
- Parameters (detrend, standardize) are computed from the images passed to transform (i.e. not computed and stored from fit)
- Whether the inverse of these parameters is applied when inverse_transform is called (not sure if they are, for detrend, standardize, and confounds).
- [ ] NiftiMapsMasker and NiftiLabelsMasker don't have docstring for fit_transforms (#603).
|
non_process
|
improve documentation of masker family perhaps the docstrings for the following classes could be improved niftimasker niftilabelsmasker niftimapsmasker multiniftimasker more specifically at the top of the respective class docstrings one might add a more generous explanation of the purpose of the class its commonalities and differences comparing to the respective other masker classes provide one simple use case when this is typically useful apart from that a new nilearn example might be useful that uses the various masker and thus gives an overview of this part of the api edited by alexandreabraham here is the list of remarks made on nifti masker doc across issues specify which data is detrended add intersect masks to niftimasker s related functions specify that only mask args and mask strategy are used for fitting it s specified in the class docstring but not on the fit function docstring gael suggested to add it in the notes section add in transform parameters detrend standardize are computed from the images passed to transform i e not computed and stored from fit whether the inverse of these parameters is applied when inverse transform is called not sure if they are for detrend standardize and confounds niftimapsmasker and niftilabelsmasker don t have docstring for fit transforms
| 0
|
6,838
| 9,979,560,449
|
IssuesEvent
|
2019-07-09 23:24:03
|
GroceriStar/food-static-files-generator
|
https://api.github.com/repos/GroceriStar/food-static-files-generator
|
closed
|
fileSystem methods separation
|
enhancement good first issue help wanted in-process
|
**Is your feature request related to a problem? Please describe.**
https://github.com/GroceriStar/food-static-files-generator/blob/master/src/fileSystem.js
https://github.com/GroceriStar/food-static-files-generator/issues/72
**Describe the solution you'd like**
move write and read methods to a separated file. it will help to make writeFile.js less complicated and shorter
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
|
1.0
|
fileSystem methods separation - **Is your feature request related to a problem? Please describe.**
https://github.com/GroceriStar/food-static-files-generator/blob/master/src/fileSystem.js
https://github.com/GroceriStar/food-static-files-generator/issues/72
**Describe the solution you'd like**
move write and read methods to a separated file. it will help to make writeFile.js less complicated and shorter
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
|
process
|
filesystem methods separation is your feature request related to a problem please describe describe the solution you d like move write and read methods to a separated file it will help to make writefile js less complicated and shorter describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
| 1
|
92,993
| 10,764,432,758
|
IssuesEvent
|
2019-11-01 08:15:47
|
tiuweehan/ped
|
https://api.github.com/repos/tiuweehan/ped
|
opened
|
Sort command is not valid (and supposed to come in v1.3)
|
severity.Medium type.DocumentationBug
|
Sort is not a valid command in the application and UG says `coming in 1.3` even though the version is 1.3

|
1.0
|
Sort command is not valid (and supposed to come in v1.3) - Sort is not a valid command in the application and UG says `coming in 1.3` even though the version is 1.3

|
non_process
|
sort command is not valid and supposed to come in sort is not a valid command in the application and ug says coming in even though the version is
| 0
|
10,663
| 13,453,185,035
|
IssuesEvent
|
2020-09-09 00:11:26
|
zaimoni/Cataclysm
|
https://api.github.com/repos/zaimoni/Cataclysm
|
opened
|
local fork: PDCurses
|
process refactor
|
For tile support. Also the reference curses implementation for Windows. catacurse.cpp/h is a hand-rolled re-implementation. License is public domain (ok).
|
1.0
|
local fork: PDCurses - For tile support. Also the reference curses implementation for Windows. catacurse.cpp/h is a hand-rolled re-implementation. License is public domain (ok).
|
process
|
local fork pdcurses for tile support also the reference curses implementation for windows catacurse cpp h is a hand rolled re implementation license is public domain ok
| 1
|
437,847
| 12,603,364,566
|
IssuesEvent
|
2020-06-11 13:22:21
|
EBISPOT/goci
|
https://api.github.com/repos/EBISPOT/goci
|
closed
|
Beta not imported into curation interface from deposition
|
Priority: High Type: Bug
|
Beta was entered in the submission template but was not imported into the curation interface - Example Buchwald J 32157176.
|
1.0
|
Beta not imported into curation interface from deposition - Beta was entered in the submission template but was not imported into the curation interface - Example Buchwald J 32157176.
|
non_process
|
beta not imported into curation interface from deposition beta was entered in the submission template but was not imported into the curation interface example buchwald j
| 0
|
256,862
| 19,474,118,292
|
IssuesEvent
|
2021-12-24 08:47:43
|
dxc-technology/halstack-style-guide
|
https://api.github.com/repos/dxc-technology/halstack-style-guide
|
reopened
|
[Docs] Component review and design improvements
|
category: documentation :notebook: task: epic :pushpin:
|
## Component reviews
- [x] #190
- [x] #224
- [x] #225
- [x] #443
- [x] #477
- [x] #478
- [x] #497
- [x] #499
- [x] #530
- [x] #542
- [x] #619
## Fixes
- [x] #322
- [x] #323
- [x] #437
- [x] #460
- [x] #474
- [x] #547
- [x] #572
- [x] #588
- [x] #645
- [x] #646
- [ ] #656
## Design improvements
- [x] #583
- [x] #607
|
1.0
|
[Docs] Component review and design improvements - ## Component reviews
- [x] #190
- [x] #224
- [x] #225
- [x] #443
- [x] #477
- [x] #478
- [x] #497
- [x] #499
- [x] #530
- [x] #542
- [x] #619
## Fixes
- [x] #322
- [x] #323
- [x] #437
- [x] #460
- [x] #474
- [x] #547
- [x] #572
- [x] #588
- [x] #645
- [x] #646
- [ ] #656
## Design improvements
- [x] #583
- [x] #607
|
non_process
|
component review and design improvements component reviews fixes design improvements
| 0
|
530,639
| 15,435,333,606
|
IssuesEvent
|
2021-03-07 08:22:29
|
magento/magento2
|
https://api.github.com/repos/magento/magento2
|
closed
|
Page layouts are hard-coded in Magento\Widget\Block\Adminhtml\Widget\Instance\Edit\Chooser\Container
|
Issue: Confirmed Issue: ready for confirmation Priority: P2 Progress: PR in progress Reported on 2.3.5 Reproduced on 2.4.x Severity: S3 Triage: Dev.Experience
|
### Preconditions (*)
Magento 2.3.5
### Steps to reproduce (*)
Have a look at the code source of `Magento\Widget\Block\Adminhtml\Widget\Instance\Edit\Chooser\Container`.
### Expected result (*)
The function `getPageLayouts()` should return the actual list of page layouts declared by the different modules.
### Actual result (*)
The function `getPageLayouts()` only returns the one provided by Magento_Theme, and they are hard-coded in the file:
```php
/**#@+
* Frontend page layouts
*/
const PAGE_LAYOUT_1COLUMN = '1column-center';
const PAGE_LAYOUT_2COLUMNS_LEFT = '2columns-left';
const PAGE_LAYOUT_2COLUMNS_RIGHT = '2columns-right';
const PAGE_LAYOUT_3COLUMNS = '3columns';
/**#@-*/
[...]
/**
* Retrieve page layouts
*
* @return array
*/
protected function getPageLayouts()
{
return [
self::PAGE_LAYOUT_1COLUMN,
self::PAGE_LAYOUT_2COLUMNS_LEFT,
self::PAGE_LAYOUT_2COLUMNS_RIGHT,
self::PAGE_LAYOUT_3COLUMNS,
];
}
```
### Workaround
I may be possible to override `Magento\Widget\Block\Adminhtml\Widget\Instance\Edit\Chooser\Container` and to inject an instance of `Magento\Framework\View\Model\PageLayout\Config\BuilderInterface` to grab the list of available page layouts from there. Untested though.
### Thoughts
Basically, it makes any container with Label declared in a non-Magento Theme page layout not selectable from the back-office to inject widgets. But the simple fact that Magento_Widget assumes that some page layouts exist is concerning. This concern is actually confirmed by the fact that one of the four hard-coded page layouts doesn't even exist: `1column-center`.
---
Please provide [Severity](https://devdocs.magento.com/guides/v2.3/contributor-guide/contributing.html#backlog) assessment for the Issue as Reporter. This information will help during Confirmation and Issue triage processes.
- [ ] Severity: **S0** _- Affects critical data or functionality and leaves users without workaround._
- [ ] Severity: **S1** _- Affects critical data or functionality and forces users to employ a workaround._
- [x] Severity: **S2** _- Affects non-critical data or functionality and forces users to employ a workaround._
- [ ] Severity: **S3** _- Affects non-critical data or functionality and does not force users to employ a workaround._
- [ ] Severity: **S4** _- Affects aesthetics, professional look and feel, “quality” or “usability”._
|
1.0
|
Page layouts are hard-coded in Magento\Widget\Block\Adminhtml\Widget\Instance\Edit\Chooser\Container - ### Preconditions (*)
Magento 2.3.5
### Steps to reproduce (*)
Have a look at the code source of `Magento\Widget\Block\Adminhtml\Widget\Instance\Edit\Chooser\Container`.
### Expected result (*)
The function `getPageLayouts()` should return the actual list of page layouts declared by the different modules.
### Actual result (*)
The function `getPageLayouts()` only returns the one provided by Magento_Theme, and they are hard-coded in the file:
```php
/**#@+
* Frontend page layouts
*/
const PAGE_LAYOUT_1COLUMN = '1column-center';
const PAGE_LAYOUT_2COLUMNS_LEFT = '2columns-left';
const PAGE_LAYOUT_2COLUMNS_RIGHT = '2columns-right';
const PAGE_LAYOUT_3COLUMNS = '3columns';
/**#@-*/
[...]
/**
* Retrieve page layouts
*
* @return array
*/
protected function getPageLayouts()
{
return [
self::PAGE_LAYOUT_1COLUMN,
self::PAGE_LAYOUT_2COLUMNS_LEFT,
self::PAGE_LAYOUT_2COLUMNS_RIGHT,
self::PAGE_LAYOUT_3COLUMNS,
];
}
```
### Workaround
I may be possible to override `Magento\Widget\Block\Adminhtml\Widget\Instance\Edit\Chooser\Container` and to inject an instance of `Magento\Framework\View\Model\PageLayout\Config\BuilderInterface` to grab the list of available page layouts from there. Untested though.
### Thoughts
Basically, it makes any container with Label declared in a non-Magento Theme page layout not selectable from the back-office to inject widgets. But the simple fact that Magento_Widget assumes that some page layouts exist is concerning. This concern is actually confirmed by the fact that one of the four hard-coded page layouts doesn't even exist: `1column-center`.
---
Please provide [Severity](https://devdocs.magento.com/guides/v2.3/contributor-guide/contributing.html#backlog) assessment for the Issue as Reporter. This information will help during Confirmation and Issue triage processes.
- [ ] Severity: **S0** _- Affects critical data or functionality and leaves users without workaround._
- [ ] Severity: **S1** _- Affects critical data or functionality and forces users to employ a workaround._
- [x] Severity: **S2** _- Affects non-critical data or functionality and forces users to employ a workaround._
- [ ] Severity: **S3** _- Affects non-critical data or functionality and does not force users to employ a workaround._
- [ ] Severity: **S4** _- Affects aesthetics, professional look and feel, “quality” or “usability”._
|
non_process
|
page layouts are hard coded in magento widget block adminhtml widget instance edit chooser container preconditions magento steps to reproduce have a look at the code source of magento widget block adminhtml widget instance edit chooser container expected result the function getpagelayouts should return the actual list of page layouts declared by the different modules actual result the function getpagelayouts only returns the one provided by magento theme and they are hard coded in the file php frontend page layouts const page layout center const page layout left left const page layout right right const page layout retrieve page layouts return array protected function getpagelayouts return self page layout self page layout left self page layout right self page layout workaround i may be possible to override magento widget block adminhtml widget instance edit chooser container and to inject an instance of magento framework view model pagelayout config builderinterface to grab the list of available page layouts from there untested though thoughts basically it makes any container with label declared in a non magento theme page layout not selectable from the back office to inject widgets but the simple fact that magento widget assumes that some page layouts exist is concerning this concern is actually confirmed by the fact that one of the four hard coded page layouts doesn t even exist center please provide assessment for the issue as reporter this information will help during confirmation and issue triage processes severity affects critical data or functionality and leaves users without workaround severity affects critical data or functionality and forces users to employ a workaround severity affects non critical data or functionality and forces users to employ a workaround severity affects non critical data or functionality and does not force users to employ a workaround severity affects aesthetics professional look and feel “quality” or “usability”
| 0
|
17,778
| 23,704,366,931
|
IssuesEvent
|
2022-08-29 22:33:47
|
googleapis/gapic-showcase
|
https://api.github.com/repos/googleapis/gapic-showcase
|
opened
|
switch docs to reference "latest" release assets instead of specific version
|
type: process
|
We should change all of the examples in the README to use the "latest" release version, specifically when downloading release assets. See https://docs.github.com/en/repositories/releasing-projects-on-github/linking-to-releases.
As part of this, we'd need to exclude the release version from the asset names e.g. `gapic-showcase-0.24.0-darwin-amd64.tar.gz` --> `gapic-showcase-darwin-amd64.tar.gz`. This would involve changes to the release asset creation workflow and notification to downstream consumers that the naming scheme changed.
|
1.0
|
switch docs to reference "latest" release assets instead of specific version - We should change all of the examples in the README to use the "latest" release version, specifically when downloading release assets. See https://docs.github.com/en/repositories/releasing-projects-on-github/linking-to-releases.
As part of this, we'd need to exclude the release version from the asset names e.g. `gapic-showcase-0.24.0-darwin-amd64.tar.gz` --> `gapic-showcase-darwin-amd64.tar.gz`. This would involve changes to the release asset creation workflow and notification to downstream consumers that the naming scheme changed.
|
process
|
switch docs to reference latest release assets instead of specific version we should change all of the examples in the readme to use the latest release version specifically when downloading release assets see as part of this we d need to exclude the release version from the asset names e g gapic showcase darwin tar gz gapic showcase darwin tar gz this would involve changes to the release asset creation workflow and notification to downstream consumers that the naming scheme changed
| 1
|
5,571
| 8,407,883,919
|
IssuesEvent
|
2018-10-11 22:34:58
|
HumanCellAtlas/dcp-community
|
https://api.github.com/repos/HumanCellAtlas/dcp-community
|
opened
|
Add recommended format for Slack Channels
|
charter-process
|
Adopt the best practice that @lauraclarke created in the Ingest Support charter:
Recommended format for Slack Channels
`[HumanCellAtlas/channel](https://humancellatlas.slack.com/messages/channel)`
- [ ] Update charter template to include sub-sections for Mailing Lists and Slack Channels now that we have mailing lists
- [ ] Update existing charters
|
1.0
|
Add recommended format for Slack Channels - Adopt the best practice that @lauraclarke created in the Ingest Support charter:
Recommended format for Slack Channels
`[HumanCellAtlas/channel](https://humancellatlas.slack.com/messages/channel)`
- [ ] Update charter template to include sub-sections for Mailing Lists and Slack Channels now that we have mailing lists
- [ ] Update existing charters
|
process
|
add recommended format for slack channels adopt the best practice that lauraclarke created in the ingest support charter recommended format for slack channels update charter template to include sub sections for mailing lists and slack channels now that we have mailing lists update existing charters
| 1
|
3,213
| 6,274,320,909
|
IssuesEvent
|
2017-07-18 01:37:39
|
gaocegege/Processing.R
|
https://api.github.com/repos/gaocegege/Processing.R
|
opened
|
Add compiler or related tech to solve call back functions
|
community/processing difficulty/high priority/p1 size/no-idea status/to-be-claimed type/bug
|
related issues: #166
There are some functions which will be defined in R but it is called in Java. Actually we have limited implementation for these functions, for example, draw and settings. But some libraries and some other built-in functions need to be supported.
Built-in functions includes: mouse related functions.
|
1.0
|
Add compiler or related tech to solve call back functions - related issues: #166
There are some functions which will be defined in R but it is called in Java. Actually we have limited implementation for these functions, for example, draw and settings. But some libraries and some other built-in functions need to be supported.
Built-in functions includes: mouse related functions.
|
process
|
add compiler or related tech to solve call back functions related issues there are some functions which will be defined in r but it is called in java actually we have limited implementation for these functions for example draw and settings but some libraries and some other built in functions need to be supported built in functions includes mouse related functions
| 1
|
13,316
| 15,784,065,090
|
IssuesEvent
|
2021-04-01 14:42:52
|
edwardsmarc/CASFRI
|
https://api.github.com/repos/edwardsmarc/CASFRI
|
opened
|
Create one version of TT_RowIsValid() and TT_HasPrecedence() for each process that requires it
|
enhancement high post-translation process
|
TT_RowIsValid() and TT_HasPrecedence() are two functions that have to be overwritten according to the context they are used. They only require to be passed the same king of argument and to return a boolean.
They are used in many contexts:
- in the basic geohistory test
- in the real data geohistory test
- in the real gridded data geohistory test
- when creating the historical table from available inventories
For now I have been trying to minimize the duplication of these function but ths has lead to just greater confusion. It's hard to know which is which and you have to execute part of a script to overwrite them to be used by another script. It would be better if each context would provide it's own version of the two version even if they duplicates other versions.
|
1.0
|
Create one version of TT_RowIsValid() and TT_HasPrecedence() for each process that requires it - TT_RowIsValid() and TT_HasPrecedence() are two functions that have to be overwritten according to the context they are used. They only require to be passed the same king of argument and to return a boolean.
They are used in many contexts:
- in the basic geohistory test
- in the real data geohistory test
- in the real gridded data geohistory test
- when creating the historical table from available inventories
For now I have been trying to minimize the duplication of these function but ths has lead to just greater confusion. It's hard to know which is which and you have to execute part of a script to overwrite them to be used by another script. It would be better if each context would provide it's own version of the two version even if they duplicates other versions.
|
process
|
create one version of tt rowisvalid and tt hasprecedence for each process that requires it tt rowisvalid and tt hasprecedence are two functions that have to be overwritten according to the context they are used they only require to be passed the same king of argument and to return a boolean they are used in many contexts in the basic geohistory test in the real data geohistory test in the real gridded data geohistory test when creating the historical table from available inventories for now i have been trying to minimize the duplication of these function but ths has lead to just greater confusion it s hard to know which is which and you have to execute part of a script to overwrite them to be used by another script it would be better if each context would provide it s own version of the two version even if they duplicates other versions
| 1
|
6,584
| 9,661,896,557
|
IssuesEvent
|
2019-05-20 19:18:33
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
opened
|
Ensure Python tests run under Windows
|
P1 team-Rules-Python type: process
|
We [currently disable](https://github.com/bazelbuild/bazel/blob/bdc6c10b674757df38982e71752aab8eddab47a0/src/test/shell/bazel/BUILD#L724-L725) some Python tests under Windows because until recently our CI machines didn't have both Python 2 and Python 3 available. Now that that's resolved, we should enable them.
P1 because this blocks fixing #7947, which in turn is needed to enable toolchains on Windows.
Separate from #6868, which concerns adding or OSSing test cases.
|
1.0
|
Ensure Python tests run under Windows - We [currently disable](https://github.com/bazelbuild/bazel/blob/bdc6c10b674757df38982e71752aab8eddab47a0/src/test/shell/bazel/BUILD#L724-L725) some Python tests under Windows because until recently our CI machines didn't have both Python 2 and Python 3 available. Now that that's resolved, we should enable them.
P1 because this blocks fixing #7947, which in turn is needed to enable toolchains on Windows.
Separate from #6868, which concerns adding or OSSing test cases.
|
process
|
ensure python tests run under windows we some python tests under windows because until recently our ci machines didn t have both python and python available now that that s resolved we should enable them because this blocks fixing which in turn is needed to enable toolchains on windows separate from which concerns adding or ossing test cases
| 1
|
4,857
| 7,745,937,541
|
IssuesEvent
|
2018-05-29 19:59:47
|
GoogleCloudPlatform/google-cloud-python
|
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python
|
reopened
|
[BigQuery] cannot be used with Python 3.7 (Pandas)
|
api: bigquery testing type: process
|
Pandas/Cython doesn't currently support 3.7. Once it does we need to enable 3.7 testing on BigQuery.
Collecting pandas>=0.17.1 (from google-cloud-monitoring==0.29.0)
Using cached https://files.pythonhosted.org/packages/08/01/803834bc8a4e708aedebb133095a88a4dad9f45bbaf5ad777d2bea543c7e/pandas-0.22.0.tar.gz
Could not find a version that satisfies the requirement Cython (from versions: )
No matching distribution found for Cython
|
1.0
|
[BigQuery] cannot be used with Python 3.7 (Pandas) - Pandas/Cython doesn't currently support 3.7. Once it does we need to enable 3.7 testing on BigQuery.
Collecting pandas>=0.17.1 (from google-cloud-monitoring==0.29.0)
Using cached https://files.pythonhosted.org/packages/08/01/803834bc8a4e708aedebb133095a88a4dad9f45bbaf5ad777d2bea543c7e/pandas-0.22.0.tar.gz
Could not find a version that satisfies the requirement Cython (from versions: )
No matching distribution found for Cython
|
process
|
cannot be used with python pandas pandas cython doesn t currently support once it does we need to enable testing on bigquery collecting pandas from google cloud monitoring using cached could not find a version that satisfies the requirement cython from versions no matching distribution found for cython
| 1
|
8,563
| 11,736,676,482
|
IssuesEvent
|
2020-03-11 13:28:55
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Mapping for South Africa
|
Pri2 automation/svc awaiting-product-team-response cxp process-automation/subsvc product-question triaged
|
Hi, what is the mapping for South Africa? Are solutions like Update Management not available in that region?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: f8f86bd0-7555-9be2-1015-76e3ab88062f
* Version Independent ID: 1d310402-fadf-d602-048a-b2e16bd86a7e
* Content: [Azure Automation and Log Analytics workspace mappings](https://docs.microsoft.com/en-us/azure/automation/how-to/region-mappings)
* Content Source: [articles/automation/how-to/region-mappings.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/how-to/region-mappings.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
1.0
|
Mapping for South Africa - Hi, what is the mapping for South Africa? Are solutions like Update Management not available in that region?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: f8f86bd0-7555-9be2-1015-76e3ab88062f
* Version Independent ID: 1d310402-fadf-d602-048a-b2e16bd86a7e
* Content: [Azure Automation and Log Analytics workspace mappings](https://docs.microsoft.com/en-us/azure/automation/how-to/region-mappings)
* Content Source: [articles/automation/how-to/region-mappings.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/how-to/region-mappings.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
process
|
mapping for south africa hi what is the mapping for south africa are solutions like update management not available in that region document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id fadf content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
| 1
|
3,208
| 6,264,803,417
|
IssuesEvent
|
2017-07-16 11:54:55
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
n.kill(19) throws error
|
child_process doc good first contribution process
|
On versions 6 and 7 of Node.js
when I started a child process with `const n = require('child_process').spawn`
if I try,
`n.kill(19)`, I get an error saying the signal is not recognized
but 19 is a valid signal
http://stackoverflow.com/questions/9951556/why-number-9-in-kill-9-command-in-unix
I would expect it to work
if I do `n.kill('SIGSTOP')`, it seems to work.
Can we (not) use numbers/integers with n.kill()?
|
2.0
|
n.kill(19) throws error - On versions 6 and 7 of Node.js
when I started a child process with `const n = require('child_process').spawn`
if I try,
`n.kill(19)`, I get an error saying the signal is not recognized
but 19 is a valid signal
http://stackoverflow.com/questions/9951556/why-number-9-in-kill-9-command-in-unix
I would expect it to work
if I do `n.kill('SIGSTOP')`, it seems to work.
Can we (not) use numbers/integers with n.kill()?
|
process
|
n kill throws error on versions and of node js when i started a child process with const n require child process spawn if i try n kill i get an error saying the signal is not recognized but is a valid signal i would expect it to work if i do n kill sigstop it seems to work can we not use numbers integers with n kill
| 1
|
12,529
| 14,972,166,098
|
IssuesEvent
|
2021-01-27 22:21:36
|
BootBlock/FileSieve
|
https://api.github.com/repos/BootBlock/FileSieve
|
opened
|
Decouple Profile processing from the UI
|
backend-core processing ui
|
To allow the simultaneous processing of multiple profiles and to allow the Trigger system to fire for more than one profile at a time, profile processing will need to be decoupled from the user interface.
Requires some thought, especially when it comes to showing the fact that multiple profiles are being processed at the same time.
|
1.0
|
Decouple Profile processing from the UI - To allow the simultaneous processing of multiple profiles and to allow the Trigger system to fire for more than one profile at a time, profile processing will need to be decoupled from the user interface.
Requires some thought, especially when it comes to showing the fact that multiple profiles are being processed at the same time.
|
process
|
decouple profile processing from the ui to allow the simultaneous processing of multiple profiles and to allow the trigger system to fire for more than one profile at a time profile processing will need to be decoupled from the user interface requires some thought especially when it comes to showing the fact that multiple profiles are being processed at the same time
| 1
|
2,226
| 3,575,814,867
|
IssuesEvent
|
2016-01-27 17:12:44
|
allo-/firefox-profilemaker
|
https://api.github.com/repos/allo-/firefox-profilemaker
|
closed
|
Migrate all forms to DynamicConfigForm
|
Infrastructure
|
All Forms should use an options dictionary and be created by ``create_configform``. Then the view can be adapted to create forms from a list of ``(id, name, options)`` tuples instead of Form objects.
|
1.0
|
Migrate all forms to DynamicConfigForm - All Forms should use an options dictionary and be created by ``create_configform``. Then the view can be adapted to create forms from a list of ``(id, name, options)`` tuples instead of Form objects.
|
non_process
|
migrate all forms to dynamicconfigform all forms should use an options dictionary and be created by create configform then the view can be adapted to create forms from a list of id name options tuples instead of form objects
| 0
|
20,122
| 26,659,350,410
|
IssuesEvent
|
2023-01-25 19:36:47
|
pb866/Kimai.jl
|
https://api.github.com/repos/pb866/Kimai.jl
|
opened
|
Allow comments in input files
|
enhancement data processing
|
Allow the possibility to add comments in the input data files of vacation and sick leave by specifying a character, e.g. `#`, in `CSV.read`.
|
1.0
|
Allow comments in input files - Allow the possibility to add comments in the input data files of vacation and sick leave by specifying a character, e.g. `#`, in `CSV.read`.
|
process
|
allow comments in input files allow the possibility to add comments in the input data files of vacation and sick leave by specifying a character e g in csv read
| 1
|
17,476
| 23,298,447,210
|
IssuesEvent
|
2022-08-07 00:19:06
|
mdsreq-fga-unb/2022.1-GDS
|
https://api.github.com/repos/mdsreq-fga-unb/2022.1-GDS
|
closed
|
Riscos
|
Planejamento Processo de Desenvolvimento
|
**Descrição**
quando será realizado o gerenciamento dos riscos? em que momento do processo?
Ou seja, quando e como a equipe irá revisar e atualizar os riscos do projeto e do produto?
|
1.0
|
Riscos - **Descrição**
quando será realizado o gerenciamento dos riscos? em que momento do processo?
Ou seja, quando e como a equipe irá revisar e atualizar os riscos do projeto e do produto?
|
process
|
riscos descrição quando será realizado o gerenciamento dos riscos em que momento do processo ou seja quando e como a equipe irá revisar e atualizar os riscos do projeto e do produto
| 1
|
9,879
| 12,886,461,010
|
IssuesEvent
|
2020-07-13 09:31:17
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
Improve tests in `cli` (currently using Mocha)
|
kind/improvement process/candidate team/typescript topic: internal topic: tests
|
Tests in https://github.com/prisma/prisma/tree/master/src/packages/cli/src/__tests__/
They run on Mocha while we are using Jest everywhere else.
Each new Mocha test now require a manual edit of the package.json.
Maybe it makes sense to keep Mocha for the integration tests but we should use Jest for the rest.
TODO
- [ ] Adding Jest along Mocha
- [ ] Fix doctor test @timsuchanek
|
1.0
|
Improve tests in `cli` (currently using Mocha) - Tests in https://github.com/prisma/prisma/tree/master/src/packages/cli/src/__tests__/
They run on Mocha while we are using Jest everywhere else.
Each new Mocha test now require a manual edit of the package.json.
Maybe it makes sense to keep Mocha for the integration tests but we should use Jest for the rest.
TODO
- [ ] Adding Jest along Mocha
- [ ] Fix doctor test @timsuchanek
|
process
|
improve tests in cli currently using mocha tests in they run on mocha while we are using jest everywhere else each new mocha test now require a manual edit of the package json maybe it makes sense to keep mocha for the integration tests but we should use jest for the rest todo adding jest along mocha fix doctor test timsuchanek
| 1
|
15,276
| 19,257,181,218
|
IssuesEvent
|
2021-12-09 12:36:46
|
km4ack/pi-build
|
https://api.github.com/repos/km4ack/pi-build
|
closed
|
EES post to outbox on Bullseye and Pat 0.12 not working
|
bug in process
|
I initially thought this was a CRONJOB issue of not posting the EES emails to the PAT outbox but I believe it is a bug in the movetopi.html script referencing a bad location for the outbox.
The EES email is created is in /var/www/html/emails
The mv command to post to the PAT outbox is referencing a folder that isn’t on Bullseye-PAT. The old system had a folder named home/pi/.wl2k. The new PAT implementation has no such folder therefore the move to outbox doesn’t work. I am not sure where the PAT In and Out boxes are in the Bullseye version of PAT.
The command mv /var/www/html/emails/*.b2f $HOME/.wl2k/mailbox/$MYCALL/out is not a good location on PAT 0.12
|
1.0
|
EES post to outbox on Bullseye and Pat 0.12 not working - I initially thought this was a CRONJOB issue of not posting the EES emails to the PAT outbox but I believe it is a bug in the movetopi.html script referencing a bad location for the outbox.
The EES email is created is in /var/www/html/emails
The mv command to post to the PAT outbox is referencing a folder that isn’t on Bullseye-PAT. The old system had a folder named home/pi/.wl2k. The new PAT implementation has no such folder therefore the move to outbox doesn’t work. I am not sure where the PAT In and Out boxes are in the Bullseye version of PAT.
The command mv /var/www/html/emails/*.b2f $HOME/.wl2k/mailbox/$MYCALL/out is not a good location on PAT 0.12
|
process
|
ees post to outbox on bullseye and pat not working i initially thought this was a cronjob issue of not posting the ees emails to the pat outbox but i believe it is a bug in the movetopi html script referencing a bad location for the outbox the ees email is created is in var www html emails the mv command to post to the pat outbox is referencing a folder that isn’t on bullseye pat the old system had a folder named home pi the new pat implementation has no such folder therefore the move to outbox doesn’t work i am not sure where the pat in and out boxes are in the bullseye version of pat the command mv var www html emails home mailbox mycall out is not a good location on pat
| 1
|
790,151
| 27,817,435,632
|
IssuesEvent
|
2023-03-18 21:06:53
|
qutebrowser/qutebrowser
|
https://api.github.com/repos/qutebrowser/qutebrowser
|
closed
|
Crashes in QStyleSheetStyle::styleHint
|
qt priority: 1 - middle bug: segfault/crash/hang
|
**Version info**:
```
qutebrowser v1.10.2
Git commit:
Backend: QtWebEngine (Chromium 77.0.3865.129)
Qt: 5.14.2
CPython: 3.8.2
PyQt: 5.14.2
sip: 5.1.2
colorama: 0.4.3
pypeg2: 2.15
jinja2: 2.11.2
pygments: 2.6.1
yaml: 5.3.1
cssutils: no
attr: 19.3.0
PyQt5.QtWebEngineWidgets: yes
PyQt5.QtWebEngine: 5.14.0
PyQt5.QtWebKitWidgets: yes
pdf.js: 2.4.456 (/usr/share/pdf.js/build/pdf.js)
sqlite: 3.31.1
QtNetwork SSL: OpenSSL 1.1.1g 21 Apr 2020
Style: QFusionStyle
Platform plugin: wayland
Platform: Linux-5.5.13-arch2-1-x86_64-with-glibc2.2.5, 64bit
Linux distribution: Arch Linux (arch)
Frozen: False
Imported from /usr/lib/python3.8/site-packages/qutebrowser
Using Python from /usr/bin/python3
Qt library executable path: /usr/lib/qt/libexec, data path: /usr/share/qt
Paths:
cache: /home/kapheine/.cache/qutebrowser
config: /home/kapheine/.config/qutebrowser
data: /home/kapheine/.local/share/qutebrowser
runtime: /run/user/1000/qutebrowser
system data: /usr/share/qutebrowser
Autoconfig loaded: yes
Config.py: /home/kapheine/.config/qutebrowser/config.py has been loaded
Uptime: 0:02:56
```
**Does the bug happen if you start with `--temp-basedir`?**:
Not sure, as it only happens occasionally. I can try running for a while in that setup, if it's useful.
**Description**
I have a coredump on my system for Qutebrowser crashing in QStyleSheetStyle:
```
>>> bt
#0 __GI_raise (sig=<optimized out>) at ../sysdeps/unix/sysv/linux/raise.c:50
#1 0x00007fea0550ad70 in <signal handler called> () at /usr/lib/libc.so.6
#2 0x000056178b0d8db0 in ()
#3 0x00007fea0082bba8 in QStyleSheetStyle::styleHint(QStyle::StyleHint, QStyleOption const*, QWidget const*, QStyleHintReturn*) const (this=0x56178721e780, sh=QStyle::SH_Widget_ShareActivation, opt=0x0, w=0x7fe974013870, shret=<optimized out>) at styles/qstylesheetstyle.cpp:5342
#4 0x00007fea007a7d9a in QWidget::isActiveWindow() const (this=this@entry=0x7fe974013870) at kernel/qwidget.cpp:6678
#5 0x00007fea007a7f4f in QWidget::palette() const (this=this@entry=0x7fe974013870) at kernel/qwidget.cpp:4356
#6 0x00007fea007c0f0b in QWidget::event(QEvent*) (this=this@entry=0x7fe974013870, event=event@entry=0x7fff498c8250) at kernel/qwidget.cpp:8788
#7 0x00007fea00871cbf in QFrame::event(QEvent*) (this=this@entry=0x7fe974013870, e=e@entry=0x7fff498c8250) at widgets/qframe.cpp:550
#8 0x00007fea00874a0a in QAbstractScrollArea::event(QEvent*) (this=this@entry=0x7fe974013870, e=e@entry=0x7fff498c8250) at widgets/qabstractscrollarea.cpp:1042
#9 0x00007fea009fcc04 in QAbstractItemView::event(QEvent*) (this=0x7fe974013870, event=0x7fff498c8250) at itemviews/qabstractitemview.cpp:1674
#10 0x00007fea00f91a97 in sipQListView::event(QEvent*) (this=0x7fe974013870, a0=0x7fff498c8250) at sipQtWidgetsQListView.cpp:1189
#11 0x00007fea0077c352 in QApplicationPrivate::notify_helper(QObject*, QEvent*) (this=this@entry=0x56178648df30, receiver=receiver@entry=0x7fe974013870, e=e@entry=0x7fff498c8250) at kernel/qapplication.cpp:3685
#12 0x00007fea00785829 in QApplication::notify(QObject*, QEvent*) (this=0x5617869eab00, receiver=0x7fe974013870, e=0x7fff498c8250) at kernel/qapplication.cpp:3431
#13 0x00007fea010f8445 in sipQApplication::notify(QObject*, QEvent*) (this=0x5617869eab00, a0=0x7fe974013870, a1=0x7fff498c8250) at sipQtWidgetsQApplication.cpp:243
#14 0x00007fea03ca34f2 in QCoreApplication::notifyInternal2(QObject*, QEvent*) (receiver=0x7fe974013870, event=0x7fff498c8250) at ../../include/QtCore/../../src/corelib/kernel/qobject.h:153
#15 0x00007fea007c0c15 in QWidget::event(QEvent*) (this=<optimized out>, event=0x7fff498c8250) at kernel/qwidget.cpp:8950
#16 0x00007fea0077c352 in QApplicationPrivate::notify_helper(QObject*, QEvent*) (this=this@entry=0x56178648df30, receiver=receiver@entry=0x561786b83790, e=e@entry=0x7fff498c8250) at kernel/qapplication.cpp:3685
#17 0x00007fea00785829 in QApplication::notify(QObject*, QEvent*) (this=0x5617869eab00, receiver=0x561786b83790, e=0x7fff498c8250) at kernel/qapplication.cpp:3431
#18 0x00007fea010f8445 in sipQApplication::notify(QObject*, QEvent*) (this=0x5617869eab00, a0=0x561786b83790, a1=0x7fff498c8250) at sipQtWidgetsQApplication.cpp:243
#19 0x00007fea03ca34f2 in QCoreApplication::notifyInternal2(QObject*, QEvent*) (receiver=0x561786b83790, event=0x7fff498c8250) at ../../include/QtCore/../../src/corelib/kernel/qobject.h:153
#20 0x00007fea00783418 in QApplication::setActiveWindow(QWidget*) (act=act@entry=0x0) at kernel/qapplication.cpp:2039
#21 0x00007fea007a6518 in QWidgetPrivate::deactivateWidgetCleanup() (this=this@entry=0x5617863c2c20) at kernel/qwidget.cpp:2305
#22 0x00007fea007b3459 in QWidgetPrivate::hide_sys() (this=this@entry=0x5617863c2c20) at kernel/qwidget.cpp:8009
#23 0x00007fea007bb232 in QWidgetPrivate::hide_helper() (this=this@entry=0x5617863c2c20) at kernel/qwidget.cpp:7952
#24 0x00007fea007bff31 in QWidgetPrivate::setVisible(bool) (this=0x5617863c2c20, visible=<optimized out>) at kernel/qwidget.cpp:8147
#25 0x00007fea007bb5a8 in QWidgetPrivate::close_helper(QWidgetPrivate::CloseMode) (mode=<optimized out>, this=0x5617863c2c20) at kernel/qwidget.cpp:8262
#26 QWidgetPrivate::close_helper(QWidgetPrivate::CloseMode) (this=0x5617863c2c20, mode=<optimized out>) at kernel/qwidget.cpp:8237
#27 0x00007fea007bbbf1 in QWidget::~QWidget() (this=0x561786b83790, __in_chrg=<optimized out>) at kernel/qwidget.cpp:1476
#28 0x00007fea011301e2 in sipQWidget::~sipQWidget() (this=0x561786b83790, __in_chrg=<optimized out>) at sipQtWidgetsQWidget.cpp:355
#29 0x00007fea01130204 in sipQWidget::~sipQWidget() (this=0x561786b83790, __in_chrg=<optimized out>) at sipQtWidgetsQWidget.cpp:358
#30 0x00007fea041f21f3 in cleanup_qobject(sipSimpleWrapper*, void*) (sw=0x7fe9cc14b940, closure=0x5617869eab00) at ../../qpy/QtCore/qpycore_public_api.cpp:66
#31 0x00007fea03f562d8 in () at /usr/lib/python3.8/site-packages/PyQt5/sip.cpython-38-x86_64-linux-gnu.so
#32 0x00007fea041f2242 in pyqt5_cleanup_qobjects() () at ../../qpy/QtCore/qpycore_public_api.cpp:78
#33 0x00007fea041f1c1f in cleanup_on_exit(PyObject*, PyObject*) () at ../../qpy/QtCore/qpycore_init.cpp:37
#34 0x00007fea05251299 in () at /usr/lib/libpython3.8.so.1.0
#35 0x00007fea0524b3fd in PyObject_Call () at /usr/lib/libpython3.8.so.1.0
#36 0x00007fea05364329 in () at /usr/lib/libpython3.8.so.1.0
#37 0x00007fea0521d328 in () at /usr/lib/libpython3.8.so.1.0
#38 0x00007fea0521d36c in Py_Exit () at /usr/lib/libpython3.8.so.1.0
#39 0x00007fea0521d385 in () at /usr/lib/libpython3.8.so.1.0
#40 0x00007fea05385172 in () at /usr/lib/libpython3.8.so.1.0
#41 0x00007fea0521ead1 in PyRun_SimpleFileExFlags () at /usr/lib/libpython3.8.so.1.0
#42 0x00007fea0538681e in Py_RunMain () at /usr/lib/libpython3.8.so.1.0
#43 0x00007fea05386909 in Py_BytesMain () at /usr/lib/libpython3.8.so.1.0
#44 0x00007fea054f6023 in __libc_start_main (main=0x561785edb020 <main>, argc=2, argv=0x7fff498c89b8, init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized out>, stack_end=0x7fff498c89a8) at ../csu/libc-start.c:308
#45 0x0000561785edb05e in _start ()
>>> frame 3
#3 0x00007fea0082bba8 in QStyleSheetStyle::styleHint (this=0x56178721e780, sh=QStyle::SH_Widget_ShareActivation, opt=0x0, w=0x7fe974013870, shret=<optimized out>) at styles/qstylesheetstyle.cpp:5342
warning: Source file is more recent than executable.
5342 return baseStyle()->styleHint(sh, opt, w, shret);
>>> list
5337 {
5338 RECURSION_GUARD(return baseStyle()->styleHint(sh, opt, w, shret))
5339 // Prevent endless loop if somebody use isActiveWindow property as selector.
5340 // QWidget::isActiveWindow uses this styleHint to determine if the window is active or not
5341 if (sh == SH_Widget_ShareActivation)
5342 return baseStyle()->styleHint(sh, opt, w, shret);
5343
5344 QRenderRule rule = renderRule(w, opt);
5345 QString s;
5346 switch (sh) {
```
Sounds like it COULD be related to #5187, but I'm a bit out of my element here.
**How to reproduce**
I'm not sure what I was doing at the time. I see a number of other crashes over the last month or so, but unfortunately my system already deleted them. I'll try to check again later, to see if the same crash is repeating.
|
1.0
|
Crashes in QStyleSheetStyle::styleHint - **Version info**:
```
qutebrowser v1.10.2
Git commit:
Backend: QtWebEngine (Chromium 77.0.3865.129)
Qt: 5.14.2
CPython: 3.8.2
PyQt: 5.14.2
sip: 5.1.2
colorama: 0.4.3
pypeg2: 2.15
jinja2: 2.11.2
pygments: 2.6.1
yaml: 5.3.1
cssutils: no
attr: 19.3.0
PyQt5.QtWebEngineWidgets: yes
PyQt5.QtWebEngine: 5.14.0
PyQt5.QtWebKitWidgets: yes
pdf.js: 2.4.456 (/usr/share/pdf.js/build/pdf.js)
sqlite: 3.31.1
QtNetwork SSL: OpenSSL 1.1.1g 21 Apr 2020
Style: QFusionStyle
Platform plugin: wayland
Platform: Linux-5.5.13-arch2-1-x86_64-with-glibc2.2.5, 64bit
Linux distribution: Arch Linux (arch)
Frozen: False
Imported from /usr/lib/python3.8/site-packages/qutebrowser
Using Python from /usr/bin/python3
Qt library executable path: /usr/lib/qt/libexec, data path: /usr/share/qt
Paths:
cache: /home/kapheine/.cache/qutebrowser
config: /home/kapheine/.config/qutebrowser
data: /home/kapheine/.local/share/qutebrowser
runtime: /run/user/1000/qutebrowser
system data: /usr/share/qutebrowser
Autoconfig loaded: yes
Config.py: /home/kapheine/.config/qutebrowser/config.py has been loaded
Uptime: 0:02:56
```
**Does the bug happen if you start with `--temp-basedir`?**:
Not sure, as it only happens occasionally. I can try running for a while in that setup, if it's useful.
**Description**
I have a coredump on my system for Qutebrowser crashing in QStyleSheetStyle:
```
>>> bt
#0 __GI_raise (sig=<optimized out>) at ../sysdeps/unix/sysv/linux/raise.c:50
#1 0x00007fea0550ad70 in <signal handler called> () at /usr/lib/libc.so.6
#2 0x000056178b0d8db0 in ()
#3 0x00007fea0082bba8 in QStyleSheetStyle::styleHint(QStyle::StyleHint, QStyleOption const*, QWidget const*, QStyleHintReturn*) const (this=0x56178721e780, sh=QStyle::SH_Widget_ShareActivation, opt=0x0, w=0x7fe974013870, shret=<optimized out>) at styles/qstylesheetstyle.cpp:5342
#4 0x00007fea007a7d9a in QWidget::isActiveWindow() const (this=this@entry=0x7fe974013870) at kernel/qwidget.cpp:6678
#5 0x00007fea007a7f4f in QWidget::palette() const (this=this@entry=0x7fe974013870) at kernel/qwidget.cpp:4356
#6 0x00007fea007c0f0b in QWidget::event(QEvent*) (this=this@entry=0x7fe974013870, event=event@entry=0x7fff498c8250) at kernel/qwidget.cpp:8788
#7 0x00007fea00871cbf in QFrame::event(QEvent*) (this=this@entry=0x7fe974013870, e=e@entry=0x7fff498c8250) at widgets/qframe.cpp:550
#8 0x00007fea00874a0a in QAbstractScrollArea::event(QEvent*) (this=this@entry=0x7fe974013870, e=e@entry=0x7fff498c8250) at widgets/qabstractscrollarea.cpp:1042
#9 0x00007fea009fcc04 in QAbstractItemView::event(QEvent*) (this=0x7fe974013870, event=0x7fff498c8250) at itemviews/qabstractitemview.cpp:1674
#10 0x00007fea00f91a97 in sipQListView::event(QEvent*) (this=0x7fe974013870, a0=0x7fff498c8250) at sipQtWidgetsQListView.cpp:1189
#11 0x00007fea0077c352 in QApplicationPrivate::notify_helper(QObject*, QEvent*) (this=this@entry=0x56178648df30, receiver=receiver@entry=0x7fe974013870, e=e@entry=0x7fff498c8250) at kernel/qapplication.cpp:3685
#12 0x00007fea00785829 in QApplication::notify(QObject*, QEvent*) (this=0x5617869eab00, receiver=0x7fe974013870, e=0x7fff498c8250) at kernel/qapplication.cpp:3431
#13 0x00007fea010f8445 in sipQApplication::notify(QObject*, QEvent*) (this=0x5617869eab00, a0=0x7fe974013870, a1=0x7fff498c8250) at sipQtWidgetsQApplication.cpp:243
#14 0x00007fea03ca34f2 in QCoreApplication::notifyInternal2(QObject*, QEvent*) (receiver=0x7fe974013870, event=0x7fff498c8250) at ../../include/QtCore/../../src/corelib/kernel/qobject.h:153
#15 0x00007fea007c0c15 in QWidget::event(QEvent*) (this=<optimized out>, event=0x7fff498c8250) at kernel/qwidget.cpp:8950
#16 0x00007fea0077c352 in QApplicationPrivate::notify_helper(QObject*, QEvent*) (this=this@entry=0x56178648df30, receiver=receiver@entry=0x561786b83790, e=e@entry=0x7fff498c8250) at kernel/qapplication.cpp:3685
#17 0x00007fea00785829 in QApplication::notify(QObject*, QEvent*) (this=0x5617869eab00, receiver=0x561786b83790, e=0x7fff498c8250) at kernel/qapplication.cpp:3431
#18 0x00007fea010f8445 in sipQApplication::notify(QObject*, QEvent*) (this=0x5617869eab00, a0=0x561786b83790, a1=0x7fff498c8250) at sipQtWidgetsQApplication.cpp:243
#19 0x00007fea03ca34f2 in QCoreApplication::notifyInternal2(QObject*, QEvent*) (receiver=0x561786b83790, event=0x7fff498c8250) at ../../include/QtCore/../../src/corelib/kernel/qobject.h:153
#20 0x00007fea00783418 in QApplication::setActiveWindow(QWidget*) (act=act@entry=0x0) at kernel/qapplication.cpp:2039
#21 0x00007fea007a6518 in QWidgetPrivate::deactivateWidgetCleanup() (this=this@entry=0x5617863c2c20) at kernel/qwidget.cpp:2305
#22 0x00007fea007b3459 in QWidgetPrivate::hide_sys() (this=this@entry=0x5617863c2c20) at kernel/qwidget.cpp:8009
#23 0x00007fea007bb232 in QWidgetPrivate::hide_helper() (this=this@entry=0x5617863c2c20) at kernel/qwidget.cpp:7952
#24 0x00007fea007bff31 in QWidgetPrivate::setVisible(bool) (this=0x5617863c2c20, visible=<optimized out>) at kernel/qwidget.cpp:8147
#25 0x00007fea007bb5a8 in QWidgetPrivate::close_helper(QWidgetPrivate::CloseMode) (mode=<optimized out>, this=0x5617863c2c20) at kernel/qwidget.cpp:8262
#26 QWidgetPrivate::close_helper(QWidgetPrivate::CloseMode) (this=0x5617863c2c20, mode=<optimized out>) at kernel/qwidget.cpp:8237
#27 0x00007fea007bbbf1 in QWidget::~QWidget() (this=0x561786b83790, __in_chrg=<optimized out>) at kernel/qwidget.cpp:1476
#28 0x00007fea011301e2 in sipQWidget::~sipQWidget() (this=0x561786b83790, __in_chrg=<optimized out>) at sipQtWidgetsQWidget.cpp:355
#29 0x00007fea01130204 in sipQWidget::~sipQWidget() (this=0x561786b83790, __in_chrg=<optimized out>) at sipQtWidgetsQWidget.cpp:358
#30 0x00007fea041f21f3 in cleanup_qobject(sipSimpleWrapper*, void*) (sw=0x7fe9cc14b940, closure=0x5617869eab00) at ../../qpy/QtCore/qpycore_public_api.cpp:66
#31 0x00007fea03f562d8 in () at /usr/lib/python3.8/site-packages/PyQt5/sip.cpython-38-x86_64-linux-gnu.so
#32 0x00007fea041f2242 in pyqt5_cleanup_qobjects() () at ../../qpy/QtCore/qpycore_public_api.cpp:78
#33 0x00007fea041f1c1f in cleanup_on_exit(PyObject*, PyObject*) () at ../../qpy/QtCore/qpycore_init.cpp:37
#34 0x00007fea05251299 in () at /usr/lib/libpython3.8.so.1.0
#35 0x00007fea0524b3fd in PyObject_Call () at /usr/lib/libpython3.8.so.1.0
#36 0x00007fea05364329 in () at /usr/lib/libpython3.8.so.1.0
#37 0x00007fea0521d328 in () at /usr/lib/libpython3.8.so.1.0
#38 0x00007fea0521d36c in Py_Exit () at /usr/lib/libpython3.8.so.1.0
#39 0x00007fea0521d385 in () at /usr/lib/libpython3.8.so.1.0
#40 0x00007fea05385172 in () at /usr/lib/libpython3.8.so.1.0
#41 0x00007fea0521ead1 in PyRun_SimpleFileExFlags () at /usr/lib/libpython3.8.so.1.0
#42 0x00007fea0538681e in Py_RunMain () at /usr/lib/libpython3.8.so.1.0
#43 0x00007fea05386909 in Py_BytesMain () at /usr/lib/libpython3.8.so.1.0
#44 0x00007fea054f6023 in __libc_start_main (main=0x561785edb020 <main>, argc=2, argv=0x7fff498c89b8, init=<optimized out>, fini=<optimized out>, rtld_fini=<optimized out>, stack_end=0x7fff498c89a8) at ../csu/libc-start.c:308
#45 0x0000561785edb05e in _start ()
>>> frame 3
#3 0x00007fea0082bba8 in QStyleSheetStyle::styleHint (this=0x56178721e780, sh=QStyle::SH_Widget_ShareActivation, opt=0x0, w=0x7fe974013870, shret=<optimized out>) at styles/qstylesheetstyle.cpp:5342
warning: Source file is more recent than executable.
5342 return baseStyle()->styleHint(sh, opt, w, shret);
>>> list
5337 {
5338 RECURSION_GUARD(return baseStyle()->styleHint(sh, opt, w, shret))
5339 // Prevent endless loop if somebody use isActiveWindow property as selector.
5340 // QWidget::isActiveWindow uses this styleHint to determine if the window is active or not
5341 if (sh == SH_Widget_ShareActivation)
5342 return baseStyle()->styleHint(sh, opt, w, shret);
5343
5344 QRenderRule rule = renderRule(w, opt);
5345 QString s;
5346 switch (sh) {
```
Sounds like it COULD be related to #5187, but I'm a bit out of my element here.
**How to reproduce**
I'm not sure what I was doing at the time. I see a number of other crashes over the last month or so, but unfortunately my system already deleted them. I'll try to check again later, to see if the same crash is repeating.
|
non_process
|
crashes in qstylesheetstyle stylehint version info qutebrowser git commit backend qtwebengine chromium qt cpython pyqt sip colorama pygments yaml cssutils no attr qtwebenginewidgets yes qtwebengine qtwebkitwidgets yes pdf js usr share pdf js build pdf js sqlite qtnetwork ssl openssl apr style qfusionstyle platform plugin wayland platform linux with linux distribution arch linux arch frozen false imported from usr lib site packages qutebrowser using python from usr bin qt library executable path usr lib qt libexec data path usr share qt paths cache home kapheine cache qutebrowser config home kapheine config qutebrowser data home kapheine local share qutebrowser runtime run user qutebrowser system data usr share qutebrowser autoconfig loaded yes config py home kapheine config qutebrowser config py has been loaded uptime does the bug happen if you start with temp basedir not sure as it only happens occasionally i can try running for a while in that setup if it s useful description i have a coredump on my system for qutebrowser crashing in qstylesheetstyle bt gi raise sig at sysdeps unix sysv linux raise c in at usr lib libc so in in qstylesheetstyle stylehint qstyle stylehint qstyleoption const qwidget const qstylehintreturn const this sh qstyle sh widget shareactivation opt w shret at styles qstylesheetstyle cpp in qwidget isactivewindow const this this entry at kernel qwidget cpp in qwidget palette const this this entry at kernel qwidget cpp in qwidget event qevent this this entry event event entry at kernel qwidget cpp in qframe event qevent this this entry e e entry at widgets qframe cpp in qabstractscrollarea event qevent this this entry e e entry at widgets qabstractscrollarea cpp in qabstractitemview event qevent this event at itemviews qabstractitemview cpp in sipqlistview event qevent this at sipqtwidgetsqlistview cpp in qapplicationprivate notify helper qobject qevent this this entry receiver receiver entry e e entry at kernel qapplication cpp in qapplication notify qobject qevent this receiver e at kernel qapplication cpp in sipqapplication notify qobject qevent this at sipqtwidgetsqapplication cpp in qcoreapplication qobject qevent receiver event at include qtcore src corelib kernel qobject h in qwidget event qevent this event at kernel qwidget cpp in qapplicationprivate notify helper qobject qevent this this entry receiver receiver entry e e entry at kernel qapplication cpp in qapplication notify qobject qevent this receiver e at kernel qapplication cpp in sipqapplication notify qobject qevent this at sipqtwidgetsqapplication cpp in qcoreapplication qobject qevent receiver event at include qtcore src corelib kernel qobject h in qapplication setactivewindow qwidget act act entry at kernel qapplication cpp in qwidgetprivate deactivatewidgetcleanup this this entry at kernel qwidget cpp in qwidgetprivate hide sys this this entry at kernel qwidget cpp in qwidgetprivate hide helper this this entry at kernel qwidget cpp in qwidgetprivate setvisible bool this visible at kernel qwidget cpp in qwidgetprivate close helper qwidgetprivate closemode mode this at kernel qwidget cpp qwidgetprivate close helper qwidgetprivate closemode this mode at kernel qwidget cpp in qwidget qwidget this in chrg at kernel qwidget cpp in sipqwidget sipqwidget this in chrg at sipqtwidgetsqwidget cpp in sipqwidget sipqwidget this in chrg at sipqtwidgetsqwidget cpp in cleanup qobject sipsimplewrapper void sw closure at qpy qtcore qpycore public api cpp in at usr lib site packages sip cpython linux gnu so in cleanup qobjects at qpy qtcore qpycore public api cpp in cleanup on exit pyobject pyobject at qpy qtcore qpycore init cpp in at usr lib so in pyobject call at usr lib so in at usr lib so in at usr lib so in py exit at usr lib so in at usr lib so in at usr lib so in pyrun simplefileexflags at usr lib so in py runmain at usr lib so in py bytesmain at usr lib so in libc start main main argc argv init fini rtld fini stack end at csu libc start c in start frame in qstylesheetstyle stylehint this sh qstyle sh widget shareactivation opt w shret at styles qstylesheetstyle cpp warning source file is more recent than executable return basestyle stylehint sh opt w shret list recursion guard return basestyle stylehint sh opt w shret prevent endless loop if somebody use isactivewindow property as selector qwidget isactivewindow uses this stylehint to determine if the window is active or not if sh sh widget shareactivation return basestyle stylehint sh opt w shret qrenderrule rule renderrule w opt qstring s switch sh sounds like it could be related to but i m a bit out of my element here how to reproduce i m not sure what i was doing at the time i see a number of other crashes over the last month or so but unfortunately my system already deleted them i ll try to check again later to see if the same crash is repeating
| 0
|
346,815
| 31,025,942,689
|
IssuesEvent
|
2023-08-10 09:08:38
|
YeolJyeongKong/fittering-BE
|
https://api.github.com/repos/YeolJyeongKong/fittering-BE
|
opened
|
로그인 기능 수정
|
🐞 BugFix ✅ Test
|
### 로그인
- 입력 정보로 조회되는 유저가 없을 때 `null`을 반환해야 하는데 예외만 날리고 있어 문제 발생
+ 입력한 이메일로 조회된 유저가 있음
- 비밀번호 틀렸음 : `일치하는 유저 정보가 없습니다.`
- 비밀번호 맞음 : `200` OK
+ 입력한 이메일로 조회된 유저가 없음
- `500` : **수정 필요!!**
- OAuth로 로그인한 사용자에게 **JWT 발급**이 되는지 점검 필요
|
1.0
|
로그인 기능 수정 - ### 로그인
- 입력 정보로 조회되는 유저가 없을 때 `null`을 반환해야 하는데 예외만 날리고 있어 문제 발생
+ 입력한 이메일로 조회된 유저가 있음
- 비밀번호 틀렸음 : `일치하는 유저 정보가 없습니다.`
- 비밀번호 맞음 : `200` OK
+ 입력한 이메일로 조회된 유저가 없음
- `500` : **수정 필요!!**
- OAuth로 로그인한 사용자에게 **JWT 발급**이 되는지 점검 필요
|
non_process
|
로그인 기능 수정 로그인 입력 정보로 조회되는 유저가 없을 때 null 을 반환해야 하는데 예외만 날리고 있어 문제 발생 입력한 이메일로 조회된 유저가 있음 비밀번호 틀렸음 일치하는 유저 정보가 없습니다 비밀번호 맞음 ok 입력한 이메일로 조회된 유저가 없음 수정 필요 oauth로 로그인한 사용자에게 jwt 발급 이 되는지 점검 필요
| 0
|
13,362
| 15,825,666,566
|
IssuesEvent
|
2021-04-06 06:13:28
|
elastic/beats
|
https://api.github.com/repos/elastic/beats
|
closed
|
Create decode_xml_wineventlog processor
|
7.13 candidate :Processors Epic Team:Security-External Integrations Theme: from_splunk_wizard enhancement
|
The goal is to separate the XML processing capability of the `winlog` input from the ability to read XML event logs from Windows. This will allow event log XML to be processed from a variety of sources. One use case is to be able to process Windows event logs that have been written to an S3 bucket as XML.
As a nice side-effect we should be to make testing of Windows modules a little easier since we can use XML instead of .evtx as the source data for tests. And these tests should be able to run on non-Windows machines.
The XML processing code is currently part of a Winlogbeat package that interacts with Windows API (using cgo). So we will want to cleanly separate this code to break any dependency in the xml event log decoder package on Windows APIs.
Example config:
```
processors:
- rename:
- from: message
to: event.original
- decode_xml_wineventlog:
field: event.original
target_field: winlog
- script:
lang: javascript
id: security
file: ${path.home}/module/security/config/winlogbeat-security.js
```
|
1.0
|
Create decode_xml_wineventlog processor - The goal is to separate the XML processing capability of the `winlog` input from the ability to read XML event logs from Windows. This will allow event log XML to be processed from a variety of sources. One use case is to be able to process Windows event logs that have been written to an S3 bucket as XML.
As a nice side-effect we should be to make testing of Windows modules a little easier since we can use XML instead of .evtx as the source data for tests. And these tests should be able to run on non-Windows machines.
The XML processing code is currently part of a Winlogbeat package that interacts with Windows API (using cgo). So we will want to cleanly separate this code to break any dependency in the xml event log decoder package on Windows APIs.
Example config:
```
processors:
- rename:
- from: message
to: event.original
- decode_xml_wineventlog:
field: event.original
target_field: winlog
- script:
lang: javascript
id: security
file: ${path.home}/module/security/config/winlogbeat-security.js
```
|
process
|
create decode xml wineventlog processor the goal is to separate the xml processing capability of the winlog input from the ability to read xml event logs from windows this will allow event log xml to be processed from a variety of sources one use case is to be able to process windows event logs that have been written to an bucket as xml as a nice side effect we should be to make testing of windows modules a little easier since we can use xml instead of evtx as the source data for tests and these tests should be able to run on non windows machines the xml processing code is currently part of a winlogbeat package that interacts with windows api using cgo so we will want to cleanly separate this code to break any dependency in the xml event log decoder package on windows apis example config processors rename from message to event original decode xml wineventlog field event original target field winlog script lang javascript id security file path home module security config winlogbeat security js
| 1
|
205,014
| 7,093,473,776
|
IssuesEvent
|
2018-01-12 20:40:16
|
hexlet-codebattle/codebattle
|
https://api.github.com/repos/hexlet-codebattle/codebattle
|
closed
|
Отключить возможность ввода текста и выбор языка для наблюдателей
|
easy high_priority
|
Третьи лица могут писать в левый редактор. Запретить!
|
1.0
|
Отключить возможность ввода текста и выбор языка для наблюдателей - Третьи лица могут писать в левый редактор. Запретить!
|
non_process
|
отключить возможность ввода текста и выбор языка для наблюдателей третьи лица могут писать в левый редактор запретить
| 0
|
21,419
| 29,359,591,378
|
IssuesEvent
|
2023-05-28 00:36:48
|
devssa/onde-codar-em-salvador
|
https://api.github.com/repos/devssa/onde-codar-em-salvador
|
closed
|
[Remoto] Data Analyst na Coodesh
|
SALVADOR DATA SCIENCE PYTHON SQL ETL REQUISITOS REMOTO PROCESSOS GITHUB INGLÊS SCALA UMA POWER BI C R TABLEAU DOCUMENTAÇÃO TESTES AUTOMATIZADOS MODELAGEM DE BANCO DE DADOS MANUTENÇÃO NEGÓCIOS DATA WAREHOUSE ENGENHARIA DE SOFTWARE MONITORAMENTO Stale
|
## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/data-analyst-124220888?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A <strong>OZ Technology</strong> está em busca de<strong> </strong><strong><ins>Data Analyst</ins></strong><strong> </strong>para compor seu time!</p>
<p>A OZ Technology garante eficiência, controle da informação e aumento da lucratividade, facilitando a vida de empresas em diversos setores com profissionalismo técnico, know-how e pronto atendimento em qualquer horário. Nossa crença enxerga o crescimento e o avanço como consequência de um ambiente integrado, acessível e open-source, visando transformar isso em realidade para todos.</p>
<p>Estamos procurando uma pessoa para controlar a coleta, o processamento e a organização dos dados para melhor modelagem, governança e performance.</p>
<p>Esta pessoa vai analisar dados quantitativos e organizar dados não estruturados, transformando-os em informações relevantes que embasarão o processo de tomada de decisão. Para fazer bem este trabalho, estamos buscando um perfil que tenha pensamento analítico e seja capaz de automatizar demandas.</p>
<p>Em última análise, seu trabalho será parte integrante de nossos esforços para garantir que a empresa adote a metodologia orientada a dados em todos os seus processos decisórios. </p>
<p><strong>Responsabilidades:</strong></p>
<ul>
<li>Liderar projetos de dados desde o mapeamento de requisitos, até a modelagem do banco e disponibilização dos dados;</li>
<li>Contribuir para o design, documentação, manutenção, monitoramento e otimização de nossos códigos e ferramentas de dados;</li>
<li>Identificar oportunidades de melhorias na nossa plataforma de ETL/ELT;</li>
<li>Traduzir as necessidades de negócios em soluções alinhadas com as ferramentas e estratégias da plataforma de dados.</li>
</ul>
## OZ Technology:
<p>A OZ Technology garante eficiência, controle da informação e aumento da lucratividade, facilitando a vida de empresas em diversos setores com profissionalismo técnico, know-how e pronto atendimento em qualquer horário. Desde sua concepção, a empresa cresceu de forma constante, baseando-se em decisões voltadas à sua missão central: a satisfação total dos nossos clientes, apresentando sempre soluções de alta tecnologia, última geração, baixo custo e que colaborem para o desenvolvimento tecnológico em âmbito nacional e mundial. Nossa crença enxerga o crescimento e o avanço como consequência de um ambiente integrado, acessível e open-source, visando transformar isso em realidade para todos.</p>
</p>
## Habilidades:
- Python
- Scala
- Microsoft SQL Server
- Looker
- Data Lake
## Local:
100% Remoto
## Requisitos:
- Ser apaixonado por dados;
- Ser capaz de escrever queries em SQL para resolver problemas de alta complexidade de forma otimizada;
- Experiência em implementação e manutenção de pipelines de dados;
- Noções sobre boas práticas de engenharia de software (Ex. Documentação, testes automatizados, código limpo, etc);
- Experiência em programação com linguagens de propósito genérico (Ex. Python, R, Scala, etc.);
- Conhecimentos sobre diferentes técnicas de arquitetura e modelagem de Banco de Dados para fins analíticos (Ex. Data Mart, Data Lake, Data Warehouse, etc.);
- Capacidade de abstrair conceitos de negócios em métricas quantitativas (Ex. KPIs, OKRs, etc);
- Boas habilidades de comunicação para lidar com as partes interessadas de áreas técnicas e áreas de negócios;
- Conhecimento em ferramentas de visualização de dados, como Looker, Power BI,Tableau, etc;
- Inglês Avançado;
- Atuar com Looker.
## Benefícios:
- Welcome Kit + Homeoffice Kit;
- Gympass (academia);
- Wellness Coach (mente, corpo e sono);
- Cíngulo (saúde mental);
- Assistência Médica;
- Assistência Odontológica;
- Seguro de Vida;
- Custeio de capacitações e certificações (conforme política interna).
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Data Analyst na OZ Technology](https://coodesh.com/vagas/data-analyst-124220888?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Regime
CLT
#### Categoria
Data Science
|
1.0
|
[Remoto] Data Analyst na Coodesh - ## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/data-analyst-124220888?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A <strong>OZ Technology</strong> está em busca de<strong> </strong><strong><ins>Data Analyst</ins></strong><strong> </strong>para compor seu time!</p>
<p>A OZ Technology garante eficiência, controle da informação e aumento da lucratividade, facilitando a vida de empresas em diversos setores com profissionalismo técnico, know-how e pronto atendimento em qualquer horário. Nossa crença enxerga o crescimento e o avanço como consequência de um ambiente integrado, acessível e open-source, visando transformar isso em realidade para todos.</p>
<p>Estamos procurando uma pessoa para controlar a coleta, o processamento e a organização dos dados para melhor modelagem, governança e performance.</p>
<p>Esta pessoa vai analisar dados quantitativos e organizar dados não estruturados, transformando-os em informações relevantes que embasarão o processo de tomada de decisão. Para fazer bem este trabalho, estamos buscando um perfil que tenha pensamento analítico e seja capaz de automatizar demandas.</p>
<p>Em última análise, seu trabalho será parte integrante de nossos esforços para garantir que a empresa adote a metodologia orientada a dados em todos os seus processos decisórios. </p>
<p><strong>Responsabilidades:</strong></p>
<ul>
<li>Liderar projetos de dados desde o mapeamento de requisitos, até a modelagem do banco e disponibilização dos dados;</li>
<li>Contribuir para o design, documentação, manutenção, monitoramento e otimização de nossos códigos e ferramentas de dados;</li>
<li>Identificar oportunidades de melhorias na nossa plataforma de ETL/ELT;</li>
<li>Traduzir as necessidades de negócios em soluções alinhadas com as ferramentas e estratégias da plataforma de dados.</li>
</ul>
## OZ Technology:
<p>A OZ Technology garante eficiência, controle da informação e aumento da lucratividade, facilitando a vida de empresas em diversos setores com profissionalismo técnico, know-how e pronto atendimento em qualquer horário. Desde sua concepção, a empresa cresceu de forma constante, baseando-se em decisões voltadas à sua missão central: a satisfação total dos nossos clientes, apresentando sempre soluções de alta tecnologia, última geração, baixo custo e que colaborem para o desenvolvimento tecnológico em âmbito nacional e mundial. Nossa crença enxerga o crescimento e o avanço como consequência de um ambiente integrado, acessível e open-source, visando transformar isso em realidade para todos.</p>
</p>
## Habilidades:
- Python
- Scala
- Microsoft SQL Server
- Looker
- Data Lake
## Local:
100% Remoto
## Requisitos:
- Ser apaixonado por dados;
- Ser capaz de escrever queries em SQL para resolver problemas de alta complexidade de forma otimizada;
- Experiência em implementação e manutenção de pipelines de dados;
- Noções sobre boas práticas de engenharia de software (Ex. Documentação, testes automatizados, código limpo, etc);
- Experiência em programação com linguagens de propósito genérico (Ex. Python, R, Scala, etc.);
- Conhecimentos sobre diferentes técnicas de arquitetura e modelagem de Banco de Dados para fins analíticos (Ex. Data Mart, Data Lake, Data Warehouse, etc.);
- Capacidade de abstrair conceitos de negócios em métricas quantitativas (Ex. KPIs, OKRs, etc);
- Boas habilidades de comunicação para lidar com as partes interessadas de áreas técnicas e áreas de negócios;
- Conhecimento em ferramentas de visualização de dados, como Looker, Power BI,Tableau, etc;
- Inglês Avançado;
- Atuar com Looker.
## Benefícios:
- Welcome Kit + Homeoffice Kit;
- Gympass (academia);
- Wellness Coach (mente, corpo e sono);
- Cíngulo (saúde mental);
- Assistência Médica;
- Assistência Odontológica;
- Seguro de Vida;
- Custeio de capacitações e certificações (conforme política interna).
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Data Analyst na OZ Technology](https://coodesh.com/vagas/data-analyst-124220888?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Regime
CLT
#### Categoria
Data Science
|
process
|
data analyst na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a oz technology está em busca de data analyst para compor seu time a oz technology garante eficiência controle da informação e aumento da lucratividade facilitando a vida de empresas em diversos setores com profissionalismo técnico know how e pronto atendimento em qualquer horário nossa crença enxerga o crescimento e o avanço como consequência de um ambiente integrado acessível e open source visando transformar isso em realidade para todos estamos procurando uma pessoa para controlar a coleta o processamento e a organização dos dados para melhor modelagem governança e performance esta pessoa vai analisar dados quantitativos e organizar dados não estruturados transformando os em informações relevantes que embasarão o processo de tomada de decisão para fazer bem este trabalho estamos buscando um perfil que tenha pensamento analítico e seja capaz de automatizar demandas em última análise seu trabalho será parte integrante de nossos esforços para garantir que a empresa adote a metodologia orientada a dados em todos os seus processos decisórios nbsp responsabilidades liderar projetos de dados desde o mapeamento de requisitos até a modelagem do banco e disponibilização dos dados contribuir para o design documentação manutenção monitoramento e otimização de nossos códigos e ferramentas de dados identificar oportunidades de melhorias na nossa plataforma de etl elt traduzir as necessidades de negócios em soluções alinhadas com as ferramentas e estratégias da plataforma de dados oz technology a oz technology garante eficiência controle da informação e aumento da lucratividade facilitando a vida de empresas em diversos setores com profissionalismo técnico know how e pronto atendimento em qualquer horário desde sua concepção a empresa cresceu de forma constante baseando se em decisões voltadas à sua missão central a satisfação total dos nossos clientes apresentando sempre soluções de alta tecnologia última geração baixo custo e que colaborem para o desenvolvimento tecnológico em âmbito nacional e mundial nossa crença enxerga o crescimento e o avanço como consequência de um ambiente integrado acessível e open source visando transformar isso em realidade para todos habilidades python scala microsoft sql server looker data lake local remoto requisitos ser apaixonado por dados ser capaz de escrever queries em sql para resolver problemas de alta complexidade de forma otimizada experiência em implementação e manutenção de pipelines de dados noções sobre boas práticas de engenharia de software ex documentação testes automatizados código limpo etc experiência em programação com linguagens de propósito genérico ex python r scala etc conhecimentos sobre diferentes técnicas de arquitetura e modelagem de banco de dados para fins analíticos ex data mart data lake data warehouse etc capacidade de abstrair conceitos de negócios em métricas quantitativas ex kpis okrs etc boas habilidades de comunicação para lidar com as partes interessadas de áreas técnicas e áreas de negócios conhecimento em ferramentas de visualização de dados como looker power bi tableau etc inglês avançado atuar com looker benefícios welcome kit homeoffice kit gympass academia wellness coach mente corpo e sono cíngulo saúde mental assistência médica assistência odontológica seguro de vida custeio de capacitações e certificações conforme política interna como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto regime clt categoria data science
| 1
|
26,374
| 2,684,343,495
|
IssuesEvent
|
2015-03-28 21:57:47
|
ConEmu/old-issues
|
https://api.github.com/repos/ConEmu/old-issues
|
opened
|
Clink was there before ConEmu
|
2–5 stars bug imported Priority-Medium
|
_From [michfi...@gmail.com](https://code.google.com/u/116044569754157729756/) on April 25, 2013 03:23:28_
OS version: Win8 x64 ConEmu version: latest *Bug description* Looks like ConEmu , probably at first start is overwriting clink's hook or something. If you have clink installed before ConEmu , it will stop working later... *Steps to reproduction* 1. install clink
2. install conemu 3. Pleae check if you can.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=1044_
|
1.0
|
Clink was there before ConEmu - _From [michfi...@gmail.com](https://code.google.com/u/116044569754157729756/) on April 25, 2013 03:23:28_
OS version: Win8 x64 ConEmu version: latest *Bug description* Looks like ConEmu , probably at first start is overwriting clink's hook or something. If you have clink installed before ConEmu , it will stop working later... *Steps to reproduction* 1. install clink
2. install conemu 3. Pleae check if you can.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=1044_
|
non_process
|
clink was there before conemu from on april os version conemu version latest bug description looks like conemu probably at first start is overwriting clink s hook or something if you have clink installed before conemu it will stop working later steps to reproduction install clink install conemu pleae check if you can original issue
| 0
|
28,581
| 11,661,735,553
|
IssuesEvent
|
2020-03-03 07:34:31
|
bas/WebGoat
|
https://api.github.com/repos/bas/WebGoat
|
opened
|
CVE-2018-19839 (Medium) detected in CSS::Sass-v3.4.11
|
security vulnerability
|
## CVE-2018-19839 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>CSS::Sassv3.4.11</b></p></summary>
<p>
<p>Library home page: <a href=https://metacpan.org/pod/CSS::Sass>https://metacpan.org/pod/CSS::Sass</a></p>
<p>Found in HEAD commit: <a href="https://github.com/bas/WebGoat/commit/e8edf255a07313048bcf3962743e41087644cb32">e8edf255a07313048bcf3962743e41087644cb32</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (60)</summary>
<p></p>
<p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p>
<p>
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/color_maps.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/sass_util.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/utf8/unchecked.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/output.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/b64/cencode.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/source_map.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/lexer.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/utf8.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/test/test_node.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/utf8_string.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/plugins.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/node.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/json.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/include/sass/base.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/environment.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/position.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/extend.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/subset_map.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/remove_placeholders.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/sass_context.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/sass.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/ast_fwd_decl.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/contrib/plugin.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/utf8/core.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/include/sass/functions.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/test/test_superselector.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/sass_functions.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/utf8_string.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/node.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/subset_map.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/base64vlq.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/listize.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/c99func.c
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/position.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/remove_placeholders.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/test/test_subset_map.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/sass_functions.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/sass2scss.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/memory/SharedPtr.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/paths.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/include/sass/context.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/color_maps.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/test/test_unification.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/sass_util.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/source_map.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/script/test-leaks.pl
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/lexer.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/memory/SharedPtr.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/json.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/to_c.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/units.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/units.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/b64/encode.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/file.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/environment.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/utf8/checked.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/listize.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/plugins.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/include/sass2scss.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/debug.hpp
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass prior to 3.5.5, the function handle_error in sass_context.cpp allows attackers to cause a denial-of-service resulting from a heap-based buffer over-read via a crafted sass file.
<p>Publish Date: 2018-12-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19839>CVE-2018-19839</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19839">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19839</a></p>
<p>Fix Resolution: 3.5.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-19839 (Medium) detected in CSS::Sass-v3.4.11 - ## CVE-2018-19839 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>CSS::Sassv3.4.11</b></p></summary>
<p>
<p>Library home page: <a href=https://metacpan.org/pod/CSS::Sass>https://metacpan.org/pod/CSS::Sass</a></p>
<p>Found in HEAD commit: <a href="https://github.com/bas/WebGoat/commit/e8edf255a07313048bcf3962743e41087644cb32">e8edf255a07313048bcf3962743e41087644cb32</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (60)</summary>
<p></p>
<p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p>
<p>
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/color_maps.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/sass_util.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/utf8/unchecked.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/output.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/b64/cencode.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/source_map.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/lexer.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/utf8.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/test/test_node.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/utf8_string.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/plugins.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/node.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/json.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/include/sass/base.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/environment.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/position.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/extend.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/subset_map.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/remove_placeholders.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/sass_context.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/sass.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/ast_fwd_decl.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/contrib/plugin.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/utf8/core.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/include/sass/functions.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/test/test_superselector.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/sass_functions.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/utf8_string.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/node.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/subset_map.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/base64vlq.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/listize.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/c99func.c
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/position.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/remove_placeholders.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/test/test_subset_map.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/sass_functions.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/sass2scss.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/memory/SharedPtr.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/paths.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/include/sass/context.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/color_maps.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/test/test_unification.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/sass_util.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/source_map.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/script/test-leaks.pl
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/lexer.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/memory/SharedPtr.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/json.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/to_c.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/units.cpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/units.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/b64/encode.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/file.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/environment.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/utf8/checked.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/listize.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/plugins.hpp
- /WebGoat/docs/node_modules/node-sass/src/libsass/include/sass2scss.h
- /WebGoat/docs/node_modules/node-sass/src/libsass/src/debug.hpp
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass prior to 3.5.5, the function handle_error in sass_context.cpp allows attackers to cause a denial-of-service resulting from a heap-based buffer over-read via a crafted sass file.
<p>Publish Date: 2018-12-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19839>CVE-2018-19839</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19839">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19839</a></p>
<p>Fix Resolution: 3.5.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in css sass cve medium severity vulnerability vulnerable library css library home page a href found in head commit a href library source files the source files were matched to this source library based on a best effort match source libraries are selected from a list of probable public libraries webgoat docs node modules node sass src libsass src color maps cpp webgoat docs node modules node sass src libsass src sass util hpp webgoat docs node modules node sass src libsass src unchecked h webgoat docs node modules node sass src libsass src output hpp webgoat docs node modules node sass src libsass src cencode h webgoat docs node modules node sass src libsass src source map cpp webgoat docs node modules node sass src libsass src lexer cpp webgoat docs node modules node sass src libsass src h webgoat docs node modules node sass src libsass test test node cpp webgoat docs node modules node sass src libsass src string cpp webgoat docs node modules node sass src libsass src plugins cpp webgoat docs node modules node sass src libsass src node hpp webgoat docs node modules node sass src libsass src json hpp webgoat docs node modules node sass src libsass include sass base h webgoat docs node modules node sass src libsass src environment cpp webgoat docs node modules node sass src libsass src position hpp webgoat docs node modules node sass src libsass src extend hpp webgoat docs node modules node sass src libsass src subset map hpp webgoat docs node modules node sass src libsass src remove placeholders cpp webgoat docs node modules node sass src libsass src sass context hpp webgoat docs node modules node sass src libsass src sass hpp webgoat docs node modules node sass src libsass src ast fwd decl cpp webgoat docs node modules node sass src libsass contrib plugin cpp webgoat docs node modules node sass src libsass src core h webgoat docs node modules node sass src libsass include sass functions h webgoat docs node modules node sass src libsass test test superselector cpp webgoat docs node modules node sass src libsass src sass functions cpp webgoat docs node modules node sass src libsass src string hpp webgoat docs node modules node sass src libsass src node cpp webgoat docs node modules node sass src libsass src subset map cpp webgoat docs node modules node sass src libsass src cpp webgoat docs node modules node sass src libsass src listize cpp webgoat docs node modules node sass src libsass src c webgoat docs node modules node sass src libsass src position cpp webgoat docs node modules node sass src libsass src remove placeholders hpp webgoat docs node modules node sass src libsass test test subset map cpp webgoat docs node modules node sass src libsass src sass functions hpp webgoat docs node modules node sass src libsass src cpp webgoat docs node modules node sass src libsass src memory sharedptr cpp webgoat docs node modules node sass src libsass src paths hpp webgoat docs node modules node sass src libsass include sass context h webgoat docs node modules node sass src libsass src color maps hpp webgoat docs node modules node sass src libsass test test unification cpp webgoat docs node modules node sass src libsass src sass util cpp webgoat docs node modules node sass src libsass src source map hpp webgoat docs node modules node sass src libsass script test leaks pl webgoat docs node modules node sass src libsass src lexer hpp webgoat docs node modules node sass src libsass src memory sharedptr hpp webgoat docs node modules node sass src libsass src json cpp webgoat docs node modules node sass src libsass src to c hpp webgoat docs node modules node sass src libsass src units cpp webgoat docs node modules node sass src libsass src units hpp webgoat docs node modules node sass src libsass src encode h webgoat docs node modules node sass src libsass src file hpp webgoat docs node modules node sass src libsass src environment hpp webgoat docs node modules node sass src libsass src checked h webgoat docs node modules node sass src libsass src listize hpp webgoat docs node modules node sass src libsass src plugins hpp webgoat docs node modules node sass src libsass include h webgoat docs node modules node sass src libsass src debug hpp vulnerability details in libsass prior to the function handle error in sass context cpp allows attackers to cause a denial of service resulting from a heap based buffer over read via a crafted sass file publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href fix resolution step up your open source security game with whitesource
| 0
|
654,768
| 21,662,200,083
|
IssuesEvent
|
2022-05-06 20:40:02
|
orbeon/orbeon-forms
|
https://api.github.com/repos/orbeon/orbeon-forms
|
closed
|
During Push to Remote, `Authorization` header not sent in PUT, resulting in a 403
|
Module: Form Runner Priority: Regression
|
As #5310 and #5314, this also regressed when we switched from using an `xf:submission` to directly calling Scala API.
|
1.0
|
During Push to Remote, `Authorization` header not sent in PUT, resulting in a 403 - As #5310 and #5314, this also regressed when we switched from using an `xf:submission` to directly calling Scala API.
|
non_process
|
during push to remote authorization header not sent in put resulting in a as and this also regressed when we switched from using an xf submission to directly calling scala api
| 0
|
23,845
| 12,134,136,383
|
IssuesEvent
|
2020-04-23 10:11:43
|
Spottedleaf/Tuinity
|
https://api.github.com/repos/Spottedleaf/Tuinity
|
closed
|
Console Error
|
confirmed performance upstream vanilla
|
Minecraft: 1.15.2
Tuinity: build 13
Console Error followed by huge TPS lag.
https://pastebin.com/raw/wVfU3zQb
I have this problem in only one server the other one i have is fine :/
|
True
|
Console Error - Minecraft: 1.15.2
Tuinity: build 13
Console Error followed by huge TPS lag.
https://pastebin.com/raw/wVfU3zQb
I have this problem in only one server the other one i have is fine :/
|
non_process
|
console error minecraft tuinity build console error followed by huge tps lag i have this problem in only one server the other one i have is fine
| 0
|
109,972
| 23,850,189,645
|
IssuesEvent
|
2022-09-06 17:10:11
|
creativecommons/wp-plugin-creativecommons
|
https://api.github.com/repos/creativecommons/wp-plugin-creativecommons
|
opened
|
[Feature] migrate from `create-guten-block` scripts to a `@wordpress/scripts` core setup
|
🟩 priority: low 🚦 status: awaiting triage ✨ goal: improvement 💻 aspect: code
|
## Problem
Previously: #189
The use of `create-guten-block` as a scaffolding for the WordPress Gutenberg blocks has a number of compatibility and stability issues. Since the blocks were originally created with `create-guten-block` core WP has grown its capacity to handle scaffolding, dev, and build. To keep block dev stable long term, `create-guten-block` needs to be replaced entirely with core functionality.
## Description
The entirety of `create-guten-block` needs to be removed, along with all its dependencies. Its implementation of block structure and file setup is not quite 1:1 compared to the `@wordpress/scripts`, which will require modification.
Additionally, the CGB setup has its own "cgb" namespace, which is not as intuitive or accurate post deprecation. The namespace should likely shift to "cc" or something similar.
As the namespace will be shifting, that will cause CSS Rules, HTML classes, and JS naming conventions to shift as well This means that any site utilizing a CGB built block will need a proper migration path to using the core scripts implementation with the new namespace and naming convention.
## Implementation
<!-- Replace the [ ] with [x] to check the box. -->
- [x] I would be interested in implementing this feature.
|
1.0
|
[Feature] migrate from `create-guten-block` scripts to a `@wordpress/scripts` core setup - ## Problem
Previously: #189
The use of `create-guten-block` as a scaffolding for the WordPress Gutenberg blocks has a number of compatibility and stability issues. Since the blocks were originally created with `create-guten-block` core WP has grown its capacity to handle scaffolding, dev, and build. To keep block dev stable long term, `create-guten-block` needs to be replaced entirely with core functionality.
## Description
The entirety of `create-guten-block` needs to be removed, along with all its dependencies. Its implementation of block structure and file setup is not quite 1:1 compared to the `@wordpress/scripts`, which will require modification.
Additionally, the CGB setup has its own "cgb" namespace, which is not as intuitive or accurate post deprecation. The namespace should likely shift to "cc" or something similar.
As the namespace will be shifting, that will cause CSS Rules, HTML classes, and JS naming conventions to shift as well This means that any site utilizing a CGB built block will need a proper migration path to using the core scripts implementation with the new namespace and naming convention.
## Implementation
<!-- Replace the [ ] with [x] to check the box. -->
- [x] I would be interested in implementing this feature.
|
non_process
|
migrate from create guten block scripts to a wordpress scripts core setup problem previously the use of create guten block as a scaffolding for the wordpress gutenberg blocks has a number of compatibility and stability issues since the blocks were originally created with create guten block core wp has grown its capacity to handle scaffolding dev and build to keep block dev stable long term create guten block needs to be replaced entirely with core functionality description the entirety of create guten block needs to be removed along with all its dependencies its implementation of block structure and file setup is not quite compared to the wordpress scripts which will require modification additionally the cgb setup has its own cgb namespace which is not as intuitive or accurate post deprecation the namespace should likely shift to cc or something similar as the namespace will be shifting that will cause css rules html classes and js naming conventions to shift as well this means that any site utilizing a cgb built block will need a proper migration path to using the core scripts implementation with the new namespace and naming convention implementation i would be interested in implementing this feature
| 0
|
125,441
| 17,836,185,198
|
IssuesEvent
|
2021-09-03 01:38:50
|
michaeldotson/todoListApi
|
https://api.github.com/repos/michaeldotson/todoListApi
|
opened
|
CVE-2021-23438 (Medium) detected in mpath-0.5.1.tgz
|
security vulnerability
|
## CVE-2021-23438 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mpath-0.5.1.tgz</b></p></summary>
<p>{G,S}et object values using MongoDB-like path notation</p>
<p>Library home page: <a href="https://registry.npmjs.org/mpath/-/mpath-0.5.1.tgz">https://registry.npmjs.org/mpath/-/mpath-0.5.1.tgz</a></p>
<p>Path to dependency file: /todoListApi/package.json</p>
<p>Path to vulnerable library: todoListApi/node_modules/mpath/package.json</p>
<p>
Dependency Hierarchy:
- mongoose-5.4.17.tgz (Root Library)
- :x: **mpath-0.5.1.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package mpath before 0.8.4. A type confusion vulnerability can lead to a bypass of CVE-2018-16490. In particular, the condition ignoreProperties.indexOf(parts[i]) !== -1 returns -1 if parts[i] is ['__proto__']. This is because the method that has been called if the input is an array is Array.prototype.indexOf() and not String.prototype.indexOf(). They behave differently depending on the type of the input.
<p>Publish Date: 2021-09-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23438>CVE-2021-23438</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23438">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23438</a></p>
<p>Release Date: 2021-09-01</p>
<p>Fix Resolution: mpath - 0.8.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-23438 (Medium) detected in mpath-0.5.1.tgz - ## CVE-2021-23438 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mpath-0.5.1.tgz</b></p></summary>
<p>{G,S}et object values using MongoDB-like path notation</p>
<p>Library home page: <a href="https://registry.npmjs.org/mpath/-/mpath-0.5.1.tgz">https://registry.npmjs.org/mpath/-/mpath-0.5.1.tgz</a></p>
<p>Path to dependency file: /todoListApi/package.json</p>
<p>Path to vulnerable library: todoListApi/node_modules/mpath/package.json</p>
<p>
Dependency Hierarchy:
- mongoose-5.4.17.tgz (Root Library)
- :x: **mpath-0.5.1.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package mpath before 0.8.4. A type confusion vulnerability can lead to a bypass of CVE-2018-16490. In particular, the condition ignoreProperties.indexOf(parts[i]) !== -1 returns -1 if parts[i] is ['__proto__']. This is because the method that has been called if the input is an array is Array.prototype.indexOf() and not String.prototype.indexOf(). They behave differently depending on the type of the input.
<p>Publish Date: 2021-09-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23438>CVE-2021-23438</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23438">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23438</a></p>
<p>Release Date: 2021-09-01</p>
<p>Fix Resolution: mpath - 0.8.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in mpath tgz cve medium severity vulnerability vulnerable library mpath tgz g s et object values using mongodb like path notation library home page a href path to dependency file todolistapi package json path to vulnerable library todolistapi node modules mpath package json dependency hierarchy mongoose tgz root library x mpath tgz vulnerable library vulnerability details this affects the package mpath before a type confusion vulnerability can lead to a bypass of cve in particular the condition ignoreproperties indexof parts returns if parts is this is because the method that has been called if the input is an array is array prototype indexof and not string prototype indexof they behave differently depending on the type of the input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution mpath step up your open source security game with whitesource
| 0
|
5,710
| 8,565,138,038
|
IssuesEvent
|
2018-11-09 18:54:24
|
easy-software-ufal/annotations_repos
|
https://api.github.com/repos/easy-software-ufal/annotations_repos
|
opened
|
Metadata autofac/Autofac resolved with missing property values even when [DefaultValue] attributes are not present.
|
C# RMA wrong processing
|
Issue: `https://github.com/autofac/Autofac/issues/231`
PR: `null`
|
1.0
|
Metadata autofac/Autofac resolved with missing property values even when [DefaultValue] attributes are not present. - Issue: `https://github.com/autofac/Autofac/issues/231`
PR: `null`
|
process
|
metadata autofac autofac resolved with missing property values even when attributes are not present issue pr null
| 1
|
21,539
| 29,837,211,933
|
IssuesEvent
|
2023-06-19 00:37:46
|
devssa/onde-codar-em-salvador
|
https://api.github.com/repos/devssa/onde-codar-em-salvador
|
closed
|
[Remoto] Fullstack Developer na Coodesh
|
SALVADOR BACK-END FRONT-END PJ MONGODB JAVASCRIPT FULL-STACK MVC HTML SCRUM SQL GIT REST DOCKER KUBERNETES DEVOPS JSON ANGULAR REACT AWS REQUISITOS REMOTO ASP.NET PROCESSOS INOVAÇÃO GITHUB KANBAN FLUTTER AZURE E-COMMERCE IONIC UMA C R APIs METODOLOGIAS ÁGEIS SAAS MANUTENÇÃO AUTOMAÇÃO DE PROCESSOS EKS Stale
|
## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/fullstack-developer-181905126?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A Tecnologia Única está em busca de Fullstack Developer .Net para compor seu time!Venha fazer parte de uma empresa que acredita e aposta em novas ideias, com um time forte, e com propósito que já soma mais de 200 pessoas. Somos uma empresa de desenvolvimento de sistemas fundada em 2004. Nossas principais soluções são para o mercado segurador e de fidelização para diversos segmentos, desde o varejo até o agronegócio. Também mantemos um ecossistema de startups, nosso pilar de inovação. Parceira de grandes players do mercado de tecnologia, entre eles Microsoft e AWS, temos a missão de impactar a vida das pessoas provendo soluções disruptivas, investindo massivamente em nosso time, sempre perseguindo e incentivando processos criativos. Sentiu a energia única e se identificou com o nosso propósito?</p>
<p>Responsabilidades:</p>
<ul>
<li>Desenvolvimento e manutenção de aplicações web;</li>
<li>Desenvolvimento de funcionalidades no Front-End;</li>
<li>Desenvolvimento de funcionalidades no Back-End.</li>
</ul>
<p>Local de trabalho: 100% Home Office OU híbrido em São Paulo/SP.</p>
<p>Aqui na única você pode ser quem é verdadeiramente, pois temos como premissa respeitar e valorizar as diferenças. Somos diversos e buscamos criar um ambiente cada vez mais inclusivo.</p>
## Tecnologia Única:
<p>Nascemos em 2004 com o propósito de trazer ao mercado soluções analíticas voltadas para o atendimento ao cliente do mercado segurador. Prestamos diversos serviços na área de fidelização. Desde o varejo até o agronegócio, incluindo plataformas de e-commerce cross indústria, bem como serviços de tecnologia como integração, automação de processos, construção de sistemas especializados. Inovação é um de nossos principais pilares e contamos com um ambiente para incubação de startups. Com mais de 150 colaboradores e faturamento de mais de R$20M, a maioria de nossas soluções é comercializada na modalidade SaaS e temos como principais parceiros de tecnologia a Microsoft e a AWS.</p>
</p>
## Habilidades:
- .NET
- .NET Core
- .NET Framework
- GIT
## Local:
100% Remoto
## Requisitos:
- Formação em Tecnologia;
- Conhecimento em metodologias ágeis (Scrum/Kanban);
- Experiência em .NET, C#, ASP.NET, APIS REST/JSON;
- Experiência com plataforma Microsoft SQL;
- Conhecimento de JavaScript, CSS, HTML para criação de Front-ends Web;
- Conhecimento em Arquitetura MVC;
- Experiência com Angular ou React;
- GIT.
## Diferenciais:
- MongoDB;
- Azure DevOps (criação/manutenção pipelines);
- AWS (Atlas DB, Lambda, SQS, EKS);
- Kubernetes;
- Docker;
- Ionic ou Flutter.
## Benefícios:
- Auxílios alimentação e saúde;
- Gympass;
- Apoio psicológico (Psicologia Viva);
- Day Off Remunerado no Aniversário;
- Dias de Descanso Remunerado;
- Bônus para Certificação.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack Developer na Tecnologia Única](https://coodesh.com/jobs/fullstack-developer-181905126?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Regime
PJ
#### Categoria
Full-Stack
|
2.0
|
[Remoto] Fullstack Developer na Coodesh - ## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/fullstack-developer-181905126?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A Tecnologia Única está em busca de Fullstack Developer .Net para compor seu time!Venha fazer parte de uma empresa que acredita e aposta em novas ideias, com um time forte, e com propósito que já soma mais de 200 pessoas. Somos uma empresa de desenvolvimento de sistemas fundada em 2004. Nossas principais soluções são para o mercado segurador e de fidelização para diversos segmentos, desde o varejo até o agronegócio. Também mantemos um ecossistema de startups, nosso pilar de inovação. Parceira de grandes players do mercado de tecnologia, entre eles Microsoft e AWS, temos a missão de impactar a vida das pessoas provendo soluções disruptivas, investindo massivamente em nosso time, sempre perseguindo e incentivando processos criativos. Sentiu a energia única e se identificou com o nosso propósito?</p>
<p>Responsabilidades:</p>
<ul>
<li>Desenvolvimento e manutenção de aplicações web;</li>
<li>Desenvolvimento de funcionalidades no Front-End;</li>
<li>Desenvolvimento de funcionalidades no Back-End.</li>
</ul>
<p>Local de trabalho: 100% Home Office OU híbrido em São Paulo/SP.</p>
<p>Aqui na única você pode ser quem é verdadeiramente, pois temos como premissa respeitar e valorizar as diferenças. Somos diversos e buscamos criar um ambiente cada vez mais inclusivo.</p>
## Tecnologia Única:
<p>Nascemos em 2004 com o propósito de trazer ao mercado soluções analíticas voltadas para o atendimento ao cliente do mercado segurador. Prestamos diversos serviços na área de fidelização. Desde o varejo até o agronegócio, incluindo plataformas de e-commerce cross indústria, bem como serviços de tecnologia como integração, automação de processos, construção de sistemas especializados. Inovação é um de nossos principais pilares e contamos com um ambiente para incubação de startups. Com mais de 150 colaboradores e faturamento de mais de R$20M, a maioria de nossas soluções é comercializada na modalidade SaaS e temos como principais parceiros de tecnologia a Microsoft e a AWS.</p>
</p>
## Habilidades:
- .NET
- .NET Core
- .NET Framework
- GIT
## Local:
100% Remoto
## Requisitos:
- Formação em Tecnologia;
- Conhecimento em metodologias ágeis (Scrum/Kanban);
- Experiência em .NET, C#, ASP.NET, APIS REST/JSON;
- Experiência com plataforma Microsoft SQL;
- Conhecimento de JavaScript, CSS, HTML para criação de Front-ends Web;
- Conhecimento em Arquitetura MVC;
- Experiência com Angular ou React;
- GIT.
## Diferenciais:
- MongoDB;
- Azure DevOps (criação/manutenção pipelines);
- AWS (Atlas DB, Lambda, SQS, EKS);
- Kubernetes;
- Docker;
- Ionic ou Flutter.
## Benefícios:
- Auxílios alimentação e saúde;
- Gympass;
- Apoio psicológico (Psicologia Viva);
- Day Off Remunerado no Aniversário;
- Dias de Descanso Remunerado;
- Bônus para Certificação.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack Developer na Tecnologia Única](https://coodesh.com/jobs/fullstack-developer-181905126?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Regime
PJ
#### Categoria
Full-Stack
|
process
|
fullstack developer na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a tecnologia única está em busca de fullstack developer net para compor seu time venha fazer parte de uma empresa que acredita e aposta em novas ideias com um time forte e com propósito que já soma mais de pessoas somos uma empresa de desenvolvimento de sistemas fundada em nossas principais soluções são para o mercado segurador e de fidelização para diversos segmentos desde o varejo até o agronegócio também mantemos um ecossistema de startups nosso pilar de inovação parceira de grandes players do mercado de tecnologia entre eles microsoft e aws temos a missão de impactar a vida das pessoas provendo soluções disruptivas investindo massivamente em nosso time sempre perseguindo e incentivando processos criativos sentiu a energia única e se identificou com o nosso propósito responsabilidades desenvolvimento e manutenção de aplicações web desenvolvimento de funcionalidades no front end desenvolvimento de funcionalidades no back end local de trabalho home office ou híbrido em são paulo sp aqui na única você pode ser quem é verdadeiramente pois temos como premissa respeitar e valorizar as diferenças somos diversos e buscamos criar um ambiente cada vez mais inclusivo tecnologia única nascemos em com o propósito de trazer ao mercado soluções analíticas voltadas para o atendimento ao cliente do mercado segurador prestamos diversos serviços na área de fidelização desde o varejo até o agronegócio incluindo plataformas de e commerce cross indústria bem como serviços de tecnologia como integração automação de processos construção de sistemas especializados inovação é um de nossos principais pilares e contamos com um ambiente para incubação de startups com mais de colaboradores e faturamento de mais de r a maioria de nossas soluções é comercializada na modalidade saas e temos como principais parceiros de tecnologia a microsoft e a aws habilidades net net core net framework git local remoto requisitos formação em tecnologia conhecimento em metodologias ágeis scrum kanban experiência em net c asp net apis rest json experiência com plataforma microsoft sql conhecimento de javascript css html para criação de front ends web conhecimento em arquitetura mvc experiência com angular ou react git diferenciais mongodb azure devops criação manutenção pipelines aws atlas db lambda sqs eks kubernetes docker ionic ou flutter benefícios auxílios alimentação e saúde gympass apoio psicológico psicologia viva day off remunerado no aniversário dias de descanso remunerado bônus para certificação como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto regime pj categoria full stack
| 1
|
121,810
| 12,134,424,466
|
IssuesEvent
|
2020-04-23 10:41:49
|
handsontable/handsontable
|
https://api.github.com/repos/handsontable/handsontable
|
opened
|
skipColumnOnPaste and skipRowOnPaste when added to table settings hang the browser down after pasting
|
Copy-paste Type: Bug Type: Documentation needed
|
### Description
<!--- Tell us what happens and what should happen -->
Options [`skipColumnOnPaste`](https://handsontable.com/docs/7.4.2/Options.html#skipColumnOnPaste) and [`skipRowOnPaste`](https://handsontable.com/docs/7.4.2/Options.html#skipRowOnPaste) when added to table settings hang the browser down after pasting.
It is said that
> When added to a column property, it skips the column on paste and pastes the data on the next column to the right.
But I guess that we should also mention that it is not supported for table level settings.
### Steps to reproduce
<!--- Provide steps to reproduce this issue -->
1. Go to https://jsfiddle.net/AMBudnik/4d3sjgn2/
2. Select any cell
3. Use <kbd>CTRL/CMD</kbd>+<kbd>C</kbd> to copy
4. Use <kbd>CTRL/CMD</kbd>+<kbd>V</kbd> to paste
**Result**
The browser does not respond.
### Demo
<!--- Provide a link to a live example on JSFiddle or Codepen or fill the following demo with your settings -->
https://jsfiddle.net/AMBudnik/4d3sjgn2/
### Your environment
* Handsontable version: 7.4.2
* Browser Name and version: Chrome 83
* Operating System: macOS Catalina
|
1.0
|
skipColumnOnPaste and skipRowOnPaste when added to table settings hang the browser down after pasting - ### Description
<!--- Tell us what happens and what should happen -->
Options [`skipColumnOnPaste`](https://handsontable.com/docs/7.4.2/Options.html#skipColumnOnPaste) and [`skipRowOnPaste`](https://handsontable.com/docs/7.4.2/Options.html#skipRowOnPaste) when added to table settings hang the browser down after pasting.
It is said that
> When added to a column property, it skips the column on paste and pastes the data on the next column to the right.
But I guess that we should also mention that it is not supported for table level settings.
### Steps to reproduce
<!--- Provide steps to reproduce this issue -->
1. Go to https://jsfiddle.net/AMBudnik/4d3sjgn2/
2. Select any cell
3. Use <kbd>CTRL/CMD</kbd>+<kbd>C</kbd> to copy
4. Use <kbd>CTRL/CMD</kbd>+<kbd>V</kbd> to paste
**Result**
The browser does not respond.
### Demo
<!--- Provide a link to a live example on JSFiddle or Codepen or fill the following demo with your settings -->
https://jsfiddle.net/AMBudnik/4d3sjgn2/
### Your environment
* Handsontable version: 7.4.2
* Browser Name and version: Chrome 83
* Operating System: macOS Catalina
|
non_process
|
skipcolumnonpaste and skiprowonpaste when added to table settings hang the browser down after pasting description options and when added to table settings hang the browser down after pasting it is said that when added to a column property it skips the column on paste and pastes the data on the next column to the right but i guess that we should also mention that it is not supported for table level settings steps to reproduce go to select any cell use ctrl cmd c to copy use ctrl cmd v to paste result the browser does not respond demo your environment handsontable version browser name and version chrome operating system macos catalina
| 0
|
558
| 2,570,755,310
|
IssuesEvent
|
2015-02-10 12:02:26
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
closed
|
hazelcast.xml variable replacement only works at the front of an xml attribute string
|
Team: Core Type: Defect
|
For example, I would like to do:
```xml
<hazelcast>
<import resource="hazelcast/group-${env}-config.xml"/>
...
```
But this tries to import a file called "qa-config.xml" instead of "hazelcast/group-qa-config.xml"
Here is a unit test that demonstrates the same problem. I am on version 3.4
```java
@Test
public void testXmlVariableReplacement() throws Exception {
String xml = "<hazelcast>\n" +
" <properties>\n" +
" <property name=\"${env}-with-suffix\">local-with-suffix</property>\n" +
" <property name=\"with-prefix-${env}\">with-prefix-local</property>\n" +
" </properties>\n" +
"</hazelcast>";
// Simulate setenv.sh service configuration
System.setProperty("env", "local");
InputStream is = IOUtils.toInputStream(xml, "UTF-8");
Config config = new XmlConfigBuilder(is).build();
Properties properties = config.getProperties();
// ok
assertEquals(properties.getProperty("local-with-suffix"), "local-with-suffix");
// this fails because "with-prefix-${env}" resolved as "local"
assertEquals(properties.getProperty("with-prefix-local"), "with-prefix-local");
}
```
|
1.0
|
hazelcast.xml variable replacement only works at the front of an xml attribute string - For example, I would like to do:
```xml
<hazelcast>
<import resource="hazelcast/group-${env}-config.xml"/>
...
```
But this tries to import a file called "qa-config.xml" instead of "hazelcast/group-qa-config.xml"
Here is a unit test that demonstrates the same problem. I am on version 3.4
```java
@Test
public void testXmlVariableReplacement() throws Exception {
String xml = "<hazelcast>\n" +
" <properties>\n" +
" <property name=\"${env}-with-suffix\">local-with-suffix</property>\n" +
" <property name=\"with-prefix-${env}\">with-prefix-local</property>\n" +
" </properties>\n" +
"</hazelcast>";
// Simulate setenv.sh service configuration
System.setProperty("env", "local");
InputStream is = IOUtils.toInputStream(xml, "UTF-8");
Config config = new XmlConfigBuilder(is).build();
Properties properties = config.getProperties();
// ok
assertEquals(properties.getProperty("local-with-suffix"), "local-with-suffix");
// this fails because "with-prefix-${env}" resolved as "local"
assertEquals(properties.getProperty("with-prefix-local"), "with-prefix-local");
}
```
|
non_process
|
hazelcast xml variable replacement only works at the front of an xml attribute string for example i would like to do xml but this tries to import a file called qa config xml instead of hazelcast group qa config xml here is a unit test that demonstrates the same problem i am on version java test public void testxmlvariablereplacement throws exception string xml n n local with suffix n with prefix local n n simulate setenv sh service configuration system setproperty env local inputstream is ioutils toinputstream xml utf config config new xmlconfigbuilder is build properties properties config getproperties ok assertequals properties getproperty local with suffix local with suffix this fails because with prefix env resolved as local assertequals properties getproperty with prefix local with prefix local
| 0
|
9,394
| 12,396,134,154
|
IssuesEvent
|
2020-05-20 19:57:27
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Temporary layers generated by processing tools from inputs with custom CRS have no CRS defined
|
Bug Processing Projections/Transformations
|
**Describe the bug**
I create a point with one of the 'ZANGI' CRS (which are not user-defined but are custom) and save it in a GeoPackage.
Then I run the Vector-GeoProcessing-Buffer tool and generate a buffer as a temporary layer.
Bug: The buffer polygon draws in completely the wrong place as it has no CRS defined. It should get the same CRS as the source layer.
If I generate the buffer directly to the GeoPackage, it takes on the correct CRS.
**How to Reproduce**
described above
**QGIS and OS versions**
QGIS version | 3.10.5-A Coruña | QGIS code revision | 984615fe1e
-- | -- | -- | --
Compiled against Qt | 5.12.3 | Running against Qt | 5.12.3
Compiled against GDAL/OGR | 2.4.1 | Running against GDAL/OGR | 2.4.1
Compiled against GEOS | 3.7.2-CAPI-1.11.2 | Running against GEOS | 3.7.2-CAPI-1.11.2 b55d2125
Compiled against SQLite | 3.28.0 | Running against SQLite | 3.28.0
PostgreSQL Client Version | 11.3 | SpatiaLite Version | 4.3.0a
QWT Version | 6.1.4 | QScintilla2 Version | 2.11.1
Compiled against PROJ | 5.2.0 | Running against PROJ | Rel. 5.2.0, September 15th, 2018
OS Version | macOS 10.15
Active python plugins | QgisModelBaker; SpreadsheetLayers; lrs; QuickOSM; VectorBender; quick_map_services; stac_browser; Discovery; qfieldsync; qchainage; OSMDownloader; parcel_plugin; lizmap; openlayers_plugin; Projestions; mmqgis; processing_circuitscape; mapswipetool_plugin; Mergin; travel_time_platform_plugin; nominatim_locator_filter; timemanager; processing; db_manager; MetaSearch
**Additional context**
<!-- Add any other context about the problem here. -->
|
1.0
|
Temporary layers generated by processing tools from inputs with custom CRS have no CRS defined - **Describe the bug**
I create a point with one of the 'ZANGI' CRS (which are not user-defined but are custom) and save it in a GeoPackage.
Then I run the Vector-GeoProcessing-Buffer tool and generate a buffer as a temporary layer.
Bug: The buffer polygon draws in completely the wrong place as it has no CRS defined. It should get the same CRS as the source layer.
If I generate the buffer directly to the GeoPackage, it takes on the correct CRS.
**How to Reproduce**
described above
**QGIS and OS versions**
QGIS version | 3.10.5-A Coruña | QGIS code revision | 984615fe1e
-- | -- | -- | --
Compiled against Qt | 5.12.3 | Running against Qt | 5.12.3
Compiled against GDAL/OGR | 2.4.1 | Running against GDAL/OGR | 2.4.1
Compiled against GEOS | 3.7.2-CAPI-1.11.2 | Running against GEOS | 3.7.2-CAPI-1.11.2 b55d2125
Compiled against SQLite | 3.28.0 | Running against SQLite | 3.28.0
PostgreSQL Client Version | 11.3 | SpatiaLite Version | 4.3.0a
QWT Version | 6.1.4 | QScintilla2 Version | 2.11.1
Compiled against PROJ | 5.2.0 | Running against PROJ | Rel. 5.2.0, September 15th, 2018
OS Version | macOS 10.15
Active python plugins | QgisModelBaker; SpreadsheetLayers; lrs; QuickOSM; VectorBender; quick_map_services; stac_browser; Discovery; qfieldsync; qchainage; OSMDownloader; parcel_plugin; lizmap; openlayers_plugin; Projestions; mmqgis; processing_circuitscape; mapswipetool_plugin; Mergin; travel_time_platform_plugin; nominatim_locator_filter; timemanager; processing; db_manager; MetaSearch
**Additional context**
<!-- Add any other context about the problem here. -->
|
process
|
temporary layers generated by processing tools from inputs with custom crs have no crs defined describe the bug i create a point with one of the zangi crs which are not user defined but are custom and save it in a geopackage then i run the vector geoprocessing buffer tool and generate a buffer as a temporary layer bug the buffer polygon draws in completely the wrong place as it has no crs defined it should get the same crs as the source layer if i generate the buffer directly to the geopackage it takes on the correct crs how to reproduce described above qgis and os versions qgis version a coruña qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel september os version macos active python plugins qgismodelbaker spreadsheetlayers lrs quickosm vectorbender quick map services stac browser discovery qfieldsync qchainage osmdownloader parcel plugin lizmap openlayers plugin projestions mmqgis processing circuitscape mapswipetool plugin mergin travel time platform plugin nominatim locator filter timemanager processing db manager metasearch additional context
| 1
|
389,186
| 11,498,672,675
|
IssuesEvent
|
2020-02-12 12:29:44
|
massenergize/frontend-portal
|
https://api.github.com/repos/massenergize/frontend-portal
|
closed
|
footer - newsletter subscription
|
enhancement priority 2
|
ISSUE 1190
Reported on Fri Dec 13 2019 19:02:16 GMT-0500 (EST)
Reported by Kaat Vander Straeten(KAATvds@gmail.com)
PRIORITY = Medium
STATUS = - Submitted
Site Url: http://community-dev.massenergize.org/wayland/contactus
Description:
In footer under newsletter signup boxes, replace "Your mail id is confidential" with "your name and email are confidential and intended solely for sending you our Newsletter".
User Role: Community User
|
1.0
|
footer - newsletter subscription - ISSUE 1190
Reported on Fri Dec 13 2019 19:02:16 GMT-0500 (EST)
Reported by Kaat Vander Straeten(KAATvds@gmail.com)
PRIORITY = Medium
STATUS = - Submitted
Site Url: http://community-dev.massenergize.org/wayland/contactus
Description:
In footer under newsletter signup boxes, replace "Your mail id is confidential" with "your name and email are confidential and intended solely for sending you our Newsletter".
User Role: Community User
|
non_process
|
footer newsletter subscription issue reported on fri dec gmt est reported by kaat vander straeten kaatvds gmail com priority medium status submitted site url description in footer under newsletter signup boxes replace your mail id is confidential with your name and email are confidential and intended solely for sending you our newsletter user role community user
| 0
|
11,366
| 14,175,827,369
|
IssuesEvent
|
2020-11-12 22:17:49
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
How to trigger pipeline on dockerhub push?
|
Pri2 devops-cicd-process/tech devops/prod doc-bug
|
"An error occurred while loading the YAML build pipeline. 'image' key is not defined for Docker hub container resource: argocd"
Trying to trigger a azure devops pipeline from an external dockerhub repositiory.
```yaml
resources:
-containers:
- container: tools
type: ACR
azureSubscription: example
resourceGroup: containerRegistry
registry: crExample
repository: tools
trigger:
tags:
include:
- "*"
- container: argocd
type: Docker
azureSubscription: argoproj
registry: argoproj
repository: argocd
trigger:
tags:
include:
- "v*"
```
First container works and triggers the pipeline correctly but the second does not.
How to fix? Could you also make an example on this site.
Use case:
We append our tools to the argocd image in a custom dockerfile. Would like to automate the release for when argocd does a release or when we rebuild our tools
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: ee4ec9d0-e0d5-4fb4-7c3e-b84abfa290c2
* Version Independent ID: 3e2b80d9-30e5-0c48-49f0-4fcdfedf5eee
* Content: [Resources - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/resources?view=azure-devops&tabs=schema)
* Content Source: [docs/pipelines/process/resources.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/resources.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
How to trigger pipeline on dockerhub push? - "An error occurred while loading the YAML build pipeline. 'image' key is not defined for Docker hub container resource: argocd"
Trying to trigger a azure devops pipeline from an external dockerhub repositiory.
```yaml
resources:
-containers:
- container: tools
type: ACR
azureSubscription: example
resourceGroup: containerRegistry
registry: crExample
repository: tools
trigger:
tags:
include:
- "*"
- container: argocd
type: Docker
azureSubscription: argoproj
registry: argoproj
repository: argocd
trigger:
tags:
include:
- "v*"
```
First container works and triggers the pipeline correctly but the second does not.
How to fix? Could you also make an example on this site.
Use case:
We append our tools to the argocd image in a custom dockerfile. Would like to automate the release for when argocd does a release or when we rebuild our tools
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: ee4ec9d0-e0d5-4fb4-7c3e-b84abfa290c2
* Version Independent ID: 3e2b80d9-30e5-0c48-49f0-4fcdfedf5eee
* Content: [Resources - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/resources?view=azure-devops&tabs=schema)
* Content Source: [docs/pipelines/process/resources.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/resources.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
how to trigger pipeline on dockerhub push an error occurred while loading the yaml build pipeline image key is not defined for docker hub container resource argocd trying to trigger a azure devops pipeline from an external dockerhub repositiory yaml resources containers container tools type acr azuresubscription example resourcegroup containerregistry registry crexample repository tools trigger tags include container argocd type docker azuresubscription argoproj registry argoproj repository argocd trigger tags include v first container works and triggers the pipeline correctly but the second does not how to fix could you also make an example on this site use case we append our tools to the argocd image in a custom dockerfile would like to automate the release for when argocd does a release or when we rebuild our tools document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
60,525
| 12,126,384,687
|
IssuesEvent
|
2020-04-22 16:56:29
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
Assertion failed 'offsetFitsInVectorMem(emitGetInsSC(id))'
|
arch-arm32 area-CodeGen-coreclr os-windows
|
Windows arm32 with:
```
set COMPlus_TieredCompilation=0
set COMPlus_JitStress=1
```
From attempt to re-enable "runtime-coreclr jitstress" testing:
https://dev.azure.com/dnceng/public/_build/results?buildId=574937&view=results
```
JIT\Methodical\explicit\coverage\expl_float_1_d\expl_float_1_d.cmd [FAIL]
Assert failure(PID 7700 [0x00001e14], Thread: 10360 [0x2878]): Assertion failed 'offsetFitsInVectorMem(emitGetInsSC(id))' in 'TestApp:Main():int' during 'Generate code' (IL size 4230)
File: F:\workspace\_work\1\s\src\coreclr\src\jit\emitarm.cpp Line: 465
Image: D:\h\w\B77C09AB\p\CoreRun.exe
```
Tests:
```
JIT\Methodical\explicit\coverage\expl_float_1_d\expl_float_1_d.cmd
JIT\Methodical\explicit\coverage\expl_float_1_r\expl_float_1_r.cmd
```
category:correctness
theme:alignment
skill-level:intermediate
cost:medium
|
1.0
|
Assertion failed 'offsetFitsInVectorMem(emitGetInsSC(id))' - Windows arm32 with:
```
set COMPlus_TieredCompilation=0
set COMPlus_JitStress=1
```
From attempt to re-enable "runtime-coreclr jitstress" testing:
https://dev.azure.com/dnceng/public/_build/results?buildId=574937&view=results
```
JIT\Methodical\explicit\coverage\expl_float_1_d\expl_float_1_d.cmd [FAIL]
Assert failure(PID 7700 [0x00001e14], Thread: 10360 [0x2878]): Assertion failed 'offsetFitsInVectorMem(emitGetInsSC(id))' in 'TestApp:Main():int' during 'Generate code' (IL size 4230)
File: F:\workspace\_work\1\s\src\coreclr\src\jit\emitarm.cpp Line: 465
Image: D:\h\w\B77C09AB\p\CoreRun.exe
```
Tests:
```
JIT\Methodical\explicit\coverage\expl_float_1_d\expl_float_1_d.cmd
JIT\Methodical\explicit\coverage\expl_float_1_r\expl_float_1_r.cmd
```
category:correctness
theme:alignment
skill-level:intermediate
cost:medium
|
non_process
|
assertion failed offsetfitsinvectormem emitgetinssc id windows with set complus tieredcompilation set complus jitstress from attempt to re enable runtime coreclr jitstress testing jit methodical explicit coverage expl float d expl float d cmd assert failure pid thread assertion failed offsetfitsinvectormem emitgetinssc id in testapp main int during generate code il size file f workspace work s src coreclr src jit emitarm cpp line image d h w p corerun exe tests jit methodical explicit coverage expl float d expl float d cmd jit methodical explicit coverage expl float r expl float r cmd category correctness theme alignment skill level intermediate cost medium
| 0
|
75,073
| 9,200,110,476
|
IssuesEvent
|
2019-03-07 16:20:33
|
GCTC-NTGC/TalentCloud
|
https://api.github.com/repos/GCTC-NTGC/TalentCloud
|
closed
|
Applicant - Bilingual prompt
|
Design Front-end
|
# Description
The lack of clarification in bilingualism at poster level has caused current processes to crash, after managers find a large number of applicants without the bilingual skills on a bilingual position. We need an improved solution to flag bilingual requirements and their full impact at the job poster level to guarantee applicants understanding of language level requirements before applying.
- [ ] This task requires a unit test.
- [ ] This task is complex and needs acceptance criteria.
# Images
Place image attachments (wireframes, etc.) here.
|
1.0
|
Applicant - Bilingual prompt - # Description
The lack of clarification in bilingualism at poster level has caused current processes to crash, after managers find a large number of applicants without the bilingual skills on a bilingual position. We need an improved solution to flag bilingual requirements and their full impact at the job poster level to guarantee applicants understanding of language level requirements before applying.
- [ ] This task requires a unit test.
- [ ] This task is complex and needs acceptance criteria.
# Images
Place image attachments (wireframes, etc.) here.
|
non_process
|
applicant bilingual prompt description the lack of clarification in bilingualism at poster level has caused current processes to crash after managers find a large number of applicants without the bilingual skills on a bilingual position we need an improved solution to flag bilingual requirements and their full impact at the job poster level to guarantee applicants understanding of language level requirements before applying this task requires a unit test this task is complex and needs acceptance criteria images place image attachments wireframes etc here
| 0
|
177,546
| 21,479,032,541
|
IssuesEvent
|
2022-04-26 15:56:08
|
andrewguest/Kubernetes-example
|
https://api.github.com/repos/andrewguest/Kubernetes-example
|
closed
|
CVE-2021-3803 (High) detected in nth-check-2.0.0.tgz, nth-check-1.0.2.tgz - autoclosed
|
security vulnerability
|
## CVE-2021-3803 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>nth-check-2.0.0.tgz</b>, <b>nth-check-1.0.2.tgz</b></p></summary>
<p>
<details><summary><b>nth-check-2.0.0.tgz</b></p></summary>
<p>Parses and compiles CSS nth-checks to highly optimized functions.</p>
<p>Library home page: <a href="https://registry.npmjs.org/nth-check/-/nth-check-2.0.0.tgz">https://registry.npmjs.org/nth-check/-/nth-check-2.0.0.tgz</a></p>
<p>Path to dependency file: /frontend/package.json</p>
<p>Path to vulnerable library: /frontend/node_modules/nth-check/package.json</p>
<p>
Dependency Hierarchy:
- nuxt-2.15.8.tgz (Root Library)
- generator-2.15.8.tgz
- node-html-parser-3.3.6.tgz
- css-select-4.1.3.tgz
- :x: **nth-check-2.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>nth-check-1.0.2.tgz</b></p></summary>
<p>performant nth-check parser & compiler</p>
<p>Library home page: <a href="https://registry.npmjs.org/nth-check/-/nth-check-1.0.2.tgz">https://registry.npmjs.org/nth-check/-/nth-check-1.0.2.tgz</a></p>
<p>Path to dependency file: /frontend/package.json</p>
<p>Path to vulnerable library: /frontend/node_modules/nth-check/package.json</p>
<p>
Dependency Hierarchy:
- nuxt-2.15.8.tgz (Root Library)
- webpack-2.15.8.tgz
- cssnano-4.1.11.tgz
- cssnano-preset-default-4.0.8.tgz
- postcss-svgo-4.0.3.tgz
- svgo-1.3.2.tgz
- css-select-2.1.0.tgz
- :x: **nth-check-1.0.2.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/andrewguest/Kubernetes-example/commit/1ce6c4a4c1e241ed5f10997bc3b47f45f804d88c">1ce6c4a4c1e241ed5f10997bc3b47f45f804d88c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
nth-check is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3803>CVE-2021-3803</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/fb55/nth-check/compare/v2.0.0...v2.0.1">https://github.com/fb55/nth-check/compare/v2.0.0...v2.0.1</a></p>
<p>Release Date: 2021-09-17</p>
<p>Fix Resolution: nth-check - v2.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-3803 (High) detected in nth-check-2.0.0.tgz, nth-check-1.0.2.tgz - autoclosed - ## CVE-2021-3803 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>nth-check-2.0.0.tgz</b>, <b>nth-check-1.0.2.tgz</b></p></summary>
<p>
<details><summary><b>nth-check-2.0.0.tgz</b></p></summary>
<p>Parses and compiles CSS nth-checks to highly optimized functions.</p>
<p>Library home page: <a href="https://registry.npmjs.org/nth-check/-/nth-check-2.0.0.tgz">https://registry.npmjs.org/nth-check/-/nth-check-2.0.0.tgz</a></p>
<p>Path to dependency file: /frontend/package.json</p>
<p>Path to vulnerable library: /frontend/node_modules/nth-check/package.json</p>
<p>
Dependency Hierarchy:
- nuxt-2.15.8.tgz (Root Library)
- generator-2.15.8.tgz
- node-html-parser-3.3.6.tgz
- css-select-4.1.3.tgz
- :x: **nth-check-2.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>nth-check-1.0.2.tgz</b></p></summary>
<p>performant nth-check parser & compiler</p>
<p>Library home page: <a href="https://registry.npmjs.org/nth-check/-/nth-check-1.0.2.tgz">https://registry.npmjs.org/nth-check/-/nth-check-1.0.2.tgz</a></p>
<p>Path to dependency file: /frontend/package.json</p>
<p>Path to vulnerable library: /frontend/node_modules/nth-check/package.json</p>
<p>
Dependency Hierarchy:
- nuxt-2.15.8.tgz (Root Library)
- webpack-2.15.8.tgz
- cssnano-4.1.11.tgz
- cssnano-preset-default-4.0.8.tgz
- postcss-svgo-4.0.3.tgz
- svgo-1.3.2.tgz
- css-select-2.1.0.tgz
- :x: **nth-check-1.0.2.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/andrewguest/Kubernetes-example/commit/1ce6c4a4c1e241ed5f10997bc3b47f45f804d88c">1ce6c4a4c1e241ed5f10997bc3b47f45f804d88c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
nth-check is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3803>CVE-2021-3803</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/fb55/nth-check/compare/v2.0.0...v2.0.1">https://github.com/fb55/nth-check/compare/v2.0.0...v2.0.1</a></p>
<p>Release Date: 2021-09-17</p>
<p>Fix Resolution: nth-check - v2.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in nth check tgz nth check tgz autoclosed cve high severity vulnerability vulnerable libraries nth check tgz nth check tgz nth check tgz parses and compiles css nth checks to highly optimized functions library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules nth check package json dependency hierarchy nuxt tgz root library generator tgz node html parser tgz css select tgz x nth check tgz vulnerable library nth check tgz performant nth check parser compiler library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules nth check package json dependency hierarchy nuxt tgz root library webpack tgz cssnano tgz cssnano preset default tgz postcss svgo tgz svgo tgz css select tgz x nth check tgz vulnerable library found in head commit a href found in base branch main vulnerability details nth check is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution nth check step up your open source security game with whitesource
| 0
|
418,274
| 28,114,215,221
|
IssuesEvent
|
2023-03-31 09:30:18
|
manushridiv/ped
|
https://api.github.com/repos/manushridiv/ped
|
opened
|
Inconsistencies in the user guide and return value
|
type.DocumentationBug severity.Low
|

The view notes function is adding a task to the list, however documentation has mentioned that the required output for the input is: There are no notes for this task!
<!--session: 1680252526668-d9d98259-62f2-40d1-aae9-119c03ca3b5d-->
<!--Version: Web v3.4.7-->
|
1.0
|
Inconsistencies in the user guide and return value - 
The view notes function is adding a task to the list, however documentation has mentioned that the required output for the input is: There are no notes for this task!
<!--session: 1680252526668-d9d98259-62f2-40d1-aae9-119c03ca3b5d-->
<!--Version: Web v3.4.7-->
|
non_process
|
inconsistencies in the user guide and return value the view notes function is adding a task to the list however documentation has mentioned that the required output for the input is there are no notes for this task
| 0
|
19,613
| 25,969,347,144
|
IssuesEvent
|
2022-12-19 09:57:51
|
googleapis/google-api-dotnet-client
|
https://api.github.com/repos/googleapis/google-api-dotnet-client
|
closed
|
1.58.0 release on NuGet has no corresponding GitHub release or Git tag
|
type: process priority: p2
|
The 1.58.0 package on NuGet was released on 01Dec, but I haven't seen the corresponding GitHub release or tag since then. Just wanted to get a sense of what changes were made before I upgrade my project to the latest version. Thanks!
|
1.0
|
1.58.0 release on NuGet has no corresponding GitHub release or Git tag - The 1.58.0 package on NuGet was released on 01Dec, but I haven't seen the corresponding GitHub release or tag since then. Just wanted to get a sense of what changes were made before I upgrade my project to the latest version. Thanks!
|
process
|
release on nuget has no corresponding github release or git tag the package on nuget was released on but i haven t seen the corresponding github release or tag since then just wanted to get a sense of what changes were made before i upgrade my project to the latest version thanks
| 1
|
3,754
| 6,733,154,225
|
IssuesEvent
|
2017-10-18 14:00:40
|
york-region-tpss/stp
|
https://api.github.com/repos/york-region-tpss/stp
|
closed
|
Price Estimate Formula
|
enhancement process workflow
|
Provide a formula to calculate this year's estimated price from previous year's price.

Note: formula should be simplified so that user should only input the increase rate. and the data should be written into database directly.
|
1.0
|
Price Estimate Formula - Provide a formula to calculate this year's estimated price from previous year's price.

Note: formula should be simplified so that user should only input the increase rate. and the data should be written into database directly.
|
process
|
price estimate formula provide a formula to calculate this year s estimated price from previous year s price note formula should be simplified so that user should only input the increase rate and the data should be written into database directly
| 1
|
14,529
| 17,630,658,472
|
IssuesEvent
|
2021-08-19 07:31:31
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
Add Maria Bamford Is Nuts!
|
suggested title in process
|
Please add as much of the following info as you can:
Title: Maria Bamford Is Nuts!
Type (film/tv show): TV Show
Film or show in which it appears: Lady Dynamite
Is the parent film/show streaming anywhere? Netflix / MuskVision
About when in the parent film/show does it appear? Season 2, Episode 2, around 15 minutes in.
Actual footage of the film/show can be seen (yes/no)? yes
|
1.0
|
Add Maria Bamford Is Nuts! - Please add as much of the following info as you can:
Title: Maria Bamford Is Nuts!
Type (film/tv show): TV Show
Film or show in which it appears: Lady Dynamite
Is the parent film/show streaming anywhere? Netflix / MuskVision
About when in the parent film/show does it appear? Season 2, Episode 2, around 15 minutes in.
Actual footage of the film/show can be seen (yes/no)? yes
|
process
|
add maria bamford is nuts please add as much of the following info as you can title maria bamford is nuts type film tv show tv show film or show in which it appears lady dynamite is the parent film show streaming anywhere netflix muskvision about when in the parent film show does it appear season episode around minutes in actual footage of the film show can be seen yes no yes
| 1
|
18,377
| 24,504,707,237
|
IssuesEvent
|
2022-10-10 15:23:43
|
openxla/stablehlo
|
https://api.github.com/repos/openxla/stablehlo
|
opened
|
Fix clang tidy variable naming mismatches
|
Process
|
Fix "argument name at declaration does not match argument name at definition" issues:
```
-bool isCompatibleForHloTypeInference(TypeRange l, TypeRange r) {
- if (l.size() != r.size()) return false;
- for (auto [lt, rt] : llvm::zip(l, r))
+bool isCompatibleForHloTypeInference(TypeRange tp1, TypeRange tp2) {
+ if (tp1.size() != tp2.size()) return false;
+ for (auto [lt, rt] : llvm::zip(tp1, tp2))
return verifyReduceScatter(*this,
- /*operand_types=*/{getOperand().getType()},
- /*result_types=*/{getType()},
- /*scatter_dimension=*/getScatterDimension());
+ /*operandTypes=*/{getOperand().getType()},
+ /*resultTypes=*/{getType()},
+ /*scatterDimension=*/getScatterDimension());
```
|
1.0
|
Fix clang tidy variable naming mismatches - Fix "argument name at declaration does not match argument name at definition" issues:
```
-bool isCompatibleForHloTypeInference(TypeRange l, TypeRange r) {
- if (l.size() != r.size()) return false;
- for (auto [lt, rt] : llvm::zip(l, r))
+bool isCompatibleForHloTypeInference(TypeRange tp1, TypeRange tp2) {
+ if (tp1.size() != tp2.size()) return false;
+ for (auto [lt, rt] : llvm::zip(tp1, tp2))
return verifyReduceScatter(*this,
- /*operand_types=*/{getOperand().getType()},
- /*result_types=*/{getType()},
- /*scatter_dimension=*/getScatterDimension());
+ /*operandTypes=*/{getOperand().getType()},
+ /*resultTypes=*/{getType()},
+ /*scatterDimension=*/getScatterDimension());
```
|
process
|
fix clang tidy variable naming mismatches fix argument name at declaration does not match argument name at definition issues bool iscompatibleforhlotypeinference typerange l typerange r if l size r size return false for auto llvm zip l r bool iscompatibleforhlotypeinference typerange typerange if size size return false for auto llvm zip return verifyreducescatter this operand types getoperand gettype result types gettype scatter dimension getscatterdimension operandtypes getoperand gettype resulttypes gettype scatterdimension getscatterdimension
| 1
|
20,740
| 27,439,840,966
|
IssuesEvent
|
2023-03-02 10:13:20
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
Ability to prefiix metrics from spanmetrics processor
|
enhancement processor/spanmetrics needs triage
|
### Component(s)
processor/spanmetrics
### Is your feature request related to a problem? Please describe.
Feature request to add an arbitrary prefix to metric names emitted by spanmetrics processor.
Currently with the 4 metric names set as `calls_total`, `latency_bucket`, `latency_count`, and `latency_sum` become harder to utilize and organize when spanmetrics processor is used for more than one source of data as they all get folded into the same 4 metrics.
Would like to be able to prefix metric names for ease of use and organization (E.G. `cicd_calls_total`)
### Describe the solution you'd like
ability to add a prefix (or fully rename) metrics emitted by spanmetrics processor.
Suggest prefixing to keep the `calls_total`, `latency_bucket`, `latency_count`, and `latency_sum` pattern intact.
Something like adding a config value (E.G. `metric_name_prefix: prefixstring_`) that would then prefix metrics such that `calls_total` would become `prefixstring_calls_total` along with `prefixstring_latency_bucket` and so on.
Without the ability to name metrics differently using the spanmetrics processor more than once quickly leads to a very large number of dimensions on the emitted metrics which may or may not be related and lead to confusion.
### Describe alternatives you've considered
Ability to fully rename metrics using transform processors.
Have tried renaming with transform processor and metrics transform processor without success on spanmetrics emitted metrics. While I was able to rename standard metrics and transform them as expected, I was not able to rename or transform spanmetrics emitted metrics. I believe this is due to the nature of spanmetrics metrics creation and feed to exporter.
### Additional context
I'm happy to contribute code for this feature request.
|
1.0
|
Ability to prefiix metrics from spanmetrics processor - ### Component(s)
processor/spanmetrics
### Is your feature request related to a problem? Please describe.
Feature request to add an arbitrary prefix to metric names emitted by spanmetrics processor.
Currently with the 4 metric names set as `calls_total`, `latency_bucket`, `latency_count`, and `latency_sum` become harder to utilize and organize when spanmetrics processor is used for more than one source of data as they all get folded into the same 4 metrics.
Would like to be able to prefix metric names for ease of use and organization (E.G. `cicd_calls_total`)
### Describe the solution you'd like
ability to add a prefix (or fully rename) metrics emitted by spanmetrics processor.
Suggest prefixing to keep the `calls_total`, `latency_bucket`, `latency_count`, and `latency_sum` pattern intact.
Something like adding a config value (E.G. `metric_name_prefix: prefixstring_`) that would then prefix metrics such that `calls_total` would become `prefixstring_calls_total` along with `prefixstring_latency_bucket` and so on.
Without the ability to name metrics differently using the spanmetrics processor more than once quickly leads to a very large number of dimensions on the emitted metrics which may or may not be related and lead to confusion.
### Describe alternatives you've considered
Ability to fully rename metrics using transform processors.
Have tried renaming with transform processor and metrics transform processor without success on spanmetrics emitted metrics. While I was able to rename standard metrics and transform them as expected, I was not able to rename or transform spanmetrics emitted metrics. I believe this is due to the nature of spanmetrics metrics creation and feed to exporter.
### Additional context
I'm happy to contribute code for this feature request.
|
process
|
ability to prefiix metrics from spanmetrics processor component s processor spanmetrics is your feature request related to a problem please describe feature request to add an arbitrary prefix to metric names emitted by spanmetrics processor currently with the metric names set as calls total latency bucket latency count and latency sum become harder to utilize and organize when spanmetrics processor is used for more than one source of data as they all get folded into the same metrics would like to be able to prefix metric names for ease of use and organization e g cicd calls total describe the solution you d like ability to add a prefix or fully rename metrics emitted by spanmetrics processor suggest prefixing to keep the calls total latency bucket latency count and latency sum pattern intact something like adding a config value e g metric name prefix prefixstring that would then prefix metrics such that calls total would become prefixstring calls total along with prefixstring latency bucket and so on without the ability to name metrics differently using the spanmetrics processor more than once quickly leads to a very large number of dimensions on the emitted metrics which may or may not be related and lead to confusion describe alternatives you ve considered ability to fully rename metrics using transform processors have tried renaming with transform processor and metrics transform processor without success on spanmetrics emitted metrics while i was able to rename standard metrics and transform them as expected i was not able to rename or transform spanmetrics emitted metrics i believe this is due to the nature of spanmetrics metrics creation and feed to exporter additional context i m happy to contribute code for this feature request
| 1
|
85,346
| 10,436,970,838
|
IssuesEvent
|
2019-09-17 20:49:03
|
toolboxtheplatform/toolbox.app
|
https://api.github.com/repos/toolboxtheplatform/toolbox.app
|
opened
|
Implement Google Analytics or similar
|
documentation
|
Explore some analytical tools to implement them for statistics.
|
1.0
|
Implement Google Analytics or similar - Explore some analytical tools to implement them for statistics.
|
non_process
|
implement google analytics or similar explore some analytical tools to implement them for statistics
| 0
|
21,431
| 29,359,594,964
|
IssuesEvent
|
2023-05-28 00:37:32
|
devssa/onde-codar-em-salvador
|
https://api.github.com/repos/devssa/onde-codar-em-salvador
|
closed
|
[Remoto] DevSecOps Engineer na Coodesh
|
SALVADOR REDES JAVASCRIPT DEVOPS ANGULAR REACT REQUISITOS REMOTO PROCESSOS GITHUB INGLÊS CI SEGURANÇA UMA C R APIs METODOLOGIAS ÁGEIS TREINAMENTO LGPD PRIVACIDADE DE DADOS CONTAINER IAC Stale
|
## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/devsecops-engineer-142805520?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A Juvo está em busca de DevSecOps Engineer para compor seu time!</p>
<p>A Juvo está aqui para ajudar e apoiar milhões de brasileiros, impactando positivamente suas vidas com nossos serviços financeiros. É nisso que acreditamos e o objetivo que buscamos a cada dia. Por que existimos? Para abrir oportunidades econômicas para o próximo bilhão de pessoas... começando pelo Brasil!</p>
<p>Incentivamos um ambiente leve, de respeito e confiança, onde as pessoas possam compartilhar suas ideias e nos ajudar a crescer. A colaboração é a nossa base! Se você se identificou como nosso propósito, vem pra Juvo!</p>
<p>Responsabilidades:</p>
<ul>
<li>Atuar como ponto focal dos times ágeis (DevSecOps Engineer) participando das definições de requisitos, análise e auxílio na resolução de vulnerabilidades apontadas pelo Pentesting e solução SAST, DAST, SCA, Container Security e IaC Security;</li>
<li>Captar e transmitir ao time, os principais gaps e necessidades de atualização/mitigação;</li>
<li>Sinergia com as equipes de Devops levando conhecimento também às equipes de desenvolvimento;</li>
<li>Definir prioridade e políticas de segurança de aplicação para novos aplicativos;</li>
<li>Implementar as políticas que virem a ser definidas;</li>
<li>Interagir com os desenvolvedores para fomentar a correção das aplicações discutindo as vulnerabilidades críticas do legado;</li>
<li>Criar um plano de correção e implementar;</li>
<li>Apoiar o Desenvolvedor que fará a correção do código conversando sobre o processo;</li>
<li>Apoiar o usuário que fará uso dos dispositivos e aplicativos da empresa para ajudá-lo a seguir as orientações encontradas nas melhores práticas de segurança do mercado;</li>
<li>Treinamento e conscientização das áreas quanto a normas, procedimentos e técnicas de Desenvolvimento Seguro e Cibersegurança;</li>
<li>Criação de processos referente à LGPD (Lei Geral de Proteção de Dados Pessoais) e com ISO270012013 ( Segurança da Informação e Privacidade de Dados ).</li>
</ul>
## Juvo:
<p>Nossa missão é ajudar e apoiar milhões de brasileiros, impactando positivamente suas jornadas com nossos serviços financeiros.</p>
<p>A Juvo foi fundada com uma missão abrangente: capacitar bilhões de pessoas sem conta bancária em todo o mundo a criar identidades financeiras e acessar serviços financeiros - tudo por meio de suas interações diárias com seus telefones celulares. Em parceria com operadoras de redes móveis, a tecnologia proprietária Identity Scoring da Juvo usa ciência de dados, aprendizado de máquina e mecânica de jogos para criar um relacionamento baseado em identidade com usuários pré-pagos anônimos, abrindo acesso a serviços financeiros móveis inatingíveis.</p><a href='https://coodesh.com/companies/juvo'>Veja mais no site</a>
## Habilidades:
- Angular
- React Native
- Javascript
- Github Actions
- Web Services
## Local:
100% Remoto
## Requisitos:
- Inglês avançado;
- Conhecimento das linguagens de programação (Angular, React Native e Javascript), testes unitários, apis documentation, serviços web, CloudSecurity e GitHub Actions;
- Experiência em conceitos e controles aplicáveis ao desenvolvimento de código seguro (SAST, DAST, IAST, SCA, etc) e também em SCA, SAST ou IAST;
- Conteineirização;
- Experiência nas principais referências em vulnerabilidade de software (OWASP, SANS, CVE, etc);
- Graduação em Ciência da Computação, Segurança da Informação, curso de Tecnólogo em TI e outros correlatos.
## Diferenciais:
- Já ter trabalhado com Metodologias Ágeis.
## Benefícios:
- Plano de Saúde Bradesco;
- Plano Odontológico;
- Seguro de Vida;
- Cartão bandeira Mastercard - Flash com R$35,00/dia útil;
- Gympass;
- "Day off" de aniversário;
- 02 "wellness days" por ano;
- Remuneração competitiva.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [DevSecOps Engineer na Juvo](https://coodesh.com/jobs/devsecops-engineer-142805520?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Regime
CLT
#### Categoria
DevOps
|
1.0
|
[Remoto] DevSecOps Engineer na Coodesh - ## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/devsecops-engineer-142805520?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A Juvo está em busca de DevSecOps Engineer para compor seu time!</p>
<p>A Juvo está aqui para ajudar e apoiar milhões de brasileiros, impactando positivamente suas vidas com nossos serviços financeiros. É nisso que acreditamos e o objetivo que buscamos a cada dia. Por que existimos? Para abrir oportunidades econômicas para o próximo bilhão de pessoas... começando pelo Brasil!</p>
<p>Incentivamos um ambiente leve, de respeito e confiança, onde as pessoas possam compartilhar suas ideias e nos ajudar a crescer. A colaboração é a nossa base! Se você se identificou como nosso propósito, vem pra Juvo!</p>
<p>Responsabilidades:</p>
<ul>
<li>Atuar como ponto focal dos times ágeis (DevSecOps Engineer) participando das definições de requisitos, análise e auxílio na resolução de vulnerabilidades apontadas pelo Pentesting e solução SAST, DAST, SCA, Container Security e IaC Security;</li>
<li>Captar e transmitir ao time, os principais gaps e necessidades de atualização/mitigação;</li>
<li>Sinergia com as equipes de Devops levando conhecimento também às equipes de desenvolvimento;</li>
<li>Definir prioridade e políticas de segurança de aplicação para novos aplicativos;</li>
<li>Implementar as políticas que virem a ser definidas;</li>
<li>Interagir com os desenvolvedores para fomentar a correção das aplicações discutindo as vulnerabilidades críticas do legado;</li>
<li>Criar um plano de correção e implementar;</li>
<li>Apoiar o Desenvolvedor que fará a correção do código conversando sobre o processo;</li>
<li>Apoiar o usuário que fará uso dos dispositivos e aplicativos da empresa para ajudá-lo a seguir as orientações encontradas nas melhores práticas de segurança do mercado;</li>
<li>Treinamento e conscientização das áreas quanto a normas, procedimentos e técnicas de Desenvolvimento Seguro e Cibersegurança;</li>
<li>Criação de processos referente à LGPD (Lei Geral de Proteção de Dados Pessoais) e com ISO270012013 ( Segurança da Informação e Privacidade de Dados ).</li>
</ul>
## Juvo:
<p>Nossa missão é ajudar e apoiar milhões de brasileiros, impactando positivamente suas jornadas com nossos serviços financeiros.</p>
<p>A Juvo foi fundada com uma missão abrangente: capacitar bilhões de pessoas sem conta bancária em todo o mundo a criar identidades financeiras e acessar serviços financeiros - tudo por meio de suas interações diárias com seus telefones celulares. Em parceria com operadoras de redes móveis, a tecnologia proprietária Identity Scoring da Juvo usa ciência de dados, aprendizado de máquina e mecânica de jogos para criar um relacionamento baseado em identidade com usuários pré-pagos anônimos, abrindo acesso a serviços financeiros móveis inatingíveis.</p><a href='https://coodesh.com/companies/juvo'>Veja mais no site</a>
## Habilidades:
- Angular
- React Native
- Javascript
- Github Actions
- Web Services
## Local:
100% Remoto
## Requisitos:
- Inglês avançado;
- Conhecimento das linguagens de programação (Angular, React Native e Javascript), testes unitários, apis documentation, serviços web, CloudSecurity e GitHub Actions;
- Experiência em conceitos e controles aplicáveis ao desenvolvimento de código seguro (SAST, DAST, IAST, SCA, etc) e também em SCA, SAST ou IAST;
- Conteineirização;
- Experiência nas principais referências em vulnerabilidade de software (OWASP, SANS, CVE, etc);
- Graduação em Ciência da Computação, Segurança da Informação, curso de Tecnólogo em TI e outros correlatos.
## Diferenciais:
- Já ter trabalhado com Metodologias Ágeis.
## Benefícios:
- Plano de Saúde Bradesco;
- Plano Odontológico;
- Seguro de Vida;
- Cartão bandeira Mastercard - Flash com R$35,00/dia útil;
- Gympass;
- "Day off" de aniversário;
- 02 "wellness days" por ano;
- Remuneração competitiva.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [DevSecOps Engineer na Juvo](https://coodesh.com/jobs/devsecops-engineer-142805520?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Regime
CLT
#### Categoria
DevOps
|
process
|
devsecops engineer na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a juvo está em busca de devsecops engineer para compor seu time a juvo está aqui para ajudar e apoiar milhões de brasileiros impactando positivamente suas vidas com nossos serviços financeiros é nisso que acreditamos e o objetivo que buscamos a cada dia por que existimos para abrir oportunidades econômicas para o próximo bilhão de pessoas começando pelo brasil incentivamos um ambiente leve de respeito e confiança onde as pessoas possam compartilhar suas ideias e nos ajudar a crescer a colaboração é a nossa base se você se identificou como nosso propósito vem pra juvo responsabilidades atuar como ponto focal dos times ágeis devsecops engineer participando das definições de requisitos análise e auxílio na resolução de vulnerabilidades apontadas pelo pentesting e solução sast dast sca container security e iac security captar e transmitir ao time os principais gaps e necessidades de atualização mitigação sinergia com as equipes de devops levando conhecimento também às equipes de desenvolvimento definir prioridade e políticas de segurança de aplicação para novos aplicativos implementar as políticas que virem a ser definidas interagir com os desenvolvedores para fomentar a correção das aplicações discutindo as vulnerabilidades críticas do legado criar um plano de correção e implementar apoiar o desenvolvedor que fará a correção do código conversando sobre o processo apoiar o usuário que fará uso dos dispositivos e aplicativos da empresa para ajudá lo a seguir as orientações encontradas nas melhores práticas de segurança do mercado treinamento e conscientização das áreas quanto a normas procedimentos e técnicas de desenvolvimento seguro e cibersegurança criação de processos referente à lgpd lei geral de proteção de dados pessoais e com segurança da informação e privacidade de dados juvo nossa missão é ajudar e apoiar milhões de brasileiros impactando positivamente suas jornadas com nossos serviços financeiros a juvo foi fundada com uma missão abrangente capacitar bilhões de pessoas sem conta bancária em todo o mundo a criar identidades financeiras e acessar serviços financeiros tudo por meio de suas interações diárias com seus telefones celulares em parceria com operadoras de redes móveis a tecnologia proprietária identity scoring da juvo usa ciência de dados aprendizado de máquina e mecânica de jogos para criar um relacionamento baseado em identidade com usuários pré pagos anônimos abrindo acesso a serviços financeiros móveis inatingíveis habilidades angular react native javascript github actions web services local remoto requisitos inglês avançado conhecimento das linguagens de programação angular react native e javascript testes unitários apis documentation serviços web cloudsecurity e github actions experiência em conceitos e controles aplicáveis ao desenvolvimento de código seguro sast dast iast sca etc e também em sca sast ou iast conteineirização experiência nas principais referências em vulnerabilidade de software owasp sans cve etc graduação em ciência da computação segurança da informação curso de tecnólogo em ti e outros correlatos diferenciais já ter trabalhado com metodologias ágeis benefícios plano de saúde bradesco plano odontológico seguro de vida cartão bandeira mastercard flash com r dia útil gympass day off de aniversário wellness days por ano remuneração competitiva como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto regime clt categoria devops
| 1
|
122,334
| 4,833,970,479
|
IssuesEvent
|
2016-11-08 12:54:44
|
Putaitu/samoosa
|
https://api.github.com/repos/Putaitu/samoosa
|
closed
|
Attach/paste images onto issues
|
estimate:3h priority:high type:improvement
|
Use the solution from [SimpleMDE for GitHub wiki](https://github.com/Putaitu/simplemde-for-github-wiki)
Upload the images to a separate branch
|
1.0
|
Attach/paste images onto issues - Use the solution from [SimpleMDE for GitHub wiki](https://github.com/Putaitu/simplemde-for-github-wiki)
Upload the images to a separate branch
|
non_process
|
attach paste images onto issues use the solution from upload the images to a separate branch
| 0
|
4,005
| 6,934,952,021
|
IssuesEvent
|
2017-12-03 01:20:06
|
bartop/tpl
|
https://api.github.com/repos/bartop/tpl
|
closed
|
Add gcc5 and gcc6 to test matrix to Travis CI
|
new feature process upgrade
|
Currently tests are only made on gcc7, but the code should also be runnable (and tested) at gcc5 and gcc6.
|
1.0
|
Add gcc5 and gcc6 to test matrix to Travis CI - Currently tests are only made on gcc7, but the code should also be runnable (and tested) at gcc5 and gcc6.
|
process
|
add and to test matrix to travis ci currently tests are only made on but the code should also be runnable and tested at and
| 1
|
19,638
| 26,004,875,962
|
IssuesEvent
|
2022-12-20 18:21:28
|
raycast/extensions
|
https://api.github.com/repos/raycast/extensions
|
closed
|
[Kill Process] Chinese process name is wrong
|
bug extension status: stalled extension: kill-process
|
### Extension
https://github.com/raycast/extensions/tree/a8a11e9408170c13a2bf502d62bb38eeece15e77/extensions/kill-process/
### Description
It looks like a wrong stdout of `exec`, I have debug it with `{encoding: 'utf-8'}`, it does not work.
`企业微信` is Kill Process extension is `M-dM-<M^AM-dM-8M^ZM-eM->M-.M-dM-?M-!`
### Steps To Reproduce
1. Install extension Kill Process
2. Use the extension to show processes
3. Some names in Chinese is wrong
### Current Behaviour
<img width="750" alt="CleanShot 2022-11-28 at 13 45 10@2x" src="https://user-images.githubusercontent.com/6803102/204202892-902a8d64-7314-43fc-8eb7-50f533ea9a63.png">
### Expected Behaviour
Please display right Chinese name of processes.
|
1.0
|
[Kill Process] Chinese process name is wrong - ### Extension
https://github.com/raycast/extensions/tree/a8a11e9408170c13a2bf502d62bb38eeece15e77/extensions/kill-process/
### Description
It looks like a wrong stdout of `exec`, I have debug it with `{encoding: 'utf-8'}`, it does not work.
`企业微信` is Kill Process extension is `M-dM-<M^AM-dM-8M^ZM-eM->M-.M-dM-?M-!`
### Steps To Reproduce
1. Install extension Kill Process
2. Use the extension to show processes
3. Some names in Chinese is wrong
### Current Behaviour
<img width="750" alt="CleanShot 2022-11-28 at 13 45 10@2x" src="https://user-images.githubusercontent.com/6803102/204202892-902a8d64-7314-43fc-8eb7-50f533ea9a63.png">
### Expected Behaviour
Please display right Chinese name of processes.
|
process
|
chinese process name is wrong extension description it looks like a wrong stdout of exec i have debug it with encoding utf it does not work 企业微信 is kill process extension is m dm m m dm m steps to reproduce install extension kill process use the extension to show processes some names in chinese is wrong current behaviour img width alt cleanshot at src expected behaviour please display right chinese name of processes
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.