Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
17,797
| 23,724,333,166
|
IssuesEvent
|
2022-08-30 18:05:55
|
darktable-org/darktable
|
https://api.github.com/repos/darktable-org/darktable
|
closed
|
Black image on Windows after color calibration detect from edges
|
scope: image processing bug: pending scope: windows support
|
**Introduction**
I first reported on Pixls.us by two windows users. I was able to replicate on my system. It is not a feature I normally use, but it should still work. Per Pixls.us users it does not occur on ubuntu.
**Did you buy darktable from an application store ?**
Build from master
**Describe the bug/issue**
Black image on Windows after color calibration detect from edges. It happens more often with detect from edges but it does happen with detect from surface too. The image flashes like it detects the correct white balance, but then changes to black. The slides are not available to tweak.

The history shows nan values

**To Reproduce**
Open image
Go to darktable
Select (ai) Detect from image edges
OpenCL On or Off yields the same results.
**Expected behavior**
No black screen.
**Which commit introduced the error**
Unknown
**Platform**
* darktable version : 4.1.0+67~gb2d742c53
* OS : Windows 11
* Memory : 16Gb
* Graphics card : Nvidia 3060
* Graphics driver : 516.59
* OpenCL installed : Yes
* OpenCL activated : Yes (tested with it OFF)
**Additional context**
https://discuss.pixls.us/t/color-calibration-dt4-0/31706
- Can you reproduce with another darktable version(s)? I did not try
- Are the steps above reproducible with a fresh edit (i.e. after discarding history)? Yes
|
1.0
|
Black image on Windows after color calibration detect from edges - **Introduction**
I first reported on Pixls.us by two windows users. I was able to replicate on my system. It is not a feature I normally use, but it should still work. Per Pixls.us users it does not occur on ubuntu.
**Did you buy darktable from an application store ?**
Build from master
**Describe the bug/issue**
Black image on Windows after color calibration detect from edges. It happens more often with detect from edges but it does happen with detect from surface too. The image flashes like it detects the correct white balance, but then changes to black. The slides are not available to tweak.

The history shows nan values

**To Reproduce**
Open image
Go to darktable
Select (ai) Detect from image edges
OpenCL On or Off yields the same results.
**Expected behavior**
No black screen.
**Which commit introduced the error**
Unknown
**Platform**
* darktable version : 4.1.0+67~gb2d742c53
* OS : Windows 11
* Memory : 16Gb
* Graphics card : Nvidia 3060
* Graphics driver : 516.59
* OpenCL installed : Yes
* OpenCL activated : Yes (tested with it OFF)
**Additional context**
https://discuss.pixls.us/t/color-calibration-dt4-0/31706
- Can you reproduce with another darktable version(s)? I did not try
- Are the steps above reproducible with a fresh edit (i.e. after discarding history)? Yes
|
process
|
black image on windows after color calibration detect from edges introduction i first reported on pixls us by two windows users i was able to replicate on my system it is not a feature i normally use but it should still work per pixls us users it does not occur on ubuntu did you buy darktable from an application store build from master describe the bug issue black image on windows after color calibration detect from edges it happens more often with detect from edges but it does happen with detect from surface too the image flashes like it detects the correct white balance but then changes to black the slides are not available to tweak the history shows nan values to reproduce open image go to darktable select ai detect from image edges opencl on or off yields the same results expected behavior no black screen which commit introduced the error unknown platform darktable version os windows memory graphics card nvidia graphics driver opencl installed yes opencl activated yes tested with it off additional context can you reproduce with another darktable version s i did not try are the steps above reproducible with a fresh edit i e after discarding history yes
| 1
|
388,736
| 11,491,858,252
|
IssuesEvent
|
2020-02-11 19:48:55
|
opencollective/opencollective
|
https://api.github.com/repos/opencollective/opencollective
|
closed
|
API crash (heap out of memory)
|
api bug complexity → unknown priority
|
**Describe the bug**
API seems to be running out of memory.
This will cause the service to crash and drop all the connections.
**To Reproduce**
Steps to reproduce the behavior:
1. Cyclic, just wait enough time.
**Screenshots**

**Additional context**
Paper-trail log: https://my.papertrailapp.com/systems/opencollective-prod-api/events?focus=1160943645058613268&selected=1160943645058613268
|
1.0
|
API crash (heap out of memory) - **Describe the bug**
API seems to be running out of memory.
This will cause the service to crash and drop all the connections.
**To Reproduce**
Steps to reproduce the behavior:
1. Cyclic, just wait enough time.
**Screenshots**

**Additional context**
Paper-trail log: https://my.papertrailapp.com/systems/opencollective-prod-api/events?focus=1160943645058613268&selected=1160943645058613268
|
non_process
|
api crash heap out of memory describe the bug api seems to be running out of memory this will cause the service to crash and drop all the connections to reproduce steps to reproduce the behavior cyclic just wait enough time screenshots additional context paper trail log
| 0
|
5,121
| 7,891,243,314
|
IssuesEvent
|
2018-06-28 11:28:37
|
Great-Hill-Corporation/quickBlocks
|
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
|
closed
|
Improperly handle what appears to be correct input to the wrong smart contract.
|
apps-ethslurp status-inprocess type-bug
|
This transaction: [0x0c49b8ff8365c43e48ed6aeb0bbe79250b386d20fab46caf1404cc6d773b53e2](https://etherscan.io/tx/0x0c49b8ff8365c43e48ed6aeb0bbe79250b386d20fab46caf1404cc6d773b53e2) sends a valid looking input data to a non-existant function in the singularDTV code (that is, this is an end user who sent the right function signature to the wrong contract).
My code reports "Field Not Found" because this looks like (and is) a valid function signature, but the parse returns the empty string. EtherScan (incorrectly, I think) reports the name of the function on the wrong contract.
The input data is 0x861731d5. The address of smart contract is 0xaec2e87e0a235266d9c5adc9deb4b2e29b54d009. This is singular DTV's contract.
|
1.0
|
Improperly handle what appears to be correct input to the wrong smart contract. - This transaction: [0x0c49b8ff8365c43e48ed6aeb0bbe79250b386d20fab46caf1404cc6d773b53e2](https://etherscan.io/tx/0x0c49b8ff8365c43e48ed6aeb0bbe79250b386d20fab46caf1404cc6d773b53e2) sends a valid looking input data to a non-existant function in the singularDTV code (that is, this is an end user who sent the right function signature to the wrong contract).
My code reports "Field Not Found" because this looks like (and is) a valid function signature, but the parse returns the empty string. EtherScan (incorrectly, I think) reports the name of the function on the wrong contract.
The input data is 0x861731d5. The address of smart contract is 0xaec2e87e0a235266d9c5adc9deb4b2e29b54d009. This is singular DTV's contract.
|
process
|
improperly handle what appears to be correct input to the wrong smart contract this transaction sends a valid looking input data to a non existant function in the singulardtv code that is this is an end user who sent the right function signature to the wrong contract my code reports field not found because this looks like and is a valid function signature but the parse returns the empty string etherscan incorrectly i think reports the name of the function on the wrong contract the input data is the address of smart contract is this is singular dtv s contract
| 1
|
4,660
| 7,496,707,519
|
IssuesEvent
|
2018-04-08 12:26:08
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Streams for extra pipes cannot be created
|
child_process
|
* **Version**: v8.4.0
* **Platform**: `Linux aki-X220t 4.12.10-1-ARCH #1 SMP PREEMPT Wed Aug 30 12:18:42 CEST 2017 x86_64 GNU/Linux`
* **Subsystem**: process
Node.js does not provide any interface to create streams for extra pipes. Such extra pipes can be opened with the following example, which is shown in the documentation of child process.
```JavaScript
// Open an extra fd=4, to interact with programs presenting a
// startd-style interface.
spawn('prg', [], { stdio: ['pipe', null, null, null, 'pipe'] });
```
https://github.com/nodejs/node/commit/e4362728974ca7f46cbf92abc09d071d195d849f#diff-c2265ae943e23284b24d3e140f18113bR473
Node.js internally has an interface for the use case, namely `createWritableStdioStream`, in `lib/internal/process/stdio.js`.
https://github.com/nodejs/node/blob/master/lib/internal/process/stdio.js
They may be exposed to solve the issue.
|
1.0
|
Streams for extra pipes cannot be created - * **Version**: v8.4.0
* **Platform**: `Linux aki-X220t 4.12.10-1-ARCH #1 SMP PREEMPT Wed Aug 30 12:18:42 CEST 2017 x86_64 GNU/Linux`
* **Subsystem**: process
Node.js does not provide any interface to create streams for extra pipes. Such extra pipes can be opened with the following example, which is shown in the documentation of child process.
```JavaScript
// Open an extra fd=4, to interact with programs presenting a
// startd-style interface.
spawn('prg', [], { stdio: ['pipe', null, null, null, 'pipe'] });
```
https://github.com/nodejs/node/commit/e4362728974ca7f46cbf92abc09d071d195d849f#diff-c2265ae943e23284b24d3e140f18113bR473
Node.js internally has an interface for the use case, namely `createWritableStdioStream`, in `lib/internal/process/stdio.js`.
https://github.com/nodejs/node/blob/master/lib/internal/process/stdio.js
They may be exposed to solve the issue.
|
process
|
streams for extra pipes cannot be created version platform linux aki arch smp preempt wed aug cest gnu linux subsystem process node js does not provide any interface to create streams for extra pipes such extra pipes can be opened with the following example which is shown in the documentation of child process javascript open an extra fd to interact with programs presenting a startd style interface spawn prg stdio node js internally has an interface for the use case namely createwritablestdiostream in lib internal process stdio js they may be exposed to solve the issue
| 1
|
14,172
| 17,088,128,141
|
IssuesEvent
|
2021-07-08 14:14:49
|
nanoframework/Home
|
https://api.github.com/repos/nanoframework/Home
|
closed
|
MDP Fail
|
Area: Metadata Processor Status: In progress Type: Bug
|
### Details about Problem
**nanoFramework area:** Visual Studio extension
**VS version<!--(if relevant)-->:** 16.9.5
**VS extension version<!--(if relevant)-->:** 2019.6.0.25
**Target<!--(if relevant)-->:**
**Firmware image version<!--(if relevant)-->:**
**Device capabilities output<!--(if relevant)-->:**
### Description
I found another edge condition where the MetaData Processor fails
### Detailed repro steps so we can see the same problem
1. Create library with this classes:
```cs
namespace MyLibrary
{
public class Class1
{
}
public class DerivedClass : Class1
{
}
public class SecondClass
{
public Class1[] arr;
}
}
```
2. Create second library and add reference to the first one
3. Add class in second library:
```cs
using MyLibrary;
using System;
namespace MDP_test
{
public class MDP_test
{
public int GetIndex(SecondClass c, DerivedClass d)
{
//this doesn't work
return Array.IndexOf(c.arr, d);
}
}
}
```
4. try compile solution. You get error:
```
Severity Code Description Project File Line Suppression State
Error Exception minimizing assembly: Can't find entry in type reference table for MyLibrary.Class1.. MDP_Fail C:\Program Files (x86)\Microsoft Visual Studio\2019\Professional\MSBuild\nanoFramework\v1.0\NFProjectSystem.MDP.targets 226
```
### Other suggested things
Interestingly, modifying the code to such a form compiles without any problem:
```cs
public int GetIndex(SecondClass c, DerivedClass d)
{
var array = c.arr;
return Array.IndexOf(array, d);
}
```
### Expected behaviour
I expect the code to compile without the need for an additional variable.
### Screenshot
<!-- if applicable/relevant -->
<!--Very helpful if you send along a few screenshots to help visualize the issue!-->
### Additional context
Attached solution:
[MDP_Fail.zip](https://github.com/nanoframework/Home/files/6478002/MDP_Fail.zip)
### Make an effort to fix the bug
Attempt to submit a [Pull Request (PR)](https://help.github.com/articles/about-pull-requests/) that fixes the bug. Include in this PR a test that verifies the fix. If you were not able to fix the bug, a PR that illustrates your partial progress will suffice.
<!-- bug-report-tools-tag DO NOT REMOVE -->
|
1.0
|
MDP Fail -
### Details about Problem
**nanoFramework area:** Visual Studio extension
**VS version<!--(if relevant)-->:** 16.9.5
**VS extension version<!--(if relevant)-->:** 2019.6.0.25
**Target<!--(if relevant)-->:**
**Firmware image version<!--(if relevant)-->:**
**Device capabilities output<!--(if relevant)-->:**
### Description
I found another edge condition where the MetaData Processor fails
### Detailed repro steps so we can see the same problem
1. Create library with this classes:
```cs
namespace MyLibrary
{
public class Class1
{
}
public class DerivedClass : Class1
{
}
public class SecondClass
{
public Class1[] arr;
}
}
```
2. Create second library and add reference to the first one
3. Add class in second library:
```cs
using MyLibrary;
using System;
namespace MDP_test
{
public class MDP_test
{
public int GetIndex(SecondClass c, DerivedClass d)
{
//this doesn't work
return Array.IndexOf(c.arr, d);
}
}
}
```
4. try compile solution. You get error:
```
Severity Code Description Project File Line Suppression State
Error Exception minimizing assembly: Can't find entry in type reference table for MyLibrary.Class1.. MDP_Fail C:\Program Files (x86)\Microsoft Visual Studio\2019\Professional\MSBuild\nanoFramework\v1.0\NFProjectSystem.MDP.targets 226
```
### Other suggested things
Interestingly, modifying the code to such a form compiles without any problem:
```cs
public int GetIndex(SecondClass c, DerivedClass d)
{
var array = c.arr;
return Array.IndexOf(array, d);
}
```
### Expected behaviour
I expect the code to compile without the need for an additional variable.
### Screenshot
<!-- if applicable/relevant -->
<!--Very helpful if you send along a few screenshots to help visualize the issue!-->
### Additional context
Attached solution:
[MDP_Fail.zip](https://github.com/nanoframework/Home/files/6478002/MDP_Fail.zip)
### Make an effort to fix the bug
Attempt to submit a [Pull Request (PR)](https://help.github.com/articles/about-pull-requests/) that fixes the bug. Include in this PR a test that verifies the fix. If you were not able to fix the bug, a PR that illustrates your partial progress will suffice.
<!-- bug-report-tools-tag DO NOT REMOVE -->
|
process
|
mdp fail details about problem nanoframework area visual studio extension vs version vs extension version target firmware image version device capabilities output description i found another edge condition where the metadata processor fails detailed repro steps so we can see the same problem create library with this classes cs namespace mylibrary public class public class derivedclass public class secondclass public arr create second library and add reference to the first one add class in second library cs using mylibrary using system namespace mdp test public class mdp test public int getindex secondclass c derivedclass d this doesn t work return array indexof c arr d try compile solution you get error severity code description project file line suppression state error exception minimizing assembly can t find entry in type reference table for mylibrary mdp fail c program files microsoft visual studio professional msbuild nanoframework nfprojectsystem mdp targets other suggested things interestingly modifying the code to such a form compiles without any problem cs public int getindex secondclass c derivedclass d var array c arr return array indexof array d expected behaviour i expect the code to compile without the need for an additional variable screenshot additional context attached solution make an effort to fix the bug attempt to submit a that fixes the bug include in this pr a test that verifies the fix if you were not able to fix the bug a pr that illustrates your partial progress will suffice
| 1
|
9,189
| 7,610,284,002
|
IssuesEvent
|
2018-05-01 07:12:00
|
OpenSID/OpenSID
|
https://api.github.com/repos/OpenSID/OpenSID
|
closed
|
Security: tambah cara untuk mengharuskan pengguna mengganti password default
|
security
|
Ternyata banyak desa online yang masih menggunakan password default. Untuk mengamankan website mereka, perlu dicari cara untuk memaksa passwordnya diganti.
Dan juga harus mengganti password paling lama dalam waktu 3 bulan.
Perlu juga ditambah cek agar password yang digunakan harus memenuhi syarat kekuatan tertentu.
|
True
|
Security: tambah cara untuk mengharuskan pengguna mengganti password default - Ternyata banyak desa online yang masih menggunakan password default. Untuk mengamankan website mereka, perlu dicari cara untuk memaksa passwordnya diganti.
Dan juga harus mengganti password paling lama dalam waktu 3 bulan.
Perlu juga ditambah cek agar password yang digunakan harus memenuhi syarat kekuatan tertentu.
|
non_process
|
security tambah cara untuk mengharuskan pengguna mengganti password default ternyata banyak desa online yang masih menggunakan password default untuk mengamankan website mereka perlu dicari cara untuk memaksa passwordnya diganti dan juga harus mengganti password paling lama dalam waktu bulan perlu juga ditambah cek agar password yang digunakan harus memenuhi syarat kekuatan tertentu
| 0
|
1,668
| 4,307,295,667
|
IssuesEvent
|
2016-07-21 08:27:28
|
opentrials/opentrials
|
https://api.github.com/repos/opentrials/opentrials
|
opened
|
Processor for FDA's labels and approval history
|
3. In Development Processors
|
We collected this data on https://github.com/opentrials/opentrials/issues/263, now we need to:
1. Merge multiple-part PDFs into a single file;
2. Push the PDFs to S3;
3. Push the PDFs to DocumentCloud;
4. Save the (so far incomplete) document to the API database, including the links to S3 and DocumentCloud.
In this step, the document won't have its contents extracted yet (that's why we're pushing to DocumentCloud), which means it'll also won't be linked to any trial. That'll happen in the DocumentCloud's collector/processor.
|
1.0
|
Processor for FDA's labels and approval history - We collected this data on https://github.com/opentrials/opentrials/issues/263, now we need to:
1. Merge multiple-part PDFs into a single file;
2. Push the PDFs to S3;
3. Push the PDFs to DocumentCloud;
4. Save the (so far incomplete) document to the API database, including the links to S3 and DocumentCloud.
In this step, the document won't have its contents extracted yet (that's why we're pushing to DocumentCloud), which means it'll also won't be linked to any trial. That'll happen in the DocumentCloud's collector/processor.
|
process
|
processor for fda s labels and approval history we collected this data on now we need to merge multiple part pdfs into a single file push the pdfs to push the pdfs to documentcloud save the so far incomplete document to the api database including the links to and documentcloud in this step the document won t have its contents extracted yet that s why we re pushing to documentcloud which means it ll also won t be linked to any trial that ll happen in the documentcloud s collector processor
| 1
|
2,473
| 5,246,311,714
|
IssuesEvent
|
2017-02-01 09:10:01
|
jlm2017/jlm-video-subtitles
|
https://api.github.com/repos/jlm2017/jlm-video-subtitles
|
reopened
|
[Subtitles] [FR] #RDLS16 - VENISE, DROIT À L'IVG, JEU VIDÉO, BD, NÉGAWATT, PRIMAIRE PS, FILLON
|
Language: French Process: [4] Ready for review (2)
|
# Video title
#RDLS16 - VENISE, DROIT À L'IVG, JEU VIDÉO, BD, NÉGAWATT, PRIMAIRE PS, FILLON
# URL
https://www.youtube.com/watch?v=6HG7IsmGD38
# Youtube subtitles language
Français
# Duration
42:25
# Subtitles URL
https://www.youtube.com/timedtext_editor?v=6HG7IsmGD38&ui=hd&action_mde_edit_form=1&ref=player&bl=vmp&lang=fr&tab=captions
|
1.0
|
[Subtitles] [FR] #RDLS16 - VENISE, DROIT À L'IVG, JEU VIDÉO, BD, NÉGAWATT, PRIMAIRE PS, FILLON - # Video title
#RDLS16 - VENISE, DROIT À L'IVG, JEU VIDÉO, BD, NÉGAWATT, PRIMAIRE PS, FILLON
# URL
https://www.youtube.com/watch?v=6HG7IsmGD38
# Youtube subtitles language
Français
# Duration
42:25
# Subtitles URL
https://www.youtube.com/timedtext_editor?v=6HG7IsmGD38&ui=hd&action_mde_edit_form=1&ref=player&bl=vmp&lang=fr&tab=captions
|
process
|
venise droit à l ivg jeu vidéo bd négawatt primaire ps fillon video title venise droit à l ivg jeu vidéo bd négawatt primaire ps fillon url youtube subtitles language français duration subtitles url
| 1
|
206,107
| 16,019,945,581
|
IssuesEvent
|
2021-04-20 21:14:59
|
ess-dive-community/essdive-model-data-archiving-guidelines
|
https://api.github.com/repos/ess-dive-community/essdive-model-data-archiving-guidelines
|
opened
|
Notes after initial uploads to repository
|
documentation
|
**Submitter**: Rob Crystal-Ornelas
I suggest the following changes:
After reading through the associated manuscript, and adding content to the GitHub repo, there are a few other additions we should consider
- [ ] Which usage license should we select. In ESS-DIVE you’ll have the option to select CC BY 1.0 or CC BY 4.0. Then, I can create a .txt file for the license so that is clearly visible in the GitHub repo.
- [ ] Which document should we encourage users to cite on our README page? The manuscript that is soon to be submitted?
- [ ] I created a small set of GitHub issue templates that can help to organize feedback. Any other templates that I should add or are the 2 that I have OK for now?
|
1.0
|
Notes after initial uploads to repository - **Submitter**: Rob Crystal-Ornelas
I suggest the following changes:
After reading through the associated manuscript, and adding content to the GitHub repo, there are a few other additions we should consider
- [ ] Which usage license should we select. In ESS-DIVE you’ll have the option to select CC BY 1.0 or CC BY 4.0. Then, I can create a .txt file for the license so that is clearly visible in the GitHub repo.
- [ ] Which document should we encourage users to cite on our README page? The manuscript that is soon to be submitted?
- [ ] I created a small set of GitHub issue templates that can help to organize feedback. Any other templates that I should add or are the 2 that I have OK for now?
|
non_process
|
notes after initial uploads to repository submitter rob crystal ornelas i suggest the following changes after reading through the associated manuscript and adding content to the github repo there are a few other additions we should consider which usage license should we select in ess dive you’ll have the option to select cc by or cc by then i can create a txt file for the license so that is clearly visible in the github repo which document should we encourage users to cite on our readme page the manuscript that is soon to be submitted i created a small set of github issue templates that can help to organize feedback any other templates that i should add or are the that i have ok for now
| 0
|
93,476
| 26,963,347,166
|
IssuesEvent
|
2023-02-08 20:03:20
|
kobotoolbox/kpi
|
https://api.github.com/repos/kobotoolbox/kpi
|
reopened
|
Improve automatic naming of question names and response values
|
enhancement formbuilder moved
|
The current naming only works with Roman ascii characters, all other label characters are replaced by underscores (and 0s). This leads to unreadable XLSForms and datasets.
Possible options:
1. Find a better way to replace non-ASCII characters for question names
2. Don't replace response names - keep them as-is (not necessary for pyxform to use ascii-only anyway)

|
1.0
|
Improve automatic naming of question names and response values - The current naming only works with Roman ascii characters, all other label characters are replaced by underscores (and 0s). This leads to unreadable XLSForms and datasets.
Possible options:
1. Find a better way to replace non-ASCII characters for question names
2. Don't replace response names - keep them as-is (not necessary for pyxform to use ascii-only anyway)

|
non_process
|
improve automatic naming of question names and response values the current naming only works with roman ascii characters all other label characters are replaced by underscores and this leads to unreadable xlsforms and datasets possible options find a better way to replace non ascii characters for question names don t replace response names keep them as is not necessary for pyxform to use ascii only anyway
| 0
|
232
| 2,659,574,961
|
IssuesEvent
|
2015-03-18 21:50:07
|
joyent/node
|
https://api.github.com/repos/joyent/node
|
opened
|
child_process.exec doc is inconsistent with actual implementation
|
child_process doc P-3
|
[child_process.exec' documentation](https://nodejs.org/api/child_process.html#child_process_child_process_exec_command_options_callback) states that the `stdout` and `stderr` parameters are instances of `Buffer` by default, but it's actually not the case. They are instances of `Buffer` only if `{ encoding: null}` is passed as options.
The documentation of v0.12.x should be fixed to reflect this behavior.
The original intent was that `stdout` and `stderr` parameters would be instances of `Buffer`. If that still holds true, then the change in behavior should be done in master post v0.12.x.
Adding to milestone 0.12.2 as it's a quick change.
|
1.0
|
child_process.exec doc is inconsistent with actual implementation - [child_process.exec' documentation](https://nodejs.org/api/child_process.html#child_process_child_process_exec_command_options_callback) states that the `stdout` and `stderr` parameters are instances of `Buffer` by default, but it's actually not the case. They are instances of `Buffer` only if `{ encoding: null}` is passed as options.
The documentation of v0.12.x should be fixed to reflect this behavior.
The original intent was that `stdout` and `stderr` parameters would be instances of `Buffer`. If that still holds true, then the change in behavior should be done in master post v0.12.x.
Adding to milestone 0.12.2 as it's a quick change.
|
process
|
child process exec doc is inconsistent with actual implementation states that the stdout and stderr parameters are instances of buffer by default but it s actually not the case they are instances of buffer only if encoding null is passed as options the documentation of x should be fixed to reflect this behavior the original intent was that stdout and stderr parameters would be instances of buffer if that still holds true then the change in behavior should be done in master post x adding to milestone as it s a quick change
| 1
|
3,615
| 6,653,858,794
|
IssuesEvent
|
2017-09-29 10:07:16
|
nuclio/nuclio
|
https://api.github.com/repos/nuclio/nuclio
|
closed
|
Missing event fields should be implemented and tested
|
area/processor priority/high
|
Both Golang and Python functions should support receiving:
1. Fields (query arguments in HTTP - e.g. `?foo=1&moo=2`)
2. Headers
3. Path
In addition, both golang and python functions should support returning headers.
The above should be implemented (some may be missing) and have proper integration tests to test this.
|
1.0
|
Missing event fields should be implemented and tested - Both Golang and Python functions should support receiving:
1. Fields (query arguments in HTTP - e.g. `?foo=1&moo=2`)
2. Headers
3. Path
In addition, both golang and python functions should support returning headers.
The above should be implemented (some may be missing) and have proper integration tests to test this.
|
process
|
missing event fields should be implemented and tested both golang and python functions should support receiving fields query arguments in http e g foo moo headers path in addition both golang and python functions should support returning headers the above should be implemented some may be missing and have proper integration tests to test this
| 1
|
22,365
| 31,080,280,281
|
IssuesEvent
|
2023-08-13 01:37:33
|
h4sh5/npm-auto-scanner
|
https://api.github.com/repos/h4sh5/npm-auto-scanner
|
opened
|
rn-wind-lint 1.0.0 has 74 guarddog issues
|
npm-install-script shady-links npm-silent-process-execution
|
```{"npm-install-script":[{"code":" \"prepare\": \"npm run build\",","location":"package/node_modules/@eslint/eslintrc/package.json:32","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"taskr release\",","location":"package/node_modules/@expo/cli/package.json:15","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/config/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/config-plugins/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/config-types/package.json:11","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/dev-server/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/env/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/image-utils/package.json:10","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/json-file/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/metro-config/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/osascript/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/package-manager/package.json:10","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn build\",","location":"package/node_modules/@expo/plist/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/prebuild-config/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"tsc\",","location":"package/node_modules/@expo/rudder-sdk-node/package.json:25","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"rm -rf build \u0026\u0026 tsc\",","location":"package/node_modules/@expo/sdk-runtime-versions/package.json:13","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/spawn-async/package.json:16","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare \u0026\u0026 npm run generate-lazy \u0026\u0026 npm run copy-vendor\",","location":"package/node_modules/@expo/vector-icons/package.json:17","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/xcpretty/package.json:20","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run build\",","location":"package/node_modules/@humanwhocodes/module-importer/package.json:34","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run build\"","location":"package/node_modules/@react-native/codegen/package.json:14","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install\"","location":"package/node_modules/@sinonjs/commons/package.json:20","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install\"","location":"package/node_modules/@sinonjs/fake-timers/package.json:29","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"node ../../scripts/prepare/index.js\",","location":"package/node_modules/@urql/core/package.json:52","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"node ../../scripts/prepare/index.js\",","location":"package/node_modules/@urql/exchange-retry/package.json:46","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"cd ..; npm run build:main\"","location":"package/node_modules/acorn/package.json:45","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn clean \u0026\u0026 yarn build\"","location":"package/node_modules/ansi-fragments/package.json:28","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run clean \u0026\u0026 npm run gen \u0026\u0026 npm run build\",","location":"package/node_modules/ast-types/package.json:36","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install\",","location":"package/node_modules/axe-core/package.json:106","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install\"","location":"package/node_modules/ci-info/package.json:34","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run compile\"","location":"package/node_modules/cliui/package.json:29","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run build\"","location":"package/node_modules/colorette/package.json:33","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install\",","location":"package/node_modules/cross-fetch/package.json:12","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"cp error-stack-parser.js dist/ \u0026\u0026 uglifyjs node_modules/stackframe/stackframe.js error-stack-parser.js -o dist/error-stack-parser.min.js --compress --mangle --source-map \\\"url=error-stack-parser.min.js.map\\\"\"","location":"package/node_modules/error-stack-parser/package.json:61","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npx husky install\",","location":"package/node_modules/eslint-plugin-n/package.json:62","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"simple-git-hooks \u0026\u0026 yarn-deduplicate --strategy fewer || exit 0\",","location":"package/node_modules/eslint-plugin-prettier/package.json:33","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install\",","location":"package/node_modules/eslint-plugin-promise/package.json:27","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"postinstall\": \"lerna bootstrap\",","location":"package/node_modules/eslint-plugin-react/node_modules/resolve/test/resolver/multirepo/package.json:8","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run build\",","location":"package/node_modules/eslint-visitor-keys/package.json:55","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-application/package.json:12","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-asset/package.json:16","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-constants/package.json:13","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-file-system/package.json:13","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-font/package.json:13","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-keep-awake/package.json:13","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-modules-autolinking/package.json:12","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-modules-core/package.json:12","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-status-bar/package.json:12","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run build\",","location":"package/node_modules/fbjs/package.json:11","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"tsc\",","location":"package/node_modules/get-caller-file/package.json:15","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"tsdx build\"","location":"package/node_modules/hermes-profile-transformer/package.json:18","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"patch-package\",","location":"package/node_modules/lightningcss/package.json:79","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run compile\",","location":"package/node_modules/logkitty/node_modules/yargs/package.json:65","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn build:source \u0026\u0026 yarn build:def\",","location":"package/node_modules/logkitty/package.json:38","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"node src/build.js \u0026\u0026 runmd --output README.md src/README_js.md\",","location":"package/node_modules/mime/package.json:36","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run build\",","location":"package/node_modules/node-fetch/package.json:19","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"del-cli dist \u0026\u0026 BABEL_ENV=publish babel src --out-dir dist --ignore /__tests__/\",","location":"package/node_modules/postcss-selector-parser/package.json:37","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run build\",","location":"package/node_modules/recast/package.json:32","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"postinstall\": \"lerna bootstrap\",","location":"package/node_modules/resolve/test/resolver/multirepo/package.json:8","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"node src/build.js\",","location":"package/node_modules/send/node_modules/mime/package.json:29","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install\",","location":"package/node_modules/simple-plist/package.json:28","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"cp stackframe.js dist/ \u0026\u0026 uglifyjs stackframe.js -o dist/stackframe.min.js --compress --mangle --source-map \\\"url=stackframe.min.js.map\\\"\"","location":"package/node_modules/stackframe/package.json:57","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"make build\",","location":"package/node_modules/structured-headers/package.json:8","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"node ./scripts/transpile-to-esm.js\",","location":"package/node_modules/tar/node_modules/minipass/package.json:35","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run build\",","location":"package/node_modules/terser/package.json:72","message":"The package.json has a script automatically running when the package is installed"},{"code":"{\"name\":\"type-detect\",\"description\":\"Improved typeof detection for node.js and the browser.\",\"keywords\":[\"type\",\"typeof\",\"types\"],\"license\":\"MIT\",\"author\":\"Jake Luer \u003cjake@alogicalparadox.com\u003e (http://alogicalparadox.com)\",\"contributors\":[\"...\":\"4.0.8\"}","location":"package/node_modules/type-detect/package.json:1","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"runmd --output=README.md README_js.md\"","location":"package/node_modules/uuid/package.json:33","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"make dist/fetch.umd.js dist/fetch.umd.js.flow\",","location":"package/node_modules/whatwg-fetch/package.json:36","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run compile\"","location":"package/node_modules/y18n/package.json:41","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run compile\",","location":"package/node_modules/yargs/package.json:95","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run compile\"","location":"package/node_modules/yargs-parser/package.json:34","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const emulatorProcess = (0, _childProcess).spawn(whichEmulator(), [\n `@${device.name}`\n ], {\n stdio: \"ignore\",\n detached: true\n });","location":"package/node_modules/@expo/cli/build/src/start/platforms/android/emulator.js:73","message":"This package is silently executing another executable"}],"shady-links":[{"code":" return yield password_prompt_1.default('devcert password (http://bit.ly/devcert-what-password?):');","location":"package/node_modules/@expo/devcert/dist/user-interface.js:9","message":"This package contains an URL to a domain with a suspicious extension"},{"code":" return await passwordPrompt('devcert password (http://bit.ly/devcert-what-password?):');","location":"package/node_modules/@expo/devcert/src/user-interface.ts:15","message":"This package contains an URL to a domain with a suspicious extension"}]}```
|
1.0
|
rn-wind-lint 1.0.0 has 74 guarddog issues - ```{"npm-install-script":[{"code":" \"prepare\": \"npm run build\",","location":"package/node_modules/@eslint/eslintrc/package.json:32","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"taskr release\",","location":"package/node_modules/@expo/cli/package.json:15","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/config/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/config-plugins/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/config-types/package.json:11","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/dev-server/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/env/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/image-utils/package.json:10","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/json-file/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/metro-config/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/osascript/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/package-manager/package.json:10","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn build\",","location":"package/node_modules/@expo/plist/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/prebuild-config/package.json:9","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"tsc\",","location":"package/node_modules/@expo/rudder-sdk-node/package.json:25","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"rm -rf build \u0026\u0026 tsc\",","location":"package/node_modules/@expo/sdk-runtime-versions/package.json:13","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/spawn-async/package.json:16","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare \u0026\u0026 npm run generate-lazy \u0026\u0026 npm run copy-vendor\",","location":"package/node_modules/@expo/vector-icons/package.json:17","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run clean \u0026\u0026 yarn build\",","location":"package/node_modules/@expo/xcpretty/package.json:20","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run build\",","location":"package/node_modules/@humanwhocodes/module-importer/package.json:34","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run build\"","location":"package/node_modules/@react-native/codegen/package.json:14","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install\"","location":"package/node_modules/@sinonjs/commons/package.json:20","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install\"","location":"package/node_modules/@sinonjs/fake-timers/package.json:29","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"node ../../scripts/prepare/index.js\",","location":"package/node_modules/@urql/core/package.json:52","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"node ../../scripts/prepare/index.js\",","location":"package/node_modules/@urql/exchange-retry/package.json:46","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"cd ..; npm run build:main\"","location":"package/node_modules/acorn/package.json:45","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn clean \u0026\u0026 yarn build\"","location":"package/node_modules/ansi-fragments/package.json:28","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run clean \u0026\u0026 npm run gen \u0026\u0026 npm run build\",","location":"package/node_modules/ast-types/package.json:36","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install\",","location":"package/node_modules/axe-core/package.json:106","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install\"","location":"package/node_modules/ci-info/package.json:34","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run compile\"","location":"package/node_modules/cliui/package.json:29","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run build\"","location":"package/node_modules/colorette/package.json:33","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install\",","location":"package/node_modules/cross-fetch/package.json:12","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"cp error-stack-parser.js dist/ \u0026\u0026 uglifyjs node_modules/stackframe/stackframe.js error-stack-parser.js -o dist/error-stack-parser.min.js --compress --mangle --source-map \\\"url=error-stack-parser.min.js.map\\\"\"","location":"package/node_modules/error-stack-parser/package.json:61","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npx husky install\",","location":"package/node_modules/eslint-plugin-n/package.json:62","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"simple-git-hooks \u0026\u0026 yarn-deduplicate --strategy fewer || exit 0\",","location":"package/node_modules/eslint-plugin-prettier/package.json:33","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install\",","location":"package/node_modules/eslint-plugin-promise/package.json:27","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"postinstall\": \"lerna bootstrap\",","location":"package/node_modules/eslint-plugin-react/node_modules/resolve/test/resolver/multirepo/package.json:8","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run build\",","location":"package/node_modules/eslint-visitor-keys/package.json:55","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-application/package.json:12","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-asset/package.json:16","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-constants/package.json:13","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-file-system/package.json:13","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-font/package.json:13","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-keep-awake/package.json:13","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-modules-autolinking/package.json:12","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-modules-core/package.json:12","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"expo-module prepare\",","location":"package/node_modules/expo-status-bar/package.json:12","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn run build\",","location":"package/node_modules/fbjs/package.json:11","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"tsc\",","location":"package/node_modules/get-caller-file/package.json:15","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"tsdx build\"","location":"package/node_modules/hermes-profile-transformer/package.json:18","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"patch-package\",","location":"package/node_modules/lightningcss/package.json:79","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run compile\",","location":"package/node_modules/logkitty/node_modules/yargs/package.json:65","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"yarn build:source \u0026\u0026 yarn build:def\",","location":"package/node_modules/logkitty/package.json:38","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"node src/build.js \u0026\u0026 runmd --output README.md src/README_js.md\",","location":"package/node_modules/mime/package.json:36","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run build\",","location":"package/node_modules/node-fetch/package.json:19","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"del-cli dist \u0026\u0026 BABEL_ENV=publish babel src --out-dir dist --ignore /__tests__/\",","location":"package/node_modules/postcss-selector-parser/package.json:37","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run build\",","location":"package/node_modules/recast/package.json:32","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"postinstall\": \"lerna bootstrap\",","location":"package/node_modules/resolve/test/resolver/multirepo/package.json:8","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"node src/build.js\",","location":"package/node_modules/send/node_modules/mime/package.json:29","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"husky install\",","location":"package/node_modules/simple-plist/package.json:28","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"cp stackframe.js dist/ \u0026\u0026 uglifyjs stackframe.js -o dist/stackframe.min.js --compress --mangle --source-map \\\"url=stackframe.min.js.map\\\"\"","location":"package/node_modules/stackframe/package.json:57","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"make build\",","location":"package/node_modules/structured-headers/package.json:8","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"node ./scripts/transpile-to-esm.js\",","location":"package/node_modules/tar/node_modules/minipass/package.json:35","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run build\",","location":"package/node_modules/terser/package.json:72","message":"The package.json has a script automatically running when the package is installed"},{"code":"{\"name\":\"type-detect\",\"description\":\"Improved typeof detection for node.js and the browser.\",\"keywords\":[\"type\",\"typeof\",\"types\"],\"license\":\"MIT\",\"author\":\"Jake Luer \u003cjake@alogicalparadox.com\u003e (http://alogicalparadox.com)\",\"contributors\":[\"...\":\"4.0.8\"}","location":"package/node_modules/type-detect/package.json:1","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"runmd --output=README.md README_js.md\"","location":"package/node_modules/uuid/package.json:33","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"make dist/fetch.umd.js dist/fetch.umd.js.flow\",","location":"package/node_modules/whatwg-fetch/package.json:36","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run compile\"","location":"package/node_modules/y18n/package.json:41","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run compile\",","location":"package/node_modules/yargs/package.json:95","message":"The package.json has a script automatically running when the package is installed"},{"code":" \"prepare\": \"npm run compile\"","location":"package/node_modules/yargs-parser/package.json:34","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" const emulatorProcess = (0, _childProcess).spawn(whichEmulator(), [\n `@${device.name}`\n ], {\n stdio: \"ignore\",\n detached: true\n });","location":"package/node_modules/@expo/cli/build/src/start/platforms/android/emulator.js:73","message":"This package is silently executing another executable"}],"shady-links":[{"code":" return yield password_prompt_1.default('devcert password (http://bit.ly/devcert-what-password?):');","location":"package/node_modules/@expo/devcert/dist/user-interface.js:9","message":"This package contains an URL to a domain with a suspicious extension"},{"code":" return await passwordPrompt('devcert password (http://bit.ly/devcert-what-password?):');","location":"package/node_modules/@expo/devcert/src/user-interface.ts:15","message":"This package contains an URL to a domain with a suspicious extension"}]}```
|
process
|
rn wind lint has guarddog issues npm install script license mit author jake luer alogicalparadox com npm silent process execution n stdio ignore n detached true n location package node modules expo cli build src start platforms android emulator js message this package is silently executing another executable shady links
| 1
|
511,768
| 14,881,102,934
|
IssuesEvent
|
2021-01-20 10:03:56
|
factly/vidcheck
|
https://api.github.com/repos/factly/vidcheck
|
closed
|
Modify fields in analysis entity in the backend from text to JSON
|
good first issue priority:medium
|
- [ ] claim
- [ ] fact
- [ ] review_sources
- [ ] claim_sources
add the following fields
- [ ] claimant - depends on https://github.com/factly/vidcheck/issues/21
- [ ] is_claim (boolean)
|
1.0
|
Modify fields in analysis entity in the backend from text to JSON - - [ ] claim
- [ ] fact
- [ ] review_sources
- [ ] claim_sources
add the following fields
- [ ] claimant - depends on https://github.com/factly/vidcheck/issues/21
- [ ] is_claim (boolean)
|
non_process
|
modify fields in analysis entity in the backend from text to json claim fact review sources claim sources add the following fields claimant depends on is claim boolean
| 0
|
256,372
| 19,410,166,151
|
IssuesEvent
|
2021-12-20 08:40:42
|
homedork-se2/Mocked-client
|
https://api.github.com/repos/homedork-se2/Mocked-client
|
closed
|
Bliss
|
documentation enhancement question
|
implementation of R7 - bliss expressions. A simple translation mode for certain labels or text fields where the text is replaced by bliss symbols. Applicable to android and web.
|
1.0
|
Bliss - implementation of R7 - bliss expressions. A simple translation mode for certain labels or text fields where the text is replaced by bliss symbols. Applicable to android and web.
|
non_process
|
bliss implementation of bliss expressions a simple translation mode for certain labels or text fields where the text is replaced by bliss symbols applicable to android and web
| 0
|
772,411
| 27,121,221,649
|
IssuesEvent
|
2023-02-15 23:06:21
|
trufflesuite/ganache-ui
|
https://api.github.com/repos/trufflesuite/ganache-ui
|
closed
|
After upgrading ganache to 1.2.3 it closes unexpectedly while running test suites
|
needs validation priority-high
|
Error: nonce generation function failed or private key is invalid
at Object.exports.ecsign (/node_modules/ganache-core/node_modules/ethereumjs-util/dist/index.js:369:23)
at Transaction.sign (/node_modules/ganache-core/node_modules/ethereumjs-tx/es5/index.js:252:23)
at Transaction.sign (/node_modules/ganache-core/lib/utils/transaction.js:260:10)
at Transaction.hash (/node_modules/ganache-core/node_modules/ethereumjs-tx/fake.js:66:12)
at Transaction.fakeHash [as hash] (/node_modules/ganache-core/lib/utils/transaction.js:17:51)
at RuntimeError.combine (/node_modules/ganache-core/lib/utils/runtimeerror.js:42:30)
at new RuntimeError (/node_modules/ganache-core/lib/utils/runtimeerror.js:19:8)
at Function.RuntimeError.fromResults (/node_modules/ganache-core/lib/utils/runtimeerror.js:89:13)
at /node_modules/ganache-core/lib/blockchain_double.js:527:30
at /node_modules/ganache-core/node_modules/ethereumjs-vm/dist/runTx.js:44:5
at /node_modules/ganache-core/node_modules/async/dist/async.js:3888:9
at /node_modules/ganache-core/node_modules/async/dist/async.js:473:16
at replenish (/node_modules/ganache-core/node_modules/async/dist/async.js:1006:25)
at iterateeCallback (/node_modules/ganache-core/node_modules/async/dist/async.js:995:17)
at /node_modules/ganache-core/node_modules/async/dist/async.js:969:16
at /node_modules/ganache-core/node_modules/async/dist/async.js:3885:13
at /node_modules/ganache-core/node_modules/ethereumjs-vm/dist/runTx.js:194:7
at /node_modules/ganache-core/node_modules/ethereumjs-vm/dist/cache.js:110:7
at /node_modules/ganache-core/node_modules/async/dist/async.js:473:16
at replenish (/node_modules/ganache-core/node_modules/async/dist/async.js:1006:25)
at /node_modules/ganache-core/node_modules/async/dist/async.js:1016:9
at eachLimit$1 (/node_modules/ganache-core/node_modules/async/dist/async.js:3196:24)
|
1.0
|
After upgrading ganache to 1.2.3 it closes unexpectedly while running test suites - Error: nonce generation function failed or private key is invalid
at Object.exports.ecsign (/node_modules/ganache-core/node_modules/ethereumjs-util/dist/index.js:369:23)
at Transaction.sign (/node_modules/ganache-core/node_modules/ethereumjs-tx/es5/index.js:252:23)
at Transaction.sign (/node_modules/ganache-core/lib/utils/transaction.js:260:10)
at Transaction.hash (/node_modules/ganache-core/node_modules/ethereumjs-tx/fake.js:66:12)
at Transaction.fakeHash [as hash] (/node_modules/ganache-core/lib/utils/transaction.js:17:51)
at RuntimeError.combine (/node_modules/ganache-core/lib/utils/runtimeerror.js:42:30)
at new RuntimeError (/node_modules/ganache-core/lib/utils/runtimeerror.js:19:8)
at Function.RuntimeError.fromResults (/node_modules/ganache-core/lib/utils/runtimeerror.js:89:13)
at /node_modules/ganache-core/lib/blockchain_double.js:527:30
at /node_modules/ganache-core/node_modules/ethereumjs-vm/dist/runTx.js:44:5
at /node_modules/ganache-core/node_modules/async/dist/async.js:3888:9
at /node_modules/ganache-core/node_modules/async/dist/async.js:473:16
at replenish (/node_modules/ganache-core/node_modules/async/dist/async.js:1006:25)
at iterateeCallback (/node_modules/ganache-core/node_modules/async/dist/async.js:995:17)
at /node_modules/ganache-core/node_modules/async/dist/async.js:969:16
at /node_modules/ganache-core/node_modules/async/dist/async.js:3885:13
at /node_modules/ganache-core/node_modules/ethereumjs-vm/dist/runTx.js:194:7
at /node_modules/ganache-core/node_modules/ethereumjs-vm/dist/cache.js:110:7
at /node_modules/ganache-core/node_modules/async/dist/async.js:473:16
at replenish (/node_modules/ganache-core/node_modules/async/dist/async.js:1006:25)
at /node_modules/ganache-core/node_modules/async/dist/async.js:1016:9
at eachLimit$1 (/node_modules/ganache-core/node_modules/async/dist/async.js:3196:24)
|
non_process
|
after upgrading ganache to it closes unexpectedly while running test suites error nonce generation function failed or private key is invalid at object exports ecsign node modules ganache core node modules ethereumjs util dist index js at transaction sign node modules ganache core node modules ethereumjs tx index js at transaction sign node modules ganache core lib utils transaction js at transaction hash node modules ganache core node modules ethereumjs tx fake js at transaction fakehash node modules ganache core lib utils transaction js at runtimeerror combine node modules ganache core lib utils runtimeerror js at new runtimeerror node modules ganache core lib utils runtimeerror js at function runtimeerror fromresults node modules ganache core lib utils runtimeerror js at node modules ganache core lib blockchain double js at node modules ganache core node modules ethereumjs vm dist runtx js at node modules ganache core node modules async dist async js at node modules ganache core node modules async dist async js at replenish node modules ganache core node modules async dist async js at iterateecallback node modules ganache core node modules async dist async js at node modules ganache core node modules async dist async js at node modules ganache core node modules async dist async js at node modules ganache core node modules ethereumjs vm dist runtx js at node modules ganache core node modules ethereumjs vm dist cache js at node modules ganache core node modules async dist async js at replenish node modules ganache core node modules async dist async js at node modules ganache core node modules async dist async js at eachlimit node modules ganache core node modules async dist async js
| 0
|
8,197
| 11,394,705,753
|
IssuesEvent
|
2020-01-30 09:54:18
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
MP: GO:0140404 | effector-mediated modulation of host innate immune response by symbiont
|
multi-species process quick fix
|
GO:0140404 | effector-mediated modulation of host innate immune response by symbiont |
should be a descendant of
GO:0140415 | effector-mediated modulation of host defenses by symbiont
|
1.0
|
MP: GO:0140404 | effector-mediated modulation of host innate immune response by symbiont - GO:0140404 | effector-mediated modulation of host innate immune response by symbiont |
should be a descendant of
GO:0140415 | effector-mediated modulation of host defenses by symbiont
|
process
|
mp go effector mediated modulation of host innate immune response by symbiont go effector mediated modulation of host innate immune response by symbiont should be a descendant of go effector mediated modulation of host defenses by symbiont
| 1
|
10,306
| 4,039,113,089
|
IssuesEvent
|
2016-05-20 02:13:50
|
dotnet/coreclr
|
https://api.github.com/repos/dotnet/coreclr
|
closed
|
TailcallVerifyWithPrefix test has assertion failure on Windows 64
|
2 - In Progress CodeGen Test-Issue
|
The JIT\opt\Tailcall\TailcallVerifyWithPrefix\TailcallVerifyWithPrefix.cmd test
fails with this message:
```
Assert failure(PID 54344 [0x0000d448], Thread: 61376 [0xefc0]):
Assertion failed 'nextStmtExpr->gtOper == GT_RETURN'
in 'TailcallVerify.Condition2:Caller1()' (IL size 55)
File: coreclr\src\jit\morph.cpp Line: 6671
Image: coreclr\bin\tests\Windows_NT.x64.debug\Tests\Core_Root\CoreRun.exe
```
|
1.0
|
TailcallVerifyWithPrefix test has assertion failure on Windows 64 - The JIT\opt\Tailcall\TailcallVerifyWithPrefix\TailcallVerifyWithPrefix.cmd test
fails with this message:
```
Assert failure(PID 54344 [0x0000d448], Thread: 61376 [0xefc0]):
Assertion failed 'nextStmtExpr->gtOper == GT_RETURN'
in 'TailcallVerify.Condition2:Caller1()' (IL size 55)
File: coreclr\src\jit\morph.cpp Line: 6671
Image: coreclr\bin\tests\Windows_NT.x64.debug\Tests\Core_Root\CoreRun.exe
```
|
non_process
|
tailcallverifywithprefix test has assertion failure on windows the jit opt tailcall tailcallverifywithprefix tailcallverifywithprefix cmd test fails with this message assert failure pid thread assertion failed nextstmtexpr gtoper gt return in tailcallverify il size file coreclr src jit morph cpp line image coreclr bin tests windows nt debug tests core root corerun exe
| 0
|
445,000
| 31,160,272,494
|
IssuesEvent
|
2023-08-16 15:31:45
|
ita-social-projects/dokazovi-rqsb
|
https://api.github.com/repos/ita-social-projects/dokazovi-rqsb
|
opened
|
[Test for Story #8] Verify that the publication has status ‘Опубліковано’ on the ‘Адміністрування’ page.
|
documentation test case
|
**Story link**
[#8 Story](https://github.com/ita-social-projects/dokazovi-rqsb/issues/15)
### Status:
Not executed
### Title:
Verify that the publication has status ‘Опубліковано’ on the ‘Адміністрування’ page.
### Description:
This test case verifies that the Admin can see publications with status ‘Опубліковано’ on the 'Адміністрування' page.
### Pre-conditions:
1. User logged in as Admin.
2. The page 'Адміністрування' is opened.
3. Must be one publication with status “На модерації”.
Step № | Test Steps | Test data | Expected result | Status (Pass/Fail/Not executed) | Notes
------------ | ------------ | ------------ | ------------ | ------------ | ------------
1 | Click on the 'Action button group' in publication row with status 'На модерації' | | The dropdown bar is appeared | Not executed |
2 | Verify if dropdown bar contains the appropriate button | | Dropdown bar contains follow button ‘Опубліковано’ | Not executed |
3 | Click on the button ‘Опублікувати’ of 'Action button group' | | The publication changed status to ‘Опубліковано’ | Not executed |
4 | Verify that published publication displayed for all users on the 'Матеріали’ page. | | That publication is displayed for all users on the 'Матеріали’ page. | Not executed |
### [Gantt Chart](https://docs.google.com/spreadsheets/d/1kK8bHKrlzOq2UbabHmQjlXlhPy_d3pRR5RD_akOjDsM/edit#gid=775577959)
|
1.0
|
[Test for Story #8] Verify that the publication has status ‘Опубліковано’ on the ‘Адміністрування’ page. - **Story link**
[#8 Story](https://github.com/ita-social-projects/dokazovi-rqsb/issues/15)
### Status:
Not executed
### Title:
Verify that the publication has status ‘Опубліковано’ on the ‘Адміністрування’ page.
### Description:
This test case verifies that the Admin can see publications with status ‘Опубліковано’ on the 'Адміністрування' page.
### Pre-conditions:
1. User logged in as Admin.
2. The page 'Адміністрування' is opened.
3. Must be one publication with status “На модерації”.
Step № | Test Steps | Test data | Expected result | Status (Pass/Fail/Not executed) | Notes
------------ | ------------ | ------------ | ------------ | ------------ | ------------
1 | Click on the 'Action button group' in publication row with status 'На модерації' | | The dropdown bar is appeared | Not executed |
2 | Verify if dropdown bar contains the appropriate button | | Dropdown bar contains follow button ‘Опубліковано’ | Not executed |
3 | Click on the button ‘Опублікувати’ of 'Action button group' | | The publication changed status to ‘Опубліковано’ | Not executed |
4 | Verify that published publication displayed for all users on the 'Матеріали’ page. | | That publication is displayed for all users on the 'Матеріали’ page. | Not executed |
### [Gantt Chart](https://docs.google.com/spreadsheets/d/1kK8bHKrlzOq2UbabHmQjlXlhPy_d3pRR5RD_akOjDsM/edit#gid=775577959)
|
non_process
|
verify that the publication has status ‘опубліковано’ on the ‘адміністрування’ page story link status not executed title verify that the publication has status ‘опубліковано’ on the ‘адміністрування’ page description this test case verifies that the admin can see publications with status ‘опубліковано’ on the адміністрування page pre conditions user logged in as admin the page адміністрування is opened must be one publication with status “на модерації” step № test steps test data expected result status pass fail not executed notes click on the action button group in publication row with status на модерації the dropdown bar is appeared not executed verify if dropdown bar contains the appropriate button dropdown bar contains follow button ‘опубліковано’ not executed click on the button ‘опублікувати’ of action button group the publication changed status to ‘опубліковано’ not executed verify that published publication displayed for all users on the матеріали’ page that publication is displayed for all users on the матеріали’ page not executed
| 0
|
21,360
| 29,191,378,752
|
IssuesEvent
|
2023-05-19 20:23:00
|
opensearch-project/data-prepper
|
https://api.github.com/repos/opensearch-project/data-prepper
|
closed
|
Support parsing User-Agent fields
|
enhancement plugin - processor
|
**Is your feature request related to a problem? Please describe.**
Many use-cases seek to parse out User-Agent headers to get useful metadata before saving to OpenSearch. This can include the type of browser, version information of the browser, the operation system, version information for the operating system, and even device information.
**Describe the solution you'd like**
Provide a processor for parsing User-Agent headers.
```
useragent:
source_key: /headers/user_agent
target_key: /user_agent
exclude_original: false
```
Following along with https://github.com/opensearch-project/observability/issues/1398, this should support ECS-compatibility.
In the example above, I might have output which looks like the following:
```
user_agent : {
name: "Safari",
version: "16.4",
device: {
name: "iPhone"
},
original: "..."
}
```
Fields:
* `source_key` - The key path to use for the User-Agent string. required
* `target_key` - The key path which will be the object containing the User-Agent metadata. Optional: if not specified, defaults to `user_agent` to match the ECS schema.
* `exclude_original` - By default, the source_key will be copied. Set this to `true` to not copy it. Optional; defaults to `false`.
**Additional context**
ECS reference for User-Agent: https://www.elastic.co/guide/en/ecs/current/ecs-user_agent.html#field-user-agent-device-name
|
1.0
|
Support parsing User-Agent fields - **Is your feature request related to a problem? Please describe.**
Many use-cases seek to parse out User-Agent headers to get useful metadata before saving to OpenSearch. This can include the type of browser, version information of the browser, the operation system, version information for the operating system, and even device information.
**Describe the solution you'd like**
Provide a processor for parsing User-Agent headers.
```
useragent:
source_key: /headers/user_agent
target_key: /user_agent
exclude_original: false
```
Following along with https://github.com/opensearch-project/observability/issues/1398, this should support ECS-compatibility.
In the example above, I might have output which looks like the following:
```
user_agent : {
name: "Safari",
version: "16.4",
device: {
name: "iPhone"
},
original: "..."
}
```
Fields:
* `source_key` - The key path to use for the User-Agent string. required
* `target_key` - The key path which will be the object containing the User-Agent metadata. Optional: if not specified, defaults to `user_agent` to match the ECS schema.
* `exclude_original` - By default, the source_key will be copied. Set this to `true` to not copy it. Optional; defaults to `false`.
**Additional context**
ECS reference for User-Agent: https://www.elastic.co/guide/en/ecs/current/ecs-user_agent.html#field-user-agent-device-name
|
process
|
support parsing user agent fields is your feature request related to a problem please describe many use cases seek to parse out user agent headers to get useful metadata before saving to opensearch this can include the type of browser version information of the browser the operation system version information for the operating system and even device information describe the solution you d like provide a processor for parsing user agent headers useragent source key headers user agent target key user agent exclude original false following along with this should support ecs compatibility in the example above i might have output which looks like the following user agent name safari version device name iphone original fields source key the key path to use for the user agent string required target key the key path which will be the object containing the user agent metadata optional if not specified defaults to user agent to match the ecs schema exclude original by default the source key will be copied set this to true to not copy it optional defaults to false additional context ecs reference for user agent
| 1
|
6,917
| 10,075,027,229
|
IssuesEvent
|
2019-07-24 13:27:04
|
material-components/material-components-ios
|
https://api.github.com/repos/material-components/material-components-ios
|
closed
|
[BottomNavigation] Deprecate `sizeThatFitsIncludesSafeArea`
|
[BottomNavigation] skill:Deprecations type:Process
|
## Feature Request/Suggestion
`sizeThatFitsIncludesSafeArea` should be deprecated once internal clients have had sufficient time to migrate to the new behavior. Migration documentation is available at https://github.com/material-components/material-components-ios/blob/d58c49c/components/BottomNavigation/README.md#deprecations
## Definition of Ready
1. `sizeThatFitsIncludesSafeArea` is internally (to Google) announced as deprecated with a migration timeline
2. [Migration documentation](https://github.com/material-components/material-components-ios/blob/d58c49c/components/BottomNavigation/README.md#deprecations) is provided to clients
## Acceptance Criteria
1. `sizeThatFitsIncludesSafeArea` is marked deprecated
<!-- Auto-generated content below, do not modify -->
---
#### Internal data
- Associated internal bug: [b/127679785](http://b/127679785)
|
1.0
|
[BottomNavigation] Deprecate `sizeThatFitsIncludesSafeArea` - ## Feature Request/Suggestion
`sizeThatFitsIncludesSafeArea` should be deprecated once internal clients have had sufficient time to migrate to the new behavior. Migration documentation is available at https://github.com/material-components/material-components-ios/blob/d58c49c/components/BottomNavigation/README.md#deprecations
## Definition of Ready
1. `sizeThatFitsIncludesSafeArea` is internally (to Google) announced as deprecated with a migration timeline
2. [Migration documentation](https://github.com/material-components/material-components-ios/blob/d58c49c/components/BottomNavigation/README.md#deprecations) is provided to clients
## Acceptance Criteria
1. `sizeThatFitsIncludesSafeArea` is marked deprecated
<!-- Auto-generated content below, do not modify -->
---
#### Internal data
- Associated internal bug: [b/127679785](http://b/127679785)
|
process
|
deprecate sizethatfitsincludessafearea feature request suggestion sizethatfitsincludessafearea should be deprecated once internal clients have had sufficient time to migrate to the new behavior migration documentation is available at definition of ready sizethatfitsincludessafearea is internally to google announced as deprecated with a migration timeline is provided to clients acceptance criteria sizethatfitsincludessafearea is marked deprecated internal data associated internal bug
| 1
|
3,163
| 6,219,827,525
|
IssuesEvent
|
2017-07-09 17:10:05
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
test: new flakiness in test-process-kill-null.js
|
CI / flaky test process regression test windows
|
<!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Version**: `master`
* **Platform**: Windows
* **Subsystem**: test,process
<!-- Enter your issue details below this comment. -->
https://github.com/nodejs/node/commit/44483b6898128b05c09635e03e4d7ceaf3efb14c introduced flakiness to `test/parallel/test-process-kill-null.js`.
Looking for advice on the best course of action:
1. Revert & reopen PR.
2. Fix.
3. Mark as flaky.
/cc @nodejs/testing @nodejs/platform-windows
|
1.0
|
test: new flakiness in test-process-kill-null.js - <!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Version**: `master`
* **Platform**: Windows
* **Subsystem**: test,process
<!-- Enter your issue details below this comment. -->
https://github.com/nodejs/node/commit/44483b6898128b05c09635e03e4d7ceaf3efb14c introduced flakiness to `test/parallel/test-process-kill-null.js`.
Looking for advice on the best course of action:
1. Revert & reopen PR.
2. Fix.
3. Mark as flaky.
/cc @nodejs/testing @nodejs/platform-windows
|
process
|
test new flakiness in test process kill null js thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify affected core module name if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able version master platform windows subsystem test process introduced flakiness to test parallel test process kill null js looking for advice on the best course of action revert reopen pr fix mark as flaky cc nodejs testing nodejs platform windows
| 1
|
978
| 3,437,912,046
|
IssuesEvent
|
2015-12-13 16:08:27
|
pwittchen/ReactiveBeacons
|
https://api.github.com/repos/pwittchen/ReactiveBeacons
|
closed
|
Release 0.4.1
|
release process
|
**Initial release notes**:
- bumped RxJava dependency to v. 1.1.0
- bumped Google Truth test dependency to v. 0.27
- bumped RxAndroid dependency in code samples to v. 1.1.0
- bumped Gradle Build Tools version to v. 1.3.1
**Things to do**:
- [x] bump library version
- [x] upload archives to Maven Central
- [x] close and release artifact on Maven Central
- [x] update `CHANGELOG.md` after Maven Sync
- [x] bump library version in `README.md`
- [x] create new GitHub release
|
1.0
|
Release 0.4.1 - **Initial release notes**:
- bumped RxJava dependency to v. 1.1.0
- bumped Google Truth test dependency to v. 0.27
- bumped RxAndroid dependency in code samples to v. 1.1.0
- bumped Gradle Build Tools version to v. 1.3.1
**Things to do**:
- [x] bump library version
- [x] upload archives to Maven Central
- [x] close and release artifact on Maven Central
- [x] update `CHANGELOG.md` after Maven Sync
- [x] bump library version in `README.md`
- [x] create new GitHub release
|
process
|
release initial release notes bumped rxjava dependency to v bumped google truth test dependency to v bumped rxandroid dependency in code samples to v bumped gradle build tools version to v things to do bump library version upload archives to maven central close and release artifact on maven central update changelog md after maven sync bump library version in readme md create new github release
| 1
|
2,218
| 5,065,054,702
|
IssuesEvent
|
2016-12-23 10:12:11
|
our-city-app/oca-backend
|
https://api.github.com/repos/our-city-app/oca-backend
|
closed
|
Loyalty for everybody
|
process_duplicate type_feature
|
OCA Loyalty-Scan should be able to scan every Qr-Code.
The goal is that OCA accepts every loyalty-card on the market with a QR-Code.
If necessary let's have a chat about this issue.
|
1.0
|
Loyalty for everybody - OCA Loyalty-Scan should be able to scan every Qr-Code.
The goal is that OCA accepts every loyalty-card on the market with a QR-Code.
If necessary let's have a chat about this issue.
|
process
|
loyalty for everybody oca loyalty scan should be able to scan every qr code the goal is that oca accepts every loyalty card on the market with a qr code if necessary let s have a chat about this issue
| 1
|
19,623
| 25,977,602,415
|
IssuesEvent
|
2022-12-19 16:02:29
|
alphagov/govuk-design-system
|
https://api.github.com/repos/alphagov/govuk-design-system
|
closed
|
Add Steve to team page on the Design System website
|
🕔 hours process
|
## What
Steve has joined the team so we are adding him to the team page on the website.
## Why
So our users know who is on the team.
## Who needs to work on this
Kelly
## Who needs to review this
Anyone on the team
## Done when
- [x] Raised a PR to add Steve to the team page
- [ ] Reviewed
- [ ] Merged
|
1.0
|
Add Steve to team page on the Design System website - ## What
Steve has joined the team so we are adding him to the team page on the website.
## Why
So our users know who is on the team.
## Who needs to work on this
Kelly
## Who needs to review this
Anyone on the team
## Done when
- [x] Raised a PR to add Steve to the team page
- [ ] Reviewed
- [ ] Merged
|
process
|
add steve to team page on the design system website what steve has joined the team so we are adding him to the team page on the website why so our users know who is on the team who needs to work on this kelly who needs to review this anyone on the team done when raised a pr to add steve to the team page reviewed merged
| 1
|
31,538
| 11,953,240,909
|
IssuesEvent
|
2020-04-03 20:29:26
|
MicrosoftDocs/vsts-docs
|
https://api.github.com/repos/MicrosoftDocs/vsts-docs
|
closed
|
North and West - incorrect axis..?
|
Pri1 devops-security/tech devops/prod doc-bug
|
The text distinguishes "between North and West Europe". Are you sure this is by intention? Shouldn't it correctly read "between **East** and West Europe" or "between North and **South** Europe"?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 52eb9f91-b952-afb6-6d17-d7092f46440d
* Version Independent ID: bb9113a5-8b8b-b20b-9c6f-09a57f53ba7c
* Content: [Azure DevOps - Data Protection - Azure DevOps](https://docs.microsoft.com/en-us/azure/devops/organizations/security/data-protection?view=azure-devops&source=docs#feedback)
* Content Source: [docs/organizations/security/data-protection.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/organizations/security/data-protection.md)
* Product: **devops**
* Technology: **devops-security**
* GitHub Login: @mlearned
* Microsoft Alias: **mlearned**
|
True
|
North and West - incorrect axis..? - The text distinguishes "between North and West Europe". Are you sure this is by intention? Shouldn't it correctly read "between **East** and West Europe" or "between North and **South** Europe"?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 52eb9f91-b952-afb6-6d17-d7092f46440d
* Version Independent ID: bb9113a5-8b8b-b20b-9c6f-09a57f53ba7c
* Content: [Azure DevOps - Data Protection - Azure DevOps](https://docs.microsoft.com/en-us/azure/devops/organizations/security/data-protection?view=azure-devops&source=docs#feedback)
* Content Source: [docs/organizations/security/data-protection.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/organizations/security/data-protection.md)
* Product: **devops**
* Technology: **devops-security**
* GitHub Login: @mlearned
* Microsoft Alias: **mlearned**
|
non_process
|
north and west incorrect axis the text distinguishes between north and west europe are you sure this is by intention shouldn t it correctly read between east and west europe or between north and south europe document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops security github login mlearned microsoft alias mlearned
| 0
|
826,377
| 31,592,949,009
|
IssuesEvent
|
2023-09-05 01:26:58
|
davidfstr/Crystal-Web-Archiver
|
https://api.github.com/repos/davidfstr/Crystal-Web-Archiver
|
opened
|
View: Long delay opening web browser
|
priority-medium type-bug topic-firstrun
|
<img width="1376" alt="Screen Shot 2023-09-04 at 8 50 27 PM" src="https://github.com/davidfstr/Crystal-Web-Archiver/assets/764688/e725fdb9-ad44-49f1-977a-75366f12ee2a">
|
1.0
|
View: Long delay opening web browser - <img width="1376" alt="Screen Shot 2023-09-04 at 8 50 27 PM" src="https://github.com/davidfstr/Crystal-Web-Archiver/assets/764688/e725fdb9-ad44-49f1-977a-75366f12ee2a">
|
non_process
|
view long delay opening web browser img width alt screen shot at pm src
| 0
|
20,568
| 27,228,376,067
|
IssuesEvent
|
2023-02-21 11:20:42
|
corona-warn-app/cwa-wishlist
|
https://api.github.com/repos/corona-warn-app/cwa-wishlist
|
closed
|
Version 3.0: Increase time until a SRS warning can be issued after the app installation
|
enhancement mirrored-to-jira Test/Share process SRS
|
## Upcoming Implementation
With version 3.0, it will be possible to warn other users without having to scan a QR code or call the hotline to get a TAN to warn others, but a warning can be issued simply by clicking some buttons in the CWA. There is a period of time implemented in the mobile apps which hinders users from issuing such a SRS warning if the app hasn't been installed for a given amount of time (Please note that I think this amount of time is security relevant, and I'm thus not exposing the value).
## Suggested change to the upcoming implementation
I think the value which has been chosen for the upcoming implementation is too short, the amount of time should be increased by at least two days.
## Expected Benefits
In my opinion, this change could drastically reduce the number of times when trolls are able to issue a SRS warning, although they are not tested positive. Especially in combination with https://github.com/corona-warn-app/cwa-wishlist/issues/871 this would be a good change and would make the implementation even more bullet-proof against trolls.
## Related issue
- https://github.com/corona-warn-app/cwa-wishlist/issues/871
---
Internal Tracking ID: [EXPOSUREAPP-14520](https://jira-ibs.wbs.net.sap/browse/EXPOSUREAPP-14520)
|
1.0
|
Version 3.0: Increase time until a SRS warning can be issued after the app installation - ## Upcoming Implementation
With version 3.0, it will be possible to warn other users without having to scan a QR code or call the hotline to get a TAN to warn others, but a warning can be issued simply by clicking some buttons in the CWA. There is a period of time implemented in the mobile apps which hinders users from issuing such a SRS warning if the app hasn't been installed for a given amount of time (Please note that I think this amount of time is security relevant, and I'm thus not exposing the value).
## Suggested change to the upcoming implementation
I think the value which has been chosen for the upcoming implementation is too short, the amount of time should be increased by at least two days.
## Expected Benefits
In my opinion, this change could drastically reduce the number of times when trolls are able to issue a SRS warning, although they are not tested positive. Especially in combination with https://github.com/corona-warn-app/cwa-wishlist/issues/871 this would be a good change and would make the implementation even more bullet-proof against trolls.
## Related issue
- https://github.com/corona-warn-app/cwa-wishlist/issues/871
---
Internal Tracking ID: [EXPOSUREAPP-14520](https://jira-ibs.wbs.net.sap/browse/EXPOSUREAPP-14520)
|
process
|
version increase time until a srs warning can be issued after the app installation upcoming implementation with version it will be possible to warn other users without having to scan a qr code or call the hotline to get a tan to warn others but a warning can be issued simply by clicking some buttons in the cwa there is a period of time implemented in the mobile apps which hinders users from issuing such a srs warning if the app hasn t been installed for a given amount of time please note that i think this amount of time is security relevant and i m thus not exposing the value suggested change to the upcoming implementation i think the value which has been chosen for the upcoming implementation is too short the amount of time should be increased by at least two days expected benefits in my opinion this change could drastically reduce the number of times when trolls are able to issue a srs warning although they are not tested positive especially in combination with this would be a good change and would make the implementation even more bullet proof against trolls related issue internal tracking id
| 1
|
8,025
| 11,209,157,743
|
IssuesEvent
|
2020-01-06 09:49:38
|
inasafe/inasafe-realtime
|
https://api.github.com/repos/inasafe/inasafe-realtime
|
closed
|
Realtime earthquake report for event outside population extent
|
bug ready realtime processor
|
### Problem
Right now realtime InaSAFE generates reports that are outside the population extent. It results in zero people affected in the analysis report.
This happened for example in event with shake ID 20161017132313

### Expected Behaviour
It should show some additional information if the event is outside the exposure (population) extent, that describe why there is no result shown in the report.
### Environment
- InaSAFE : Realtime Earthquake
- QGIS :
- OS :
- Data used :
- Exposure :
- Hazard :
- Aggregation :
### Proposed Solution

|
1.0
|
Realtime earthquake report for event outside population extent - ### Problem
Right now realtime InaSAFE generates reports that are outside the population extent. It results in zero people affected in the analysis report.
This happened for example in event with shake ID 20161017132313

### Expected Behaviour
It should show some additional information if the event is outside the exposure (population) extent, that describe why there is no result shown in the report.
### Environment
- InaSAFE : Realtime Earthquake
- QGIS :
- OS :
- Data used :
- Exposure :
- Hazard :
- Aggregation :
### Proposed Solution

|
process
|
realtime earthquake report for event outside population extent problem right now realtime inasafe generates reports that are outside the population extent it results in zero people affected in the analysis report this happened for example in event with shake id expected behaviour it should show some additional information if the event is outside the exposure population extent that describe why there is no result shown in the report environment inasafe realtime earthquake qgis os data used exposure hazard aggregation proposed solution
| 1
|
14,007
| 16,813,663,014
|
IssuesEvent
|
2021-06-17 03:22:54
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Processing r.neighbors
|
Feature Request Processing
|
Would be nice if user was given some notification that the size of the neighborhood needs to be odd prior to executing the tool. Maybe to add in the description/help. Idea is to prevent user to fail to execute based on this fairly easy to overlook mistake

|
1.0
|
Processing r.neighbors - Would be nice if user was given some notification that the size of the neighborhood needs to be odd prior to executing the tool. Maybe to add in the description/help. Idea is to prevent user to fail to execute based on this fairly easy to overlook mistake

|
process
|
processing r neighbors would be nice if user was given some notification that the size of the neighborhood needs to be odd prior to executing the tool maybe to add in the description help idea is to prevent user to fail to execute based on this fairly easy to overlook mistake
| 1
|
1,103
| 3,576,879,848
|
IssuesEvent
|
2016-01-27 21:12:40
|
neuropoly/spinalcordtoolbox
|
https://api.github.com/repos/neuropoly/spinalcordtoolbox
|
opened
|
ValueError: min() arg is an empty sequence
|
bug priority: high sct_process_segmentation
|
data:
~~~
folder_shared/sct_issues/20160127_process_segmentation
~~~
command+output:
~~~
sct_process_segmentation -i mt1_crop_gmseg.nii.gz -p csa -t label/template/MNI-Poly-AMU_level.nii.gz -vert 2
Check parameters:
.. segmentation file: mt1_crop_gmseg.nii.gz
Create temporary folder...
mkdir tmp.160127161013_552240/
Copying input data to tmp folder and convert to nii...
sct_convert -i /Users/julien/data/biospective/errsm_40/mt/mt1_crop_gmseg.nii.gz -o tmp.160127161013_552240/segmentation.nii.gz
Change orientation to RPI...
sct_image -i segmentation.nii.gz -setorient RPI -o segmentation_RPI.nii.gz
Open segmentation volume...
Get data dimensions...
54 x 60 x 22
Smooth centerline/segmentation...
.. Get center of mass of the centerline/segmentation...
.. Smoothing algo = hanning
.. Windows length = 50
Compute CSA...
Smooth CSA across slices...
.. No smoothing!
Write text file...
z=0: 9.06391104924 mm^2
z=1: 14.2616509747 mm^2
z=2: 15.2012901933 mm^2
z=3: 15.5559849775 mm^2
z=4: 16.1036781824 mm^2
z=5: 15.6591779559 mm^2
z=6: 16.2573781044 mm^2
z=7: 15.690932775 mm^2
z=8: 13.9837083455 mm^2
z=9: 12.9248975033 mm^2
z=10: 12.5429819143 mm^2
z=11: 12.0464303047 mm^2
z=12: 12.1140937945 mm^2
z=13: 11.7674229091 mm^2
z=14: 11.931274159 mm^2
z=15: 11.5055300595 mm^2
z=16: 11.3108929668 mm^2
z=17: 11.4525950518 mm^2
z=18: 11.3685918608 mm^2
z=19: 11.4825773038 mm^2
z=20: 13.6705556668 mm^2
Create volume of CSA values...
sct_image -i csa_volume_RPI.nii.gz -setorient RPI -o csa_volume.nii.gz
Generate output files...
WARNING: File /Users/julien/data/biospective/errsm_40/mt/csa_volume.nii.gz already exists. Deleting it...
File created: /Users/julien/data/biospective/errsm_40/mt/csa_volume.nii.gz
Selected vertebral levels... 2
Find slices corresponding to vertebral levels based on the centerline...
Traceback (most recent call last):
File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 818, in <module>
main(sys.argv[1:])
File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 235, in main
compute_csa(fname_segmentation, verbose, remove_temp_files, step, smoothing_param, figure_fit, param.file_csa_volume, slices, vert_lev, fname_vertebral_labeling, algo_fitting = param.algo_fitting, type_window= param.type_window, window_length=param.window_length)
File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 583, in compute_csa
slices, vert_levels_list, warning = get_slices_matching_with_vertebral_levels_based_centerline(vert_levels, im_vertebral_labeling.data, x_centerline_fit, y_centerline_fit, z_centerline)
File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 702, in get_slices_matching_with_vertebral_levels_based_centerline
slices = str(min(matching_slices_centerline_vert_labeling))+':'+str(max(matching_slices_centerline_vert_labeling))
ValueError: min() arg is an empty sequence
~~~
|
1.0
|
ValueError: min() arg is an empty sequence - data:
~~~
folder_shared/sct_issues/20160127_process_segmentation
~~~
command+output:
~~~
sct_process_segmentation -i mt1_crop_gmseg.nii.gz -p csa -t label/template/MNI-Poly-AMU_level.nii.gz -vert 2
Check parameters:
.. segmentation file: mt1_crop_gmseg.nii.gz
Create temporary folder...
mkdir tmp.160127161013_552240/
Copying input data to tmp folder and convert to nii...
sct_convert -i /Users/julien/data/biospective/errsm_40/mt/mt1_crop_gmseg.nii.gz -o tmp.160127161013_552240/segmentation.nii.gz
Change orientation to RPI...
sct_image -i segmentation.nii.gz -setorient RPI -o segmentation_RPI.nii.gz
Open segmentation volume...
Get data dimensions...
54 x 60 x 22
Smooth centerline/segmentation...
.. Get center of mass of the centerline/segmentation...
.. Smoothing algo = hanning
.. Windows length = 50
Compute CSA...
Smooth CSA across slices...
.. No smoothing!
Write text file...
z=0: 9.06391104924 mm^2
z=1: 14.2616509747 mm^2
z=2: 15.2012901933 mm^2
z=3: 15.5559849775 mm^2
z=4: 16.1036781824 mm^2
z=5: 15.6591779559 mm^2
z=6: 16.2573781044 mm^2
z=7: 15.690932775 mm^2
z=8: 13.9837083455 mm^2
z=9: 12.9248975033 mm^2
z=10: 12.5429819143 mm^2
z=11: 12.0464303047 mm^2
z=12: 12.1140937945 mm^2
z=13: 11.7674229091 mm^2
z=14: 11.931274159 mm^2
z=15: 11.5055300595 mm^2
z=16: 11.3108929668 mm^2
z=17: 11.4525950518 mm^2
z=18: 11.3685918608 mm^2
z=19: 11.4825773038 mm^2
z=20: 13.6705556668 mm^2
Create volume of CSA values...
sct_image -i csa_volume_RPI.nii.gz -setorient RPI -o csa_volume.nii.gz
Generate output files...
WARNING: File /Users/julien/data/biospective/errsm_40/mt/csa_volume.nii.gz already exists. Deleting it...
File created: /Users/julien/data/biospective/errsm_40/mt/csa_volume.nii.gz
Selected vertebral levels... 2
Find slices corresponding to vertebral levels based on the centerline...
Traceback (most recent call last):
File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 818, in <module>
main(sys.argv[1:])
File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 235, in main
compute_csa(fname_segmentation, verbose, remove_temp_files, step, smoothing_param, figure_fit, param.file_csa_volume, slices, vert_lev, fname_vertebral_labeling, algo_fitting = param.algo_fitting, type_window= param.type_window, window_length=param.window_length)
File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 583, in compute_csa
slices, vert_levels_list, warning = get_slices_matching_with_vertebral_levels_based_centerline(vert_levels, im_vertebral_labeling.data, x_centerline_fit, y_centerline_fit, z_centerline)
File "/Users/julien/code/spinalcordtoolbox/bin/sct_process_segmentation", line 702, in get_slices_matching_with_vertebral_levels_based_centerline
slices = str(min(matching_slices_centerline_vert_labeling))+':'+str(max(matching_slices_centerline_vert_labeling))
ValueError: min() arg is an empty sequence
~~~
|
process
|
valueerror min arg is an empty sequence data folder shared sct issues process segmentation command output sct process segmentation i crop gmseg nii gz p csa t label template mni poly amu level nii gz vert check parameters segmentation file crop gmseg nii gz create temporary folder mkdir tmp copying input data to tmp folder and convert to nii sct convert i users julien data biospective errsm mt crop gmseg nii gz o tmp segmentation nii gz change orientation to rpi sct image i segmentation nii gz setorient rpi o segmentation rpi nii gz open segmentation volume get data dimensions x x smooth centerline segmentation get center of mass of the centerline segmentation smoothing algo hanning windows length compute csa smooth csa across slices no smoothing write text file z mm z mm z mm z mm z mm z mm z mm z mm z mm z mm z mm z mm z mm z mm z mm z mm z mm z mm z mm z mm z mm create volume of csa values sct image i csa volume rpi nii gz setorient rpi o csa volume nii gz generate output files warning file users julien data biospective errsm mt csa volume nii gz already exists deleting it file created users julien data biospective errsm mt csa volume nii gz selected vertebral levels find slices corresponding to vertebral levels based on the centerline traceback most recent call last file users julien code spinalcordtoolbox bin sct process segmentation line in main sys argv file users julien code spinalcordtoolbox bin sct process segmentation line in main compute csa fname segmentation verbose remove temp files step smoothing param figure fit param file csa volume slices vert lev fname vertebral labeling algo fitting param algo fitting type window param type window window length param window length file users julien code spinalcordtoolbox bin sct process segmentation line in compute csa slices vert levels list warning get slices matching with vertebral levels based centerline vert levels im vertebral labeling data x centerline fit y centerline fit z centerline file users julien code spinalcordtoolbox bin sct process segmentation line in get slices matching with vertebral levels based centerline slices str min matching slices centerline vert labeling str max matching slices centerline vert labeling valueerror min arg is an empty sequence
| 1
|
19,371
| 25,499,009,129
|
IssuesEvent
|
2022-11-28 00:56:52
|
processing/processing4
|
https://api.github.com/repos/processing/processing4
|
closed
|
size(863,863/2) not working as first line of setup() function in a simple sketch
|
preprocessor
|
<!--- ** For coding questions, ask the forum: https://discourse.processing.org ** -->
<!--- ** This page is only for bugs in the software & feature requests ** -->
When running one of Dan Shiffman's examples, it failed on the size(863,863/2); line in the setup(). The error said to look at the documentation for the size() function. The only requirement was for the size() function to be in the setup() function as the first line and for it to be in the processing sketch named the same.
However, it appears the size() can't have arithmetic expressions as part of the argument list.
<!--- ** If your code won't start, that's a better question for the forum. ** -->
<!--- ** If Processing won't start, post on the forum where you can get help. ** -->
<!--- ** Also be sure to read the troubleshooting page first: ** -->
<!--- ** https://github.com/processing/processing/wiki/Troubleshooting ** -->
<!--- ** Before posting, please search Issues for duplicates ** -->
## Description
<!--- Use a title that describes what is happening. -->
<!--- Give a description of the proposed change. -->
When running one of Dan Shiffman's mandelbrot example sketch, it failed on the size(863,863/2); statement in the setup() function. The error said to look at the documentation for the size() function.
The documentation for size():
- size() function must be in the setup() function as the first line
- and for size() to be in the main processing sketch.
However, it appears the size() can't have arithmetic expressions as part of the argument list.
## Expected Behavior
<!--- Bug? Tell us what you were expecting. -->
<!--- Improvement? Tell us how you’d like it to work. -->
expected the size(863,863/2); to work or with integer variables as arguments
## Current Behavior
<!--- Explain the difference from current behavior. -->
sketch won't run, error messages says size() can't be used here, changing it to size(863,431); got it to work.
## Steps to Reproduce
<!--- Provide an unambiguous set of steps to reproduce. -->
<!--- Including code will make it more likely to be fixed. -->
1. include a size(863,863/2) as the first statement in the setup() function of a simple sketch
2.
```java
void setup() {
size(863,863/2);
}
void draw() {
}
```
3. run program
4. should see error below sketch code
## Your Environment
<!--- Include details about your environment. -->
<!--- Thousands of people use Processing every day and may not have this issue, -->
<!--- so this gives us clues about why you’re running into a problem. -->
* Processing version: 4.0.1
* Operating System and OS version: MacOS 10.13 Ventura
* Other information: Running on a macmini with Apple M1 chip
## Possible Causes / Solutions
<!--- Optionally, if you have a diagnosis or fix in mind, please share. -->
workaround is to do the sizing math and keep size(863,431); arguments as simple integer constants.
provide a better error message that points to the arguments as not being simple integers.
|
1.0
|
size(863,863/2) not working as first line of setup() function in a simple sketch - <!--- ** For coding questions, ask the forum: https://discourse.processing.org ** -->
<!--- ** This page is only for bugs in the software & feature requests ** -->
When running one of Dan Shiffman's examples, it failed on the size(863,863/2); line in the setup(). The error said to look at the documentation for the size() function. The only requirement was for the size() function to be in the setup() function as the first line and for it to be in the processing sketch named the same.
However, it appears the size() can't have arithmetic expressions as part of the argument list.
<!--- ** If your code won't start, that's a better question for the forum. ** -->
<!--- ** If Processing won't start, post on the forum where you can get help. ** -->
<!--- ** Also be sure to read the troubleshooting page first: ** -->
<!--- ** https://github.com/processing/processing/wiki/Troubleshooting ** -->
<!--- ** Before posting, please search Issues for duplicates ** -->
## Description
<!--- Use a title that describes what is happening. -->
<!--- Give a description of the proposed change. -->
When running one of Dan Shiffman's mandelbrot example sketch, it failed on the size(863,863/2); statement in the setup() function. The error said to look at the documentation for the size() function.
The documentation for size():
- size() function must be in the setup() function as the first line
- and for size() to be in the main processing sketch.
However, it appears the size() can't have arithmetic expressions as part of the argument list.
## Expected Behavior
<!--- Bug? Tell us what you were expecting. -->
<!--- Improvement? Tell us how you’d like it to work. -->
expected the size(863,863/2); to work or with integer variables as arguments
## Current Behavior
<!--- Explain the difference from current behavior. -->
sketch won't run, error messages says size() can't be used here, changing it to size(863,431); got it to work.
## Steps to Reproduce
<!--- Provide an unambiguous set of steps to reproduce. -->
<!--- Including code will make it more likely to be fixed. -->
1. include a size(863,863/2) as the first statement in the setup() function of a simple sketch
2.
```java
void setup() {
size(863,863/2);
}
void draw() {
}
```
3. run program
4. should see error below sketch code
## Your Environment
<!--- Include details about your environment. -->
<!--- Thousands of people use Processing every day and may not have this issue, -->
<!--- so this gives us clues about why you’re running into a problem. -->
* Processing version: 4.0.1
* Operating System and OS version: MacOS 10.13 Ventura
* Other information: Running on a macmini with Apple M1 chip
## Possible Causes / Solutions
<!--- Optionally, if you have a diagnosis or fix in mind, please share. -->
workaround is to do the sizing math and keep size(863,431); arguments as simple integer constants.
provide a better error message that points to the arguments as not being simple integers.
|
process
|
size not working as first line of setup function in a simple sketch when running one of dan shiffman s examples it failed on the size line in the setup the error said to look at the documentation for the size function the only requirement was for the size function to be in the setup function as the first line and for it to be in the processing sketch named the same however it appears the size can t have arithmetic expressions as part of the argument list description when running one of dan shiffman s mandelbrot example sketch it failed on the size statement in the setup function the error said to look at the documentation for the size function the documentation for size size function must be in the setup function as the first line and for size to be in the main processing sketch however it appears the size can t have arithmetic expressions as part of the argument list expected behavior expected the size to work or with integer variables as arguments current behavior sketch won t run error messages says size can t be used here changing it to size got it to work steps to reproduce include a size as the first statement in the setup function of a simple sketch java void setup size void draw run program should see error below sketch code your environment processing version operating system and os version macos ventura other information running on a macmini with apple chip possible causes solutions workaround is to do the sizing math and keep size arguments as simple integer constants provide a better error message that points to the arguments as not being simple integers
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.