Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16
values | text_combine stringlengths 96 262k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
562,509 | 16,662,346,976 | IssuesEvent | 2021-06-06 15:00:36 | reepo-bots/reepo-cop-beta | https://api.github.com/repos/reepo-bots/reepo-cop-beta | closed | [U] Setup Deployment Workflow | 🔥 priority.Urgent | There needs to be an automated workflow that triggers on mergers to `master` branch that deploys features directly to [reepo-cop](https://github.com/reepo-bots/reepo-cop) | 1.0 | [U] Setup Deployment Workflow - There needs to be an automated workflow that triggers on mergers to `master` branch that deploys features directly to [reepo-cop](https://github.com/reepo-bots/reepo-cop) | priority | setup deployment workflow there needs to be an automated workflow that triggers on mergers to master branch that deploys features directly to | 1 |
271,771 | 23,629,811,963 | IssuesEvent | 2022-08-25 08:22:38 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | closed | [CI] IndexShardTests testFlushOnIdleConcurrentFlushDoesNotWait failing | >test-failure :Distributed/Engine Team:Distributed | This looks to be a newly added test in #89430 and failed for me in a PR.
**Build scan:**
https://gradle-enterprise.elastic.co/s/ctyyqap6ywxok/tests/:server:test/org.elasticsearch.index.shard.IndexShardTests/testFlushOnIdleConcurrentFlushDoesNotWait
**Reproduction line:**
`./gradlew ':server:test' --tests "org.elasticsearch.index.shard.IndexShardTests.testFlushOnIdleConcurrentFlushDoesNotWait" -Dtests.seed=3B825C4D0772201A -Dtests.locale=ru-RU -Dtests.timezone=Australia/Broken_Hill -Druntime.java=17`
**Applicable branches:**
main
**Reproduces locally?:**
No
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.index.shard.IndexShardTests&tests.test=testFlushOnIdleConcurrentFlushDoesNotWait
**Failure excerpt:**
```
java.lang.AssertionError: expected to see should see first flush getting lock immediately but did not
Expected: <true>
but: was <false>
at __randomizedtesting.SeedInfo.seed([3B825C4D0772201A:B93D2E3091554306]:0)
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18)
at org.elasticsearch.test.MockLogAppender$SeenEventExpectation.assertMatched(MockLogAppender.java:121)
at org.elasticsearch.test.MockLogAppender.assertAllExpectationsMatched(MockLogAppender.java:55)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1104)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1077)
at org.elasticsearch.index.shard.IndexShardTests.testFlushOnIdleConcurrentFlushDoesNotWait(IndexShardTests.java:3939)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:568)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850)
at java.lang.Thread.run(Thread.java:833)
``` | 1.0 | [CI] IndexShardTests testFlushOnIdleConcurrentFlushDoesNotWait failing - This looks to be a newly added test in #89430 and failed for me in a PR.
**Build scan:**
https://gradle-enterprise.elastic.co/s/ctyyqap6ywxok/tests/:server:test/org.elasticsearch.index.shard.IndexShardTests/testFlushOnIdleConcurrentFlushDoesNotWait
**Reproduction line:**
`./gradlew ':server:test' --tests "org.elasticsearch.index.shard.IndexShardTests.testFlushOnIdleConcurrentFlushDoesNotWait" -Dtests.seed=3B825C4D0772201A -Dtests.locale=ru-RU -Dtests.timezone=Australia/Broken_Hill -Druntime.java=17`
**Applicable branches:**
main
**Reproduces locally?:**
No
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.index.shard.IndexShardTests&tests.test=testFlushOnIdleConcurrentFlushDoesNotWait
**Failure excerpt:**
```
java.lang.AssertionError: expected to see should see first flush getting lock immediately but did not
Expected: <true>
but: was <false>
at __randomizedtesting.SeedInfo.seed([3B825C4D0772201A:B93D2E3091554306]:0)
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18)
at org.elasticsearch.test.MockLogAppender$SeenEventExpectation.assertMatched(MockLogAppender.java:121)
at org.elasticsearch.test.MockLogAppender.assertAllExpectationsMatched(MockLogAppender.java:55)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1104)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1077)
at org.elasticsearch.index.shard.IndexShardTests.testFlushOnIdleConcurrentFlushDoesNotWait(IndexShardTests.java:3939)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:568)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850)
at java.lang.Thread.run(Thread.java:833)
``` | non_priority | indexshardtests testflushonidleconcurrentflushdoesnotwait failing this looks to be a newly added test in and failed for me in a pr build scan reproduction line gradlew server test tests org elasticsearch index shard indexshardtests testflushonidleconcurrentflushdoesnotwait dtests seed dtests locale ru ru dtests timezone australia broken hill druntime java applicable branches main reproduces locally no failure history failure excerpt java lang assertionerror expected to see should see first flush getting lock immediately but did not expected but was at randomizedtesting seedinfo seed at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch test mocklogappender seeneventexpectation assertmatched mocklogappender java at org elasticsearch test mocklogappender assertallexpectationsmatched mocklogappender java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch index shard indexshardtests testflushonidleconcurrentflushdoesnotwait indexshardtests java at jdk internal reflect nativemethodaccessorimpl nativemethodaccessorimpl java at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol lambda forktimeoutingtask threadleakcontrol java at java lang thread run thread java | 0 |
336,162 | 24,488,683,029 | IssuesEvent | 2022-10-09 19:35:52 | ekanshsinghal/se-group26-proj1 | https://api.github.com/repos/ekanshsinghal/se-group26-proj1 | closed | Generate docs for frontend and backend | documentation | Doc generation for python and react js
Python can use Sphinx, | 1.0 | Generate docs for frontend and backend - Doc generation for python and react js
Python can use Sphinx, | non_priority | generate docs for frontend and backend doc generation for python and react js python can use sphinx | 0 |
378,038 | 11,194,763,160 | IssuesEvent | 2020-01-03 02:48:12 | DarshanShet777/Model-Airport | https://api.github.com/repos/DarshanShet777/Model-Airport | closed | Finance: Missing Items In Inventory | Low Priority Issue | Inventory updates necessary:
- [ ] Missing DeWalt 37 piece driver bit set. Mfr Model "#DW2176"
- [ ] Missing DeWalt 14 piece drill bit set. 1/16"-3/8". "#DW1162"
- [ ] Update name of "Aluminum Refrigerator LED Fixture Holder" to "Aluminum Refrigerator LED Bracket" in inventory.
- [ ] Missing variety of nuts and bolts.
- [ ] Missing screws.
- [ ] Missing Dremel Multi Pro Kit 3966
- [ ] Missing other power tool: Bosch "#34614" Power Tool 14.4V | 1.0 | Finance: Missing Items In Inventory - Inventory updates necessary:
- [ ] Missing DeWalt 37 piece driver bit set. Mfr Model "#DW2176"
- [ ] Missing DeWalt 14 piece drill bit set. 1/16"-3/8". "#DW1162"
- [ ] Update name of "Aluminum Refrigerator LED Fixture Holder" to "Aluminum Refrigerator LED Bracket" in inventory.
- [ ] Missing variety of nuts and bolts.
- [ ] Missing screws.
- [ ] Missing Dremel Multi Pro Kit 3966
- [ ] Missing other power tool: Bosch "#34614" Power Tool 14.4V | priority | finance missing items in inventory inventory updates necessary missing dewalt piece driver bit set mfr model missing dewalt piece drill bit set update name of aluminum refrigerator led fixture holder to aluminum refrigerator led bracket in inventory missing variety of nuts and bolts missing screws missing dremel multi pro kit missing other power tool bosch power tool | 1 |
122,405 | 12,150,715,104 | IssuesEvent | 2020-04-24 18:30:09 | monoDriveIO/documentation | https://api.github.com/repos/monoDriveIO/documentation | closed | Create documentation for import and export of OpenDrive | documentation enhancement | Document the process for importing and exporting OpenDrive data into the Scene Edtior. | 1.0 | Create documentation for import and export of OpenDrive - Document the process for importing and exporting OpenDrive data into the Scene Edtior. | non_priority | create documentation for import and export of opendrive document the process for importing and exporting opendrive data into the scene edtior | 0 |
135,903 | 11,028,796,761 | IssuesEvent | 2019-12-06 12:33:32 | firecracker-microvm/firecracker | https://api.github.com/repos/firecracker-microvm/firecracker | opened | CI skips test_drives.py::test_rescan_dev | Priority: High Quality: Automated Testing Quality: Bug | Looking at logs for various PRs, it seems our CI skips `test_rescan_dev` from `integration_tests/functional/test_drives.py` on `x86_64` for no apparent reason. The test is not skipped when running on a local dev machine.
The relevant part in the CI log is something like:
```
integration_tests/functional/test_drives.py resize2fs 1.44.1 (24-Mar-2018)
Resizing the filesystem on /tmp/firecracker_test_session/17f55cd2-26ee-4ebb-be10-a2cae034e07e/fsfiles/scratch.ext4 to 524288 (1k) blocks.
The filesystem on /tmp/firecracker_test_session/17f55cd2-26ee-4ebb-be10-a2cae034e07e/fsfiles/scratch.ext4 is now 524288 (1k) blocks long.
.s....
``` | 1.0 | CI skips test_drives.py::test_rescan_dev - Looking at logs for various PRs, it seems our CI skips `test_rescan_dev` from `integration_tests/functional/test_drives.py` on `x86_64` for no apparent reason. The test is not skipped when running on a local dev machine.
The relevant part in the CI log is something like:
```
integration_tests/functional/test_drives.py resize2fs 1.44.1 (24-Mar-2018)
Resizing the filesystem on /tmp/firecracker_test_session/17f55cd2-26ee-4ebb-be10-a2cae034e07e/fsfiles/scratch.ext4 to 524288 (1k) blocks.
The filesystem on /tmp/firecracker_test_session/17f55cd2-26ee-4ebb-be10-a2cae034e07e/fsfiles/scratch.ext4 is now 524288 (1k) blocks long.
.s....
``` | non_priority | ci skips test drives py test rescan dev looking at logs for various prs it seems our ci skips test rescan dev from integration tests functional test drives py on for no apparent reason the test is not skipped when running on a local dev machine the relevant part in the ci log is something like integration tests functional test drives py mar resizing the filesystem on tmp firecracker test session fsfiles scratch to blocks the filesystem on tmp firecracker test session fsfiles scratch is now blocks long s | 0 |
63,924 | 3,202,435,102 | IssuesEvent | 2015-10-02 14:03:52 | JuliaOpt/JuMP.jl | https://api.github.com/repos/JuliaOpt/JuMP.jl | closed | completing variable indices | low priority | The following code fails in the last command.
******************
m = Model()
@defVar(m,x[1,1:2])
@addConstraint(m,x[1,1]+x[1,2]<=1)
@defVar(m,x[1,0]>=0)
@addConstraint(m,x[1,1]+x[1,2]+x[1,0]<=1)
ERROR: key not found: 1
in getindex at dict.jl:515
in getindex at no file
******************
This code could be useful when dealing with socp constraints. There is an easy fix bellow, but adding new indices as above could be useful (if it is feasible).
******************
m = Model()
@defVar(m,x[1,i=0:2]>= ( i == 0 ? 0 : -Inf))
@addConstraint(m,x[1,1]+x[1,2]<=1)
@addConstraint(m,x[1,1]+x[1,2]+x[1,0]<=1)
****************** | 1.0 | completing variable indices - The following code fails in the last command.
******************
m = Model()
@defVar(m,x[1,1:2])
@addConstraint(m,x[1,1]+x[1,2]<=1)
@defVar(m,x[1,0]>=0)
@addConstraint(m,x[1,1]+x[1,2]+x[1,0]<=1)
ERROR: key not found: 1
in getindex at dict.jl:515
in getindex at no file
******************
This code could be useful when dealing with socp constraints. There is an easy fix bellow, but adding new indices as above could be useful (if it is feasible).
******************
m = Model()
@defVar(m,x[1,i=0:2]>= ( i == 0 ? 0 : -Inf))
@addConstraint(m,x[1,1]+x[1,2]<=1)
@addConstraint(m,x[1,1]+x[1,2]+x[1,0]<=1)
****************** | priority | completing variable indices the following code fails in the last command m model defvar m x addconstraint m x x defvar m x addconstraint m x x x error key not found in getindex at dict jl in getindex at no file this code could be useful when dealing with socp constraints there is an easy fix bellow but adding new indices as above could be useful if it is feasible m model defvar m x i inf addconstraint m x x addconstraint m x x x | 1 |
629,614 | 20,048,294,333 | IssuesEvent | 2022-02-03 01:02:12 | WordPress/openverse-catalog | https://api.github.com/repos/WordPress/openverse-catalog | opened | Reactivate provider DAGs | 🟧 priority: high 🚧 status: blocked 🛠 goal: fix 🕹 aspect: interface 🔧 tech: airflow | ## Description
<!-- Fully describe the benefit of the change (E.g., improve speed, robustness, etc.) -->
Since migrating from Creative Commons, we have had very few providers running consistently. We would like to get these DAGs in a consistently operational state so that we are regularly ingesting data.
## Providers
<!-- Describe any alternative solutions you have considered -->
- [ ] Brooklyn Museum
- [ ] Cleveland Museum
- [ ] Europeana
- [ ] Finnish Museums
- [ ] Flickr
- [ ] Freesound
- [ ] Jamendo
- [ ] Metropolitan Museum
- [ ] Museum Victoria
- [ ] NYPL
- [ ] Phylopic
- [ ] Raw Pixel
- [ ] Science Museum
- [ ] Smithsonian
- [ ] Staten Museum
- [ ] Stocksnap
- [ ] Walters
- [ ] Wikimedia commons
- [ ] WordPress Photo Directory
## Additional context
<!-- Add any other context about the feature request. -->
This work is currently blocked pending #285
## Implementation
<!-- Replace the [ ] with [x] to check the box. -->
- [ ] 🙋 I would be interested in implementing this feature.
| 1.0 | Reactivate provider DAGs - ## Description
<!-- Fully describe the benefit of the change (E.g., improve speed, robustness, etc.) -->
Since migrating from Creative Commons, we have had very few providers running consistently. We would like to get these DAGs in a consistently operational state so that we are regularly ingesting data.
## Providers
<!-- Describe any alternative solutions you have considered -->
- [ ] Brooklyn Museum
- [ ] Cleveland Museum
- [ ] Europeana
- [ ] Finnish Museums
- [ ] Flickr
- [ ] Freesound
- [ ] Jamendo
- [ ] Metropolitan Museum
- [ ] Museum Victoria
- [ ] NYPL
- [ ] Phylopic
- [ ] Raw Pixel
- [ ] Science Museum
- [ ] Smithsonian
- [ ] Staten Museum
- [ ] Stocksnap
- [ ] Walters
- [ ] Wikimedia commons
- [ ] WordPress Photo Directory
## Additional context
<!-- Add any other context about the feature request. -->
This work is currently blocked pending #285
## Implementation
<!-- Replace the [ ] with [x] to check the box. -->
- [ ] 🙋 I would be interested in implementing this feature.
| priority | reactivate provider dags description since migrating from creative commons we have had very few providers running consistently we would like to get these dags in a consistently operational state so that we are regularly ingesting data providers brooklyn museum cleveland museum europeana finnish museums flickr freesound jamendo metropolitan museum museum victoria nypl phylopic raw pixel science museum smithsonian staten museum stocksnap walters wikimedia commons wordpress photo directory additional context this work is currently blocked pending implementation 🙋 i would be interested in implementing this feature | 1 |
190,423 | 6,818,271,038 | IssuesEvent | 2017-11-07 04:18:40 | openshift/origin | https://api.github.com/repos/openshift/origin | closed | signature_import_controller: log spam overwhelms master controller logs | component/image kind/bug priority/P0 | After upgrading to Origin 3.7.0-rc.0 we are seeing this spam in the logs from the master controllers:
```
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.381556 58933 signature_import_controller.go:109] error syncing image sha256:aaf622099aa45f1fea8343510237d5e0fd1678ca4388f0b72b6e1c29f4454846, it will be retried: unauthorized: authentication required
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.445446 58933 signature_import_controller.go:109] error syncing image sha256:dd2eaf9b1797fb2a66929131ba901099b54fcf228010d276e6796a79cbcde515, it will be retried: unauthorized: authentication required
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.445847 58933 signature_import_controller.go:109] error syncing image sha256:782bdc4d52dc103a0a5b6d8735df9e51a7f5bbcc3972b1753b8a42ff5097afde, it will be retried: unauthorized: authentication required
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.486213 58933 signature_import_controller.go:109] error syncing image sha256:19b2d6a18090bd4bece3715d671f9424e0b0faec502fc923bb5f5bd68412ac9a, it will be retried: unauthorized: authentication required
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.546640 58933 signature_import_controller.go:109] error syncing image sha256:5170c454d09f24d5d42a159a7dcbcca7eab8f7dc102f245a5867c9e0524fabb9, it will be retried: unauthorized: authentication required
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.548303 58933 signature_import_controller.go:109] error syncing image sha256:fd744f3268b2bfb1e307a5de2001baa506939a490d76bdc321b831e5e55588ce, it will be retried: unauthorized: authentication required
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.656923 58933 signature_import_controller.go:109] error syncing image sha256:6724afedd005c11643fa1595bcd156dd7a805bbc78f191d7516d5236f6859457, it will be retried: received unexpected HTTP status: 500 Internal Server Error
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.658089 58933 signature_import_controller.go:109] error syncing image sha256:296efc3be851fdc4bd03cbc6f519b6b7455b4806acff5c170d10e0f350268d5d, it will be retried: received unexpected HTTP status: 500 Internal Server Error
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.666556 58933 signature_import_controller.go:109] error syncing image sha256:7a6f635ae92f23d0bb6cd763240ddb2cbea67c5658237d0bad0ceb52a5dd9312, it will be retried: unauthorized: authentication required
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.691758 58933 signature_import_controller.go:109] error syncing image sha256:659c68eea74c988a4e5a5a93a8032863e952c4b5eaa3929253151d134e994cf6, it will be retried: unauthorized: authentication required
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.703032 58933 signature_import_controller.go:109] error syncing image sha256:7036e07397e54b34a422560481a77f12220215137c9a93e9c81c7981116adeeb, it will be retried: received unexpected HTTP status: 500 Internal Server Error
```
The entire journal we have:
```
# journalctl --all --lines all --unit origin-master-controllers.service --no-pager | wc -l
512193
```
Of that,
```
# journalctl --all --lines all --unit origin-master-controllers.service --no-pager | grep signature_import_controller | wc -l
454678
```
This is 89% of all logging coming out of the controllers and means we have less than 24 hours of historical data for the logs before `journald` starts rotating us out.
/kind bug
/cc @mfojtik @smarterclayton @eparis | 1.0 | signature_import_controller: log spam overwhelms master controller logs - After upgrading to Origin 3.7.0-rc.0 we are seeing this spam in the logs from the master controllers:
```
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.381556 58933 signature_import_controller.go:109] error syncing image sha256:aaf622099aa45f1fea8343510237d5e0fd1678ca4388f0b72b6e1c29f4454846, it will be retried: unauthorized: authentication required
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.445446 58933 signature_import_controller.go:109] error syncing image sha256:dd2eaf9b1797fb2a66929131ba901099b54fcf228010d276e6796a79cbcde515, it will be retried: unauthorized: authentication required
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.445847 58933 signature_import_controller.go:109] error syncing image sha256:782bdc4d52dc103a0a5b6d8735df9e51a7f5bbcc3972b1753b8a42ff5097afde, it will be retried: unauthorized: authentication required
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.486213 58933 signature_import_controller.go:109] error syncing image sha256:19b2d6a18090bd4bece3715d671f9424e0b0faec502fc923bb5f5bd68412ac9a, it will be retried: unauthorized: authentication required
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.546640 58933 signature_import_controller.go:109] error syncing image sha256:5170c454d09f24d5d42a159a7dcbcca7eab8f7dc102f245a5867c9e0524fabb9, it will be retried: unauthorized: authentication required
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.548303 58933 signature_import_controller.go:109] error syncing image sha256:fd744f3268b2bfb1e307a5de2001baa506939a490d76bdc321b831e5e55588ce, it will be retried: unauthorized: authentication required
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.656923 58933 signature_import_controller.go:109] error syncing image sha256:6724afedd005c11643fa1595bcd156dd7a805bbc78f191d7516d5236f6859457, it will be retried: received unexpected HTTP status: 500 Internal Server Error
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.658089 58933 signature_import_controller.go:109] error syncing image sha256:296efc3be851fdc4bd03cbc6f519b6b7455b4806acff5c170d10e0f350268d5d, it will be retried: received unexpected HTTP status: 500 Internal Server Error
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.666556 58933 signature_import_controller.go:109] error syncing image sha256:7a6f635ae92f23d0bb6cd763240ddb2cbea67c5658237d0bad0ceb52a5dd9312, it will be retried: unauthorized: authentication required
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.691758 58933 signature_import_controller.go:109] error syncing image sha256:659c68eea74c988a4e5a5a93a8032863e952c4b5eaa3929253151d134e994cf6, it will be retried: unauthorized: authentication required
Nov 03 17:22:42 origin-ci-ig-m-11v4 origin-master-controllers[58933]: E1103 17:22:42.703032 58933 signature_import_controller.go:109] error syncing image sha256:7036e07397e54b34a422560481a77f12220215137c9a93e9c81c7981116adeeb, it will be retried: received unexpected HTTP status: 500 Internal Server Error
```
The entire journal we have:
```
# journalctl --all --lines all --unit origin-master-controllers.service --no-pager | wc -l
512193
```
Of that,
```
# journalctl --all --lines all --unit origin-master-controllers.service --no-pager | grep signature_import_controller | wc -l
454678
```
This is 89% of all logging coming out of the controllers and means we have less than 24 hours of historical data for the logs before `journald` starts rotating us out.
/kind bug
/cc @mfojtik @smarterclayton @eparis | priority | signature import controller log spam overwhelms master controller logs after upgrading to origin rc we are seeing this spam in the logs from the master controllers nov origin ci ig m origin master controllers signature import controller go error syncing image it will be retried unauthorized authentication required nov origin ci ig m origin master controllers signature import controller go error syncing image it will be retried unauthorized authentication required nov origin ci ig m origin master controllers signature import controller go error syncing image it will be retried unauthorized authentication required nov origin ci ig m origin master controllers signature import controller go error syncing image it will be retried unauthorized authentication required nov origin ci ig m origin master controllers signature import controller go error syncing image it will be retried unauthorized authentication required nov origin ci ig m origin master controllers signature import controller go error syncing image it will be retried unauthorized authentication required nov origin ci ig m origin master controllers signature import controller go error syncing image it will be retried received unexpected http status internal server error nov origin ci ig m origin master controllers signature import controller go error syncing image it will be retried received unexpected http status internal server error nov origin ci ig m origin master controllers signature import controller go error syncing image it will be retried unauthorized authentication required nov origin ci ig m origin master controllers signature import controller go error syncing image it will be retried unauthorized authentication required nov origin ci ig m origin master controllers signature import controller go error syncing image it will be retried received unexpected http status internal server error the entire journal we have journalctl all lines all unit origin master controllers service no pager wc l of that journalctl all lines all unit origin master controllers service no pager grep signature import controller wc l this is of all logging coming out of the controllers and means we have less than hours of historical data for the logs before journald starts rotating us out kind bug cc mfojtik smarterclayton eparis | 1 |
375,456 | 26,162,633,756 | IssuesEvent | 2022-12-31 20:25:07 | dpalmasan/py-logic | https://api.github.com/repos/dpalmasan/py-logic | closed | Add Contribution guidelines | documentation | Add guidelines just in case someone gets interested in the project, and wants to contribute. A `CONTRIBUTING.md` file will need to be created. | 1.0 | Add Contribution guidelines - Add guidelines just in case someone gets interested in the project, and wants to contribute. A `CONTRIBUTING.md` file will need to be created. | non_priority | add contribution guidelines add guidelines just in case someone gets interested in the project and wants to contribute a contributing md file will need to be created | 0 |
15,394 | 5,954,591,396 | IssuesEvent | 2017-05-27 19:21:37 | openMVG/openMVG | https://api.github.com/repos/openMVG/openMVG | closed | OpenMVGConfig.cmake incomplete with Visual studio | build | After building and installing openMVG, I include it in my project like this:
```
find_package(openMVG)
target_link_libraries(myproject ${OPENMVG_LIBRARIES})
```
and I got the following errors:
1: `LINK : fatal error LNK1104: cannot open file 'Threads::Threads.lib'`
It seems https://github.com/openMVG/openMVG/commit/73e990c2b6b7d3dc16a706b92be7c64d7be0984f introduced the dependency Threads::Threads, which is unnecessary, at least on Windows.
2: Missing vl_sift symbols.
There is no line `target_link_libraries(openMVG_features vlsift)`. Is this an oversight, or is there some legal reason for this? | 1.0 | OpenMVGConfig.cmake incomplete with Visual studio - After building and installing openMVG, I include it in my project like this:
```
find_package(openMVG)
target_link_libraries(myproject ${OPENMVG_LIBRARIES})
```
and I got the following errors:
1: `LINK : fatal error LNK1104: cannot open file 'Threads::Threads.lib'`
It seems https://github.com/openMVG/openMVG/commit/73e990c2b6b7d3dc16a706b92be7c64d7be0984f introduced the dependency Threads::Threads, which is unnecessary, at least on Windows.
2: Missing vl_sift symbols.
There is no line `target_link_libraries(openMVG_features vlsift)`. Is this an oversight, or is there some legal reason for this? | non_priority | openmvgconfig cmake incomplete with visual studio after building and installing openmvg i include it in my project like this find package openmvg target link libraries myproject openmvg libraries and i got the following errors link fatal error cannot open file threads threads lib it seems introduced the dependency threads threads which is unnecessary at least on windows missing vl sift symbols there is no line target link libraries openmvg features vlsift is this an oversight or is there some legal reason for this | 0 |
107,290 | 13,450,089,021 | IssuesEvent | 2020-09-08 17:57:08 | chapel-lang/chapel | https://api.github.com/repos/chapel-lang/chapel | closed | Should storing an array view in a variable make it non-view? | area: Compiler area: Language type: Bug type: Design | (Related to #5474 and #12178).
Generally, we have the behavior that storing a slice into a (non-ref) variable causes a copy of the slice to be made. Similarly for returning a slice from a function by value. (But we never created syntax for the alternative - see #12178).
In working on #16180, I was surprised that copying a reindex/rank change array left one with a reindex/rank change array. (However it does copy the elements). I was expecting that an array view would no longer be an array view once a copy of it is made. We've had this behavior since 1.18 (and the test program below didn't compile in 1.17 or 1.16).
Here is a test program to demonstrate:
``` chapel
proc returnIt(arg) { return arg; }
proc main() {
var A: [1..2] real;
var AA: [1..2, 1..2] real;
{
writeln();
writeln("slice");
const ref rB = A[1..1];
writeln("to const ref is view: ", rB.isSliceArrayView());
var B = A[1..1];
writeln("to var is view: ", B.isSliceArrayView());
const ref rReturned = returnIt(A[1..1]);
writeln("returned is view: ", rReturned.isSliceArrayView());
const ref rD = A[1..1].domain;
writeln(".domain to const ref is view: ", rD.isSliceDomainView());
var D = A[1..1].domain;
writeln(".domain to var is view: ", D.isSliceDomainView());
}
{
writeln();
writeln("reindex");
const ref rB = A.reindex(3..4);
writeln("to const ref is view: ", rB.isReindexArrayView());
var B = A.reindex(3..4);
writeln("to var is view: ", B.isReindexArrayView());
const ref rReturned = returnIt(A.reindex(3..4));
writeln("returned is view: ", rReturned.isReindexArrayView());
const ref rD = A.reindex(3..4).domain;
writeln(".domain to const ref is view: ", rD.isReindexDomainView());
var D = A.reindex(3..4).domain;
writeln(".domain to var is view: ", D.isReindexDomainView());
}
{
writeln();
writeln("rank change");
const ref rB = AA[1, ..];
writeln("to const ref is view: ", rB.isRankChangeArrayView());
var B = AA[1, ..];
writeln("to var is view: ", B.isRankChangeArrayView());
const ref rReturned = returnIt(AA[1, ..]);
writeln("returned is view: ", rReturned.isRankChangeArrayView());
const ref rD = AA[1, ..].domain;
writeln(".domain to const ref is view: ", rD.isRankChangeDomainView());
var D = AA[1, ..].domain;
writeln(".domain to var is view: ", D.isRankChangeDomainView());
}
}
```
It outputs
```
slice
to const ref is view: true
to var is view: false
returned is view: false
.domain to const ref is view: false
.domain to var is view: false
reindex
to const ref is view: true
to var is view: true
returned is view: true
.domain to const ref is view: true
.domain to var is view: true
rank change
to const ref is view: true
to var is view: true
returned is view: true
.domain to const ref is view: true
.domain to var is view: true
```
however I would expect only the "to const ref" rows to print out `true`. | 1.0 | Should storing an array view in a variable make it non-view? - (Related to #5474 and #12178).
Generally, we have the behavior that storing a slice into a (non-ref) variable causes a copy of the slice to be made. Similarly for returning a slice from a function by value. (But we never created syntax for the alternative - see #12178).
In working on #16180, I was surprised that copying a reindex/rank change array left one with a reindex/rank change array. (However it does copy the elements). I was expecting that an array view would no longer be an array view once a copy of it is made. We've had this behavior since 1.18 (and the test program below didn't compile in 1.17 or 1.16).
Here is a test program to demonstrate:
``` chapel
proc returnIt(arg) { return arg; }
proc main() {
var A: [1..2] real;
var AA: [1..2, 1..2] real;
{
writeln();
writeln("slice");
const ref rB = A[1..1];
writeln("to const ref is view: ", rB.isSliceArrayView());
var B = A[1..1];
writeln("to var is view: ", B.isSliceArrayView());
const ref rReturned = returnIt(A[1..1]);
writeln("returned is view: ", rReturned.isSliceArrayView());
const ref rD = A[1..1].domain;
writeln(".domain to const ref is view: ", rD.isSliceDomainView());
var D = A[1..1].domain;
writeln(".domain to var is view: ", D.isSliceDomainView());
}
{
writeln();
writeln("reindex");
const ref rB = A.reindex(3..4);
writeln("to const ref is view: ", rB.isReindexArrayView());
var B = A.reindex(3..4);
writeln("to var is view: ", B.isReindexArrayView());
const ref rReturned = returnIt(A.reindex(3..4));
writeln("returned is view: ", rReturned.isReindexArrayView());
const ref rD = A.reindex(3..4).domain;
writeln(".domain to const ref is view: ", rD.isReindexDomainView());
var D = A.reindex(3..4).domain;
writeln(".domain to var is view: ", D.isReindexDomainView());
}
{
writeln();
writeln("rank change");
const ref rB = AA[1, ..];
writeln("to const ref is view: ", rB.isRankChangeArrayView());
var B = AA[1, ..];
writeln("to var is view: ", B.isRankChangeArrayView());
const ref rReturned = returnIt(AA[1, ..]);
writeln("returned is view: ", rReturned.isRankChangeArrayView());
const ref rD = AA[1, ..].domain;
writeln(".domain to const ref is view: ", rD.isRankChangeDomainView());
var D = AA[1, ..].domain;
writeln(".domain to var is view: ", D.isRankChangeDomainView());
}
}
```
It outputs
```
slice
to const ref is view: true
to var is view: false
returned is view: false
.domain to const ref is view: false
.domain to var is view: false
reindex
to const ref is view: true
to var is view: true
returned is view: true
.domain to const ref is view: true
.domain to var is view: true
rank change
to const ref is view: true
to var is view: true
returned is view: true
.domain to const ref is view: true
.domain to var is view: true
```
however I would expect only the "to const ref" rows to print out `true`. | non_priority | should storing an array view in a variable make it non view related to and generally we have the behavior that storing a slice into a non ref variable causes a copy of the slice to be made similarly for returning a slice from a function by value but we never created syntax for the alternative see in working on i was surprised that copying a reindex rank change array left one with a reindex rank change array however it does copy the elements i was expecting that an array view would no longer be an array view once a copy of it is made we ve had this behavior since and the test program below didn t compile in or here is a test program to demonstrate chapel proc returnit arg return arg proc main var a real var aa real writeln writeln slice const ref rb a writeln to const ref is view rb isslicearrayview var b a writeln to var is view b isslicearrayview const ref rreturned returnit a writeln returned is view rreturned isslicearrayview const ref rd a domain writeln domain to const ref is view rd isslicedomainview var d a domain writeln domain to var is view d isslicedomainview writeln writeln reindex const ref rb a reindex writeln to const ref is view rb isreindexarrayview var b a reindex writeln to var is view b isreindexarrayview const ref rreturned returnit a reindex writeln returned is view rreturned isreindexarrayview const ref rd a reindex domain writeln domain to const ref is view rd isreindexdomainview var d a reindex domain writeln domain to var is view d isreindexdomainview writeln writeln rank change const ref rb aa writeln to const ref is view rb isrankchangearrayview var b aa writeln to var is view b isrankchangearrayview const ref rreturned returnit aa writeln returned is view rreturned isrankchangearrayview const ref rd aa domain writeln domain to const ref is view rd isrankchangedomainview var d aa domain writeln domain to var is view d isrankchangedomainview it outputs slice to const ref is view true to var is view false returned is view false domain to const ref is view false domain to var is view false reindex to const ref is view true to var is view true returned is view true domain to const ref is view true domain to var is view true rank change to const ref is view true to var is view true returned is view true domain to const ref is view true domain to var is view true however i would expect only the to const ref rows to print out true | 0 |
67,905 | 14,892,016,142 | IssuesEvent | 2021-01-21 01:49:41 | RG4421/codebuilder-cloud-plugin | https://api.github.com/repos/RG4421/codebuilder-cloud-plugin | opened | CVE-2020-36180 (Medium) detected in jackson-databind-2.8.11.1.jar | security vulnerability | ## CVE-2020-36180 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.11.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: codebuilder-cloud-plugin/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.11.1/jackson-databind-2.8.11.1.jar</p>
<p>
Dependency Hierarchy:
- aws-java-sdk-1.11.403.jar (Root Library)
- jackson2-api-2.8.11.2.jar
- :x: **jackson-databind-2.8.11.1.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.cpdsadapter.DriverAdapterCPDS.
<p>Publish Date: 2021-01-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36180>CVE-2020-36180</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.1","isTransitiveDependency":true,"dependencyTree":"org.jenkins-ci.plugins:aws-java-sdk:1.11.403;org.jenkins-ci.plugins:jackson2-api:2.8.11.2;com.fasterxml.jackson.core:jackson-databind:2.8.11.1","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2020-36180","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.cpdsadapter.DriverAdapterCPDS.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36180","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> --> | True | CVE-2020-36180 (Medium) detected in jackson-databind-2.8.11.1.jar - ## CVE-2020-36180 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.11.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: codebuilder-cloud-plugin/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.11.1/jackson-databind-2.8.11.1.jar</p>
<p>
Dependency Hierarchy:
- aws-java-sdk-1.11.403.jar (Root Library)
- jackson2-api-2.8.11.2.jar
- :x: **jackson-databind-2.8.11.1.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.cpdsadapter.DriverAdapterCPDS.
<p>Publish Date: 2021-01-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36180>CVE-2020-36180</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.1","isTransitiveDependency":true,"dependencyTree":"org.jenkins-ci.plugins:aws-java-sdk:1.11.403;org.jenkins-ci.plugins:jackson2-api:2.8.11.2;com.fasterxml.jackson.core:jackson-databind:2.8.11.1","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2020-36180","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.cpdsadapter.DriverAdapterCPDS.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36180","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in jackson databind jar cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file codebuilder cloud plugin pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy aws java sdk jar root library api jar x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons cpdsadapter driveradaptercpds publish date url a href cvss score details base score metrics not available isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons cpdsadapter driveradaptercpds vulnerabilityurl | 0 |
310,719 | 26,737,688,428 | IssuesEvent | 2023-01-30 10:38:10 | rizinorg/rizin | https://api.github.com/repos/rizinorg/rizin | closed | rz-bin -z -N 1 on ELF fails to find strings of length < 4 | good first issue test-required | ### Work environment
| Questions | Answers
|------------------------------------------------------|--------------------
| OS/arch/bits | Kali x86 64
| File format of the file you reverse | ELF
| Architecture/bits of the file | x86/64
| `rizin -v` full output, **not truncated** | rizin 0.5.0 @ linux-x86-64 commit: f1b2e7bd758f24dfab2577d615d3b86b4c31a60b
### Expected behavior
If the minimum string length specified to `rz-bin` using `-N` is `x`, then strings of length at least `x` should be found.
### Actual behavior
No strings of length less than 4 are found, seemingly regardless of the argument provided to `-N`. An example of strings which are commonly of length less than 4 are `scanf()` conversion specifications (e.g. "%d").
### Steps to reproduce the behavior
The source code of `some_elf` in [binary.zip](https://github.com/rizinorg/rizin/files/9229265/binary.zip):
```c
int main() { char *a = "1", *b = "22", *c = "333", *d = "4444"; }
```
Run `rz-bin -z -N 1 some_elf` and you'll see that only the string "4444" is discovered as compared to all four of them.
### Identified cause
The call to find strings in the binary in this case is made from `rz_bin_object_set_items()` in [librz/bin/bobj.c on line 503](https://github.com/rizinorg/rizin/blob/f1b2e7bd758f24dfab2577d615d3b86b4c31a60b/librz/bin/bobj.c#L503). The string function called is in [librz/bin/p/bin_elf.inc on line 1965](https://github.com/rizinorg/rizin/blob/f1b2e7bd758f24dfab2577d615d3b86b4c31a60b/librz/bin/p/bin_elf.inc#L1965). This function simply wraps a call to `rz_bin_file_strings(bf, 4, false)`. The `4` is a hard coded minimal string length. I believe the issue affects all file formats which has a plugin. [The `strings` function pointer in the RzBinPlugin struct currently only takes a `RzBinFile *` argument](https://github.com/rizinorg/rizin/blob/f1b2e7bd758f24dfab2577d615d3b86b4c31a60b/librz/include/rz_bin.h#L531). Shouldn't it perhaps also take a minimum string length argument, so that the call from `rz_bin_object_set_items()` can pass along the correct minimum string length?
| 1.0 | rz-bin -z -N 1 on ELF fails to find strings of length < 4 - ### Work environment
| Questions | Answers
|------------------------------------------------------|--------------------
| OS/arch/bits | Kali x86 64
| File format of the file you reverse | ELF
| Architecture/bits of the file | x86/64
| `rizin -v` full output, **not truncated** | rizin 0.5.0 @ linux-x86-64 commit: f1b2e7bd758f24dfab2577d615d3b86b4c31a60b
### Expected behavior
If the minimum string length specified to `rz-bin` using `-N` is `x`, then strings of length at least `x` should be found.
### Actual behavior
No strings of length less than 4 are found, seemingly regardless of the argument provided to `-N`. An example of strings which are commonly of length less than 4 are `scanf()` conversion specifications (e.g. "%d").
### Steps to reproduce the behavior
The source code of `some_elf` in [binary.zip](https://github.com/rizinorg/rizin/files/9229265/binary.zip):
```c
int main() { char *a = "1", *b = "22", *c = "333", *d = "4444"; }
```
Run `rz-bin -z -N 1 some_elf` and you'll see that only the string "4444" is discovered as compared to all four of them.
### Identified cause
The call to find strings in the binary in this case is made from `rz_bin_object_set_items()` in [librz/bin/bobj.c on line 503](https://github.com/rizinorg/rizin/blob/f1b2e7bd758f24dfab2577d615d3b86b4c31a60b/librz/bin/bobj.c#L503). The string function called is in [librz/bin/p/bin_elf.inc on line 1965](https://github.com/rizinorg/rizin/blob/f1b2e7bd758f24dfab2577d615d3b86b4c31a60b/librz/bin/p/bin_elf.inc#L1965). This function simply wraps a call to `rz_bin_file_strings(bf, 4, false)`. The `4` is a hard coded minimal string length. I believe the issue affects all file formats which has a plugin. [The `strings` function pointer in the RzBinPlugin struct currently only takes a `RzBinFile *` argument](https://github.com/rizinorg/rizin/blob/f1b2e7bd758f24dfab2577d615d3b86b4c31a60b/librz/include/rz_bin.h#L531). Shouldn't it perhaps also take a minimum string length argument, so that the call from `rz_bin_object_set_items()` can pass along the correct minimum string length?
| non_priority | rz bin z n on elf fails to find strings of length work environment questions answers os arch bits kali file format of the file you reverse elf architecture bits of the file rizin v full output not truncated rizin linux commit expected behavior if the minimum string length specified to rz bin using n is x then strings of length at least x should be found actual behavior no strings of length less than are found seemingly regardless of the argument provided to n an example of strings which are commonly of length less than are scanf conversion specifications e g d steps to reproduce the behavior the source code of some elf in c int main char a b c d run rz bin z n some elf and you ll see that only the string is discovered as compared to all four of them identified cause the call to find strings in the binary in this case is made from rz bin object set items in the string function called is in this function simply wraps a call to rz bin file strings bf false the is a hard coded minimal string length i believe the issue affects all file formats which has a plugin shouldn t it perhaps also take a minimum string length argument so that the call from rz bin object set items can pass along the correct minimum string length | 0 |
929 | 3,236,771,466 | IssuesEvent | 2015-10-14 08:04:52 | IPM-France/glpi_plugin_kiosks | https://api.github.com/repos/IPM-France/glpi_plugin_kiosks | closed | Logs pour le WS GeoLoc | f_WebServices m_GlpiKiosks p3 | Le WS geoLoc utilisé par la CNAM fait maintenant des logs dans le fichier *pk-geolog.log* du répertoire des logs Glpi. Il serait bien de mettre en place une rotation quotidienne de ce fichier ... | 1.0 | Logs pour le WS GeoLoc - Le WS geoLoc utilisé par la CNAM fait maintenant des logs dans le fichier *pk-geolog.log* du répertoire des logs Glpi. Il serait bien de mettre en place une rotation quotidienne de ce fichier ... | non_priority | logs pour le ws geoloc le ws geoloc utilisé par la cnam fait maintenant des logs dans le fichier pk geolog log du répertoire des logs glpi il serait bien de mettre en place une rotation quotidienne de ce fichier | 0 |
19,805 | 26,188,048,933 | IssuesEvent | 2023-01-03 04:59:03 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | AssignSystemIdentity option in the Set-AzAutomationAccount does not exists | automation/svc triaged cxp process-automation/subsvc Pri1 docs-bug |
[Enter feedback here]
AssignSystemIdentity option in the Set-AzAutomationAccount does not work and this open does not exists in the Microsoft doc as well: https://learn.microsoft.com/en-us/powershell/module/az.automation/set-azautomationaccount?view=azps-9.2.0
There is a documentation bug in this page
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: d9ce2612-b600-3fca-3315-a7836ef91c96
* Version Independent ID: 78766eed-c3c6-ce60-7620-17b99f3d9d5e
* Content: [Using a system-assigned managed identity for an Azure Automation account](https://learn.microsoft.com/en-us/azure/automation/enable-managed-identity-for-automation)
* Content Source: [articles/automation/enable-managed-identity-for-automation.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/enable-managed-identity-for-automation.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SnehaSudhirG
* Microsoft Alias: **sudhirsneha** | 1.0 | AssignSystemIdentity option in the Set-AzAutomationAccount does not exists -
[Enter feedback here]
AssignSystemIdentity option in the Set-AzAutomationAccount does not work and this open does not exists in the Microsoft doc as well: https://learn.microsoft.com/en-us/powershell/module/az.automation/set-azautomationaccount?view=azps-9.2.0
There is a documentation bug in this page
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: d9ce2612-b600-3fca-3315-a7836ef91c96
* Version Independent ID: 78766eed-c3c6-ce60-7620-17b99f3d9d5e
* Content: [Using a system-assigned managed identity for an Azure Automation account](https://learn.microsoft.com/en-us/azure/automation/enable-managed-identity-for-automation)
* Content Source: [articles/automation/enable-managed-identity-for-automation.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/enable-managed-identity-for-automation.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SnehaSudhirG
* Microsoft Alias: **sudhirsneha** | non_priority | assignsystemidentity option in the set azautomationaccount does not exists assignsystemidentity option in the set azautomationaccount does not work and this open does not exists in the microsoft doc as well there is a documentation bug in this page document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login snehasudhirg microsoft alias sudhirsneha | 0 |
222,320 | 24,697,560,176 | IssuesEvent | 2022-10-19 13:13:04 | webc-zss2/traefik | https://api.github.com/repos/webc-zss2/traefik | opened | CVE-2022-3517 (High) detected in minimatch-3.0.4.tgz | security vulnerability | ## CVE-2022-3517 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimatch-3.0.4.tgz</b></p></summary>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz">https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz</a></p>
<p>Path to dependency file: /webui/package.json</p>
<p>Path to vulnerable library: /webui/node_modules/minimatch/package.json</p>
<p>
Dependency Hierarchy:
- app-1.2.4.tgz (Root Library)
- webpack-dev-server-3.9.0.tgz
- chokidar-2.1.8.tgz
- fsevents-1.2.9.tgz
- node-pre-gyp-0.12.0.tgz
- npm-packlist-1.4.1.tgz
- ignore-walk-3.0.1.tgz
- :x: **minimatch-3.0.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/webc-zss2/traefik/commit/5fb0a687f6f4614ef22e971903247af1c371d3c8">5fb0a687f6f4614ef22e971903247af1c371d3c8</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service.
<p>Publish Date: 2022-10-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-3517>CVE-2022-3517</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-17</p>
<p>Fix Resolution: minimatch - 3.0.5</p>
</p>
</details>
<p></p>
| True | CVE-2022-3517 (High) detected in minimatch-3.0.4.tgz - ## CVE-2022-3517 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimatch-3.0.4.tgz</b></p></summary>
<p>a glob matcher in javascript</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz">https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz</a></p>
<p>Path to dependency file: /webui/package.json</p>
<p>Path to vulnerable library: /webui/node_modules/minimatch/package.json</p>
<p>
Dependency Hierarchy:
- app-1.2.4.tgz (Root Library)
- webpack-dev-server-3.9.0.tgz
- chokidar-2.1.8.tgz
- fsevents-1.2.9.tgz
- node-pre-gyp-0.12.0.tgz
- npm-packlist-1.4.1.tgz
- ignore-walk-3.0.1.tgz
- :x: **minimatch-3.0.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/webc-zss2/traefik/commit/5fb0a687f6f4614ef22e971903247af1c371d3c8">5fb0a687f6f4614ef22e971903247af1c371d3c8</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service.
<p>Publish Date: 2022-10-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-3517>CVE-2022-3517</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-17</p>
<p>Fix Resolution: minimatch - 3.0.5</p>
</p>
</details>
<p></p>
| non_priority | cve high detected in minimatch tgz cve high severity vulnerability vulnerable library minimatch tgz a glob matcher in javascript library home page a href path to dependency file webui package json path to vulnerable library webui node modules minimatch package json dependency hierarchy app tgz root library webpack dev server tgz chokidar tgz fsevents tgz node pre gyp tgz npm packlist tgz ignore walk tgz x minimatch tgz vulnerable library found in head commit a href found in base branch master vulnerability details a vulnerability was found in the minimatch package this flaw allows a regular expression denial of service redos when calling the braceexpand function with specific arguments resulting in a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution minimatch | 0 |
35,832 | 8,014,581,854 | IssuesEvent | 2018-07-25 07:07:56 | mozilla-mobile/focus-ios | https://api.github.com/repos/mozilla-mobile/focus-ios | closed | Style: Use type inference when possible | P5 code health | To mimic Firefox iOS's style guide. Good for code health. | 1.0 | Style: Use type inference when possible - To mimic Firefox iOS's style guide. Good for code health. | non_priority | style use type inference when possible to mimic firefox ios s style guide good for code health | 0 |
116,117 | 17,358,354,874 | IssuesEvent | 2021-07-29 16:59:12 | Tim-sandbox/webgoat-trng | https://api.github.com/repos/Tim-sandbox/webgoat-trng | opened | CVE-2020-7793 (High) detected in ua-parser-js-0.7.17.tgz | security vulnerability | ## CVE-2020-7793 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.17.tgz</b></p></summary>
<p>Lightweight JavaScript-based user-agent string parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz</a></p>
<p>Path to dependency file: webgoat-trng/docs/package.json</p>
<p>Path to vulnerable library: webgoat-trng/docs/node_modules/ua-parser-js/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.26.3.tgz (Root Library)
- :x: **ua-parser-js-0.7.17.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Tim-sandbox/webgoat-trng/commit/35c20cdfd5acad60536ca6b0402f619e18097690">35c20cdfd5acad60536ca6b0402f619e18097690</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package ua-parser-js before 0.7.23 are vulnerable to Regular Expression Denial of Service (ReDoS) in multiple regexes (see linked commit for more info).
<p>Publish Date: 2020-12-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7793>CVE-2020-7793</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/faisalman/ua-parser-js/commit/6d1f26df051ba681463ef109d36c9cf0f7e32b18">https://github.com/faisalman/ua-parser-js/commit/6d1f26df051ba681463ef109d36c9cf0f7e32b18</a></p>
<p>Release Date: 2020-12-11</p>
<p>Fix Resolution: 0.7.23</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"ua-parser-js","packageVersion":"0.7.17","packageFilePaths":["/docs/package.json"],"isTransitiveDependency":true,"dependencyTree":"browser-sync:2.26.3;ua-parser-js:0.7.17","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.7.23"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-7793","vulnerabilityDetails":"The package ua-parser-js before 0.7.23 are vulnerable to Regular Expression Denial of Service (ReDoS) in multiple regexes (see linked commit for more info).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7793","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-7793 (High) detected in ua-parser-js-0.7.17.tgz - ## CVE-2020-7793 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.17.tgz</b></p></summary>
<p>Lightweight JavaScript-based user-agent string parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz</a></p>
<p>Path to dependency file: webgoat-trng/docs/package.json</p>
<p>Path to vulnerable library: webgoat-trng/docs/node_modules/ua-parser-js/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.26.3.tgz (Root Library)
- :x: **ua-parser-js-0.7.17.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Tim-sandbox/webgoat-trng/commit/35c20cdfd5acad60536ca6b0402f619e18097690">35c20cdfd5acad60536ca6b0402f619e18097690</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package ua-parser-js before 0.7.23 are vulnerable to Regular Expression Denial of Service (ReDoS) in multiple regexes (see linked commit for more info).
<p>Publish Date: 2020-12-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7793>CVE-2020-7793</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/faisalman/ua-parser-js/commit/6d1f26df051ba681463ef109d36c9cf0f7e32b18">https://github.com/faisalman/ua-parser-js/commit/6d1f26df051ba681463ef109d36c9cf0f7e32b18</a></p>
<p>Release Date: 2020-12-11</p>
<p>Fix Resolution: 0.7.23</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"ua-parser-js","packageVersion":"0.7.17","packageFilePaths":["/docs/package.json"],"isTransitiveDependency":true,"dependencyTree":"browser-sync:2.26.3;ua-parser-js:0.7.17","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.7.23"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-7793","vulnerabilityDetails":"The package ua-parser-js before 0.7.23 are vulnerable to Regular Expression Denial of Service (ReDoS) in multiple regexes (see linked commit for more info).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7793","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in ua parser js tgz cve high severity vulnerability vulnerable library ua parser js tgz lightweight javascript based user agent string parser library home page a href path to dependency file webgoat trng docs package json path to vulnerable library webgoat trng docs node modules ua parser js package json dependency hierarchy browser sync tgz root library x ua parser js tgz vulnerable library found in head commit a href found in base branch main vulnerability details the package ua parser js before are vulnerable to regular expression denial of service redos in multiple regexes see linked commit for more info publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree browser sync ua parser js isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails the package ua parser js before are vulnerable to regular expression denial of service redos in multiple regexes see linked commit for more info vulnerabilityurl | 0 |
417,534 | 12,167,067,921 | IssuesEvent | 2020-04-27 10:16:20 | hazelcast/hazelcast-aws | https://api.github.com/repos/hazelcast/hazelcast-aws | closed | Fargate support | Estimation: XL Priority: High Type: Enhancement | Hi,
I'm trying to deploy Hazelcast cluster on AWS Fargate but I couldn't find any documentation about the process...
Does Hazelcast suport Fargate or it only supports ECS?
All the best! | 1.0 | Fargate support - Hi,
I'm trying to deploy Hazelcast cluster on AWS Fargate but I couldn't find any documentation about the process...
Does Hazelcast suport Fargate or it only supports ECS?
All the best! | priority | fargate support hi i m trying to deploy hazelcast cluster on aws fargate but i couldn t find any documentation about the process does hazelcast suport fargate or it only supports ecs all the best | 1 |
79,047 | 3,520,016,113 | IssuesEvent | 2016-01-12 19:05:32 | IQSS/dataverse | https://api.github.com/repos/IQSS/dataverse | closed | Ingest: clean up temp files from /tmp. | Component: File Upload & Handling Priority: Medium Status: QA Type: Suggestion |
Ingest leaves a bunch of files with names like: columnBufferFile1042015997543016021bytes
They are not cleaned up and probably should be, otherwise it would become an ops issue. | 1.0 | Ingest: clean up temp files from /tmp. -
Ingest leaves a bunch of files with names like: columnBufferFile1042015997543016021bytes
They are not cleaned up and probably should be, otherwise it would become an ops issue. | priority | ingest clean up temp files from tmp ingest leaves a bunch of files with names like they are not cleaned up and probably should be otherwise it would become an ops issue | 1 |
122,668 | 4,838,768,007 | IssuesEvent | 2016-11-09 05:58:13 | gama-platform/gama | https://api.github.com/repos/gama-platform/gama | closed | Invisible error when parameter with slider editor is initialized with a non-constant value (another variable) | > Bug Affects Stability Concerns GAML OS All Priority High Version All | ### Steps to reproduce
1.
Define variable in global block
e.g. int numberOfInfectedPeopleInit <- 2147 update: when people count (each.isInfected);
2.
Define parameter in experiment block
e.g. parameter "Number of Infected People" value: numberOfInfectedPeopleInit min: 1 max: 2147;
and of course some output display
e.g. display geom3D ...
3.
Run experiment and environment should displays some "display" contains one is named geom3D
### Expected behavior
The running screen should
### Actual behavior
The running environment is blank.
### System and version
GAMA 1.7
JDK 1.8
OS: MacOS Serrier
| 1.0 | Invisible error when parameter with slider editor is initialized with a non-constant value (another variable) - ### Steps to reproduce
1.
Define variable in global block
e.g. int numberOfInfectedPeopleInit <- 2147 update: when people count (each.isInfected);
2.
Define parameter in experiment block
e.g. parameter "Number of Infected People" value: numberOfInfectedPeopleInit min: 1 max: 2147;
and of course some output display
e.g. display geom3D ...
3.
Run experiment and environment should displays some "display" contains one is named geom3D
### Expected behavior
The running screen should
### Actual behavior
The running environment is blank.
### System and version
GAMA 1.7
JDK 1.8
OS: MacOS Serrier
| priority | invisible error when parameter with slider editor is initialized with a non constant value another variable steps to reproduce define variable in global block e g int numberofinfectedpeopleinit update when people count each isinfected define parameter in experiment block e g parameter number of infected people value numberofinfectedpeopleinit min max and of course some output display e g display run experiment and environment should displays some display contains one is named expected behavior the running screen should actual behavior the running environment is blank system and version gama jdk os macos serrier | 1 |
720,163 | 24,781,756,324 | IssuesEvent | 2022-10-24 06:05:46 | airqo-platform/AirQo-frontend | https://api.github.com/repos/airqo-platform/AirQo-frontend | closed | "Expand" the deploy/Not Deployed State with more options | enhancement good first issue netmanager feature-request priority-high marked-for-fixing | **Is your feature request related to a problem? Please describe.**
At the moment, the user managing the network is not able able to easily track the status of a newly created device in the production assembly line accordingly. The current implementation of netmanager does not offer options to monitor the hardware production process.
**Describe the solution you'd like**
Expanding the deploy/not deployed state with more options.
In assembly line => In-stock and ready => Deployed => Un deployed -> Recalled for maintenance -> Decommissioned.
**Describe alternatives you've considered**
- Utlliization of the "new" `status` field for the deployment status of the device instead of using the `isActive` field
- We need to review the UI colors we shall use for the different statuses. Currently we are use green for` deployed` and red for `undeployed`.
**Additional context**
This was picked up as a feature request from the recently concluded hardware demo.
API details:
We are using the update endpoint for devices where the updated field is `status`
There is already a device status field on the platform which we can take advantage of for this. Screenshot follows shortly:
<img width="1241" alt="Screenshot 2021-08-30 at 12 47 37" src="https://user-images.githubusercontent.com/1590213/131321134-60be582a-741a-42ec-a458-b241706a8072.png">
| 1.0 | "Expand" the deploy/Not Deployed State with more options - **Is your feature request related to a problem? Please describe.**
At the moment, the user managing the network is not able able to easily track the status of a newly created device in the production assembly line accordingly. The current implementation of netmanager does not offer options to monitor the hardware production process.
**Describe the solution you'd like**
Expanding the deploy/not deployed state with more options.
In assembly line => In-stock and ready => Deployed => Un deployed -> Recalled for maintenance -> Decommissioned.
**Describe alternatives you've considered**
- Utlliization of the "new" `status` field for the deployment status of the device instead of using the `isActive` field
- We need to review the UI colors we shall use for the different statuses. Currently we are use green for` deployed` and red for `undeployed`.
**Additional context**
This was picked up as a feature request from the recently concluded hardware demo.
API details:
We are using the update endpoint for devices where the updated field is `status`
There is already a device status field on the platform which we can take advantage of for this. Screenshot follows shortly:
<img width="1241" alt="Screenshot 2021-08-30 at 12 47 37" src="https://user-images.githubusercontent.com/1590213/131321134-60be582a-741a-42ec-a458-b241706a8072.png">
| priority | expand the deploy not deployed state with more options is your feature request related to a problem please describe at the moment the user managing the network is not able able to easily track the status of a newly created device in the production assembly line accordingly the current implementation of netmanager does not offer options to monitor the hardware production process describe the solution you d like expanding the deploy not deployed state with more options in assembly line in stock and ready deployed un deployed recalled for maintenance decommissioned describe alternatives you ve considered utlliization of the new status field for the deployment status of the device instead of using the isactive field we need to review the ui colors we shall use for the different statuses currently we are use green for deployed and red for undeployed additional context this was picked up as a feature request from the recently concluded hardware demo api details we are using the update endpoint for devices where the updated field is status there is already a device status field on the platform which we can take advantage of for this screenshot follows shortly img width alt screenshot at src | 1 |
84,638 | 24,369,213,353 | IssuesEvent | 2022-10-03 17:43:17 | xamarin/xamarin-android | https://api.github.com/repos/xamarin/xamarin-android | opened | Access denied when attempting to build a project | Area: App+Library Build needs-triage | ### Android application type
Android for .NET (net6.0-android, etc.)
### Affected platform version
Rider 2022.2.3, .NET6.0.401
### Description
When attempting to build a project targeting Android I receive the following exception. I have attempted to move the project to different folders and changing the folder permissions to be more open. I haven't found anything online regarding this issue.
` Xamarin.Android.Common.targets(1055, 2): [XACIC7019] System.UnauthorizedAccessException: Access to the path is denied.
---> System.IO.IOException: Operation not permitted
--- End of inner exception stack trace ---
at Interop.ThrowExceptionForIoErrno(ErrorInfo errorInfo, String path, Boolean isDirectory, Func`2 errorRewriter)
at System.IO.FileSystem.CopyFile(String sourceFullPath, String destFullPath, Boolean overwrite)
at System.IO.File.Copy(String sourceFileName, String destFileName, Boolean overwrite)
at Microsoft.Android.Build.Tasks.Files.CopyIfChanged(String source, String destination) in /Users/runner/work/1/s/xamarin-android/external/xamarin-android-tools/src/Microsoft.Android.Build.BaseTasks/Files.cs:line 125
at Xamarin.Android.Tasks.CopyIfChanged.RunTask()
at Microsoft.Android.Build.Tasks.AndroidTask.Execute() in /Users/runner/work/1/s/xamarin-android/external/xamarin-android-tools/src/Microsoft.Android.Build.BaseTasks/AndroidTask.cs:line 17`
### Steps to Reproduce
1. Create a new Android project
2. Attempt to build the project
### Did you find any workaround?
[msbuild.binlog.txt](https://github.com/xamarin/xamarin-android/files/9699687/msbuild.binlog.txt)
### Relevant log output
_No response_ | 1.0 | Access denied when attempting to build a project - ### Android application type
Android for .NET (net6.0-android, etc.)
### Affected platform version
Rider 2022.2.3, .NET6.0.401
### Description
When attempting to build a project targeting Android I receive the following exception. I have attempted to move the project to different folders and changing the folder permissions to be more open. I haven't found anything online regarding this issue.
` Xamarin.Android.Common.targets(1055, 2): [XACIC7019] System.UnauthorizedAccessException: Access to the path is denied.
---> System.IO.IOException: Operation not permitted
--- End of inner exception stack trace ---
at Interop.ThrowExceptionForIoErrno(ErrorInfo errorInfo, String path, Boolean isDirectory, Func`2 errorRewriter)
at System.IO.FileSystem.CopyFile(String sourceFullPath, String destFullPath, Boolean overwrite)
at System.IO.File.Copy(String sourceFileName, String destFileName, Boolean overwrite)
at Microsoft.Android.Build.Tasks.Files.CopyIfChanged(String source, String destination) in /Users/runner/work/1/s/xamarin-android/external/xamarin-android-tools/src/Microsoft.Android.Build.BaseTasks/Files.cs:line 125
at Xamarin.Android.Tasks.CopyIfChanged.RunTask()
at Microsoft.Android.Build.Tasks.AndroidTask.Execute() in /Users/runner/work/1/s/xamarin-android/external/xamarin-android-tools/src/Microsoft.Android.Build.BaseTasks/AndroidTask.cs:line 17`
### Steps to Reproduce
1. Create a new Android project
2. Attempt to build the project
### Did you find any workaround?
[msbuild.binlog.txt](https://github.com/xamarin/xamarin-android/files/9699687/msbuild.binlog.txt)
### Relevant log output
_No response_ | non_priority | access denied when attempting to build a project android application type android for net android etc affected platform version rider description when attempting to build a project targeting android i receive the following exception i have attempted to move the project to different folders and changing the folder permissions to be more open i haven t found anything online regarding this issue xamarin android common targets system unauthorizedaccessexception access to the path is denied system io ioexception operation not permitted end of inner exception stack trace at interop throwexceptionforioerrno errorinfo errorinfo string path boolean isdirectory func errorrewriter at system io filesystem copyfile string sourcefullpath string destfullpath boolean overwrite at system io file copy string sourcefilename string destfilename boolean overwrite at microsoft android build tasks files copyifchanged string source string destination in users runner work s xamarin android external xamarin android tools src microsoft android build basetasks files cs line at xamarin android tasks copyifchanged runtask at microsoft android build tasks androidtask execute in users runner work s xamarin android external xamarin android tools src microsoft android build basetasks androidtask cs line steps to reproduce create a new android project attempt to build the project did you find any workaround relevant log output no response | 0 |
12,467 | 3,615,793,174 | IssuesEvent | 2016-02-07 00:37:45 | yocheah/Fluxnet-Ameriflux | https://api.github.com/repos/yocheah/Fluxnet-Ameriflux | closed | Documentation of known issues | FLUXNET April Release High Priority Website (Documentation) | Known issues in Dec FLUXNET 2015 release.
- On Ameriflux, we need to know what has changed between April 2016 release and December release
- Also list of data issues: such as aggregation, etc in December 2015 data. | 1.0 | Documentation of known issues - Known issues in Dec FLUXNET 2015 release.
- On Ameriflux, we need to know what has changed between April 2016 release and December release
- Also list of data issues: such as aggregation, etc in December 2015 data. | non_priority | documentation of known issues known issues in dec fluxnet release on ameriflux we need to know what has changed between april release and december release also list of data issues such as aggregation etc in december data | 0 |
411,885 | 12,033,194,972 | IssuesEvent | 2020-04-13 13:47:14 | deora-earth/tealgarden | https://api.github.com/repos/deora-earth/tealgarden | closed | Setup Open Source Analytics tool (PP) | 02 Medium Priority 350 Deora RepPoints | <!--
# Simple Summary
This policy allows to write out rewards to complete required tasks. Completed tasks are payed by the deora council to the claiming member.
# How to create a new bounty?
1. To start you'll have to fill out the bounty form below.
- If the bounty spans across multiple repositories, consider splitting it in a smaller per-repo bounties if possible.
- If the bounty is larger than M, then the best known expert in the bounty matter should be consulted and included in an
"Expert" field in the bounty description.
2. Communicate the bounty to the organisation by submitting the following form:
https://forms.gle/STSNjTBGygNtTUwLA
- The bounty will get published on the deora communication channel.
# Bounty sizes
XS / 50 to 200 / DAI
S / 200 to 350 / DAI
M / 350 to 550 / DAI
L / 550 to 900 / DAI
XL / 900 to 1400 / DAI
You can specify the range individually under #Roles
# Pair programming
If 2 people claim the bounty together, the payout increases by 1.5x.
# Bounty Challenge
Once a bounty is assigned, the worker is asked to start working immediately on the issue.
If the worker feels blocked in execution, he/she has to communicate the tensions to the gardener.
Only if tensions are not reported and the bounty get's no further attention, anyone can challenge the bounty or takeover.
Bounties should be delivered within time, even if work is left to be performed. Leftover work can be tackled by submitting a new bounty with support by the organisation.
Bounty forking: complexity of bounties that has been undersized can be forked out by a new bounty submission.
**START DESCRIBING YOUR BOUNTY HERE:**
-->
# Bounty
The website is live and ready for their first hungry user. To see who, why and how they are visiting we need a tracking tool that matches our values
- [matomo](https://matomo.org/matomo-on-premise/)
- [openwebanalytics](http://www.openwebanalytics.com/)
- [countly](https://count.ly/pricing)
## Scope
- Check the tools and find the best
- set it up
## Deliverables
- working tracking tool, maybe a PR
## Gain for the project
good data analytics
## Roles
bounty gardener: @cyan-one / 10%
bounty worker: @johannbarbie @cyan-one / 80%
bounty reviewer: name / 10%
| 1.0 | Setup Open Source Analytics tool (PP) - <!--
# Simple Summary
This policy allows to write out rewards to complete required tasks. Completed tasks are payed by the deora council to the claiming member.
# How to create a new bounty?
1. To start you'll have to fill out the bounty form below.
- If the bounty spans across multiple repositories, consider splitting it in a smaller per-repo bounties if possible.
- If the bounty is larger than M, then the best known expert in the bounty matter should be consulted and included in an
"Expert" field in the bounty description.
2. Communicate the bounty to the organisation by submitting the following form:
https://forms.gle/STSNjTBGygNtTUwLA
- The bounty will get published on the deora communication channel.
# Bounty sizes
XS / 50 to 200 / DAI
S / 200 to 350 / DAI
M / 350 to 550 / DAI
L / 550 to 900 / DAI
XL / 900 to 1400 / DAI
You can specify the range individually under #Roles
# Pair programming
If 2 people claim the bounty together, the payout increases by 1.5x.
# Bounty Challenge
Once a bounty is assigned, the worker is asked to start working immediately on the issue.
If the worker feels blocked in execution, he/she has to communicate the tensions to the gardener.
Only if tensions are not reported and the bounty get's no further attention, anyone can challenge the bounty or takeover.
Bounties should be delivered within time, even if work is left to be performed. Leftover work can be tackled by submitting a new bounty with support by the organisation.
Bounty forking: complexity of bounties that has been undersized can be forked out by a new bounty submission.
**START DESCRIBING YOUR BOUNTY HERE:**
-->
# Bounty
The website is live and ready for their first hungry user. To see who, why and how they are visiting we need a tracking tool that matches our values
- [matomo](https://matomo.org/matomo-on-premise/)
- [openwebanalytics](http://www.openwebanalytics.com/)
- [countly](https://count.ly/pricing)
## Scope
- Check the tools and find the best
- set it up
## Deliverables
- working tracking tool, maybe a PR
## Gain for the project
good data analytics
## Roles
bounty gardener: @cyan-one / 10%
bounty worker: @johannbarbie @cyan-one / 80%
bounty reviewer: name / 10%
| priority | setup open source analytics tool pp simple summary this policy allows to write out rewards to complete required tasks completed tasks are payed by the deora council to the claiming member how to create a new bounty to start you ll have to fill out the bounty form below if the bounty spans across multiple repositories consider splitting it in a smaller per repo bounties if possible if the bounty is larger than m then the best known expert in the bounty matter should be consulted and included in an expert field in the bounty description communicate the bounty to the organisation by submitting the following form the bounty will get published on the deora communication channel bounty sizes xs to dai s to dai m to dai l to dai xl to dai you can specify the range individually under roles pair programming if people claim the bounty together the payout increases by bounty challenge once a bounty is assigned the worker is asked to start working immediately on the issue if the worker feels blocked in execution he she has to communicate the tensions to the gardener only if tensions are not reported and the bounty get s no further attention anyone can challenge the bounty or takeover bounties should be delivered within time even if work is left to be performed leftover work can be tackled by submitting a new bounty with support by the organisation bounty forking complexity of bounties that has been undersized can be forked out by a new bounty submission start describing your bounty here bounty the website is live and ready for their first hungry user to see who why and how they are visiting we need a tracking tool that matches our values scope check the tools and find the best set it up deliverables working tracking tool maybe a pr gain for the project good data analytics roles bounty gardener cyan one bounty worker johannbarbie cyan one bounty reviewer name | 1 |
721,755 | 24,836,662,172 | IssuesEvent | 2022-10-26 09:22:25 | mlcommons/mobile_app_open | https://api.github.com/repos/mlcommons/mobile_app_open | closed | Save CPU model in result | priority:high release:3.0 size:big | Currently we save very limited info about app environment: os type, os version, manufacturer, model id.
We should at least also save CPU model. | 1.0 | Save CPU model in result - Currently we save very limited info about app environment: os type, os version, manufacturer, model id.
We should at least also save CPU model. | priority | save cpu model in result currently we save very limited info about app environment os type os version manufacturer model id we should at least also save cpu model | 1 |
401,987 | 11,800,590,408 | IssuesEvent | 2020-03-18 17:49:09 | mozilla/addons | https://api.github.com/repos/mozilla/addons | closed | Cannot Upload New Version of Add-On Built With WebExtensions | component: developer hub priority: p4 state: stale | I have an existing add-on that I developed a few years ago using the SDK (package.json, etc.). I rewrote the add-on using WebExtensions (manifest.json, etc.). I tested the new version successfully and everything worked fine. When I went to AMO to upload the new version, 'All Platforms' was checked by default, and I could check and uncheck that box or the boxes next to any of the other platforms (Windows, Linux, Mac, Android, etc.). I selected a file to upload and eventually was shown a message informing me that the file was uploaded and validated successfully with 0 errors. However, all of the Platform checkboxes got automatically unchecked after the file upload, and furthermore, they were all disabled. If I click the 'Add Version' button on that form to continue, I just get an error message telling me 'Need at least one platform.' Since I can't check any of the Platform checkboxes because they are now disabled, I can't continue. As a test, I went ahead and selected my older version's file to upload (the version that was developed using the SDK), and I didn't experience the unchecking and disabling of the Platform checkboxes.



| 1.0 | Cannot Upload New Version of Add-On Built With WebExtensions - I have an existing add-on that I developed a few years ago using the SDK (package.json, etc.). I rewrote the add-on using WebExtensions (manifest.json, etc.). I tested the new version successfully and everything worked fine. When I went to AMO to upload the new version, 'All Platforms' was checked by default, and I could check and uncheck that box or the boxes next to any of the other platforms (Windows, Linux, Mac, Android, etc.). I selected a file to upload and eventually was shown a message informing me that the file was uploaded and validated successfully with 0 errors. However, all of the Platform checkboxes got automatically unchecked after the file upload, and furthermore, they were all disabled. If I click the 'Add Version' button on that form to continue, I just get an error message telling me 'Need at least one platform.' Since I can't check any of the Platform checkboxes because they are now disabled, I can't continue. As a test, I went ahead and selected my older version's file to upload (the version that was developed using the SDK), and I didn't experience the unchecking and disabling of the Platform checkboxes.



| priority | cannot upload new version of add on built with webextensions i have an existing add on that i developed a few years ago using the sdk package json etc i rewrote the add on using webextensions manifest json etc i tested the new version successfully and everything worked fine when i went to amo to upload the new version all platforms was checked by default and i could check and uncheck that box or the boxes next to any of the other platforms windows linux mac android etc i selected a file to upload and eventually was shown a message informing me that the file was uploaded and validated successfully with errors however all of the platform checkboxes got automatically unchecked after the file upload and furthermore they were all disabled if i click the add version button on that form to continue i just get an error message telling me need at least one platform since i can t check any of the platform checkboxes because they are now disabled i can t continue as a test i went ahead and selected my older version s file to upload the version that was developed using the sdk and i didn t experience the unchecking and disabling of the platform checkboxes | 1 |
414,276 | 12,101,442,244 | IssuesEvent | 2020-04-20 15:13:13 | codesmithtools/Templates | https://api.github.com/repos/codesmithtools/Templates | closed | Add GuidAttribute to PK Accessors | Framework-PLINQO Priority-High Type-Enhancement auto-migrated | ```
Please modify the PLINQO templates to automatically add the "GuidAttribute"
attribute to the accessor for the primary key columns were the ColumnSchema
does not report a default value. It is my opinion that this will increase the
usability of the templates and save the user additional time.
Currently, CS/PLINQO users must either manually stuff a value for the key
property, before doing an Insert, when the PK is a uniqueidentifier and the
column does not have a default value. The other option is for them to manually
alter the Metadata class and add the "GuidAttribute" attribute to allow
framework to automatically set the value. Since these are the only two options
for user, and since the framework will handle the latter automatically, it only
make sense that it be added to the template.
Thanx,
Fred
```
Original issue reported on code.google.com by `fred.lac...@gmail.com` on 1 Dec 2010 at 12:02
| 1.0 | Add GuidAttribute to PK Accessors - ```
Please modify the PLINQO templates to automatically add the "GuidAttribute"
attribute to the accessor for the primary key columns were the ColumnSchema
does not report a default value. It is my opinion that this will increase the
usability of the templates and save the user additional time.
Currently, CS/PLINQO users must either manually stuff a value for the key
property, before doing an Insert, when the PK is a uniqueidentifier and the
column does not have a default value. The other option is for them to manually
alter the Metadata class and add the "GuidAttribute" attribute to allow
framework to automatically set the value. Since these are the only two options
for user, and since the framework will handle the latter automatically, it only
make sense that it be added to the template.
Thanx,
Fred
```
Original issue reported on code.google.com by `fred.lac...@gmail.com` on 1 Dec 2010 at 12:02
| priority | add guidattribute to pk accessors please modify the plinqo templates to automatically add the guidattribute attribute to the accessor for the primary key columns were the columnschema does not report a default value it is my opinion that this will increase the usability of the templates and save the user additional time currently cs plinqo users must either manually stuff a value for the key property before doing an insert when the pk is a uniqueidentifier and the column does not have a default value the other option is for them to manually alter the metadata class and add the guidattribute attribute to allow framework to automatically set the value since these are the only two options for user and since the framework will handle the latter automatically it only make sense that it be added to the template thanx fred original issue reported on code google com by fred lac gmail com on dec at | 1 |
28,919 | 11,706,022,387 | IssuesEvent | 2020-03-07 19:26:28 | vlaship/spark | https://api.github.com/repos/vlaship/spark | opened | CVE-2018-10237 (Medium) detected in guava-16.0.1.jar | security vulnerability | ## CVE-2018-10237 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>guava-16.0.1.jar</b></p></summary>
<p>Guava is a suite of core and expanded libraries that include
utility classes, google's collections, io classes, and much
much more.
Guava has only one code dependency - javax.annotation,
per the JSR-305 spec.</p>
<p>Library home page: <a href="http://code.google.com/">http://code.google.com/</a></p>
<p>Path to dependency file: /tmp/ws-scm/spark/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.google.guava/guava/16.0.1/5fa98cd1a63c99a44dd8d3b77e4762b066a5d0c5/guava-16.0.1.jar,/root/.gradle/caches/modules-2/files-2.1/com.google.guava/guava/16.0.1/5fa98cd1a63c99a44dd8d3b77e4762b066a5d0c5/guava-16.0.1.jar</p>
<p>
Dependency Hierarchy:
- spark-sql_2.12-2.4.4.jar (Root Library)
- spark-core_2.12-2.4.4.jar
- hadoop-client-2.6.5.jar
- hadoop-hdfs-2.6.5.jar
- :x: **guava-16.0.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/vlaship/spark/commit/2d517982709d6cf8718c3af48ef079795b448bd9">2d517982709d6cf8718c3af48ef079795b448bd9</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Unbounded memory allocation in Google Guava 11.0 through 24.x before 24.1.1 allows remote attackers to conduct denial of service attacks against servers that depend on this library and deserialize attacker-provided data, because the AtomicDoubleArray class (when serialized with Java serialization) and the CompoundOrdering class (when serialized with GWT serialization) perform eager allocation without appropriate checks on what a client has sent and whether the data size is reasonable.
<p>Publish Date: 2018-04-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-10237>CVE-2018-10237</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-10237">https://nvd.nist.gov/vuln/detail/CVE-2018-10237</a></p>
<p>Release Date: 2018-04-26</p>
<p>Fix Resolution: 24.1.1-jre, 24.1.1-android</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-10237 (Medium) detected in guava-16.0.1.jar - ## CVE-2018-10237 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>guava-16.0.1.jar</b></p></summary>
<p>Guava is a suite of core and expanded libraries that include
utility classes, google's collections, io classes, and much
much more.
Guava has only one code dependency - javax.annotation,
per the JSR-305 spec.</p>
<p>Library home page: <a href="http://code.google.com/">http://code.google.com/</a></p>
<p>Path to dependency file: /tmp/ws-scm/spark/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.google.guava/guava/16.0.1/5fa98cd1a63c99a44dd8d3b77e4762b066a5d0c5/guava-16.0.1.jar,/root/.gradle/caches/modules-2/files-2.1/com.google.guava/guava/16.0.1/5fa98cd1a63c99a44dd8d3b77e4762b066a5d0c5/guava-16.0.1.jar</p>
<p>
Dependency Hierarchy:
- spark-sql_2.12-2.4.4.jar (Root Library)
- spark-core_2.12-2.4.4.jar
- hadoop-client-2.6.5.jar
- hadoop-hdfs-2.6.5.jar
- :x: **guava-16.0.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/vlaship/spark/commit/2d517982709d6cf8718c3af48ef079795b448bd9">2d517982709d6cf8718c3af48ef079795b448bd9</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Unbounded memory allocation in Google Guava 11.0 through 24.x before 24.1.1 allows remote attackers to conduct denial of service attacks against servers that depend on this library and deserialize attacker-provided data, because the AtomicDoubleArray class (when serialized with Java serialization) and the CompoundOrdering class (when serialized with GWT serialization) perform eager allocation without appropriate checks on what a client has sent and whether the data size is reasonable.
<p>Publish Date: 2018-04-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-10237>CVE-2018-10237</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-10237">https://nvd.nist.gov/vuln/detail/CVE-2018-10237</a></p>
<p>Release Date: 2018-04-26</p>
<p>Fix Resolution: 24.1.1-jre, 24.1.1-android</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in guava jar cve medium severity vulnerability vulnerable library guava jar guava is a suite of core and expanded libraries that include utility classes google s collections io classes and much much more guava has only one code dependency javax annotation per the jsr spec library home page a href path to dependency file tmp ws scm spark build gradle path to vulnerable library root gradle caches modules files com google guava guava guava jar root gradle caches modules files com google guava guava guava jar dependency hierarchy spark sql jar root library spark core jar hadoop client jar hadoop hdfs jar x guava jar vulnerable library found in head commit a href vulnerability details unbounded memory allocation in google guava through x before allows remote attackers to conduct denial of service attacks against servers that depend on this library and deserialize attacker provided data because the atomicdoublearray class when serialized with java serialization and the compoundordering class when serialized with gwt serialization perform eager allocation without appropriate checks on what a client has sent and whether the data size is reasonable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jre android step up your open source security game with whitesource | 0 |
68,442 | 14,923,644,442 | IssuesEvent | 2021-01-23 19:52:55 | NixOS/nixpkgs | https://api.github.com/repos/NixOS/nixpkgs | closed | Vulnerability roundup 85: imagemagick-7.0.10-17: 1 advisory [7.1] | 1.severity: security | [search](https://search.nix.gsc.io/?q=imagemagick&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=imagemagick+in%3Apath&type=Code)
* [ ] [CVE-2020-13902](https://nvd.nist.gov/vuln/detail/CVE-2020-13902) CVSSv3=7.1 (nixos-unstable)
Scanned versions: nixos-unstable: 22c98819ccd. May contain false positives.
| True | Vulnerability roundup 85: imagemagick-7.0.10-17: 1 advisory [7.1] - [search](https://search.nix.gsc.io/?q=imagemagick&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=imagemagick+in%3Apath&type=Code)
* [ ] [CVE-2020-13902](https://nvd.nist.gov/vuln/detail/CVE-2020-13902) CVSSv3=7.1 (nixos-unstable)
Scanned versions: nixos-unstable: 22c98819ccd. May contain false positives.
| non_priority | vulnerability roundup imagemagick advisory nixos unstable scanned versions nixos unstable may contain false positives | 0 |
139,994 | 12,883,398,703 | IssuesEvent | 2020-07-12 21:55:25 | mtkennerly/realia | https://api.github.com/repos/mtkennerly/realia | closed | Could use some doc comments | documentation | This is how docs.rs currently looks like


That's a bit empty in there... | 1.0 | Could use some doc comments - This is how docs.rs currently looks like


That's a bit empty in there... | non_priority | could use some doc comments this is how docs rs currently looks like that s a bit empty in there | 0 |
101,732 | 8,796,547,433 | IssuesEvent | 2018-12-23 08:36:58 | acacode/kinka | https://api.github.com/repos/acacode/kinka | closed | Testing engine and base unit tests | Testing enhancement | This ticket more needed for creating a testing engine (configs, karma, mocha)
And also needed to create some unit tests | 1.0 | Testing engine and base unit tests - This ticket more needed for creating a testing engine (configs, karma, mocha)
And also needed to create some unit tests | non_priority | testing engine and base unit tests this ticket more needed for creating a testing engine configs karma mocha and also needed to create some unit tests | 0 |
25,473 | 18,774,396,833 | IssuesEvent | 2021-11-07 12:15:53 | nekoguys/game_competition | https://api.github.com/repos/nekoguys/game_competition | opened | Актуализировать состояние тикетов, почистить борду | infrastructure sprint-1 (08.11 - present) | Надо посмотреть, что уже сделано, что не будет сделано вообще, что не будет сделано в ближайшее время.
Всем тикетам надо проставить лейблы, проверить актуальность, назначить исполнителей, если требуется.
__Обязательно:__ прочтите пожалуйста вики с инструкцией, там есть useful советики
_Дополнительно:_ хотел бы услышать замечания и предложения по работе борды, если такие будут :) | 1.0 | Актуализировать состояние тикетов, почистить борду - Надо посмотреть, что уже сделано, что не будет сделано вообще, что не будет сделано в ближайшее время.
Всем тикетам надо проставить лейблы, проверить актуальность, назначить исполнителей, если требуется.
__Обязательно:__ прочтите пожалуйста вики с инструкцией, там есть useful советики
_Дополнительно:_ хотел бы услышать замечания и предложения по работе борды, если такие будут :) | non_priority | актуализировать состояние тикетов почистить борду надо посмотреть что уже сделано что не будет сделано вообще что не будет сделано в ближайшее время всем тикетам надо проставить лейблы проверить актуальность назначить исполнителей если требуется обязательно прочтите пожалуйста вики с инструкцией там есть useful советики дополнительно хотел бы услышать замечания и предложения по работе борды если такие будут | 0 |
140,875 | 12,950,680,723 | IssuesEvent | 2020-07-19 14:07:20 | alvarosaburido/vue-dynamic-forms | https://api.github.com/repos/alvarosaburido/vue-dynamic-forms | opened | Login Demo on documentation | documentation |
**Describe the solution you'd like**
How to implement a login form with the library.
| 1.0 | Login Demo on documentation -
**Describe the solution you'd like**
How to implement a login form with the library.
| non_priority | login demo on documentation describe the solution you d like how to implement a login form with the library | 0 |
64,184 | 15,819,743,654 | IssuesEvent | 2021-04-05 17:57:05 | ironmansoftware/issues | https://api.github.com/repos/ironmansoftware/issues | closed | Linux PSModulePath for environment adds semi colon | Fixed in Nightly Build PowerShell Universal bug completed | #### Description
When setting up an environment (either through the GUI or environments.ps1 path) using Linux and pwsh 7, setting the PSModulePath add a semi colon to the path you are adding.
#### Additional Information
Example Code:
`New-PSUEnvironment -Name "pwsh" -Path "/usr/bin/pwsh" -PSModulePath @('/home/.PowerShellUniversal/Repository/Modules') `
Output in Universal Automation

Work around
Add a colon to the beginning of your path in the GUI or ps1 file.
`New-PSUEnvironment -Name "pwsh" -Path "/usr/bin/pwsh" -PSModulePath @(':/home/.PowerShellUniversal/Repository/Modules')`

#### Version Information
- Universal: 1.5.X
- Docker
- PowerShell: 7
- Operating System: Any Linux | 1.0 | Linux PSModulePath for environment adds semi colon - #### Description
When setting up an environment (either through the GUI or environments.ps1 path) using Linux and pwsh 7, setting the PSModulePath add a semi colon to the path you are adding.
#### Additional Information
Example Code:
`New-PSUEnvironment -Name "pwsh" -Path "/usr/bin/pwsh" -PSModulePath @('/home/.PowerShellUniversal/Repository/Modules') `
Output in Universal Automation

Work around
Add a colon to the beginning of your path in the GUI or ps1 file.
`New-PSUEnvironment -Name "pwsh" -Path "/usr/bin/pwsh" -PSModulePath @(':/home/.PowerShellUniversal/Repository/Modules')`

#### Version Information
- Universal: 1.5.X
- Docker
- PowerShell: 7
- Operating System: Any Linux | non_priority | linux psmodulepath for environment adds semi colon description when setting up an environment either through the gui or environments path using linux and pwsh setting the psmodulepath add a semi colon to the path you are adding additional information example code new psuenvironment name pwsh path usr bin pwsh psmodulepath home powershelluniversal repository modules output in universal automation work around add a colon to the beginning of your path in the gui or file new psuenvironment name pwsh path usr bin pwsh psmodulepath home powershelluniversal repository modules version information universal x docker powershell operating system any linux | 0 |
466,177 | 13,398,148,727 | IssuesEvent | 2020-09-03 12:44:59 | googleapis/google-cloud-php | https://api.github.com/repos/googleapis/google-cloud-php | closed | Synthesis failed for pubsub | api: pubsub autosynth failure priority: p1 type: bug | Hello! Autosynth couldn't regenerate pubsub. :broken_heart:
Here's the output from running `synth.py`:
```
eb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 4))
Using cached https://files.pythonhosted.org/packages/30/9e/f663a2aa66a09d838042ae1a2c5659828bb9b41ea3a6efa20a20fd92b121/Jinja2-2.11.2-py2.py3-none-any.whl
Saved ./Jinja2-2.11.2-py2.py3-none-any.whl
Collecting MarkupSafe==1.1.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 5))
Using cached https://files.pythonhosted.org/packages/b2/5f/23e0023be6bb885d00ffbefad2942bc51a620328ee910f64abe5a8d18dd1/MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl
Saved ./MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl
Collecting protobuf==3.13.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 6))
Using cached https://files.pythonhosted.org/packages/30/79/510974552cebff2ba04038544799450defe75e96ea5f1675dbf72cc8744f/protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl
Saved ./protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl
Collecting pypandoc==1.5 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 7))
Using cached https://files.pythonhosted.org/packages/d6/b7/5050dc1769c8a93d3ec7c4bd55be161991c94b8b235f88bf7c764449e708/pypandoc-1.5.tar.gz
Complete output from command python setup.py egg_info:
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmpfs/tmp/tmprfrbzoct/setuptools-tmp/setuptools/__init__.py", line 6, in <module>
import distutils.core
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/_distutils_hack/__init__.py", line 83, in create_module
return importlib.import_module('setuptools._distutils')
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
ModuleNotFoundError: No module named 'setuptools._distutils'
----------------------------------------
(Command "python setup.py egg_info" failed with error code 1 in /tmpfs/tmp/pip-build-unbd5bic/pypandoc/
)
ERROR: no such package '@gapic_generator_python_pip_deps//': pip_import failed: Collecting click==7.1.2 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 1))
Using cached https://files.pythonhosted.org/packages/d2/3d/fa76db83bf75c4f8d338c2fd15c8d33fdd7ad23a9b5e57eb6c5de26b430e/click-7.1.2-py2.py3-none-any.whl
Saved ./click-7.1.2-py2.py3-none-any.whl
Collecting google-api-core==1.22.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 2))
Using cached https://files.pythonhosted.org/packages/e0/2d/7c6c75013105e1d2b6eaa1bf18a56995be1dbc673c38885aea31136e9918/google_api_core-1.22.1-py2.py3-none-any.whl
Saved ./google_api_core-1.22.1-py2.py3-none-any.whl
Collecting googleapis-common-protos==1.52.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 3))
Using cached https://files.pythonhosted.org/packages/03/74/3956721ea1eb4bcf7502a311fdaa60b85bd751de4e57d1943afe9b334141/googleapis_common_protos-1.52.0-py2.py3-none-any.whl
Saved ./googleapis_common_protos-1.52.0-py2.py3-none-any.whl
Collecting jinja2==2.11.2 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 4))
Using cached https://files.pythonhosted.org/packages/30/9e/f663a2aa66a09d838042ae1a2c5659828bb9b41ea3a6efa20a20fd92b121/Jinja2-2.11.2-py2.py3-none-any.whl
Saved ./Jinja2-2.11.2-py2.py3-none-any.whl
Collecting MarkupSafe==1.1.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 5))
Using cached https://files.pythonhosted.org/packages/b2/5f/23e0023be6bb885d00ffbefad2942bc51a620328ee910f64abe5a8d18dd1/MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl
Saved ./MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl
Collecting protobuf==3.13.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 6))
Using cached https://files.pythonhosted.org/packages/30/79/510974552cebff2ba04038544799450defe75e96ea5f1675dbf72cc8744f/protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl
Saved ./protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl
Collecting pypandoc==1.5 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 7))
Using cached https://files.pythonhosted.org/packages/d6/b7/5050dc1769c8a93d3ec7c4bd55be161991c94b8b235f88bf7c764449e708/pypandoc-1.5.tar.gz
Complete output from command python setup.py egg_info:
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmpfs/tmp/tmprfrbzoct/setuptools-tmp/setuptools/__init__.py", line 6, in <module>
import distutils.core
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/_distutils_hack/__init__.py", line 83, in create_module
return importlib.import_module('setuptools._distutils')
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
ModuleNotFoundError: No module named 'setuptools._distutils'
----------------------------------------
(Command "python setup.py egg_info" failed with error code 1 in /tmpfs/tmp/pip-build-unbd5bic/pypandoc/
)
INFO: Elapsed time: 2.307s
INFO: 0 processes.
FAILED: Build did NOT complete successfully (0 packages loaded)
FAILED: Build did NOT complete successfully (0 packages loaded)
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/home/kbuilder/.cache/synthtool/google-cloud-php/PubSub/synth.py", line 29, in <module>
bazel_target='//google/pubsub/v1:google-cloud-pubsub-v1-php',
File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 58, in php_library
return self._generate_code(service, version, "php", **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 183, in _generate_code
shell.run(bazel_run_args)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['bazel', '--max_idle_secs=240', 'build', '//google/pubsub/v1:google-cloud-pubsub-v1-php']' returned non-zero exit status 1.
2020-09-02 03:03:31,938 autosynth [ERROR] > Synthesis failed
2020-09-02 03:03:31,938 autosynth [DEBUG] > Running: git reset --hard HEAD
HEAD is now at 8038912e chore(deps): update dependency google/cloud-tools to ^0.11.0 (#3363)
2020-09-02 03:03:31,972 autosynth [DEBUG] > Running: git checkout autosynth-pubsub
Switched to branch 'autosynth-pubsub'
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 690, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 539, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 670, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 375, in synthesize_loop
has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 273, in synthesize_version_in_new_branch
synthesizer.synthesize(synth_log_path, self.environ)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/c2163f87-9245-498a-8edb-c7613ed58dfe/targets/github%2Fsynthtool;config=default/tests;query=google-cloud-php;failed=false).
| 1.0 | Synthesis failed for pubsub - Hello! Autosynth couldn't regenerate pubsub. :broken_heart:
Here's the output from running `synth.py`:
```
eb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 4))
Using cached https://files.pythonhosted.org/packages/30/9e/f663a2aa66a09d838042ae1a2c5659828bb9b41ea3a6efa20a20fd92b121/Jinja2-2.11.2-py2.py3-none-any.whl
Saved ./Jinja2-2.11.2-py2.py3-none-any.whl
Collecting MarkupSafe==1.1.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 5))
Using cached https://files.pythonhosted.org/packages/b2/5f/23e0023be6bb885d00ffbefad2942bc51a620328ee910f64abe5a8d18dd1/MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl
Saved ./MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl
Collecting protobuf==3.13.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 6))
Using cached https://files.pythonhosted.org/packages/30/79/510974552cebff2ba04038544799450defe75e96ea5f1675dbf72cc8744f/protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl
Saved ./protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl
Collecting pypandoc==1.5 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 7))
Using cached https://files.pythonhosted.org/packages/d6/b7/5050dc1769c8a93d3ec7c4bd55be161991c94b8b235f88bf7c764449e708/pypandoc-1.5.tar.gz
Complete output from command python setup.py egg_info:
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmpfs/tmp/tmprfrbzoct/setuptools-tmp/setuptools/__init__.py", line 6, in <module>
import distutils.core
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/_distutils_hack/__init__.py", line 83, in create_module
return importlib.import_module('setuptools._distutils')
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
ModuleNotFoundError: No module named 'setuptools._distutils'
----------------------------------------
(Command "python setup.py egg_info" failed with error code 1 in /tmpfs/tmp/pip-build-unbd5bic/pypandoc/
)
ERROR: no such package '@gapic_generator_python_pip_deps//': pip_import failed: Collecting click==7.1.2 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 1))
Using cached https://files.pythonhosted.org/packages/d2/3d/fa76db83bf75c4f8d338c2fd15c8d33fdd7ad23a9b5e57eb6c5de26b430e/click-7.1.2-py2.py3-none-any.whl
Saved ./click-7.1.2-py2.py3-none-any.whl
Collecting google-api-core==1.22.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 2))
Using cached https://files.pythonhosted.org/packages/e0/2d/7c6c75013105e1d2b6eaa1bf18a56995be1dbc673c38885aea31136e9918/google_api_core-1.22.1-py2.py3-none-any.whl
Saved ./google_api_core-1.22.1-py2.py3-none-any.whl
Collecting googleapis-common-protos==1.52.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 3))
Using cached https://files.pythonhosted.org/packages/03/74/3956721ea1eb4bcf7502a311fdaa60b85bd751de4e57d1943afe9b334141/googleapis_common_protos-1.52.0-py2.py3-none-any.whl
Saved ./googleapis_common_protos-1.52.0-py2.py3-none-any.whl
Collecting jinja2==2.11.2 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 4))
Using cached https://files.pythonhosted.org/packages/30/9e/f663a2aa66a09d838042ae1a2c5659828bb9b41ea3a6efa20a20fd92b121/Jinja2-2.11.2-py2.py3-none-any.whl
Saved ./Jinja2-2.11.2-py2.py3-none-any.whl
Collecting MarkupSafe==1.1.1 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 5))
Using cached https://files.pythonhosted.org/packages/b2/5f/23e0023be6bb885d00ffbefad2942bc51a620328ee910f64abe5a8d18dd1/MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl
Saved ./MarkupSafe-1.1.1-cp36-cp36m-manylinux1_x86_64.whl
Collecting protobuf==3.13.0 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 6))
Using cached https://files.pythonhosted.org/packages/30/79/510974552cebff2ba04038544799450defe75e96ea5f1675dbf72cc8744f/protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl
Saved ./protobuf-3.13.0-cp36-cp36m-manylinux1_x86_64.whl
Collecting pypandoc==1.5 (from -r /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/gapic_generator_python/requirements.txt (line 7))
Using cached https://files.pythonhosted.org/packages/d6/b7/5050dc1769c8a93d3ec7c4bd55be161991c94b8b235f88bf7c764449e708/pypandoc-1.5.tar.gz
Complete output from command python setup.py egg_info:
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmpfs/tmp/tmprfrbzoct/setuptools-tmp/setuptools/__init__.py", line 6, in <module>
import distutils.core
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/_distutils_hack/__init__.py", line 83, in create_module
return importlib.import_module('setuptools._distutils')
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/importlib/__init__.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
ModuleNotFoundError: No module named 'setuptools._distutils'
----------------------------------------
(Command "python setup.py egg_info" failed with error code 1 in /tmpfs/tmp/pip-build-unbd5bic/pypandoc/
)
INFO: Elapsed time: 2.307s
INFO: 0 processes.
FAILED: Build did NOT complete successfully (0 packages loaded)
FAILED: Build did NOT complete successfully (0 packages loaded)
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/home/kbuilder/.cache/synthtool/google-cloud-php/PubSub/synth.py", line 29, in <module>
bazel_target='//google/pubsub/v1:google-cloud-pubsub-v1-php',
File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 58, in php_library
return self._generate_code(service, version, "php", **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 183, in _generate_code
shell.run(bazel_run_args)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['bazel', '--max_idle_secs=240', 'build', '//google/pubsub/v1:google-cloud-pubsub-v1-php']' returned non-zero exit status 1.
2020-09-02 03:03:31,938 autosynth [ERROR] > Synthesis failed
2020-09-02 03:03:31,938 autosynth [DEBUG] > Running: git reset --hard HEAD
HEAD is now at 8038912e chore(deps): update dependency google/cloud-tools to ^0.11.0 (#3363)
2020-09-02 03:03:31,972 autosynth [DEBUG] > Running: git checkout autosynth-pubsub
Switched to branch 'autosynth-pubsub'
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 690, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 539, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 670, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 375, in synthesize_loop
has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 273, in synthesize_version_in_new_branch
synthesizer.synthesize(synth_log_path, self.environ)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/c2163f87-9245-498a-8edb-c7613ed58dfe/targets/github%2Fsynthtool;config=default/tests;query=google-cloud-php;failed=false).
| priority | synthesis failed for pubsub hello autosynth couldn t regenerate pubsub broken heart here s the output from running synth py external gapic generator python requirements txt line using cached saved none any whl collecting markupsafe from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved markupsafe whl collecting protobuf from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved protobuf whl collecting pypandoc from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached complete output from command python setup py egg info traceback most recent call last file line in file tmpfs tmp tmprfrbzoct setuptools tmp setuptools init py line in import distutils core file tmpfs src github synthtool env lib site packages distutils hack init py line in create module return importlib import module setuptools distutils file home kbuilder pyenv versions lib importlib init py line in import module return bootstrap gcd import name package level modulenotfounderror no module named setuptools distutils command python setup py egg info failed with error code in tmpfs tmp pip build pypandoc error no such package gapic generator python pip deps pip import failed collecting click from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved click none any whl collecting google api core from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved google api core none any whl collecting googleapis common protos from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved googleapis common protos none any whl collecting from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved none any whl collecting markupsafe from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved markupsafe whl collecting protobuf from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached saved protobuf whl collecting pypandoc from r home kbuilder cache bazel bazel kbuilder external gapic generator python requirements txt line using cached complete output from command python setup py egg info traceback most recent call last file line in file tmpfs tmp tmprfrbzoct setuptools tmp setuptools init py line in import distutils core file tmpfs src github synthtool env lib site packages distutils hack init py line in create module return importlib import module setuptools distutils file home kbuilder pyenv versions lib importlib init py line in import module return bootstrap gcd import name package level modulenotfounderror no module named setuptools distutils command python setup py egg info failed with error code in tmpfs tmp pip build pypandoc info elapsed time info processes failed build did not complete successfully packages loaded failed build did not complete successfully packages loaded traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file home kbuilder cache synthtool google cloud php pubsub synth py line in bazel target google pubsub google cloud pubsub php file tmpfs src github synthtool synthtool gcp gapic bazel py line in php library return self generate code service version php kwargs file tmpfs src github synthtool synthtool gcp gapic bazel py line in generate code shell run bazel run args file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status autosynth synthesis failed autosynth running git reset hard head head is now at chore deps update dependency google cloud tools to autosynth running git checkout autosynth pubsub switched to branch autosynth pubsub traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop has changes toolbox synthesize version in new branch synthesizer youngest file tmpfs src github synthtool autosynth synth py line in synthesize version in new branch synthesizer synthesize synth log path self environ file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log | 1 |
300,400 | 9,210,344,910 | IssuesEvent | 2019-03-09 04:07:21 | WEBsinjyuku/meister-hackers | https://api.github.com/repos/WEBsinjyuku/meister-hackers | closed | 投稿画面(edit) validateエラー時のレイアウトが変わる | high priority バグ | validateエラー時、募集ステータスとキャンセルボタンが非表示になる。
【edit初期画面】

【Validateエラー時】

| 1.0 | 投稿画面(edit) validateエラー時のレイアウトが変わる - validateエラー時、募集ステータスとキャンセルボタンが非表示になる。
【edit初期画面】

【Validateエラー時】

| priority | 投稿画面(edit) validateエラー時のレイアウトが変わる validateエラー時、募集ステータスとキャンセルボタンが非表示になる。 【edit初期画面】 【validateエラー時】 | 1 |
4,234 | 2,714,575,083 | IssuesEvent | 2015-04-10 05:39:47 | SatelliteQE/robottelo | https://api.github.com/repos/SatelliteQE/robottelo | closed | Fix tests.foreman.ui.test_repository.Repos.test_create_repo_3 | Tests UI | Looks like at the time of adding the testcase in the UI automation, adding upstream repo name "Upstream Repository Name" was not mandatory, which now is required I think. Without which submission fails and the search for it too. | 1.0 | Fix tests.foreman.ui.test_repository.Repos.test_create_repo_3 - Looks like at the time of adding the testcase in the UI automation, adding upstream repo name "Upstream Repository Name" was not mandatory, which now is required I think. Without which submission fails and the search for it too. | non_priority | fix tests foreman ui test repository repos test create repo looks like at the time of adding the testcase in the ui automation adding upstream repo name upstream repository name was not mandatory which now is required i think without which submission fails and the search for it too | 0 |
408,685 | 11,950,815,769 | IssuesEvent | 2020-04-03 15:48:53 | levylabpitt/Multichannel-Lockin | https://api.github.com/repos/levylabpitt/Multichannel-Lockin | opened | Generalize simulation mode | Priority: Medium Status: Pending Type: Enhancement | - How to better simulate analog inputs?
- How to have more external inputs (Temperatre, Magnet Field, etc) that are not directly controlled by the Multichannel Lockin itself? | 1.0 | Generalize simulation mode - - How to better simulate analog inputs?
- How to have more external inputs (Temperatre, Magnet Field, etc) that are not directly controlled by the Multichannel Lockin itself? | priority | generalize simulation mode how to better simulate analog inputs how to have more external inputs temperatre magnet field etc that are not directly controlled by the multichannel lockin itself | 1 |
230,016 | 17,599,038,706 | IssuesEvent | 2021-08-17 09:28:16 | alan-turing-institute/sktime | https://api.github.com/repos/alan-turing-institute/sktime | closed | [DOC] fix incorrect docstrings of dataset loaders in `datasets` module | documentation module:classification | The docstrings for the loaders of TSC datasets in the `datasets` module are incorrect.
The return object is conditional on the splitting (X/y, train/test) of the data, yet the docstring insists it's always two objects that are returned - this is not accurate and should be fixed. | 1.0 | [DOC] fix incorrect docstrings of dataset loaders in `datasets` module - The docstrings for the loaders of TSC datasets in the `datasets` module are incorrect.
The return object is conditional on the splitting (X/y, train/test) of the data, yet the docstring insists it's always two objects that are returned - this is not accurate and should be fixed. | non_priority | fix incorrect docstrings of dataset loaders in datasets module the docstrings for the loaders of tsc datasets in the datasets module are incorrect the return object is conditional on the splitting x y train test of the data yet the docstring insists it s always two objects that are returned this is not accurate and should be fixed | 0 |
169,797 | 6,417,592,820 | IssuesEvent | 2017-08-08 17:05:35 | GoogleCloudPlatform/google-cloud-python | https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python | closed | Allow returning values from transactional functions | api: spanner priority: p1 status: acknowledged type: bug | Somebody asks in the Slack:
> Question about spanner python APIs, specifically transactions.
> In a read-write transaction, is it possible to return the updated
> value without hitting the database again?
>
> let's say I have the following transaction (pretty much the example from the docs)
> ```def f(transaction):
> second_album_result = transaction.read(
> table='Albums', columns=('MarketingBudget',),
> keyset=second_album_keyset, limit=1)
> second_album_row = list(second_album_result)[0]
> second_album_budget = second_album_row[0]
> second_album_budget += 10000000
> transaction.update(table=mytable, columns=('MarketingBudget'),
> keyset=second_album_keyset, values=[(second_album_budget)]
> )
> return second_album_budget
> database.run_in_transaction(f)
> ```
> is the `return` at all possible?
I believe that this is currently not possible in the Python library, because when you pass a closure to `run_in_transaction`, the return value always gets ignored, and it just returns the commit timestamp.
However this _is_ possible in the Cloud Spanner Java library, and being be able to return whatever you want from the body of a transactional function in the case that a transaction successfully commits is extremely useful. | 1.0 | Allow returning values from transactional functions - Somebody asks in the Slack:
> Question about spanner python APIs, specifically transactions.
> In a read-write transaction, is it possible to return the updated
> value without hitting the database again?
>
> let's say I have the following transaction (pretty much the example from the docs)
> ```def f(transaction):
> second_album_result = transaction.read(
> table='Albums', columns=('MarketingBudget',),
> keyset=second_album_keyset, limit=1)
> second_album_row = list(second_album_result)[0]
> second_album_budget = second_album_row[0]
> second_album_budget += 10000000
> transaction.update(table=mytable, columns=('MarketingBudget'),
> keyset=second_album_keyset, values=[(second_album_budget)]
> )
> return second_album_budget
> database.run_in_transaction(f)
> ```
> is the `return` at all possible?
I believe that this is currently not possible in the Python library, because when you pass a closure to `run_in_transaction`, the return value always gets ignored, and it just returns the commit timestamp.
However this _is_ possible in the Cloud Spanner Java library, and being be able to return whatever you want from the body of a transactional function in the case that a transaction successfully commits is extremely useful. | priority | allow returning values from transactional functions somebody asks in the slack question about spanner python apis specifically transactions in a read write transaction is it possible to return the updated value without hitting the database again let s say i have the following transaction pretty much the example from the docs def f transaction second album result transaction read table albums columns marketingbudget keyset second album keyset limit second album row list second album result second album budget second album row second album budget transaction update table mytable columns marketingbudget keyset second album keyset values return second album budget database run in transaction f is the return at all possible i believe that this is currently not possible in the python library because when you pass a closure to run in transaction the return value always gets ignored and it just returns the commit timestamp however this is possible in the cloud spanner java library and being be able to return whatever you want from the body of a transactional function in the case that a transaction successfully commits is extremely useful | 1 |
715,716 | 24,607,525,688 | IssuesEvent | 2022-10-14 17:43:38 | duckduckgo/p5-app-duckpan | https://api.github.com/repos/duckduckgo/p5-app-duckpan | closed | DuckPAN should warn about, or not follow redirects for API calls (to mimic our NGINX) | Suggestion Priority: Medium | We've run into issues with NGINX related to HTTP APIs that redirect to HTTPS or API calls that first return a 302 and point elsewhere.
DuckPAN currently follows these without problems, but our NGINX setup does not.
To aid developers, we should point this out to them so they notice ahead of time
| 1.0 | DuckPAN should warn about, or not follow redirects for API calls (to mimic our NGINX) - We've run into issues with NGINX related to HTTP APIs that redirect to HTTPS or API calls that first return a 302 and point elsewhere.
DuckPAN currently follows these without problems, but our NGINX setup does not.
To aid developers, we should point this out to them so they notice ahead of time
| priority | duckpan should warn about or not follow redirects for api calls to mimic our nginx we ve run into issues with nginx related to http apis that redirect to https or api calls that first return a and point elsewhere duckpan currently follows these without problems but our nginx setup does not to aid developers we should point this out to them so they notice ahead of time | 1 |
380,635 | 11,268,848,673 | IssuesEvent | 2020-01-14 07:26:49 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | imgur.com - see bug description | browser-firefox engine-gecko nsfw os-linux priority-critical | <!-- @browser: Firefox 71.0 -->
<!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:72.0) Gecko/20100101 Firefox/72.0 -->
<!-- @reported_with: -->
**URL**: https://imgur.com/
**Browser / Version**: Firefox 71.0
**Operating System**: Linux
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: Age confirmation check is broken by tracking protection
**Steps to Reproduce**:
- Open a private browsing window
- Go to https://www.redditery.com/r/example
- Replace "example" by the name of a NSFW subreddit you are comfortable with to test this bug. Here is a list http://redditlist.com/nsfw
- click on a image hosted at imgur.com
- if it's a direct image link, just remove the ".jpg" extension from the image
- You should be prompted about age confirmation
- On Chromium, you accept and you are taken to the page
- On Firefox your are again prompted about age confirmation
- Again and again
- disable tracking protection on the tab
- you are not stuck anymore
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | imgur.com - see bug description - <!-- @browser: Firefox 71.0 -->
<!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:72.0) Gecko/20100101 Firefox/72.0 -->
<!-- @reported_with: -->
**URL**: https://imgur.com/
**Browser / Version**: Firefox 71.0
**Operating System**: Linux
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: Age confirmation check is broken by tracking protection
**Steps to Reproduce**:
- Open a private browsing window
- Go to https://www.redditery.com/r/example
- Replace "example" by the name of a NSFW subreddit you are comfortable with to test this bug. Here is a list http://redditlist.com/nsfw
- click on a image hosted at imgur.com
- if it's a direct image link, just remove the ".jpg" extension from the image
- You should be prompted about age confirmation
- On Chromium, you accept and you are taken to the page
- On Firefox your are again prompted about age confirmation
- Again and again
- disable tracking protection on the tab
- you are not stuck anymore
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | imgur com see bug description url browser version firefox operating system linux tested another browser yes problem type something else description age confirmation check is broken by tracking protection steps to reproduce open a private browsing window go to replace example by the name of a nsfw subreddit you are comfortable with to test this bug here is a list click on a image hosted at imgur com if it s a direct image link just remove the jpg extension from the image you should be prompted about age confirmation on chromium you accept and you are taken to the page on firefox your are again prompted about age confirmation again and again disable tracking protection on the tab you are not stuck anymore browser configuration none from with ❤️ | 1 |
306,521 | 9,396,014,488 | IssuesEvent | 2019-04-08 05:38:56 | Brycey92/Galaxy-Craft-Issues | https://api.github.com/repos/Brycey92/Galaxy-Craft-Issues | closed | Potential nether portal dupe bug - (portal creation currently disabled) | help wanted hotfixed missing info priority | **Pack version**
1.0.3
**Describe the bug**
Some using SpongeForge have reported that throwing items through a nether portal will result in multiple of the item appearing on the other side. I was unable to reproduce the issue, but like #28, it may only happen when there's lag on the server. I'm not sure about the exact steps to reproduce the issue. If anyone can add more details, it would be very helpful!
**To reproduce**
1. ?
**Expected behavior**
Items thrown through a nether portal should appear on the other side in the quantity they entered.
If we can't fix this, we may need to leave nether portal creation banned and rely on commands or command blocks to teleport players to and from the nether. | 1.0 | Potential nether portal dupe bug - (portal creation currently disabled) - **Pack version**
1.0.3
**Describe the bug**
Some using SpongeForge have reported that throwing items through a nether portal will result in multiple of the item appearing on the other side. I was unable to reproduce the issue, but like #28, it may only happen when there's lag on the server. I'm not sure about the exact steps to reproduce the issue. If anyone can add more details, it would be very helpful!
**To reproduce**
1. ?
**Expected behavior**
Items thrown through a nether portal should appear on the other side in the quantity they entered.
If we can't fix this, we may need to leave nether portal creation banned and rely on commands or command blocks to teleport players to and from the nether. | priority | potential nether portal dupe bug portal creation currently disabled pack version describe the bug some using spongeforge have reported that throwing items through a nether portal will result in multiple of the item appearing on the other side i was unable to reproduce the issue but like it may only happen when there s lag on the server i m not sure about the exact steps to reproduce the issue if anyone can add more details it would be very helpful to reproduce expected behavior items thrown through a nether portal should appear on the other side in the quantity they entered if we can t fix this we may need to leave nether portal creation banned and rely on commands or command blocks to teleport players to and from the nether | 1 |
46,260 | 2,951,891,098 | IssuesEvent | 2015-07-07 03:59:11 | edwardyh/GlobalVRGameJam2015 | https://api.github.com/repos/edwardyh/GlobalVRGameJam2015 | opened | As a player, I want to see awesome graphics so that I can enjoy my time staying in the game. | Priority 1 | Acceptance criteria:
- Good graphics (this is a must).
Testing criteria:
- Ensure that none of us hates it.
Task breakdown:
- Make good graphics with any tools you like. | 1.0 | As a player, I want to see awesome graphics so that I can enjoy my time staying in the game. - Acceptance criteria:
- Good graphics (this is a must).
Testing criteria:
- Ensure that none of us hates it.
Task breakdown:
- Make good graphics with any tools you like. | priority | as a player i want to see awesome graphics so that i can enjoy my time staying in the game acceptance criteria good graphics this is a must testing criteria ensure that none of us hates it task breakdown make good graphics with any tools you like | 1 |
722,530 | 24,865,971,163 | IssuesEvent | 2022-10-27 11:59:24 | AY2223S1-CS2103T-T10-1/tp | https://api.github.com/repos/AY2223S1-CS2103T-T10-1/tp | closed | As a student, I can add the association between a contact and module | type.Story priority.Medium | so that I can track which of my friends could potentially become my study buddy. | 1.0 | As a student, I can add the association between a contact and module - so that I can track which of my friends could potentially become my study buddy. | priority | as a student i can add the association between a contact and module so that i can track which of my friends could potentially become my study buddy | 1 |
594,855 | 18,055,809,870 | IssuesEvent | 2021-09-20 08:05:43 | ita-social-projects/TeachUA | https://api.github.com/repos/ita-social-projects/TeachUA | opened | [Footer] Wrong color of labels | bug UI Priority: Low | **Environment:** Windows 10, Google Chrome Version 92.0.4515.107 (Official Build) (64-bit)
**Reproducible:** always
**Build found:** last commit
**Preconditions**
1. Go to https://speak-ukrainian.org.ua/dev/
**Steps to reproduce**
1. Scroll down to the footer
**Actual result**
Color of 'Наші партнери' and 'Як допомогти проєкту?' labels is 002766

**Expected result**
Color of 'Наші партнери' and 'Як допомогти проєкту?' labels is #2D4C68

| 1.0 | [Footer] Wrong color of labels - **Environment:** Windows 10, Google Chrome Version 92.0.4515.107 (Official Build) (64-bit)
**Reproducible:** always
**Build found:** last commit
**Preconditions**
1. Go to https://speak-ukrainian.org.ua/dev/
**Steps to reproduce**
1. Scroll down to the footer
**Actual result**
Color of 'Наші партнери' and 'Як допомогти проєкту?' labels is 002766

**Expected result**
Color of 'Наші партнери' and 'Як допомогти проєкту?' labels is #2D4C68

| priority | wrong color of labels environment windows google chrome version official build bit reproducible always build found last commit preconditions go to steps to reproduce scroll down to the footer actual result color of наші партнери and як допомогти проєкту labels is expected result color of наші партнери and як допомогти проєкту labels is | 1 |
229,106 | 17,513,219,830 | IssuesEvent | 2021-08-11 02:01:29 | flutter/flutter | https://api.github.com/repos/flutter/flutter | closed | integration_test: Can't run XCode "Product" --> "Test" | a: tests platform-ios documentation passed first triage integration_test | ## Description
I tried to set up XCTest with the `integration_test` package. What I did is:
1. Create a new fresh flutter project with `flutter create`
2. Add integration test (I copied the widget test as integration test)
3. Run integration test with `flutter drive --driver=test_driver/integration_test.dart --target=integration_test/app_test.dart -d 63B1E047-86E5-462C-9B49-16E6FD9F131D` --> everything worked ✅
4. Now I followed the steps to set tup the XCTests: https://github.com/flutter/flutter/tree/master/packages/integration_test#ios-device-testing
5. For this I created a new target, which I called `RunnerTests`
6. Added this to my `Podfile`:
```ruby
target 'RunnerTests' do
inherit! :search_paths
end
```
7. Replaced `RunnerTests.m` with:
```swift
@import XCTest;
@import integration_test;
INTEGRATION_TEST_IOS_RUNNER(RunnerTests)
```
8. And finally I tried to run the test via XCode: "Product" --> "Test" (or the shortcut: Command + U)
## Problem
XCode started building, but I got this expection:
```
can't link with a main executable file '/Users/nils/Library/Developer/Xcode/DerivedData/Runner-agxmbwcpuxvqzdbzeonhnekntuhm/Build/Products/Debug-iphonesimulator/Runner.app/Runner' for architecture arm64
clang: error: linker command failed with exit code 1 (use -v to see invocation)
```

## Tested device
Running build from MacBook Pro M1, macOS 11.3. Tried to build for iOS Simulator.
## Code to reproduce
I published all my Code in this Repo: https://github.com/nilsreichardt/integration_test_problem
<details>
<summary>Flutter Doctor</summary>
```
Doctor summary (to see all details, run flutter doctor -v):
[✓] Flutter (Channel stable, 2.2.2-0.0.pre.1, on macOS 11.3 20E232 darwin-arm, locale en-DE)
[!] Android toolchain - develop for Android devices (Android SDK version 29.0.3)
✗ Android license status unknown.
Run `flutter doctor --android-licenses` to accept the SDK licenses.
See https://flutter.dev/docs/get-started/install/macos#android-setup for more details.
[✓] Xcode - develop for iOS and macOS
[✓] Chrome - develop for the web
[✓] Android Studio (version 4.2)
[✓] IntelliJ IDEA Ultimate Edition (version 2021.1.2)
[✓] VS Code (version 1.58.0-insider)
[✓] Connected device (4 available)
! Error: iPhone von Roland is not connected. Xcode will continue when iPhone von Roland is connected. (code -13)
! Doctor found issues in 1 category.
```
<!--
Run `flutter analyze` and attach any output of that command below.
If there are any analysis errors, try resolving them before filing this issue.
-->
```
```
<!-- Finally, paste the output of running `flutter doctor -v` here. -->
```
```
</details>
| 1.0 | integration_test: Can't run XCode "Product" --> "Test" - ## Description
I tried to set up XCTest with the `integration_test` package. What I did is:
1. Create a new fresh flutter project with `flutter create`
2. Add integration test (I copied the widget test as integration test)
3. Run integration test with `flutter drive --driver=test_driver/integration_test.dart --target=integration_test/app_test.dart -d 63B1E047-86E5-462C-9B49-16E6FD9F131D` --> everything worked ✅
4. Now I followed the steps to set tup the XCTests: https://github.com/flutter/flutter/tree/master/packages/integration_test#ios-device-testing
5. For this I created a new target, which I called `RunnerTests`
6. Added this to my `Podfile`:
```ruby
target 'RunnerTests' do
inherit! :search_paths
end
```
7. Replaced `RunnerTests.m` with:
```swift
@import XCTest;
@import integration_test;
INTEGRATION_TEST_IOS_RUNNER(RunnerTests)
```
8. And finally I tried to run the test via XCode: "Product" --> "Test" (or the shortcut: Command + U)
## Problem
XCode started building, but I got this expection:
```
can't link with a main executable file '/Users/nils/Library/Developer/Xcode/DerivedData/Runner-agxmbwcpuxvqzdbzeonhnekntuhm/Build/Products/Debug-iphonesimulator/Runner.app/Runner' for architecture arm64
clang: error: linker command failed with exit code 1 (use -v to see invocation)
```

## Tested device
Running build from MacBook Pro M1, macOS 11.3. Tried to build for iOS Simulator.
## Code to reproduce
I published all my Code in this Repo: https://github.com/nilsreichardt/integration_test_problem
<details>
<summary>Flutter Doctor</summary>
```
Doctor summary (to see all details, run flutter doctor -v):
[✓] Flutter (Channel stable, 2.2.2-0.0.pre.1, on macOS 11.3 20E232 darwin-arm, locale en-DE)
[!] Android toolchain - develop for Android devices (Android SDK version 29.0.3)
✗ Android license status unknown.
Run `flutter doctor --android-licenses` to accept the SDK licenses.
See https://flutter.dev/docs/get-started/install/macos#android-setup for more details.
[✓] Xcode - develop for iOS and macOS
[✓] Chrome - develop for the web
[✓] Android Studio (version 4.2)
[✓] IntelliJ IDEA Ultimate Edition (version 2021.1.2)
[✓] VS Code (version 1.58.0-insider)
[✓] Connected device (4 available)
! Error: iPhone von Roland is not connected. Xcode will continue when iPhone von Roland is connected. (code -13)
! Doctor found issues in 1 category.
```
<!--
Run `flutter analyze` and attach any output of that command below.
If there are any analysis errors, try resolving them before filing this issue.
-->
```
```
<!-- Finally, paste the output of running `flutter doctor -v` here. -->
```
```
</details>
| non_priority | integration test can t run xcode product test description i tried to set up xctest with the integration test package what i did is create a new fresh flutter project with flutter create add integration test i copied the widget test as integration test run integration test with flutter drive driver test driver integration test dart target integration test app test dart d everything worked ✅ now i followed the steps to set tup the xctests for this i created a new target which i called runnertests added this to my podfile ruby target runnertests do inherit search paths end replaced runnertests m with swift import xctest import integration test integration test ios runner runnertests and finally i tried to run the test via xcode product test or the shortcut command u problem xcode started building but i got this expection can t link with a main executable file users nils library developer xcode deriveddata runner agxmbwcpuxvqzdbzeonhnekntuhm build products debug iphonesimulator runner app runner for architecture clang error linker command failed with exit code use v to see invocation tested device running build from macbook pro macos tried to build for ios simulator code to reproduce i published all my code in this repo flutter doctor doctor summary to see all details run flutter doctor v flutter channel stable pre on macos darwin arm locale en de android toolchain develop for android devices android sdk version ✗ android license status unknown run flutter doctor android licenses to accept the sdk licenses see for more details xcode develop for ios and macos chrome develop for the web android studio version intellij idea ultimate edition version vs code version insider connected device available error iphone von roland is not connected xcode will continue when iphone von roland is connected code doctor found issues in category run flutter analyze and attach any output of that command below if there are any analysis errors try resolving them before filing this issue | 0 |
86,160 | 3,703,191,793 | IssuesEvent | 2016-02-29 19:31:10 | GalliumOS/galliumos-distro | https://api.github.com/repos/GalliumOS/galliumos-distro | closed | Disk usage panel indicator | enhancement priority:low | A redditor (/u/adrinux) pointed out that this would be useful, especially on Chromebooks with tiny SSDs. | 1.0 | Disk usage panel indicator - A redditor (/u/adrinux) pointed out that this would be useful, especially on Chromebooks with tiny SSDs. | priority | disk usage panel indicator a redditor u adrinux pointed out that this would be useful especially on chromebooks with tiny ssds | 1 |
237,381 | 19,620,271,543 | IssuesEvent | 2022-01-07 05:02:15 | istio/istio | https://api.github.com/repos/istio/istio | closed | TestBadRemoteSecret causes future tests to fail once secrets are removed | kind/test failure lifecycle/stale feature/Multi-cluster | https://prow.istio.io/view/gs/istio-prow/pr-logs/pull/istio_istio/33701/integ-pilot-multicluster-tests_istio/1409935357113274368
test fails at 18:10:00. Its missing endpoints for some service
istiod logs show the cluster update at same time: https://storage.googleapis.com/istio-prow/pr-logs/pull/istio_istio/33701/integ-pilot-multicluster-tests_istio/1409935357113274368/artifacts/pilot-de882957a19240769ee1a3c93/TestTraffic/services/case_2_service_port_match_in_cluster_cross-network-primary/_test_context/istio-state096177769/primary/istiod-8ccfbd58c-k2gk2_discovery.log
It looks like a single push (from debounce) is handling removal of the clusters and the addition of the new service
I am fairly sure I have seen this before, but couldn't find an issue...
cc @stevenctl | 1.0 | TestBadRemoteSecret causes future tests to fail once secrets are removed - https://prow.istio.io/view/gs/istio-prow/pr-logs/pull/istio_istio/33701/integ-pilot-multicluster-tests_istio/1409935357113274368
test fails at 18:10:00. Its missing endpoints for some service
istiod logs show the cluster update at same time: https://storage.googleapis.com/istio-prow/pr-logs/pull/istio_istio/33701/integ-pilot-multicluster-tests_istio/1409935357113274368/artifacts/pilot-de882957a19240769ee1a3c93/TestTraffic/services/case_2_service_port_match_in_cluster_cross-network-primary/_test_context/istio-state096177769/primary/istiod-8ccfbd58c-k2gk2_discovery.log
It looks like a single push (from debounce) is handling removal of the clusters and the addition of the new service
I am fairly sure I have seen this before, but couldn't find an issue...
cc @stevenctl | non_priority | testbadremotesecret causes future tests to fail once secrets are removed test fails at its missing endpoints for some service istiod logs show the cluster update at same time it looks like a single push from debounce is handling removal of the clusters and the addition of the new service i am fairly sure i have seen this before but couldn t find an issue cc stevenctl | 0 |
90,647 | 26,160,526,543 | IssuesEvent | 2022-12-31 12:53:36 | Traben-0/Entity_Texture_Features | https://api.github.com/repos/Traben-0/Entity_Texture_Features | closed | [BUG] Broken texture paths option does the opposite | bug Fixed or added in dev build (released) | **Describe the issue**
- What happened? The option "allow broken texture paths" prevents resource packs with broken paths from being loaded, aka the error will be logged and all resource packs will be disabled.
- What were you doing when it happened? Enabling resource packs that have broken paths
**Screenshots**
N/A
**Crash log**
https://mclo.gs/mtcTZPC
https://mclo.gs/vyGm5s2
**Your setup: (please complete the following information):**
- Minecraft version: 1.19.3
- ModLoader: Fabric
- ETF Version 4.2.0.1
- Does the issue persist with only ETF installed? Not tested yet
- Have you checked the FAQ in the Readme, or Discord, to see if it is a known issue? Yes
See also:
https://discord.com/channels/859124104644788234/859124571005648946/1056922668764508240
https://discord.com/channels/859124104644788234/859124571005648946/1054991864161374269 | 1.0 | [BUG] Broken texture paths option does the opposite - **Describe the issue**
- What happened? The option "allow broken texture paths" prevents resource packs with broken paths from being loaded, aka the error will be logged and all resource packs will be disabled.
- What were you doing when it happened? Enabling resource packs that have broken paths
**Screenshots**
N/A
**Crash log**
https://mclo.gs/mtcTZPC
https://mclo.gs/vyGm5s2
**Your setup: (please complete the following information):**
- Minecraft version: 1.19.3
- ModLoader: Fabric
- ETF Version 4.2.0.1
- Does the issue persist with only ETF installed? Not tested yet
- Have you checked the FAQ in the Readme, or Discord, to see if it is a known issue? Yes
See also:
https://discord.com/channels/859124104644788234/859124571005648946/1056922668764508240
https://discord.com/channels/859124104644788234/859124571005648946/1054991864161374269 | non_priority | broken texture paths option does the opposite describe the issue what happened the option allow broken texture paths prevents resource packs with broken paths from being loaded aka the error will be logged and all resource packs will be disabled what were you doing when it happened enabling resource packs that have broken paths screenshots n a crash log your setup please complete the following information minecraft version modloader fabric etf version does the issue persist with only etf installed not tested yet have you checked the faq in the readme or discord to see if it is a known issue yes see also | 0 |
185,501 | 15,023,857,908 | IssuesEvent | 2021-02-01 18:48:57 | freqtrade/freqtrade | https://api.github.com/repos/freqtrade/freqtrade | closed | Derived strategies | Documentation Question | <!--
Have you searched for similar issues before posting it?
Did you have a VERY good look at the [documentation](https://www.freqtrade.io/en/latest/) and are sure that the question is not explained there
Yes
Please do not use the question template to report bugs or to request new features.
-->
## Describe your environment
* Operating system: Ubuntu 18.04
* Python Version: 3.7.5 (`python -V`)
* CCXT version: 1.40.86 (`pip freeze | grep ccxt`)
* Freqtrade Version: 2020.12 (`freqtrade -V` or `docker-compose run --rm freqtrade -V` for Freqtrade running in docker)
## Your question
How do I import MyAwesomeStrategy in order to write my derived strategy MyAwesomeStrategy2, like in [this documentation page](https://www.freqtrade.io/en/latest/strategy-advanced/#derived-strategies)? Assuming both in user_data/strategies/
Thank you.
#### Edited - some follow up
I've tried ```from user_data.strategies.myawesomestrategy import MyAwesomeStrategy```, getting the ```No module named 'user_data'``` error.
However, if I ```import ipdb``` before the above from...import statement, everything went alright.
This is very strange to me. There must be a more freqtrade-style way of deriving strategies without using a third party module.
| 1.0 | Derived strategies - <!--
Have you searched for similar issues before posting it?
Did you have a VERY good look at the [documentation](https://www.freqtrade.io/en/latest/) and are sure that the question is not explained there
Yes
Please do not use the question template to report bugs or to request new features.
-->
## Describe your environment
* Operating system: Ubuntu 18.04
* Python Version: 3.7.5 (`python -V`)
* CCXT version: 1.40.86 (`pip freeze | grep ccxt`)
* Freqtrade Version: 2020.12 (`freqtrade -V` or `docker-compose run --rm freqtrade -V` for Freqtrade running in docker)
## Your question
How do I import MyAwesomeStrategy in order to write my derived strategy MyAwesomeStrategy2, like in [this documentation page](https://www.freqtrade.io/en/latest/strategy-advanced/#derived-strategies)? Assuming both in user_data/strategies/
Thank you.
#### Edited - some follow up
I've tried ```from user_data.strategies.myawesomestrategy import MyAwesomeStrategy```, getting the ```No module named 'user_data'``` error.
However, if I ```import ipdb``` before the above from...import statement, everything went alright.
This is very strange to me. There must be a more freqtrade-style way of deriving strategies without using a third party module.
| non_priority | derived strategies have you searched for similar issues before posting it did you have a very good look at the and are sure that the question is not explained there yes please do not use the question template to report bugs or to request new features describe your environment operating system ubuntu python version python v ccxt version pip freeze grep ccxt freqtrade version freqtrade v or docker compose run rm freqtrade v for freqtrade running in docker your question how do i import myawesomestrategy in order to write my derived strategy like in assuming both in user data strategies thank you edited some follow up i ve tried from user data strategies myawesomestrategy import myawesomestrategy getting the no module named user data error however if i import ipdb before the above from import statement everything went alright this is very strange to me there must be a more freqtrade style way of deriving strategies without using a third party module | 0 |
451,616 | 13,039,333,608 | IssuesEvent | 2020-07-28 16:34:22 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | [Coverity CID :211508] Unchecked return value in tests/kernel/mem_protect/futex/src/main.c | Coverity bug priority: low |
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/565a61dd79b0f061056c9b3fe0adba11cbeca6ab/tests/kernel/mem_protect/futex/src/main.c#L479
Category: Error handling issues
Function: `futex_wake`
Component: Tests
CID: [211508](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=211508)
Details:
```
473 void futex_wake(void *p1, void *p2, void *p3)
474 {
475 int32_t atomic_ret_val;
476
477 k_futex_wake(&simple_futex, false);
478
>>> CID 211508: Error handling issues (CHECKED_RETURN)
>>> Calling "k_futex_wait" without checking return value (as is done elsewhere 9 out of 10 times).
479 k_futex_wait(&simple_futex, 13, K_FOREVER);
480
481 /* Test user can write to the futex value
482 * Use assertion to verify substraction correctness
483 * Initial value was 13, after atomic_sub() must be 12
484 */
```
Please fix or provide comments in coverity using the link:
https://scan9.coverity.com/reports.htm#v32951/p12996.
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
| 1.0 | [Coverity CID :211508] Unchecked return value in tests/kernel/mem_protect/futex/src/main.c -
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/565a61dd79b0f061056c9b3fe0adba11cbeca6ab/tests/kernel/mem_protect/futex/src/main.c#L479
Category: Error handling issues
Function: `futex_wake`
Component: Tests
CID: [211508](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=211508)
Details:
```
473 void futex_wake(void *p1, void *p2, void *p3)
474 {
475 int32_t atomic_ret_val;
476
477 k_futex_wake(&simple_futex, false);
478
>>> CID 211508: Error handling issues (CHECKED_RETURN)
>>> Calling "k_futex_wait" without checking return value (as is done elsewhere 9 out of 10 times).
479 k_futex_wait(&simple_futex, 13, K_FOREVER);
480
481 /* Test user can write to the futex value
482 * Use assertion to verify substraction correctness
483 * Initial value was 13, after atomic_sub() must be 12
484 */
```
Please fix or provide comments in coverity using the link:
https://scan9.coverity.com/reports.htm#v32951/p12996.
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
| priority | unchecked return value in tests kernel mem protect futex src main c static code scan issues found in file category error handling issues function futex wake component tests cid details void futex wake void void void t atomic ret val k futex wake simple futex false cid error handling issues checked return calling k futex wait without checking return value as is done elsewhere out of times k futex wait simple futex k forever test user can write to the futex value use assertion to verify substraction correctness initial value was after atomic sub must be please fix or provide comments in coverity using the link note this issue was created automatically priority was set based on classification of the file affected and the impact field in coverity assignees were set using the codeowners file | 1 |
216,074 | 7,301,019,212 | IssuesEvent | 2018-02-27 02:40:18 | ChromatixAU/phpcs-config-chromatix | https://api.github.com/repos/ChromatixAU/phpcs-config-chromatix | closed | Linting errors due to unavailable wpcom_vip_ functions | bandaided bug good first issue medium priority | Latest v0.0.15 giving linting error :- [phpcs] get_category_link() is prohibited, please use wpcom_vip_get_term_link() instead. | 1.0 | Linting errors due to unavailable wpcom_vip_ functions - Latest v0.0.15 giving linting error :- [phpcs] get_category_link() is prohibited, please use wpcom_vip_get_term_link() instead. | priority | linting errors due to unavailable wpcom vip functions latest giving linting error get category link is prohibited please use wpcom vip get term link instead | 1 |
138,022 | 11,172,981,022 | IssuesEvent | 2019-12-29 11:12:39 | Vachok/ftpplus | https://api.github.com/repos/Vachok/ftpplus | closed | realRunTest | Medium TestQuality ad bug resolution_Dupe | Execute ACLParserTest::realRunTest**realRunTest**
*ACLParserTest*
*null TimeoutException - null ACLParserTest.java: ru.vachok.networker.ad.usermanagement.ACLParserTest.realRunTest(ACLParserTest.java:59) expected [null] but found [java.util.concurrent.TimeoutException]*
*java.lang.AssertionError* | 1.0 | realRunTest - Execute ACLParserTest::realRunTest**realRunTest**
*ACLParserTest*
*null TimeoutException - null ACLParserTest.java: ru.vachok.networker.ad.usermanagement.ACLParserTest.realRunTest(ACLParserTest.java:59) expected [null] but found [java.util.concurrent.TimeoutException]*
*java.lang.AssertionError* | non_priority | realruntest execute aclparsertest realruntest realruntest aclparsertest null timeoutexception null aclparsertest java ru vachok networker ad usermanagement aclparsertest realruntest aclparsertest java expected but found java lang assertionerror | 0 |
275,517 | 23,919,558,793 | IssuesEvent | 2022-09-09 15:31:58 | envelope-zero/frontend | https://api.github.com/repos/envelope-zero/frontend | closed | Test: Accounts | testing | What should be tested?
- creating internal account with all fields (esp. "off budget")
- creating external accounts with all fields (make sure there is no "off budget")
- switching list: internal vs. external | 1.0 | Test: Accounts - What should be tested?
- creating internal account with all fields (esp. "off budget")
- creating external accounts with all fields (make sure there is no "off budget")
- switching list: internal vs. external | non_priority | test accounts what should be tested creating internal account with all fields esp off budget creating external accounts with all fields make sure there is no off budget switching list internal vs external | 0 |
124,359 | 10,310,020,986 | IssuesEvent | 2019-08-29 14:24:07 | ansible/awx | https://api.github.com/repos/ansible/awx | closed | Editing workflow breaks extra_vars | component:ui priority:high state:needs_test type:bug | ##### ISSUE TYPE
- Bug Report
##### SUMMARY
Editing workflow generates lots of issues with extra vars on job templates
##### ENVIRONMENT
* AWX version:6.0.0
* AWX install method: docker
* Ansible version: 2.8.1
* Operating System: Win
* Web Browser: FF/Chrome
##### STEPS TO REPRODUCE
Edit existing workflow template.
##### EXPECTED RESULTS
Everything is working as on v3
##### ACTUAL RESULTS
Workflow designer to some jobs adds pipe to extra vars, like this:
```
|
{}
```
Causing error: Failed to parse variables. Attempted to parse yaml. Parser did not return an object.
On other jobs it splits extra vars content char by char generating list of numbered, one char vars, like this:
https://gist.github.com/blindrood/156a73d31654114508bff7307ef2fdd1
##### ADDITIONAL INFORMATION
| 1.0 | Editing workflow breaks extra_vars - ##### ISSUE TYPE
- Bug Report
##### SUMMARY
Editing workflow generates lots of issues with extra vars on job templates
##### ENVIRONMENT
* AWX version:6.0.0
* AWX install method: docker
* Ansible version: 2.8.1
* Operating System: Win
* Web Browser: FF/Chrome
##### STEPS TO REPRODUCE
Edit existing workflow template.
##### EXPECTED RESULTS
Everything is working as on v3
##### ACTUAL RESULTS
Workflow designer to some jobs adds pipe to extra vars, like this:
```
|
{}
```
Causing error: Failed to parse variables. Attempted to parse yaml. Parser did not return an object.
On other jobs it splits extra vars content char by char generating list of numbered, one char vars, like this:
https://gist.github.com/blindrood/156a73d31654114508bff7307ef2fdd1
##### ADDITIONAL INFORMATION
| non_priority | editing workflow breaks extra vars issue type bug report summary editing workflow generates lots of issues with extra vars on job templates environment awx version awx install method docker ansible version operating system win web browser ff chrome steps to reproduce edit existing workflow template expected results everything is working as on actual results workflow designer to some jobs adds pipe to extra vars like this causing error failed to parse variables attempted to parse yaml parser did not return an object on other jobs it splits extra vars content char by char generating list of numbered one char vars like this additional information | 0 |
70,140 | 13,432,703,400 | IssuesEvent | 2020-09-07 08:48:25 | Regalis11/Barotrauma | https://api.github.com/repos/Regalis11/Barotrauma | closed | Aliens are unable to attack in Ruins | Bug Code | - [x] I have searched the issue tracker to check if the issue has already been reported.
**Description**
Aliens are unable to attack in Ruins. Been noticing this for a while now.
**Version**
~0.10.502 | 1.0 | Aliens are unable to attack in Ruins - - [x] I have searched the issue tracker to check if the issue has already been reported.
**Description**
Aliens are unable to attack in Ruins. Been noticing this for a while now.
**Version**
~0.10.502 | non_priority | aliens are unable to attack in ruins i have searched the issue tracker to check if the issue has already been reported description aliens are unable to attack in ruins been noticing this for a while now version | 0 |
407,188 | 11,907,803,662 | IssuesEvent | 2020-03-30 23:11:45 | BeeStation/NSV13 | https://api.github.com/repos/BeeStation/NSV13 | opened | [ENTERPRISE] Turbolift improperly instances areas for it's aircraft elevator | Enterprise FUCK Map Issue Multi-Z Priority: High | <!-- Write **BELOW** The Headers and **ABOVE** The comments else it may not be viewable -->
## Round ID:
FUCK
<!--- **INCLUDE THE ROUND ID**
If you discovered this issue from playing tgstation hosted servers:
[Round ID]: # (It can be found in the Status panel or retrieved from https://atlantaned.space/statbus/round.php ! The round id let's us look up valuable information and logs for the round the bug happened.)-->
## Testmerges:
Irrelevant
<!-- If you're certain the issue is to be caused by a test merge [OOC tab -> Show Server Revision], report it in the pull request's comment section rather than on the tracker(If you're unsure you can refer to the issue number by prefixing said number with #. The issue number can be found beside the title after submitting it to the tracker).If no testmerges are active, feel free to remove this section. -->
## Reproduction:
HOW HARD IS IT TO MAKE A NEW AREA

<!-- Explain your issue in detail, including the steps to reproduce it. Issues without proper reproduction steps or explanation are open to being ignored/closed by maintainers.-->
<!-- **For Admins:** Oddities induced by var-edits and other admin tools are not necessarily bugs. Verify that your issues occur under regular circumstances before reporting them. -->
| 1.0 | [ENTERPRISE] Turbolift improperly instances areas for it's aircraft elevator - <!-- Write **BELOW** The Headers and **ABOVE** The comments else it may not be viewable -->
## Round ID:
FUCK
<!--- **INCLUDE THE ROUND ID**
If you discovered this issue from playing tgstation hosted servers:
[Round ID]: # (It can be found in the Status panel or retrieved from https://atlantaned.space/statbus/round.php ! The round id let's us look up valuable information and logs for the round the bug happened.)-->
## Testmerges:
Irrelevant
<!-- If you're certain the issue is to be caused by a test merge [OOC tab -> Show Server Revision], report it in the pull request's comment section rather than on the tracker(If you're unsure you can refer to the issue number by prefixing said number with #. The issue number can be found beside the title after submitting it to the tracker).If no testmerges are active, feel free to remove this section. -->
## Reproduction:
HOW HARD IS IT TO MAKE A NEW AREA

<!-- Explain your issue in detail, including the steps to reproduce it. Issues without proper reproduction steps or explanation are open to being ignored/closed by maintainers.-->
<!-- **For Admins:** Oddities induced by var-edits and other admin tools are not necessarily bugs. Verify that your issues occur under regular circumstances before reporting them. -->
| priority | turbolift improperly instances areas for it s aircraft elevator round id fuck include the round id if you discovered this issue from playing tgstation hosted servers it can be found in the status panel or retrieved from the round id let s us look up valuable information and logs for the round the bug happened testmerges irrelevant reproduction how hard is it to make a new area | 1 |
32,407 | 2,754,291,739 | IssuesEvent | 2015-04-25 14:42:55 | raspibo/eventman | https://api.github.com/repos/raspibo/eventman | opened | import persons using eventbrite APIs | enhancement priority: low | Evenbrite provides some easy to use API, that can be used to import registered persons.
See https://developer.eventbrite.com/ | 1.0 | import persons using eventbrite APIs - Evenbrite provides some easy to use API, that can be used to import registered persons.
See https://developer.eventbrite.com/ | priority | import persons using eventbrite apis evenbrite provides some easy to use api that can be used to import registered persons see | 1 |
585,466 | 17,498,134,160 | IssuesEvent | 2021-08-10 05:23:00 | vmware/singleton | https://api.github.com/repos/vmware/singleton | closed | [BUG] [S3 Service] [V1]API returns wrong when prodcut name has been configured in bundle.json ,details below | kind/bug priority/medium | **Describe the bug**
API returns wrong when prodcut name has been configured in bundle.json ,details below
**To Reproduce**
\Service\API\V1\translation-source-api\createSourceUsingPOST_1
\Service\API\V1\translation-source-api\getTranslationBySourceUsingGET
\Service\API\V1\translation-product-component-key-api\getStringTranslationExclusiveComponentUsingGET
\Service\API\V1\translation-product-component-key-api\createSourceUsingPOST
\Service\API\V1\translation-product-component-key-api\getStringTranslationContainComponentUsingGET
\Service\API\V1\translation-product-component-api\getMultipleComponentsTranslationUsingGET_1
Request such as :

{component=default, source=Partner Name, locale=zh-Hans, version=1.1.0, pseudo=false, productName=vmWarevip, key=Partner_Name}

**Expected behavior**

**Screenshots**
If applicable, add screenshots to help you explain your issue
**Desktop (please complete the following information):**
- OS: [e.g. iOS]
- Browser Name [e.g. chrome, safari]
- Browser Version [e.g. 22]
**Additional context**
Add any other context about the problem here.
| 1.0 | [BUG] [S3 Service] [V1]API returns wrong when prodcut name has been configured in bundle.json ,details below - **Describe the bug**
API returns wrong when prodcut name has been configured in bundle.json ,details below
**To Reproduce**
\Service\API\V1\translation-source-api\createSourceUsingPOST_1
\Service\API\V1\translation-source-api\getTranslationBySourceUsingGET
\Service\API\V1\translation-product-component-key-api\getStringTranslationExclusiveComponentUsingGET
\Service\API\V1\translation-product-component-key-api\createSourceUsingPOST
\Service\API\V1\translation-product-component-key-api\getStringTranslationContainComponentUsingGET
\Service\API\V1\translation-product-component-api\getMultipleComponentsTranslationUsingGET_1
Request such as :

{component=default, source=Partner Name, locale=zh-Hans, version=1.1.0, pseudo=false, productName=vmWarevip, key=Partner_Name}

**Expected behavior**

**Screenshots**
If applicable, add screenshots to help you explain your issue
**Desktop (please complete the following information):**
- OS: [e.g. iOS]
- Browser Name [e.g. chrome, safari]
- Browser Version [e.g. 22]
**Additional context**
Add any other context about the problem here.
| priority | api returns wrong when prodcut name has been configured in bundle json details below describe the bug api returns wrong when prodcut name has been configured in bundle json details below to reproduce service api translation source api createsourceusingpost service api translation source api gettranslationbysourceusingget service api translation product component key api getstringtranslationexclusivecomponentusingget service api translation product component key api createsourceusingpost service api translation product component key api getstringtranslationcontaincomponentusingget service api translation product component api getmultiplecomponentstranslationusingget request such as component default source partner name locale zh hans version pseudo false productname vmwarevip key partner name expected behavior screenshots if applicable add screenshots to help you explain your issue desktop please complete the following information os browser name browser version additional context add any other context about the problem here | 1 |
186,575 | 21,944,959,354 | IssuesEvent | 2022-05-23 22:48:43 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Add rich type for reading X509 Authority Information Access | api-approved area-System.Security in-pr | One of the standard X.509 extension types defined by RFC 3280 (and successive RFCs) is the "Authority Information Access" extension. The use case is largely to inform chain building engines where any OCSP endpoints might be, and where one might find a copy of the issuing CA's certificate.
With a lot of handwaving, this extension can be thought of an order series of 3-tuples:
* What kind of purpose is this for? (OCSP endpoint, where to download a CA cert, some future/private purpose)
* What type of data is present? (the most useful type here is URI, other options include an LDAP X.500 identifier, or an email address, or an IP address... most of which don't make sense in context)
* What is the value? (legal encodings are determined by the data type)
For a typical .NET caller, the most important information is the URI-typed values for OCSP or the issuing CA. Rather than muddy the type with all of the complexities of modelling everything, the type is limited mostly to these common cases and anything more complicated can be accomplished with `AsnReader`.
The Enumerate methods all ignore non-conforming data. A `(OCSP, emailAddress, ...)` value will simply be skipped over. Exceptions are only thrown when the encoded representation doesn't decode at all.
```C#
namespace System.Security.Cryptography.X509Certificates
{
public sealed partial class X509AuthorityInformationAccessExtension : X509Extension
{
public X509AuthorityInformationAccessExtension() { }
public X509AuthorityInformationAccessExtension(byte[] rawData, bool critical = false) { }
public X509AuthorityInformationAccessExtension(ReadOnlySpan<byte> rawData, bool critical = false) { }
public X509AuthorityInformationAccessExtension(IEnumerable<string> ocspUris, IEnumerable<string> caIssuersUris, bool critical = false) { }
public override void CopyFrom(AsnEncodedData asnEncodedData) { }
public IEnumerable<string> EnumerateCAIssuersUris() { throw null; }
public IEnumerable<string> EnumerateOcspUris() { throw null; }
public IEnumerable<string> EnumerateUris(string accessMethodOid) { throw null; }
}
}
``` | True | Add rich type for reading X509 Authority Information Access - One of the standard X.509 extension types defined by RFC 3280 (and successive RFCs) is the "Authority Information Access" extension. The use case is largely to inform chain building engines where any OCSP endpoints might be, and where one might find a copy of the issuing CA's certificate.
With a lot of handwaving, this extension can be thought of an order series of 3-tuples:
* What kind of purpose is this for? (OCSP endpoint, where to download a CA cert, some future/private purpose)
* What type of data is present? (the most useful type here is URI, other options include an LDAP X.500 identifier, or an email address, or an IP address... most of which don't make sense in context)
* What is the value? (legal encodings are determined by the data type)
For a typical .NET caller, the most important information is the URI-typed values for OCSP or the issuing CA. Rather than muddy the type with all of the complexities of modelling everything, the type is limited mostly to these common cases and anything more complicated can be accomplished with `AsnReader`.
The Enumerate methods all ignore non-conforming data. A `(OCSP, emailAddress, ...)` value will simply be skipped over. Exceptions are only thrown when the encoded representation doesn't decode at all.
```C#
namespace System.Security.Cryptography.X509Certificates
{
public sealed partial class X509AuthorityInformationAccessExtension : X509Extension
{
public X509AuthorityInformationAccessExtension() { }
public X509AuthorityInformationAccessExtension(byte[] rawData, bool critical = false) { }
public X509AuthorityInformationAccessExtension(ReadOnlySpan<byte> rawData, bool critical = false) { }
public X509AuthorityInformationAccessExtension(IEnumerable<string> ocspUris, IEnumerable<string> caIssuersUris, bool critical = false) { }
public override void CopyFrom(AsnEncodedData asnEncodedData) { }
public IEnumerable<string> EnumerateCAIssuersUris() { throw null; }
public IEnumerable<string> EnumerateOcspUris() { throw null; }
public IEnumerable<string> EnumerateUris(string accessMethodOid) { throw null; }
}
}
``` | non_priority | add rich type for reading authority information access one of the standard x extension types defined by rfc and successive rfcs is the authority information access extension the use case is largely to inform chain building engines where any ocsp endpoints might be and where one might find a copy of the issuing ca s certificate with a lot of handwaving this extension can be thought of an order series of tuples what kind of purpose is this for ocsp endpoint where to download a ca cert some future private purpose what type of data is present the most useful type here is uri other options include an ldap x identifier or an email address or an ip address most of which don t make sense in context what is the value legal encodings are determined by the data type for a typical net caller the most important information is the uri typed values for ocsp or the issuing ca rather than muddy the type with all of the complexities of modelling everything the type is limited mostly to these common cases and anything more complicated can be accomplished with asnreader the enumerate methods all ignore non conforming data a ocsp emailaddress value will simply be skipped over exceptions are only thrown when the encoded representation doesn t decode at all c namespace system security cryptography public sealed partial class public public byte rawdata bool critical false public readonlyspan rawdata bool critical false public ienumerable ocspuris ienumerable caissuersuris bool critical false public override void copyfrom asnencodeddata asnencodeddata public ienumerable enumeratecaissuersuris throw null public ienumerable enumerateocspuris throw null public ienumerable enumerateuris string accessmethodoid throw null | 0 |
209,273 | 23,705,985,737 | IssuesEvent | 2022-08-30 01:11:14 | nidhi7598/linux-4.1.15 | https://api.github.com/repos/nidhi7598/linux-4.1.15 | closed | CVE-2017-16527 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed | security vulnerability | ## CVE-2017-16527 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.1.15/commit/51d583e7739ca814844fa8c1f3d517c972c1f323">51d583e7739ca814844fa8c1f3d517c972c1f323</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
sound/usb/mixer.c in the Linux kernel before 4.13.8 allows local users to cause a denial of service (snd_usb_mixer_interrupt use-after-free and system crash) or possibly have unspecified other impact via a crafted USB device.
<p>Publish Date: 2017-11-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16527>CVE-2017-16527</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Physical
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-16527">https://nvd.nist.gov/vuln/detail/CVE-2017-16527</a></p>
<p>Release Date: 2017-11-04</p>
<p>Fix Resolution: 4.13.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2017-16527 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed - ## CVE-2017-16527 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.1.15/commit/51d583e7739ca814844fa8c1f3d517c972c1f323">51d583e7739ca814844fa8c1f3d517c972c1f323</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
sound/usb/mixer.c in the Linux kernel before 4.13.8 allows local users to cause a denial of service (snd_usb_mixer_interrupt use-after-free and system crash) or possibly have unspecified other impact via a crafted USB device.
<p>Publish Date: 2017-11-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16527>CVE-2017-16527</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Physical
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-16527">https://nvd.nist.gov/vuln/detail/CVE-2017-16527</a></p>
<p>Release Date: 2017-11-04</p>
<p>Fix Resolution: 4.13.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in linux stable autoclosed cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details sound usb mixer c in the linux kernel before allows local users to cause a denial of service snd usb mixer interrupt use after free and system crash or possibly have unspecified other impact via a crafted usb device publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
557,736 | 16,517,492,234 | IssuesEvent | 2021-05-26 11:18:20 | PropertyTools/PropertyTools | https://api.github.com/repos/PropertyTools/PropertyTools | closed | Paste isn't working for DataGrid | DataGrid bug help wanted high priority | pasting multiple rows in DataGrid isn't working, it shows default value(s) for each row(s). The number of rows are created as per source but not the values. Paste is working fine when pasting for the second time. | 1.0 | Paste isn't working for DataGrid - pasting multiple rows in DataGrid isn't working, it shows default value(s) for each row(s). The number of rows are created as per source but not the values. Paste is working fine when pasting for the second time. | priority | paste isn t working for datagrid pasting multiple rows in datagrid isn t working it shows default value s for each row s the number of rows are created as per source but not the values paste is working fine when pasting for the second time | 1 |
65,570 | 27,145,488,512 | IssuesEvent | 2023-02-16 19:38:16 | hachyderm/community | https://api.github.com/repos/hachyderm/community | closed | CORS error when trying to play audio hosted on media.hachyderm.io | Reporting Service Issue Infra Team | ### What would you like to discuss with us or let us know?
After adding an MP3 (https://hachyderm.io/@tvdv/109574002737602560), I noticed that both Safari and Chrome refuse to load the media from media.hachyderm.io because the main hachyderm.io domain is not allowed by its Access-Control-Allow-Origin headers.
<img width="1470" alt="Screenshot 2022-12-25 at 12 11 50" src="https://user-images.githubusercontent.com/4490/209465891-ac62f276-afa5-4e9a-83da-299494fbfb41.png">
<img width="1361" alt="Screenshot 2022-12-25 at 12 12 52" src="https://user-images.githubusercontent.com/4490/209465894-339ae631-a271-4b42-8268-de31ceeb2539.png">
| 1.0 | CORS error when trying to play audio hosted on media.hachyderm.io - ### What would you like to discuss with us or let us know?
After adding an MP3 (https://hachyderm.io/@tvdv/109574002737602560), I noticed that both Safari and Chrome refuse to load the media from media.hachyderm.io because the main hachyderm.io domain is not allowed by its Access-Control-Allow-Origin headers.
<img width="1470" alt="Screenshot 2022-12-25 at 12 11 50" src="https://user-images.githubusercontent.com/4490/209465891-ac62f276-afa5-4e9a-83da-299494fbfb41.png">
<img width="1361" alt="Screenshot 2022-12-25 at 12 12 52" src="https://user-images.githubusercontent.com/4490/209465894-339ae631-a271-4b42-8268-de31ceeb2539.png">
| non_priority | cors error when trying to play audio hosted on media hachyderm io what would you like to discuss with us or let us know after adding an i noticed that both safari and chrome refuse to load the media from media hachyderm io because the main hachyderm io domain is not allowed by its access control allow origin headers img width alt screenshot at src img width alt screenshot at src | 0 |
26,976 | 6,812,943,005 | IssuesEvent | 2017-11-06 06:50:09 | BTDF/DeploymentFramework | https://api.github.com/repos/BTDF/DeploymentFramework | closed | Tweak NTFS permission assignment on FILE adapter path setup so assigned perms exactly match the Windows Security dialog box | bug CodePlexMigrationInitiated General Impact: Medium Release 5.0 | Tweak NTFS permission assignment on FILE adapter path setup so assigned perms exactly match the Windows Security dialog box. Specific change - set PropagationFlags to None.
#### This work item was migrated from CodePlex
CodePlex work item ID: '6255'
Assigned to: 'tfabraham'
Vote count: '0'
| 1.0 | Tweak NTFS permission assignment on FILE adapter path setup so assigned perms exactly match the Windows Security dialog box - Tweak NTFS permission assignment on FILE adapter path setup so assigned perms exactly match the Windows Security dialog box. Specific change - set PropagationFlags to None.
#### This work item was migrated from CodePlex
CodePlex work item ID: '6255'
Assigned to: 'tfabraham'
Vote count: '0'
| non_priority | tweak ntfs permission assignment on file adapter path setup so assigned perms exactly match the windows security dialog box tweak ntfs permission assignment on file adapter path setup so assigned perms exactly match the windows security dialog box specific change set propagationflags to none this work item was migrated from codeplex codeplex work item id assigned to tfabraham vote count | 0 |
363,821 | 10,755,638,281 | IssuesEvent | 2019-10-31 09:34:10 | AY1920S1-CS2103T-T17-2/main | https://api.github.com/repos/AY1920S1-CS2103T-T17-2/main | closed | As a user I want a travel diary | priority.Low type.Story | Description: As a traveller, I want to record down all my travelling stories so that I can have my sweet memories stored.
| 1.0 | As a user I want a travel diary - Description: As a traveller, I want to record down all my travelling stories so that I can have my sweet memories stored.
| priority | as a user i want a travel diary description as a traveller i want to record down all my travelling stories so that i can have my sweet memories stored | 1 |
434,246 | 12,515,944,446 | IssuesEvent | 2020-06-03 08:34:36 | eclipse/codewind | https://api.github.com/repos/eclipse/codewind | closed | Insane number of socket messages when cancelling a load run | area/portal kind/bug priority/hot | **Codewind version:** latest
**OS:** N/A
**Che version:**N/A
**IDE extension version:**N/A
**IDE version:**N/A
**Kubernetes cluster:**N/A
**Description:**
Cancelling load causes an insane amount of socket messages to be emitted which makes all the screen status values flash

**Steps to reproduce:**
**Workaround:**
<!-- Did you find a way to work around the bug? If so, please describe how you worked around it.-->
| 1.0 | Insane number of socket messages when cancelling a load run - **Codewind version:** latest
**OS:** N/A
**Che version:**N/A
**IDE extension version:**N/A
**IDE version:**N/A
**Kubernetes cluster:**N/A
**Description:**
Cancelling load causes an insane amount of socket messages to be emitted which makes all the screen status values flash

**Steps to reproduce:**
**Workaround:**
<!-- Did you find a way to work around the bug? If so, please describe how you worked around it.-->
| priority | insane number of socket messages when cancelling a load run codewind version latest os n a che version n a ide extension version n a ide version n a kubernetes cluster n a description cancelling load causes an insane amount of socket messages to be emitted which makes all the screen status values flash steps to reproduce workaround | 1 |
543,522 | 15,883,374,282 | IssuesEvent | 2021-04-09 17:17:55 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | Ask Password Option: Confusing and repetitive steps for the user | Affected/5.4.0-Update1 Priority/Highest docs | In [1] the expiry time is set in the first step and the email verifcation is enabled in a latter step. In order to get this working the user has to restart the server, log in to management console twice. This maybe tiresome for the user who wants to try out this particular flow. Hence to reduce effort on user side to enable configs, please provide the configs in one go.
[1] https://docs.wso2.com/display/IS540/Creating+Users+using+the+Ask+Password+Option | 1.0 | Ask Password Option: Confusing and repetitive steps for the user - In [1] the expiry time is set in the first step and the email verifcation is enabled in a latter step. In order to get this working the user has to restart the server, log in to management console twice. This maybe tiresome for the user who wants to try out this particular flow. Hence to reduce effort on user side to enable configs, please provide the configs in one go.
[1] https://docs.wso2.com/display/IS540/Creating+Users+using+the+Ask+Password+Option | priority | ask password option confusing and repetitive steps for the user in the expiry time is set in the first step and the email verifcation is enabled in a latter step in order to get this working the user has to restart the server log in to management console twice this maybe tiresome for the user who wants to try out this particular flow hence to reduce effort on user side to enable configs please provide the configs in one go | 1 |
618,537 | 19,473,927,128 | IssuesEvent | 2021-12-24 08:26:01 | GEWIS/gewisweb | https://api.github.com/repos/GEWIS/gewisweb | opened | Member search page is exposed | Type: Bug Priority: Low For: Backend Status: Confirmed | You will not be able to search any members, but the page is still exposed. Would be preferable to also have it behind an ACL check. | 1.0 | Member search page is exposed - You will not be able to search any members, but the page is still exposed. Would be preferable to also have it behind an ACL check. | priority | member search page is exposed you will not be able to search any members but the page is still exposed would be preferable to also have it behind an acl check | 1 |
304,862 | 9,336,788,253 | IssuesEvent | 2019-03-28 22:19:54 | nationalparkservice/mojn-sv-database | https://api.github.com/repos/nationalparkservice/mojn-sv-database | closed | Photo list not showing up after photo data entered | high priority | The data show up in the subform query so maybe the subform isn't being made visible when it should? | 1.0 | Photo list not showing up after photo data entered - The data show up in the subform query so maybe the subform isn't being made visible when it should? | priority | photo list not showing up after photo data entered the data show up in the subform query so maybe the subform isn t being made visible when it should | 1 |
583,217 | 17,380,029,041 | IssuesEvent | 2021-07-31 14:07:17 | ShapeGroup/kimera-frontend-framework | https://api.github.com/repos/ShapeGroup/kimera-frontend-framework | opened | Modals open at start do not close on click | FIXED ;) bug --low priority improvements | _Create a modal with [status-active], when loading the page try to close the modal: it doesn't work._
The bug is known and is fixed for the next update | 1.0 | Modals open at start do not close on click - _Create a modal with [status-active], when loading the page try to close the modal: it doesn't work._
The bug is known and is fixed for the next update | priority | modals open at start do not close on click create a modal with when loading the page try to close the modal it doesn t work the bug is known and is fixed for the next update | 1 |
161,608 | 6,131,099,150 | IssuesEvent | 2017-06-24 12:31:29 | timtrice/rrricanes | https://api.github.com/repos/timtrice/rrricanes | opened | Remove `Adv` variable from `update` | Bugs High Priority | Updates do not have advisory numbers. Remove this variable from dataset. | 1.0 | Remove `Adv` variable from `update` - Updates do not have advisory numbers. Remove this variable from dataset. | priority | remove adv variable from update updates do not have advisory numbers remove this variable from dataset | 1 |
39,360 | 2,853,861,195 | IssuesEvent | 2015-06-01 20:46:21 | Chatterbox/Chatterbox | https://api.github.com/repos/Chatterbox/Chatterbox | opened | Sponge support | enhancement high priority todo | This was always going to be a thing we were doing, and now it's time to make ze ticket.
So! What do we need to do to make Sponge support happen?
- [ ] Research [Sponge API](https://docs.spongepowered.org/en/plugin/index.html)
- [ ] Abstract anything that is currently Bukkit-specific
- [ ] Do the thing? | 1.0 | Sponge support - This was always going to be a thing we were doing, and now it's time to make ze ticket.
So! What do we need to do to make Sponge support happen?
- [ ] Research [Sponge API](https://docs.spongepowered.org/en/plugin/index.html)
- [ ] Abstract anything that is currently Bukkit-specific
- [ ] Do the thing? | priority | sponge support this was always going to be a thing we were doing and now it s time to make ze ticket so what do we need to do to make sponge support happen research abstract anything that is currently bukkit specific do the thing | 1 |
5,802 | 8,247,685,493 | IssuesEvent | 2018-09-11 16:11:47 | cornellius-gp/gpytorch | https://api.github.com/repos/cornellius-gp/gpytorch | opened | For consistency: attrs should be `num_*`, not `n_*` | compatibility | We have a mix of `n_samples`, `num_dims`, etc. for different attributes. For consistency, we should use just one of these. I propose `num_` since it seems to be more commonly used with PyTorch.
We can accept both (for now) and deprecate `n_*`. | True | For consistency: attrs should be `num_*`, not `n_*` - We have a mix of `n_samples`, `num_dims`, etc. for different attributes. For consistency, we should use just one of these. I propose `num_` since it seems to be more commonly used with PyTorch.
We can accept both (for now) and deprecate `n_*`. | non_priority | for consistency attrs should be num not n we have a mix of n samples num dims etc for different attributes for consistency we should use just one of these i propose num since it seems to be more commonly used with pytorch we can accept both for now and deprecate n | 0 |
134,896 | 5,239,293,229 | IssuesEvent | 2017-01-31 09:21:16 | DiCarloLab-Delft/PycQED_py3 | https://api.github.com/repos/DiCarloLab-Delft/PycQED_py3 | closed | Codeword parameter in QWG should start counting at 0 and not at 1 | enhancement priority: could/low | This is to be consistent with the bits of the codeword trigger.
Easy to fix no @wvlothuizen required. | 1.0 | Codeword parameter in QWG should start counting at 0 and not at 1 - This is to be consistent with the bits of the codeword trigger.
Easy to fix no @wvlothuizen required. | priority | codeword parameter in qwg should start counting at and not at this is to be consistent with the bits of the codeword trigger easy to fix no wvlothuizen required | 1 |
31,972 | 4,309,322,589 | IssuesEvent | 2016-07-21 15:39:20 | jupyter/jupyterlab | https://api.github.com/repos/jupyter/jupyterlab | opened | Terminal cursor should be hidden if the terminal does not have focus | cat:Design and UX type:Enhancement | Even if some other input field has focus, the terminal cursor will still blink. This makes it *appear* that the terminal has focus, when in fact it does not. We should disable/hide terminal cursors unless the terminal actually has input focus. | 1.0 | Terminal cursor should be hidden if the terminal does not have focus - Even if some other input field has focus, the terminal cursor will still blink. This makes it *appear* that the terminal has focus, when in fact it does not. We should disable/hide terminal cursors unless the terminal actually has input focus. | non_priority | terminal cursor should be hidden if the terminal does not have focus even if some other input field has focus the terminal cursor will still blink this makes it appear that the terminal has focus when in fact it does not we should disable hide terminal cursors unless the terminal actually has input focus | 0 |
269,004 | 8,419,038,073 | IssuesEvent | 2018-10-15 04:34:23 | CS2103-AY1819S1-F10-1/main | https://api.github.com/repos/CS2103-AY1819S1-F10-1/main | opened | Add leave application storage | priority.High status.Ongoing type.Epic | Implement the functionality of saving leave application data to XML file | 1.0 | Add leave application storage - Implement the functionality of saving leave application data to XML file | priority | add leave application storage implement the functionality of saving leave application data to xml file | 1 |
733,348 | 25,302,796,334 | IssuesEvent | 2022-11-17 12:01:51 | geosolutions-it/MapStore2 | https://api.github.com/repos/geosolutions-it/MapStore2 | opened | Issue with jenkins jobs | bug Priority: Blocker DevOps | ## Description
The builds of MapStore do not seems to run anymore.
List of failures:
http://build.geosolutionsgroup.com/view/MapStore/job/MapStore/view/MapStore/job/MapStore2-QA-Build/33/
http://build.geosolutionsgroup.com/view/MapStore/job/MapStore/job/MapStore2-Stable-Releaser/5/
http://build.geosolutionsgroup.com/view/MapStore/job/MapStore/view/MapStore%20DEV/job/MapStore2-Dev-Build/178/
| 1.0 | Issue with jenkins jobs - ## Description
The builds of MapStore do not seems to run anymore.
List of failures:
http://build.geosolutionsgroup.com/view/MapStore/job/MapStore/view/MapStore/job/MapStore2-QA-Build/33/
http://build.geosolutionsgroup.com/view/MapStore/job/MapStore/job/MapStore2-Stable-Releaser/5/
http://build.geosolutionsgroup.com/view/MapStore/job/MapStore/view/MapStore%20DEV/job/MapStore2-Dev-Build/178/
| priority | issue with jenkins jobs description the builds of mapstore do not seems to run anymore list of failures | 1 |
787,264 | 27,712,272,584 | IssuesEvent | 2023-03-14 14:55:27 | TykTechnologies/tyk | https://api.github.com/repos/TykTechnologies/tyk | closed | Create swagger documentation for Tyk Dashboard API | dashboard customer request Priority: Low | **Do you want to request a *feature* or report a *bug*?**
feature
**What is the current behavior?**
none
**What is the expected behavior?**
Swagger documentation of the Tyk Dashboard API. This makes it easier to use to within a CI/CD tool. | 1.0 | Create swagger documentation for Tyk Dashboard API - **Do you want to request a *feature* or report a *bug*?**
feature
**What is the current behavior?**
none
**What is the expected behavior?**
Swagger documentation of the Tyk Dashboard API. This makes it easier to use to within a CI/CD tool. | priority | create swagger documentation for tyk dashboard api do you want to request a feature or report a bug feature what is the current behavior none what is the expected behavior swagger documentation of the tyk dashboard api this makes it easier to use to within a ci cd tool | 1 |
153,513 | 5,893,503,327 | IssuesEvent | 2017-05-17 22:09:50 | AZMAG/map-LandUse | https://api.github.com/repos/AZMAG/map-LandUse | closed | Make Excel Button Standard throughout Viewer | optimization Priority: High | We currently have several different Excel buttons in the viewer. We need to standardize the look.
We need to use the Kendo button
1: Make sure the sprite works
2: maybe change the color to Excel Green | 1.0 | Make Excel Button Standard throughout Viewer - We currently have several different Excel buttons in the viewer. We need to standardize the look.
We need to use the Kendo button
1: Make sure the sprite works
2: maybe change the color to Excel Green | priority | make excel button standard throughout viewer we currently have several different excel buttons in the viewer we need to standardize the look we need to use the kendo button make sure the sprite works maybe change the color to excel green | 1 |
67,092 | 8,998,862,043 | IssuesEvent | 2019-02-03 02:39:44 | Pyomo/pyomo | https://api.github.com/repos/Pyomo/pyomo | closed | Clarify documentation for updating a Param value | documentation | The [online documentation](https://pyomo.readthedocs.io/en/latest/working_models.html#changing-the-model-or-data-and-re-solving) recommends that we do repeated solutions with different parameter values by updating instance data like `instance.ParamName[idx].value`. However, at least in Pyomo 5.1.1 or 5.6.1, `instance.ParamName[idx]` is a simple value with no `.value` attribute, so attempting this gives an `AttributeError`. Here is some example code:
import pyomo.environ as pyo
i = pyo.ConcreteModel()
i.param = pyo.Param(['a', 'b'], initialize=lambda m, k: 1)
i.param.display()
i.param['a'].value = 2 # AttributeError: 'int' object has no attribute 'value'
type(i.param['a']) # int
I don't know if this worked in some earlier version of Pyomo, but it doesn't seem to now. So I assume the online documentation needs to be updated. | 1.0 | Clarify documentation for updating a Param value - The [online documentation](https://pyomo.readthedocs.io/en/latest/working_models.html#changing-the-model-or-data-and-re-solving) recommends that we do repeated solutions with different parameter values by updating instance data like `instance.ParamName[idx].value`. However, at least in Pyomo 5.1.1 or 5.6.1, `instance.ParamName[idx]` is a simple value with no `.value` attribute, so attempting this gives an `AttributeError`. Here is some example code:
import pyomo.environ as pyo
i = pyo.ConcreteModel()
i.param = pyo.Param(['a', 'b'], initialize=lambda m, k: 1)
i.param.display()
i.param['a'].value = 2 # AttributeError: 'int' object has no attribute 'value'
type(i.param['a']) # int
I don't know if this worked in some earlier version of Pyomo, but it doesn't seem to now. So I assume the online documentation needs to be updated. | non_priority | clarify documentation for updating a param value the recommends that we do repeated solutions with different parameter values by updating instance data like instance paramname value however at least in pyomo or instance paramname is a simple value with no value attribute so attempting this gives an attributeerror here is some example code import pyomo environ as pyo i pyo concretemodel i param pyo param initialize lambda m k i param display i param value attributeerror int object has no attribute value type i param int i don t know if this worked in some earlier version of pyomo but it doesn t seem to now so i assume the online documentation needs to be updated | 0 |
794,024 | 28,019,784,034 | IssuesEvent | 2023-03-28 03:48:48 | AY2223S2-CS2113T-T09-1/tp | https://api.github.com/repos/AY2223S2-CS2113T-T09-1/tp | closed | View command can see into the future | bug priority:Low severity:Low | To remove the user's ability to see transactions that has not happened yet.
To integrate with view -all | 1.0 | View command can see into the future - To remove the user's ability to see transactions that has not happened yet.
To integrate with view -all | priority | view command can see into the future to remove the user s ability to see transactions that has not happened yet to integrate with view all | 1 |
4,527 | 6,667,607,982 | IssuesEvent | 2017-10-03 13:13:58 | BlueBrain/nexus-kg | https://api.github.com/repos/BlueBrain/nexus-kg | closed | Implicit ToUnqualifiedStringOps and ToConfiguredUnqualifiedStringOps can throw exceptions | bug service | https://github.com/BlueBrain/nexus-kg/blob/3550c934758d9b46da4477c16f189f617743e6a0/modules/indexing/src/main/scala/ch/epfl/bluebrain/nexus/kg/indexing/Qualifier.scala#L130:L149
The implicit classes handle the string argument as a safe uri, but converting them to Uri can fail. The unapply implementations need to wrap the `String` to `Uri` conversion into a `Try` and treat failures as a `None` return. | 1.0 | Implicit ToUnqualifiedStringOps and ToConfiguredUnqualifiedStringOps can throw exceptions - https://github.com/BlueBrain/nexus-kg/blob/3550c934758d9b46da4477c16f189f617743e6a0/modules/indexing/src/main/scala/ch/epfl/bluebrain/nexus/kg/indexing/Qualifier.scala#L130:L149
The implicit classes handle the string argument as a safe uri, but converting them to Uri can fail. The unapply implementations need to wrap the `String` to `Uri` conversion into a `Try` and treat failures as a `None` return. | non_priority | implicit tounqualifiedstringops and toconfiguredunqualifiedstringops can throw exceptions the implicit classes handle the string argument as a safe uri but converting them to uri can fail the unapply implementations need to wrap the string to uri conversion into a try and treat failures as a none return | 0 |
596,536 | 18,105,897,123 | IssuesEvent | 2021-09-22 19:06:49 | Signbank/Global-signbank | https://api.github.com/repos/Signbank/Global-signbank | closed | I can't upload a video for a new sign | bug high priority blocking | When I try to upload a video, I receive the following message
<img width="1792" alt="Screenshot 2021-09-01 at 14 45 01" src="https://user-images.githubusercontent.com/63103864/131682956-913663cf-8ed9-4527-98b6-29f3fcbc3e31.png">
| 1.0 | I can't upload a video for a new sign - When I try to upload a video, I receive the following message
<img width="1792" alt="Screenshot 2021-09-01 at 14 45 01" src="https://user-images.githubusercontent.com/63103864/131682956-913663cf-8ed9-4527-98b6-29f3fcbc3e31.png">
| priority | i can t upload a video for a new sign when i try to upload a video i receive the following message img width alt screenshot at src | 1 |
526,607 | 15,296,783,203 | IssuesEvent | 2021-02-24 07:25:36 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | Add page element ID to my account page | Priority/Highest Severity/Major bug myaccount | **Describe the Issue:**
For test automation required unique ID for page elements. Therefor it required add data-testid for every page element on my account page. | 1.0 | Add page element ID to my account page - **Describe the Issue:**
For test automation required unique ID for page elements. Therefor it required add data-testid for every page element on my account page. | priority | add page element id to my account page describe the issue for test automation required unique id for page elements therefor it required add data testid for every page element on my account page | 1 |
661,816 | 22,089,809,189 | IssuesEvent | 2022-06-01 04:25:25 | Railcraft/Railcraft | https://api.github.com/repos/Railcraft/Railcraft | closed | Tunnel bore drops when placed | need more info needs verification medium priority carts | When i place the tunnel bore it don´t apear, well, it do the animation of placement but when is in the rail is like it was broken for a player and go as item to the floor.
Thanks for all | 1.0 | Tunnel bore drops when placed - When i place the tunnel bore it don´t apear, well, it do the animation of placement but when is in the rail is like it was broken for a player and go as item to the floor.
Thanks for all | priority | tunnel bore drops when placed when i place the tunnel bore it don´t apear well it do the animation of placement but when is in the rail is like it was broken for a player and go as item to the floor thanks for all | 1 |
201,785 | 23,039,651,229 | IssuesEvent | 2022-07-23 01:08:23 | rvvergara/my-first-rails-app | https://api.github.com/repos/rvvergara/my-first-rails-app | opened | CVE-2022-31163 (High) detected in tzinfo-1.2.5.gem | security vulnerability | ## CVE-2022-31163 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tzinfo-1.2.5.gem</b></p></summary>
<p>TZInfo provides daylight savings aware transformations between times in different time zones.</p>
<p>Library home page: <a href="https://rubygems.org/gems/tzinfo-1.2.5.gem">https://rubygems.org/gems/tzinfo-1.2.5.gem</a></p>
<p>
Dependency Hierarchy:
- sass-rails-5.0.7.gem (Root Library)
- railties-5.2.2.1.gem
- actionpack-5.2.2.1.gem
- rails-dom-testing-2.0.3.gem
- activesupport-5.2.2.1.gem
- :x: **tzinfo-1.2.5.gem** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TZInfo is a Ruby library that provides access to time zone data and allows times to be converted using time zone rules. Versions prior to 0.36.1, as well as those prior to 1.2.10 when used with the Ruby data source tzinfo-data, are vulnerable to relative path traversal. With the Ruby data source, time zones are defined in Ruby files. There is one file per time zone. Time zone files are loaded with `require` on demand. In the affected versions, `TZInfo::Timezone.get` fails to validate time zone identifiers correctly, allowing a new line character within the identifier. With Ruby version 1.9.3 and later, `TZInfo::Timezone.get` can be made to load unintended files with `require`, executing them within the Ruby process. Versions 0.3.61 and 1.2.10 include fixes to correctly validate time zone identifiers. Versions 2.0.0 and later are not vulnerable. Version 0.3.61 can still load arbitrary files from the Ruby load path if their name follows the rules for a valid time zone identifier and the file has a prefix of `tzinfo/definition` within a directory in the load path. Applications should ensure that untrusted files are not placed in a directory on the load path. As a workaround, the time zone identifier can be validated before passing to `TZInfo::Timezone.get` by ensuring it matches the regular expression `\A[A-Za-z0-9+\-_]+(?:\/[A-Za-z0-9+\-_]+)*\z`.
<p>Publish Date: 2022-07-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-31163>CVE-2022-31163</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tzinfo/tzinfo/security/advisories/GHSA-5cm2-9h8c-rvfx">https://github.com/tzinfo/tzinfo/security/advisories/GHSA-5cm2-9h8c-rvfx</a></p>
<p>Release Date: 2022-07-22</p>
<p>Fix Resolution: tzinfo - 0.3.61,1.2.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-31163 (High) detected in tzinfo-1.2.5.gem - ## CVE-2022-31163 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tzinfo-1.2.5.gem</b></p></summary>
<p>TZInfo provides daylight savings aware transformations between times in different time zones.</p>
<p>Library home page: <a href="https://rubygems.org/gems/tzinfo-1.2.5.gem">https://rubygems.org/gems/tzinfo-1.2.5.gem</a></p>
<p>
Dependency Hierarchy:
- sass-rails-5.0.7.gem (Root Library)
- railties-5.2.2.1.gem
- actionpack-5.2.2.1.gem
- rails-dom-testing-2.0.3.gem
- activesupport-5.2.2.1.gem
- :x: **tzinfo-1.2.5.gem** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TZInfo is a Ruby library that provides access to time zone data and allows times to be converted using time zone rules. Versions prior to 0.36.1, as well as those prior to 1.2.10 when used with the Ruby data source tzinfo-data, are vulnerable to relative path traversal. With the Ruby data source, time zones are defined in Ruby files. There is one file per time zone. Time zone files are loaded with `require` on demand. In the affected versions, `TZInfo::Timezone.get` fails to validate time zone identifiers correctly, allowing a new line character within the identifier. With Ruby version 1.9.3 and later, `TZInfo::Timezone.get` can be made to load unintended files with `require`, executing them within the Ruby process. Versions 0.3.61 and 1.2.10 include fixes to correctly validate time zone identifiers. Versions 2.0.0 and later are not vulnerable. Version 0.3.61 can still load arbitrary files from the Ruby load path if their name follows the rules for a valid time zone identifier and the file has a prefix of `tzinfo/definition` within a directory in the load path. Applications should ensure that untrusted files are not placed in a directory on the load path. As a workaround, the time zone identifier can be validated before passing to `TZInfo::Timezone.get` by ensuring it matches the regular expression `\A[A-Za-z0-9+\-_]+(?:\/[A-Za-z0-9+\-_]+)*\z`.
<p>Publish Date: 2022-07-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-31163>CVE-2022-31163</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tzinfo/tzinfo/security/advisories/GHSA-5cm2-9h8c-rvfx">https://github.com/tzinfo/tzinfo/security/advisories/GHSA-5cm2-9h8c-rvfx</a></p>
<p>Release Date: 2022-07-22</p>
<p>Fix Resolution: tzinfo - 0.3.61,1.2.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in tzinfo gem cve high severity vulnerability vulnerable library tzinfo gem tzinfo provides daylight savings aware transformations between times in different time zones library home page a href dependency hierarchy sass rails gem root library railties gem actionpack gem rails dom testing gem activesupport gem x tzinfo gem vulnerable library found in base branch master vulnerability details tzinfo is a ruby library that provides access to time zone data and allows times to be converted using time zone rules versions prior to as well as those prior to when used with the ruby data source tzinfo data are vulnerable to relative path traversal with the ruby data source time zones are defined in ruby files there is one file per time zone time zone files are loaded with require on demand in the affected versions tzinfo timezone get fails to validate time zone identifiers correctly allowing a new line character within the identifier with ruby version and later tzinfo timezone get can be made to load unintended files with require executing them within the ruby process versions and include fixes to correctly validate time zone identifiers versions and later are not vulnerable version can still load arbitrary files from the ruby load path if their name follows the rules for a valid time zone identifier and the file has a prefix of tzinfo definition within a directory in the load path applications should ensure that untrusted files are not placed in a directory on the load path as a workaround the time zone identifier can be validated before passing to tzinfo timezone get by ensuring it matches the regular expression a z publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tzinfo step up your open source security game with mend | 0 |
314,155 | 9,593,393,362 | IssuesEvent | 2019-05-09 11:25:03 | luna/luna-studio | https://api.github.com/repos/luna/luna-studio | opened | Update LTS in Luna Studio Runner | Category: Tooling Change: Breaking Difficulty: Core Contributor Priority: Low Type: Enhancement | ### Summary
Currently [`luna-studio-runner`](https://github.com/luna/luna-studio/blob/master/runner/) depends on a very old LTS (`lts-9.2`) which prevents it from building on NIx, and complicates building it in general. This should be updated to use the same LTS as
### Value
- This will greatly simplify the task of building Luna Studio as it will rely only on a single version of GHC and a single version of GHCJS, rather than multiple of the former.
- This update will allow LS to be seamlessly built using nix as discussed in https://github.com/luna/luna-manager/issues/127.
### Specification
- Separate out the LTS and dependency configuration of Luna Studio's backend into a snapshot overlay, as is done in [`luna`](https://github.com/luna/luna/blob/master/config/snapshot.yaml).
- Use this config with the LS backend.
- Use this config with the `luna-studio-runner` package.
- Fix the broken git dependencies by updating to the copies contained in the [`luna/lib`](https://github.com/luna/luna/tree/master/lib) directory in the Luna tree. Some of these are updates to the existing dependencies, while others have been replaced by newer libraries with similar APIs.
- Rework the portions of the `luna-studio-runner` code that depended on the older versions of these libraries.
### Acceptance Criteria & Test Cases
- `luna-studio-runner` builds using the newer LTS.
- It still works as intended.
| 1.0 | Update LTS in Luna Studio Runner - ### Summary
Currently [`luna-studio-runner`](https://github.com/luna/luna-studio/blob/master/runner/) depends on a very old LTS (`lts-9.2`) which prevents it from building on NIx, and complicates building it in general. This should be updated to use the same LTS as
### Value
- This will greatly simplify the task of building Luna Studio as it will rely only on a single version of GHC and a single version of GHCJS, rather than multiple of the former.
- This update will allow LS to be seamlessly built using nix as discussed in https://github.com/luna/luna-manager/issues/127.
### Specification
- Separate out the LTS and dependency configuration of Luna Studio's backend into a snapshot overlay, as is done in [`luna`](https://github.com/luna/luna/blob/master/config/snapshot.yaml).
- Use this config with the LS backend.
- Use this config with the `luna-studio-runner` package.
- Fix the broken git dependencies by updating to the copies contained in the [`luna/lib`](https://github.com/luna/luna/tree/master/lib) directory in the Luna tree. Some of these are updates to the existing dependencies, while others have been replaced by newer libraries with similar APIs.
- Rework the portions of the `luna-studio-runner` code that depended on the older versions of these libraries.
### Acceptance Criteria & Test Cases
- `luna-studio-runner` builds using the newer LTS.
- It still works as intended.
| priority | update lts in luna studio runner summary currently depends on a very old lts lts which prevents it from building on nix and complicates building it in general this should be updated to use the same lts as value this will greatly simplify the task of building luna studio as it will rely only on a single version of ghc and a single version of ghcjs rather than multiple of the former this update will allow ls to be seamlessly built using nix as discussed in specification separate out the lts and dependency configuration of luna studio s backend into a snapshot overlay as is done in use this config with the ls backend use this config with the luna studio runner package fix the broken git dependencies by updating to the copies contained in the directory in the luna tree some of these are updates to the existing dependencies while others have been replaced by newer libraries with similar apis rework the portions of the luna studio runner code that depended on the older versions of these libraries acceptance criteria test cases luna studio runner builds using the newer lts it still works as intended | 1 |
498,050 | 14,399,396,076 | IssuesEvent | 2020-12-03 10:52:09 | stanvanrooy/instauto | https://api.github.com/repos/stanvanrooy/instauto | opened | Implement pagination for direct messaging inbox | Priority: Medium Status: Available Type: Enchancement | The current implementation is only able to retrieve the first 'page' of your inbox. You should be able to iterate over all results instead. | 1.0 | Implement pagination for direct messaging inbox - The current implementation is only able to retrieve the first 'page' of your inbox. You should be able to iterate over all results instead. | priority | implement pagination for direct messaging inbox the current implementation is only able to retrieve the first page of your inbox you should be able to iterate over all results instead | 1 |
88,142 | 11,034,910,706 | IssuesEvent | 2019-12-07 09:37:05 | Students-of-the-city-of-Kostroma/Ray-of-hope | https://api.github.com/repos/Students-of-the-city-of-Kostroma/Ray-of-hope | closed | Отсутствует выбор даты завершения мероприятия на макете публикации постов вида "Мероприятие" для Android | Android Bug Design O3 PR4 Sprint 7 | Макет не соответствует требованиям описанным в [спецификации](https://docs.google.com/document/d/1p0aFXnl3jeQWKu_cQHuSFoc59WKs-PMpChTdkxWICBA/edit), так как отсутствует кнопка (поле) "Дата завершения мероприятия".
Шаги для воспроизведения бага:
1. Открыть по ссылке [таск](https://github.com/Students-of-the-city-of-Kostroma/Ray-of-hope/issues/103)
2. Открыть макет "Новое мероприятие"
Ожидаемый результат:
Макет соответствует [спецификации.](https://docs.google.com/document/d/1p0aFXnl3jeQWKu_cQHuSFoc59WKs-PMpChTdkxWICBA/edit)
Фактический результат:
Макет не соответствует [спецификации.](https://docs.google.com/document/d/1p0aFXnl3jeQWKu_cQHuSFoc59WKs-PMpChTdkxWICBA/edit) | 1.0 | Отсутствует выбор даты завершения мероприятия на макете публикации постов вида "Мероприятие" для Android - Макет не соответствует требованиям описанным в [спецификации](https://docs.google.com/document/d/1p0aFXnl3jeQWKu_cQHuSFoc59WKs-PMpChTdkxWICBA/edit), так как отсутствует кнопка (поле) "Дата завершения мероприятия".
Шаги для воспроизведения бага:
1. Открыть по ссылке [таск](https://github.com/Students-of-the-city-of-Kostroma/Ray-of-hope/issues/103)
2. Открыть макет "Новое мероприятие"
Ожидаемый результат:
Макет соответствует [спецификации.](https://docs.google.com/document/d/1p0aFXnl3jeQWKu_cQHuSFoc59WKs-PMpChTdkxWICBA/edit)
Фактический результат:
Макет не соответствует [спецификации.](https://docs.google.com/document/d/1p0aFXnl3jeQWKu_cQHuSFoc59WKs-PMpChTdkxWICBA/edit) | non_priority | отсутствует выбор даты завершения мероприятия на макете публикации постов вида мероприятие для android макет не соответствует требованиям описанным в так как отсутствует кнопка поле дата завершения мероприятия шаги для воспроизведения бага открыть по ссылке открыть макет новое мероприятие ожидаемый результат макет соответствует фактический результат макет не соответствует | 0 |
310,755 | 9,523,722,347 | IssuesEvent | 2019-04-27 19:42:30 | bbottema/simple-java-mail | https://api.github.com/repos/bbottema/simple-java-mail | closed | ThreadPoolExecutor terminated when sending an async mail | Priority-Medium bug | Hello,
In production I had this error with simple-java-mail 4.4.5:
```
java.util.concurrent.RejectedExecutionException: Task sendMail process rejected from java.util.concurrent.ThreadPoolExecutor@79627f99[Terminated, pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 3]
at java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2047)
at java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:823)
at java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1369)
at org.simplejavamail.mailer.internal.mailsender.MailSender.send(MailSender.java:196)
at org.simplejavamail.mailer.Mailer.sendMail(Mailer.java:364)
```
The call was: `mailer.sendMail(email, true);`
By looking at these classes:
- the 4.4.5 version: https://github.com/bbottema/simple-java-mail/blob/4.4.5/src/main/java/org/simplejavamail/mailer/internal/mailsender/MailSender.java
- the latest version: https://github.com/bbottema/simple-java-mail/blob/develop/modules/simple-java-mail/src/main/java/org/simplejavamail/mailer/internal/mailsender/MailSenderImpl.java
There is indeed a bug:
- in the synchronized `checkShutDownRunningProcesses` method, the pool is shutdown if there is no more mail to send,
- in the synchronized `send` method, the pool is started if it is null or shutdown.
But the thing is, calling the `shutdown ` method on a `ThreadPoolExecutor` is not a synchronized operation, it just *Initiates an orderly shutdown*.
My main question is actually: why shutting down the pool if there is no mail to send?
In the comment before the shutdown is is written *shutdown the threadpool, or else the Mailer will keep any JVM alive forever*. But actually by default, the thread factory in the `ThreadPoolExecutor` build non daemon threads, so it will not keep the JVM alive. | 1.0 | ThreadPoolExecutor terminated when sending an async mail - Hello,
In production I had this error with simple-java-mail 4.4.5:
```
java.util.concurrent.RejectedExecutionException: Task sendMail process rejected from java.util.concurrent.ThreadPoolExecutor@79627f99[Terminated, pool size = 0, active threads = 0, queued tasks = 0, completed tasks = 3]
at java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2047)
at java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:823)
at java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1369)
at org.simplejavamail.mailer.internal.mailsender.MailSender.send(MailSender.java:196)
at org.simplejavamail.mailer.Mailer.sendMail(Mailer.java:364)
```
The call was: `mailer.sendMail(email, true);`
By looking at these classes:
- the 4.4.5 version: https://github.com/bbottema/simple-java-mail/blob/4.4.5/src/main/java/org/simplejavamail/mailer/internal/mailsender/MailSender.java
- the latest version: https://github.com/bbottema/simple-java-mail/blob/develop/modules/simple-java-mail/src/main/java/org/simplejavamail/mailer/internal/mailsender/MailSenderImpl.java
There is indeed a bug:
- in the synchronized `checkShutDownRunningProcesses` method, the pool is shutdown if there is no more mail to send,
- in the synchronized `send` method, the pool is started if it is null or shutdown.
But the thing is, calling the `shutdown ` method on a `ThreadPoolExecutor` is not a synchronized operation, it just *Initiates an orderly shutdown*.
My main question is actually: why shutting down the pool if there is no mail to send?
In the comment before the shutdown is is written *shutdown the threadpool, or else the Mailer will keep any JVM alive forever*. But actually by default, the thread factory in the `ThreadPoolExecutor` build non daemon threads, so it will not keep the JVM alive. | priority | threadpoolexecutor terminated when sending an async mail hello in production i had this error with simple java mail java util concurrent rejectedexecutionexception task sendmail process rejected from java util concurrent threadpoolexecutor at java util concurrent threadpoolexecutor abortpolicy rejectedexecution threadpoolexecutor java at java util concurrent threadpoolexecutor reject threadpoolexecutor java at java util concurrent threadpoolexecutor execute threadpoolexecutor java at org simplejavamail mailer internal mailsender mailsender send mailsender java at org simplejavamail mailer mailer sendmail mailer java the call was mailer sendmail email true by looking at these classes the version the latest version there is indeed a bug in the synchronized checkshutdownrunningprocesses method the pool is shutdown if there is no more mail to send in the synchronized send method the pool is started if it is null or shutdown but the thing is calling the shutdown method on a threadpoolexecutor is not a synchronized operation it just initiates an orderly shutdown my main question is actually why shutting down the pool if there is no mail to send in the comment before the shutdown is is written shutdown the threadpool or else the mailer will keep any jvm alive forever but actually by default the thread factory in the threadpoolexecutor build non daemon threads so it will not keep the jvm alive | 1 |
517,260 | 14,998,157,332 | IssuesEvent | 2021-01-29 17:58:49 | IMA-WorldHealth/bhima | https://api.github.com/repos/IMA-WorldHealth/bhima | closed | Feature Request: allow users to merge lots | Stock priority:urgent | We need a way to merge lots. There are many lots with the same name at IMCK that should probably be the same lot. This would probably avoid quite a few mistakes. | 1.0 | Feature Request: allow users to merge lots - We need a way to merge lots. There are many lots with the same name at IMCK that should probably be the same lot. This would probably avoid quite a few mistakes. | priority | feature request allow users to merge lots we need a way to merge lots there are many lots with the same name at imck that should probably be the same lot this would probably avoid quite a few mistakes | 1 |
289,461 | 31,933,012,183 | IssuesEvent | 2023-09-19 08:41:41 | Trinadh465/linux-4.1.15_CVE-2023-4128 | https://api.github.com/repos/Trinadh465/linux-4.1.15_CVE-2023-4128 | opened | CVE-2021-3348 (High) detected in linux-stable-rtv4.1.33 | Mend: dependency security vulnerability | ## CVE-2021-3348 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-4128/commit/0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8">0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/nbd.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/nbd.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
nbd_add_socket in drivers/block/nbd.c in the Linux kernel through 5.10.12 has an ndb_queue_rq use-after-free that could be triggered by local attackers (with access to the nbd device) via an I/O request at a certain point during device setup, aka CID-b98e762e3d71.
<p>Publish Date: 2021-02-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3348>CVE-2021-3348</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-3348">https://www.linuxkernelcves.com/cves/CVE-2021-3348</a></p>
<p>Release Date: 2021-02-01</p>
<p>Fix Resolution: v4.14.219,4.19.173,v5.4.95,v5.10.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-3348 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2021-3348 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-4128/commit/0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8">0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/nbd.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/nbd.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
nbd_add_socket in drivers/block/nbd.c in the Linux kernel through 5.10.12 has an ndb_queue_rq use-after-free that could be triggered by local attackers (with access to the nbd device) via an I/O request at a certain point during device setup, aka CID-b98e762e3d71.
<p>Publish Date: 2021-02-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3348>CVE-2021-3348</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-3348">https://www.linuxkernelcves.com/cves/CVE-2021-3348</a></p>
<p>Release Date: 2021-02-01</p>
<p>Fix Resolution: v4.14.219,4.19.173,v5.4.95,v5.10.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers block nbd c drivers block nbd c vulnerability details nbd add socket in drivers block nbd c in the linux kernel through has an ndb queue rq use after free that could be triggered by local attackers with access to the nbd device via an i o request at a certain point during device setup aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
222,667 | 17,087,276,263 | IssuesEvent | 2021-07-08 13:23:35 | osoc21/Final-Improved-String-pHarser | https://api.github.com/repos/osoc21/Final-Improved-String-pHarser | closed | Dockerfile & documentation made for a development environment | documentation enhancement | Before deployment, care should be taken that the Dockerfile uses the development flags only when specifically ran in debug/dev, and that the documentation reflects any changes that that brings (different ports, setup...). | 1.0 | Dockerfile & documentation made for a development environment - Before deployment, care should be taken that the Dockerfile uses the development flags only when specifically ran in debug/dev, and that the documentation reflects any changes that that brings (different ports, setup...). | non_priority | dockerfile documentation made for a development environment before deployment care should be taken that the dockerfile uses the development flags only when specifically ran in debug dev and that the documentation reflects any changes that that brings different ports setup | 0 |
21,171 | 3,687,260,332 | IssuesEvent | 2016-02-25 07:15:47 | narendra-addweb/SwitchedOn | https://api.github.com/repos/narendra-addweb/SwitchedOn | closed | Homepage About us -> image | Complete Design | The image on the left side looks smaller than given in the mock up images | 1.0 | Homepage About us -> image - The image on the left side looks smaller than given in the mock up images | non_priority | homepage about us image the image on the left side looks smaller than given in the mock up images | 0 |
235,541 | 25,955,204,687 | IssuesEvent | 2022-12-18 05:32:03 | Dima2022/JS-Demo | https://api.github.com/repos/Dima2022/JS-Demo | closed | WS-2019-0032 (High) detected in js-yaml-3.6.1.tgz, js-yaml-3.5.5.tgz - autoclosed | security vulnerability | ## WS-2019-0032 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>js-yaml-3.6.1.tgz</b>, <b>js-yaml-3.5.5.tgz</b></p></summary>
<p>
<details><summary><b>js-yaml-3.6.1.tgz</b></p></summary>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.6.1.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.6.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/coveralls/node_modules/js-yaml/package.json</p>
<p>
Dependency Hierarchy:
- grunt-if-0.2.0.tgz (Root Library)
- grunt-contrib-nodeunit-1.0.0.tgz
- nodeunit-0.9.5.tgz
- tap-7.1.2.tgz
- coveralls-2.13.3.tgz
- :x: **js-yaml-3.6.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>js-yaml-3.5.5.tgz</b></p></summary>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.5.5.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.5.5.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/js-yaml/package.json</p>
<p>
Dependency Hierarchy:
- grunt-if-0.2.0.tgz (Root Library)
- grunt-contrib-nodeunit-1.0.0.tgz
- nodeunit-0.9.5.tgz
- tap-7.1.2.tgz
- :x: **js-yaml-3.5.5.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/Dima2022/JS-Demo/commit/4edf46ace164b01728ef7066c6a8e7464b89143a">4edf46ace164b01728ef7066c6a8e7464b89143a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions js-yaml prior to 3.13.0 are vulnerable to Denial of Service. By parsing a carefully-crafted YAML file, the node process stalls and may exhaust system resources leading to a Denial of Service.
<p>Publish Date: 2019-03-20
<p>URL: <a href=https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235>WS-2019-0032</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/788/versions">https://www.npmjs.com/advisories/788/versions</a></p>
<p>Release Date: 2019-03-20</p>
<p>Fix Resolution: js-yaml - 3.13.0</p>
</p>
</details>
<p></p>
| True | WS-2019-0032 (High) detected in js-yaml-3.6.1.tgz, js-yaml-3.5.5.tgz - autoclosed - ## WS-2019-0032 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>js-yaml-3.6.1.tgz</b>, <b>js-yaml-3.5.5.tgz</b></p></summary>
<p>
<details><summary><b>js-yaml-3.6.1.tgz</b></p></summary>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.6.1.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.6.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/coveralls/node_modules/js-yaml/package.json</p>
<p>
Dependency Hierarchy:
- grunt-if-0.2.0.tgz (Root Library)
- grunt-contrib-nodeunit-1.0.0.tgz
- nodeunit-0.9.5.tgz
- tap-7.1.2.tgz
- coveralls-2.13.3.tgz
- :x: **js-yaml-3.6.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>js-yaml-3.5.5.tgz</b></p></summary>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.5.5.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.5.5.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/js-yaml/package.json</p>
<p>
Dependency Hierarchy:
- grunt-if-0.2.0.tgz (Root Library)
- grunt-contrib-nodeunit-1.0.0.tgz
- nodeunit-0.9.5.tgz
- tap-7.1.2.tgz
- :x: **js-yaml-3.5.5.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/Dima2022/JS-Demo/commit/4edf46ace164b01728ef7066c6a8e7464b89143a">4edf46ace164b01728ef7066c6a8e7464b89143a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions js-yaml prior to 3.13.0 are vulnerable to Denial of Service. By parsing a carefully-crafted YAML file, the node process stalls and may exhaust system resources leading to a Denial of Service.
<p>Publish Date: 2019-03-20
<p>URL: <a href=https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235>WS-2019-0032</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/788/versions">https://www.npmjs.com/advisories/788/versions</a></p>
<p>Release Date: 2019-03-20</p>
<p>Fix Resolution: js-yaml - 3.13.0</p>
</p>
</details>
<p></p>
| non_priority | ws high detected in js yaml tgz js yaml tgz autoclosed ws high severity vulnerability vulnerable libraries js yaml tgz js yaml tgz js yaml tgz yaml parser and serializer library home page a href path to dependency file package json path to vulnerable library node modules coveralls node modules js yaml package json dependency hierarchy grunt if tgz root library grunt contrib nodeunit tgz nodeunit tgz tap tgz coveralls tgz x js yaml tgz vulnerable library js yaml tgz yaml parser and serializer library home page a href path to dependency file package json path to vulnerable library node modules js yaml package json dependency hierarchy grunt if tgz root library grunt contrib nodeunit tgz nodeunit tgz tap tgz x js yaml tgz vulnerable library found in head commit a href found in base branch master vulnerability details versions js yaml prior to are vulnerable to denial of service by parsing a carefully crafted yaml file the node process stalls and may exhaust system resources leading to a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution js yaml | 0 |
86,091 | 8,020,153,432 | IssuesEvent | 2018-07-26 03:05:54 | CGCookie/retopoflow | https://api.github.com/repos/CGCookie/retopoflow | closed | Bridge Tool | Ready for Testing tool: contours | The Bridge Tool is needed in order to manually fill two Contour strokes, such as this:

| 1.0 | Bridge Tool - The Bridge Tool is needed in order to manually fill two Contour strokes, such as this:

| non_priority | bridge tool the bridge tool is needed in order to manually fill two contour strokes such as this | 0 |
97,539 | 28,314,199,526 | IssuesEvent | 2023-04-10 18:05:24 | PowerShell/vscode-powershell | https://api.github.com/repos/PowerShell/vscode-powershell | closed | Ubuntu CI Tests do not appear to be running | Issue-Enhancement Area-Build & Release | ### Prerequisites
- [X] I have written a descriptive issue title.
- [X] I have searched all [issues](https://github.com/PowerShell/vscode-powershell/issues?q=is%3Aissue) to ensure it has not already been reported.
### Summary
Ubuntu is completing 5 minutes before the others and looking at the output pane it does not appear that tests are running.

### Proposed Design
_No response_ | 1.0 | Ubuntu CI Tests do not appear to be running - ### Prerequisites
- [X] I have written a descriptive issue title.
- [X] I have searched all [issues](https://github.com/PowerShell/vscode-powershell/issues?q=is%3Aissue) to ensure it has not already been reported.
### Summary
Ubuntu is completing 5 minutes before the others and looking at the output pane it does not appear that tests are running.

### Proposed Design
_No response_ | non_priority | ubuntu ci tests do not appear to be running prerequisites i have written a descriptive issue title i have searched all to ensure it has not already been reported summary ubuntu is completing minutes before the others and looking at the output pane it does not appear that tests are running proposed design no response | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.