Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 4
112
| repo_url
stringlengths 33
141
| action
stringclasses 3
values | title
stringlengths 1
1.02k
| labels
stringlengths 4
1.54k
| body
stringlengths 1
262k
| index
stringclasses 17
values | text_combine
stringlengths 95
262k
| label
stringclasses 2
values | text
stringlengths 96
252k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
268,248
| 8,405,323,677
|
IssuesEvent
|
2018-10-11 14:59:29
|
CS2103-AY1819S1-F10-2/main
|
https://api.github.com/repos/CS2103-AY1819S1-F10-2/main
|
opened
|
I can register my bicycles within the app
|
priority.High type.Story
|
- Implement the Bike class.
- Let the AddressBook class contain a set of Bikes.
- Implement an addBike command.
|
1.0
|
I can register my bicycles within the app - - Implement the Bike class.
- Let the AddressBook class contain a set of Bikes.
- Implement an addBike command.
|
non_test
|
i can register my bicycles within the app implement the bike class let the addressbook class contain a set of bikes implement an addbike command
| 0
|
338,247
| 30,288,514,465
|
IssuesEvent
|
2023-07-09 01:21:11
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
reopened
|
Fix creation.test_frombuffer
|
Sub Task Ivy API Experimental Failing Test
|
| | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4902117212/jobs/8753770307"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4902117212/jobs/8753770307"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4902117212/jobs/8753770307"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4902117212/jobs/8753770307"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5497189625/jobs/10017714381"><img src=https://img.shields.io/badge/-success-success></a>
|
1.0
|
Fix creation.test_frombuffer - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4902117212/jobs/8753770307"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4902117212/jobs/8753770307"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4902117212/jobs/8753770307"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4902117212/jobs/8753770307"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5497189625/jobs/10017714381"><img src=https://img.shields.io/badge/-success-success></a>
|
test
|
fix creation test frombuffer tensorflow a href src torch a href src jax a href src numpy a href src paddle a href src
| 1
|
247,000
| 20,955,364,259
|
IssuesEvent
|
2022-03-27 02:54:28
|
azuline/repertoire
|
https://api.github.com/repos/azuline/repertoire
|
closed
|
Storybook for components
|
frontend testing
|
For the `components/` directory. Consider also testing the components with Storybook.
|
1.0
|
Storybook for components - For the `components/` directory. Consider also testing the components with Storybook.
|
test
|
storybook for components for the components directory consider also testing the components with storybook
| 1
|
192,464
| 14,617,547,397
|
IssuesEvent
|
2020-12-22 14:56:36
|
github-vet/rangeloop-pointer-findings
|
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
|
closed
|
praparn/kubernetes_201907: WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go; 5 LoC
|
fresh test tiny
|
Found a possible issue in [praparn/kubernetes_201907](https://www.github.com/praparn/kubernetes_201907) at [WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go](https://github.com/praparn/kubernetes_201907/blob/309486c01dfaa48464c7dce2ac5ec6742df92eef/WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go#L56-L60)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to m at line 58 may start a goroutine
[Click here to see the code in its original context.](https://github.com/praparn/kubernetes_201907/blob/309486c01dfaa48464c7dce2ac5ec6742df92eef/WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go#L56-L60)
<details>
<summary>Click here to show the 5 line(s) of Go which triggered the analyzer.</summary>
```go
for _, m := range metas[5:] {
var buf bytes.Buffer
testutil.Ok(t, json.NewEncoder(&buf).Encode(&m))
testutil.Ok(t, bkt.Upload(ctx, path.Join(m.ULID.String(), block.MetaFilename), &buf))
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 309486c01dfaa48464c7dce2ac5ec6742df92eef
|
1.0
|
praparn/kubernetes_201907: WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go; 5 LoC -
Found a possible issue in [praparn/kubernetes_201907](https://www.github.com/praparn/kubernetes_201907) at [WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go](https://github.com/praparn/kubernetes_201907/blob/309486c01dfaa48464c7dce2ac5ec6742df92eef/WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go#L56-L60)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to m at line 58 may start a goroutine
[Click here to see the code in its original context.](https://github.com/praparn/kubernetes_201907/blob/309486c01dfaa48464c7dce2ac5ec6742df92eef/WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go#L56-L60)
<details>
<summary>Click here to show the 5 line(s) of Go which triggered the analyzer.</summary>
```go
for _, m := range metas[5:] {
var buf bytes.Buffer
testutil.Ok(t, json.NewEncoder(&buf).Encode(&m))
testutil.Ok(t, bkt.Upload(ctx, path.Join(m.ULID.String(), block.MetaFilename), &buf))
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 309486c01dfaa48464c7dce2ac5ec6742df92eef
|
test
|
praparn kubernetes workshop monitor prometheus operator vendor github com improbable eng thanos pkg compact compact test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to m at line may start a goroutine click here to show the line s of go which triggered the analyzer go for m range metas var buf bytes buffer testutil ok t json newencoder buf encode m testutil ok t bkt upload ctx path join m ulid string block metafilename buf leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 1
|
17,491
| 3,619,229,431
|
IssuesEvent
|
2016-02-08 15:17:18
|
Microsoft/vscode
|
https://api.github.com/repos/Microsoft/vscode
|
opened
|
Test working files picker
|
testplan-item
|
You can now quick open working files.
* verify the picker is well integrated into our quick open story and either shows you all working files or the ones matching your search if searching
* verify the order matches the order in the explorer unless you search, where we then use fuzzy matching/scoring
* verify you see untitled files as well as dirty indicators for dirty files
|
1.0
|
Test working files picker - You can now quick open working files.
* verify the picker is well integrated into our quick open story and either shows you all working files or the ones matching your search if searching
* verify the order matches the order in the explorer unless you search, where we then use fuzzy matching/scoring
* verify you see untitled files as well as dirty indicators for dirty files
|
test
|
test working files picker you can now quick open working files verify the picker is well integrated into our quick open story and either shows you all working files or the ones matching your search if searching verify the order matches the order in the explorer unless you search where we then use fuzzy matching scoring verify you see untitled files as well as dirty indicators for dirty files
| 1
|
7,682
| 9,935,667,963
|
IssuesEvent
|
2019-07-02 17:07:44
|
Direwolf20-MC/BuildingGadgets
|
https://api.github.com/repos/Direwolf20-MC/BuildingGadgets
|
closed
|
Gadget's drop Items, when Dim with bound Inv is deleted
|
1.12 bug incompatibility
|
A user reported this on Direcord:
- bound to a chest in an rftools dim
- delete said dim
- Gadget's will now drop Items instead of putting them into the player inventory
Notice that this happened on a dedicated Server - I don't know whether this can be reproduced using SP-mode.
I presume he was using 2.6.8 but I didn't ask either.
|
True
|
Gadget's drop Items, when Dim with bound Inv is deleted - A user reported this on Direcord:
- bound to a chest in an rftools dim
- delete said dim
- Gadget's will now drop Items instead of putting them into the player inventory
Notice that this happened on a dedicated Server - I don't know whether this can be reproduced using SP-mode.
I presume he was using 2.6.8 but I didn't ask either.
|
non_test
|
gadget s drop items when dim with bound inv is deleted a user reported this on direcord bound to a chest in an rftools dim delete said dim gadget s will now drop items instead of putting them into the player inventory notice that this happened on a dedicated server i don t know whether this can be reproduced using sp mode i presume he was using but i didn t ask either
| 0
|
90,492
| 11,405,758,977
|
IssuesEvent
|
2020-01-31 12:53:32
|
fecgov/fec-cms
|
https://api.github.com/repos/fecgov/fec-cms
|
opened
|
Design homepage change to link to map
|
High priority Work: UX/Design
|
**What we're after:**
We're rebuilding the presidential data map on .gov and we need to create a pathway from the homepage to that map.
### Completion criteria
- [ ] Mockup design(s) for the homepage to link to the map visualization
- [ ] Review with stakeholders
- [ ] Make necessary changes and move to an implementation issue
|
1.0
|
Design homepage change to link to map - **What we're after:**
We're rebuilding the presidential data map on .gov and we need to create a pathway from the homepage to that map.
### Completion criteria
- [ ] Mockup design(s) for the homepage to link to the map visualization
- [ ] Review with stakeholders
- [ ] Make necessary changes and move to an implementation issue
|
non_test
|
design homepage change to link to map what we re after we re rebuilding the presidential data map on gov and we need to create a pathway from the homepage to that map completion criteria mockup design s for the homepage to link to the map visualization review with stakeholders make necessary changes and move to an implementation issue
| 0
|
137,539
| 11,140,404,631
|
IssuesEvent
|
2019-12-21 14:08:27
|
dexpenses/dexpenses-extract
|
https://api.github.com/repos/dexpenses/dexpenses-extract
|
closed
|
Implement test receipt ec/bs-l-osteria-credit
|
enhancement test-data
|
Receipt to implement:

|
1.0
|
Implement test receipt ec/bs-l-osteria-credit - Receipt to implement:

|
test
|
implement test receipt ec bs l osteria credit receipt to implement ec bs l osteria credit
| 1
|
284,167
| 24,581,251,578
|
IssuesEvent
|
2022-10-13 15:46:31
|
IntellectualSites/PlotSquared
|
https://api.github.com/repos/IntellectualSites/PlotSquared
|
closed
|
Road schematic only paste´s top blocks not blocks under the top layer
|
Requires Testing
|
### Server Implementation
Paper
### Server Version
1.19.2
### Describe the bug
Hey there,
I´m trying to build a proper road schematic.
I built it around the plot but not only did I modify the top layer.
There are blocks beneath the top layer because I want to put beacons into the intersections with diamonds beneath.
When I create the road schematic and try it out, only the top layer is put onto roads.
Is this intended?
Sincerely cynicalmax
### To Reproduce
1. Create plotworld
2. Modify road with blocks beneath toplayer
3. Create road schematic
4. Debugpaste the road schematic
5. done
### Expected behaviour
Paste the whole road from Y:1 between the plots
### Screenshots / Videos
_No response_
### Error log (if applicable)
_No response_
### Plot Debugpaste
https://athion.net/ISPaster/paste/view/21a47b95c5764a2589496e861cd28564
### PlotSquared Version
6.9.4 Premium
### Checklist
- [X] I have included a Plot debugpaste.
- [X] I am using the newest build from https://www.spigotmc.org/resources/77506/ and the issue still persists.
### Anything else?
_No response_
|
1.0
|
Road schematic only paste´s top blocks not blocks under the top layer - ### Server Implementation
Paper
### Server Version
1.19.2
### Describe the bug
Hey there,
I´m trying to build a proper road schematic.
I built it around the plot but not only did I modify the top layer.
There are blocks beneath the top layer because I want to put beacons into the intersections with diamonds beneath.
When I create the road schematic and try it out, only the top layer is put onto roads.
Is this intended?
Sincerely cynicalmax
### To Reproduce
1. Create plotworld
2. Modify road with blocks beneath toplayer
3. Create road schematic
4. Debugpaste the road schematic
5. done
### Expected behaviour
Paste the whole road from Y:1 between the plots
### Screenshots / Videos
_No response_
### Error log (if applicable)
_No response_
### Plot Debugpaste
https://athion.net/ISPaster/paste/view/21a47b95c5764a2589496e861cd28564
### PlotSquared Version
6.9.4 Premium
### Checklist
- [X] I have included a Plot debugpaste.
- [X] I am using the newest build from https://www.spigotmc.org/resources/77506/ and the issue still persists.
### Anything else?
_No response_
|
test
|
road schematic only paste´s top blocks not blocks under the top layer server implementation paper server version describe the bug hey there i´m trying to build a proper road schematic i built it around the plot but not only did i modify the top layer there are blocks beneath the top layer because i want to put beacons into the intersections with diamonds beneath when i create the road schematic and try it out only the top layer is put onto roads is this intended sincerely cynicalmax to reproduce create plotworld modify road with blocks beneath toplayer create road schematic debugpaste the road schematic done expected behaviour paste the whole road from y between the plots screenshots videos no response error log if applicable no response plot debugpaste plotsquared version premium checklist i have included a plot debugpaste i am using the newest build from and the issue still persists anything else no response
| 1
|
114,903
| 9,764,285,076
|
IssuesEvent
|
2019-06-05 15:28:01
|
IBM/kui
|
https://api.github.com/repos/IBM/kui
|
closed
|
npm test with all layers miss k8s1 and k8s2
|
bug tests
|
LAYERS is hardcoded for non-openwhisk developers running tests.
|
1.0
|
npm test with all layers miss k8s1 and k8s2 - LAYERS is hardcoded for non-openwhisk developers running tests.
|
test
|
npm test with all layers miss and layers is hardcoded for non openwhisk developers running tests
| 1
|
72,500
| 13,879,638,912
|
IssuesEvent
|
2020-10-17 15:19:23
|
z3t0/Arduino-IRremote
|
https://api.github.com/repos/z3t0/Arduino-IRremote
|
closed
|
Mitsubishi AC Protocol Creation
|
Write Code
|
**Board:** Arduino Uno
**Library Version:** 2.1.0
**Protocol:** Creation of Mitsubishi AC Protocol
If anyone is up to it, would you be so kind as to create a Mitsubishi **Electric** AC protocol? The current Mitsubishi protocol is for a Mitsubishi TV (16 bits) and, thus, doesn't work for an AC (32 bits)
Or, could you kindly guide me on what I would have to do create my own protocol?
|
1.0
|
Mitsubishi AC Protocol Creation - **Board:** Arduino Uno
**Library Version:** 2.1.0
**Protocol:** Creation of Mitsubishi AC Protocol
If anyone is up to it, would you be so kind as to create a Mitsubishi **Electric** AC protocol? The current Mitsubishi protocol is for a Mitsubishi TV (16 bits) and, thus, doesn't work for an AC (32 bits)
Or, could you kindly guide me on what I would have to do create my own protocol?
|
non_test
|
mitsubishi ac protocol creation board arduino uno library version protocol creation of mitsubishi ac protocol if anyone is up to it would you be so kind as to create a mitsubishi electric ac protocol the current mitsubishi protocol is for a mitsubishi tv bits and thus doesn t work for an ac bits or could you kindly guide me on what i would have to do create my own protocol
| 0
|
13,451
| 9,975,252,459
|
IssuesEvent
|
2019-07-09 12:40:47
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Gpu Sharing on Azure Kubernetes
|
azure-container-service/svc container-service/svc cxp product-question triaged
|
Although there is currently no official support on k8s to share GPU resources among pods, there is an open source solution (https://github.com/AliyunContainerService/gpushare-scheduler-extender). The problem is that in order to apply the solution there is the need of direct access to the master node, and from what I understood the master node is controlled entirely by Azure.
So my question is: is there already a solution to do gpu sharing on Azure Kubernetes? If not is it possible to do modification on the master node?
I feel that the only solution I have for now is to create the VMs and set up both the master and worker nodes myself.
Thanks in advance for any help/clarification.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: deaac79e-bd40-2d95-7288-e15371726e91
* Version Independent ID: 9fd80908-20f1-7970-3bda-5a4283e4af21
* Content: [Use GPUs on Azure Kubernetes Service (AKS)](https://docs.microsoft.com/en-us/azure/aks/gpu-cluster)
* Content Source: [articles/aks/gpu-cluster.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/gpu-cluster.md)
* Service: **container-service**
* GitHub Login: @zr-msft
* Microsoft Alias: **zarhoads**
|
2.0
|
Gpu Sharing on Azure Kubernetes - Although there is currently no official support on k8s to share GPU resources among pods, there is an open source solution (https://github.com/AliyunContainerService/gpushare-scheduler-extender). The problem is that in order to apply the solution there is the need of direct access to the master node, and from what I understood the master node is controlled entirely by Azure.
So my question is: is there already a solution to do gpu sharing on Azure Kubernetes? If not is it possible to do modification on the master node?
I feel that the only solution I have for now is to create the VMs and set up both the master and worker nodes myself.
Thanks in advance for any help/clarification.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: deaac79e-bd40-2d95-7288-e15371726e91
* Version Independent ID: 9fd80908-20f1-7970-3bda-5a4283e4af21
* Content: [Use GPUs on Azure Kubernetes Service (AKS)](https://docs.microsoft.com/en-us/azure/aks/gpu-cluster)
* Content Source: [articles/aks/gpu-cluster.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/gpu-cluster.md)
* Service: **container-service**
* GitHub Login: @zr-msft
* Microsoft Alias: **zarhoads**
|
non_test
|
gpu sharing on azure kubernetes although there is currently no official support on to share gpu resources among pods there is an open source solution the problem is that in order to apply the solution there is the need of direct access to the master node and from what i understood the master node is controlled entirely by azure so my question is is there already a solution to do gpu sharing on azure kubernetes if not is it possible to do modification on the master node i feel that the only solution i have for now is to create the vms and set up both the master and worker nodes myself thanks in advance for any help clarification document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login zr msft microsoft alias zarhoads
| 0
|
57,880
| 16,114,546,169
|
IssuesEvent
|
2021-04-28 05:01:30
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Can't package dmg electron app on MacOS
|
A-Electron A-Packaging P2 S-Minor T-Defect Z-Platform-Specific
|
### Description
Can't package electron app to dmg on MacOS High Sierra v10.13.2, process hangs.
### Steps to reproduce
- git clone https://github.com/vector-im/riot-web.git && cd riot-web
- npm install
- npm run build
- npm install electron
- node_modules/.bin/build -m --x64
Describe how what happens differs from what you expected.
The process just hangs. But even before trying to build the app open with ```npm run electron``` but get only a white window.

### Version information
0.13.3
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
- **Platform**: web (in-browser) or desktop?
Desktop app, MacOS High Sierra 10.13.2
P.S: I'm pretty much a noob so it is absolutely possible that I'm doing something wrong, but following the instructions provided in the README.
|
1.0
|
Can't package dmg electron app on MacOS - ### Description
Can't package electron app to dmg on MacOS High Sierra v10.13.2, process hangs.
### Steps to reproduce
- git clone https://github.com/vector-im/riot-web.git && cd riot-web
- npm install
- npm run build
- npm install electron
- node_modules/.bin/build -m --x64
Describe how what happens differs from what you expected.
The process just hangs. But even before trying to build the app open with ```npm run electron``` but get only a white window.

### Version information
0.13.3
<!-- IMPORTANT: please answer the following questions, to help us narrow down the problem -->
- **Platform**: web (in-browser) or desktop?
Desktop app, MacOS High Sierra 10.13.2
P.S: I'm pretty much a noob so it is absolutely possible that I'm doing something wrong, but following the instructions provided in the README.
|
non_test
|
can t package dmg electron app on macos description can t package electron app to dmg on macos high sierra process hangs steps to reproduce git clone cd riot web npm install npm run build npm install electron node modules bin build m describe how what happens differs from what you expected the process just hangs but even before trying to build the app open with npm run electron but get only a white window version information platform web in browser or desktop desktop app macos high sierra p s i m pretty much a noob so it is absolutely possible that i m doing something wrong but following the instructions provided in the readme
| 0
|
335,411
| 30,028,728,556
|
IssuesEvent
|
2023-06-27 08:11:09
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Cypress: Notifications toast cypress tests failing
|
A-Developer-Experience Z-Flaky-Test
|
### Steps to reproduce
https://github.com/matrix-org/matrix-react-sdk/actions/runs/5386706635/jobs/9777756363#step:8:10104
https://cloud.cypress.io/projects/ppvnzg/runs/15772/test-results?actions=%5B%5D&browsers=%5B%5D&groups=%5B%5D&isFlaky=%5B%5D&modificationDateRange=%7B%22startDate%22%3A%221970-01-01%22%2C%22endDate%22%3A%222038-01-19%22%7D&orderBy=EXECUTION_ORDER&oses=%5B%5D&specs=%5B%5D&statuses=%5B%7B%22value%22%3A%22FAILED%22%2C%22label%22%3A%22FAILED%22%7D%5D&testingTypesEnum=%5B%5D&utm_source=github
### Outcome
x
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
_No response_
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
Cypress: Notifications toast cypress tests failing - ### Steps to reproduce
https://github.com/matrix-org/matrix-react-sdk/actions/runs/5386706635/jobs/9777756363#step:8:10104
https://cloud.cypress.io/projects/ppvnzg/runs/15772/test-results?actions=%5B%5D&browsers=%5B%5D&groups=%5B%5D&isFlaky=%5B%5D&modificationDateRange=%7B%22startDate%22%3A%221970-01-01%22%2C%22endDate%22%3A%222038-01-19%22%7D&orderBy=EXECUTION_ORDER&oses=%5B%5D&specs=%5B%5D&statuses=%5B%7B%22value%22%3A%22FAILED%22%2C%22label%22%3A%22FAILED%22%7D%5D&testingTypesEnum=%5B%5D&utm_source=github
### Outcome
x
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
_No response_
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
No
|
test
|
cypress notifications toast cypress tests failing steps to reproduce outcome x operating system no response browser information no response url for webapp no response application version no response homeserver no response will you send logs no
| 1
|
278,284
| 24,142,940,858
|
IssuesEvent
|
2022-09-21 16:10:10
|
TahmidU/mosaic
|
https://api.github.com/repos/TahmidU/mosaic
|
closed
|
Jest RTL tests
|
test
|
# TODO
## Components
- [x] Atom/Button (https://github.com/TahmidU/mosaic/pull/177)
- [x] Atom/Card (https://github.com/TahmidU/mosaic/pull/178)
- [x] Atom/Checkbox (https://github.com/TahmidU/mosaic/pull/179)
- [x] Atom/DropdownList (https://github.com/TahmidU/mosaic/pull/180)
- [x] Atom/ExternalLink (https://github.com/TahmidU/mosaic/pull/181)
- [x] Atom/Input (https://github.com/TahmidU/mosaic/pull/182, https://github.com/TahmidU/mosaic/pull/184)
- [x] Atom/ProgressiveCircle (https://github.com/TahmidU/mosaic/pull/183)
- [x] Atom/SelectTitleList (https://github.com/TahmidU/mosaic/pull/185)
- [x] Molecules/CastCard (https://github.com/TahmidU/mosaic/pull/186)
- [x] Molecules/TypeFilter (https://github.com/TahmidU/mosaic/pull/187)
- [x] Molecules/Filter (https://github.com/TahmidU/mosaic/pull/190)
- [x] Molecules/HorizonalList (https://github.com/TahmidU/mosaic/pull/191, https://github.com/TahmidU/mosaic/pull/201)
- [x] Molecules/MediaCard (https://github.com/TahmidU/mosaic/pull/193)
- [x] Molecules/MovieCard (https://github.com/TahmidU/mosaic/pull/194)
- [x] Molecules/ReviewStat (Already tested)
- [x] Molecules/ScrollDownBtn (No point testing this)
- [x] Molecules/SearchCard (https://github.com/TahmidU/mosaic/pull/195)
- [x] Molecules/SlideButton (https://github.com/TahmidU/mosaic/pull/196)
- [x] Molecules/Socials (https://github.com/TahmidU/mosaic/pull/197)
- [x] Molecules/StarRating (https://github.com/TahmidU/mosaic/pull/198)
- [x] Molecules/SubList (No point testing it, it's just atom/SelectTitleList and atom/DropdownList)
- [x] Molecules/VideoModal (https://github.com/TahmidU/mosaic/pull/199)
- [x] Organisms/Carousel (https://github.com/TahmidU/mosaic/pull/200)
- [x] Organisms/CastList (No point testing this, it's basically testing HorizontalList again)
- [x] Organisms/DetailsCard (https://github.com/TahmidU/mosaic/pull/202)
- [x] Organisms/Footer (No point testing this)
- [x] Organisms/Layout (No point testing this)
- [x] Organisms/MediaList (No point testing this, it's basically testing HorizontalList again)
- [x] Organisms/MovieList (No point testing this, it's basically testing HorizontalList again)
- [x] Organisms/Navbar (No point testing this)
- [x] Organisms/SearchInput (https://github.com/TahmidU/mosaic/pull/204)
## Utils
- [x] MathUtils (https://github.com/TahmidU/mosaic/pull/209)
- [x] RoutingUtils (https://github.com/TahmidU/mosaic/pull/207)
- [x] TestUtils (https://github.com/TahmidU/mosaic/pull/210)
- [x] TextUtils (https://github.com/TahmidU/mosaic/pull/206)
|
1.0
|
Jest RTL tests - # TODO
## Components
- [x] Atom/Button (https://github.com/TahmidU/mosaic/pull/177)
- [x] Atom/Card (https://github.com/TahmidU/mosaic/pull/178)
- [x] Atom/Checkbox (https://github.com/TahmidU/mosaic/pull/179)
- [x] Atom/DropdownList (https://github.com/TahmidU/mosaic/pull/180)
- [x] Atom/ExternalLink (https://github.com/TahmidU/mosaic/pull/181)
- [x] Atom/Input (https://github.com/TahmidU/mosaic/pull/182, https://github.com/TahmidU/mosaic/pull/184)
- [x] Atom/ProgressiveCircle (https://github.com/TahmidU/mosaic/pull/183)
- [x] Atom/SelectTitleList (https://github.com/TahmidU/mosaic/pull/185)
- [x] Molecules/CastCard (https://github.com/TahmidU/mosaic/pull/186)
- [x] Molecules/TypeFilter (https://github.com/TahmidU/mosaic/pull/187)
- [x] Molecules/Filter (https://github.com/TahmidU/mosaic/pull/190)
- [x] Molecules/HorizonalList (https://github.com/TahmidU/mosaic/pull/191, https://github.com/TahmidU/mosaic/pull/201)
- [x] Molecules/MediaCard (https://github.com/TahmidU/mosaic/pull/193)
- [x] Molecules/MovieCard (https://github.com/TahmidU/mosaic/pull/194)
- [x] Molecules/ReviewStat (Already tested)
- [x] Molecules/ScrollDownBtn (No point testing this)
- [x] Molecules/SearchCard (https://github.com/TahmidU/mosaic/pull/195)
- [x] Molecules/SlideButton (https://github.com/TahmidU/mosaic/pull/196)
- [x] Molecules/Socials (https://github.com/TahmidU/mosaic/pull/197)
- [x] Molecules/StarRating (https://github.com/TahmidU/mosaic/pull/198)
- [x] Molecules/SubList (No point testing it, it's just atom/SelectTitleList and atom/DropdownList)
- [x] Molecules/VideoModal (https://github.com/TahmidU/mosaic/pull/199)
- [x] Organisms/Carousel (https://github.com/TahmidU/mosaic/pull/200)
- [x] Organisms/CastList (No point testing this, it's basically testing HorizontalList again)
- [x] Organisms/DetailsCard (https://github.com/TahmidU/mosaic/pull/202)
- [x] Organisms/Footer (No point testing this)
- [x] Organisms/Layout (No point testing this)
- [x] Organisms/MediaList (No point testing this, it's basically testing HorizontalList again)
- [x] Organisms/MovieList (No point testing this, it's basically testing HorizontalList again)
- [x] Organisms/Navbar (No point testing this)
- [x] Organisms/SearchInput (https://github.com/TahmidU/mosaic/pull/204)
## Utils
- [x] MathUtils (https://github.com/TahmidU/mosaic/pull/209)
- [x] RoutingUtils (https://github.com/TahmidU/mosaic/pull/207)
- [x] TestUtils (https://github.com/TahmidU/mosaic/pull/210)
- [x] TextUtils (https://github.com/TahmidU/mosaic/pull/206)
|
test
|
jest rtl tests todo components atom button atom card atom checkbox atom dropdownlist atom externallink atom input atom progressivecircle atom selecttitlelist molecules castcard molecules typefilter molecules filter molecules horizonallist molecules mediacard molecules moviecard molecules reviewstat already tested molecules scrolldownbtn no point testing this molecules searchcard molecules slidebutton molecules socials molecules starrating molecules sublist no point testing it it s just atom selecttitlelist and atom dropdownlist molecules videomodal organisms carousel organisms castlist no point testing this it s basically testing horizontallist again organisms detailscard organisms footer no point testing this organisms layout no point testing this organisms medialist no point testing this it s basically testing horizontallist again organisms movielist no point testing this it s basically testing horizontallist again organisms navbar no point testing this organisms searchinput utils mathutils routingutils testutils textutils
| 1
|
661,865
| 22,092,675,306
|
IssuesEvent
|
2022-06-01 07:30:51
|
kubevela/kubevela
|
https://api.github.com/repos/kubevela/kubevela
|
closed
|
Fail to delete applications due lack of dependency between components when using terraform
|
priority/important-soon cloud-resource area/addon
|
**Describe the bug**
<!--
A clear and concise description of what the bug is.
-->
When deleting a bunch of components in the same application using the terraform addon, the controller decides to delete them in order of creation, however, depending of the provider (in my case is Azure) it requires some order (delete all resources under a resource group and after this deleting the resource group itself).
**To Reproduce**
<!--
Steps to reproduce the behavior:
1. The YAML files of Component/Trait I used.
2. The YAML file of Application I applied.
3. Other operations I have done.
4. See error
-->
1. Configure the terraform and terraform-azure addons
```
vela addon install terraform
vela addon install terraform-azure
vela provider add terraform-azure --ARM_CLIENT_ID $ARM_CLIENT_ID --ARM_CLIENT_SECRET $ARM_CLIENT_SECRET --ARM_SUBSCRIPTION_ID $ARM_SUBSCRIPTION_ID --ARM_TENANT_ID $ARM_TENANT_ID --name azure -n vela-system
```
2. Create a Virtual Machine ComponentDefinition
```
apiVersion: core.oam.dev/v1beta1
kind: ComponentDefinition
metadata:
annotations:
definition.oam.dev/description: Terraform configuration for Azure Cloud Virtual Machine object
labels:
type: terraform
name: azure-virtual-machine
namespace: vela-system
spec:
workload:
definition:
apiVersion: terraform.core.oam.dev/v1beta1
kind: Configuration
type: configurations.terraform.core.oam.dev
schematic:
terraform:
configuration: |
provider "azurerm" {
features {}
}
data "azurerm_subnet" "vnet" {
name = var.subnet_name
virtual_network_name = var.vnet_name
resource_group_name = var.subnet_rg_name
}
module "server" {
source = "Azure/compute/azurerm"
resource_group_name = var.resource_group
nb_instances = var.instance_count
vm_os_simple = var.os
public_ip_dns = [var.hostname]
vnet_subnet_id = data.azurerm_subnet.vnet.id
location = var.location
vm_hostname = var.hostname
vm_size = var.vm_size
ssh_key = ""
ssh_key_values = [var.ssh_key]
}
variable "resource_group" {
description = "Resource group name"
type = string
default = "example"
}
variable "subnet_name" {
description = "Subnet Name"
type = string
}
variable "vnet_name" {
description = "Virtual Network Name"
type = string
}
variable "subnet_rg_name" {
description = "Resource group of the Subnet"
type = string
}
variable "instance_count" {
description = "Specify the number of vm instances."
type = number
default = 1
}
variable "hostname" {
description = "VM instance name"
type = string
default = "linux-server"
}
variable "location" {
description = "Azure Location"
type = string
default = "westeurope"
}
variable "os" {
description = "Specify UbuntuServer, WindowsServer, RHEL, openSUSE-Leap, CentOS, Debian, CoreOS or SLES"
type = string
default = "UbuntuServer"
}
variable "vm_size" {
description = "VM size"
type = string
default = "Standard_B1s"
}
variable "ssh_key" {
description = "SSH Public Key"
type = string
}
output "vm_public_name" {
value = module.server.public_ip_dns_name
}
output "vm_public_ip" {
value = module.server.public_ip_address
}
deleteResource: true
providerRef:
name: azure
namespace: default
```
3. Edit the current azure-subnet (the current one doesn't work)
```
kubectl patch componentDefinitions azure-subnet -n vela-system --type merge -p '{"spec":{"schematic":{"terraform": {"configuration": "https://github.com/alisson276/terraform-modules.git"}}}}'
```
4. Create a file called first-vm.yaml:
```
apiVersion: core.oam.dev/v1beta1
kind: Application
metadata:
name: vm-1
spec:
components:
- name: example-rg
type: azure-resource-group
properties:
location: westeurope
name: example1
writeConnectionSecretToRef:
name: example-rg
- name: example-virtual-network
type: azure-subnet
properties:
vnet_name: vnet-1
create_network: true
resource_group_name: example1
subnet_prefixes:
- 10.0.0.0/24
subnet_names:
- subnet-1
writeConnectionSecretToRef:
name: example-vnet
- name: vm-1-1
type: azure-virtual-machine
properties:
instance_count: 1
subnet_name: subnet-1
vnet_name: vnet-1
subnet_rg_name: example1
resource_group: example
hostname: kubevela
writeConnectionSecretToRef:
name: vm1-conn
```
5. Apply it:
```
vela up -f first-vm.yaml
```
6. Wait untill all resource are healthy
7. vela delete vm-1
**Expected behavior**
<!--
A clear and concise description of what you expected to happen.
-->
As you can see in the components, the second component is a subnet under example1 resource group (the first component). The problem is: the virtual machine is in another resource group (example) and it has a NIC associate to it, which is linked to the subnet. So if you try to delete the first component created first, the resource group, it fails, because the vnet (the second resource) is under it and can't be deleted because the third resource (the vm) wasn't destroyed yet.
The idea is to have a way to refer it as `depends_on` (saying that the vm depends on the vnet and the vnet depends on the resource group) or simply delete them in the reverse order that they were created.
**Screenshots**
<!--
If applicable, add screenshots to help explain your problem.
-->
**KubeVela Version**
<!--
Describe your KubeVela controller or CLI version information.
-->
CLI Version: master
Core Version: 1.3.6
GitRevision: git-c8264b8c
GolangVersion: go1.18.2
**Cluster information**
<!--
Describe your kubernetes cluster information.
- Kubernetes Version [e.g. 1.16.9]
-->
Kubernetes version: 1.22.9
**Additional context**
<!--
Add any other context about the problem here.
-->
|
1.0
|
Fail to delete applications due lack of dependency between components when using terraform - **Describe the bug**
<!--
A clear and concise description of what the bug is.
-->
When deleting a bunch of components in the same application using the terraform addon, the controller decides to delete them in order of creation, however, depending of the provider (in my case is Azure) it requires some order (delete all resources under a resource group and after this deleting the resource group itself).
**To Reproduce**
<!--
Steps to reproduce the behavior:
1. The YAML files of Component/Trait I used.
2. The YAML file of Application I applied.
3. Other operations I have done.
4. See error
-->
1. Configure the terraform and terraform-azure addons
```
vela addon install terraform
vela addon install terraform-azure
vela provider add terraform-azure --ARM_CLIENT_ID $ARM_CLIENT_ID --ARM_CLIENT_SECRET $ARM_CLIENT_SECRET --ARM_SUBSCRIPTION_ID $ARM_SUBSCRIPTION_ID --ARM_TENANT_ID $ARM_TENANT_ID --name azure -n vela-system
```
2. Create a Virtual Machine ComponentDefinition
```
apiVersion: core.oam.dev/v1beta1
kind: ComponentDefinition
metadata:
annotations:
definition.oam.dev/description: Terraform configuration for Azure Cloud Virtual Machine object
labels:
type: terraform
name: azure-virtual-machine
namespace: vela-system
spec:
workload:
definition:
apiVersion: terraform.core.oam.dev/v1beta1
kind: Configuration
type: configurations.terraform.core.oam.dev
schematic:
terraform:
configuration: |
provider "azurerm" {
features {}
}
data "azurerm_subnet" "vnet" {
name = var.subnet_name
virtual_network_name = var.vnet_name
resource_group_name = var.subnet_rg_name
}
module "server" {
source = "Azure/compute/azurerm"
resource_group_name = var.resource_group
nb_instances = var.instance_count
vm_os_simple = var.os
public_ip_dns = [var.hostname]
vnet_subnet_id = data.azurerm_subnet.vnet.id
location = var.location
vm_hostname = var.hostname
vm_size = var.vm_size
ssh_key = ""
ssh_key_values = [var.ssh_key]
}
variable "resource_group" {
description = "Resource group name"
type = string
default = "example"
}
variable "subnet_name" {
description = "Subnet Name"
type = string
}
variable "vnet_name" {
description = "Virtual Network Name"
type = string
}
variable "subnet_rg_name" {
description = "Resource group of the Subnet"
type = string
}
variable "instance_count" {
description = "Specify the number of vm instances."
type = number
default = 1
}
variable "hostname" {
description = "VM instance name"
type = string
default = "linux-server"
}
variable "location" {
description = "Azure Location"
type = string
default = "westeurope"
}
variable "os" {
description = "Specify UbuntuServer, WindowsServer, RHEL, openSUSE-Leap, CentOS, Debian, CoreOS or SLES"
type = string
default = "UbuntuServer"
}
variable "vm_size" {
description = "VM size"
type = string
default = "Standard_B1s"
}
variable "ssh_key" {
description = "SSH Public Key"
type = string
}
output "vm_public_name" {
value = module.server.public_ip_dns_name
}
output "vm_public_ip" {
value = module.server.public_ip_address
}
deleteResource: true
providerRef:
name: azure
namespace: default
```
3. Edit the current azure-subnet (the current one doesn't work)
```
kubectl patch componentDefinitions azure-subnet -n vela-system --type merge -p '{"spec":{"schematic":{"terraform": {"configuration": "https://github.com/alisson276/terraform-modules.git"}}}}'
```
4. Create a file called first-vm.yaml:
```
apiVersion: core.oam.dev/v1beta1
kind: Application
metadata:
name: vm-1
spec:
components:
- name: example-rg
type: azure-resource-group
properties:
location: westeurope
name: example1
writeConnectionSecretToRef:
name: example-rg
- name: example-virtual-network
type: azure-subnet
properties:
vnet_name: vnet-1
create_network: true
resource_group_name: example1
subnet_prefixes:
- 10.0.0.0/24
subnet_names:
- subnet-1
writeConnectionSecretToRef:
name: example-vnet
- name: vm-1-1
type: azure-virtual-machine
properties:
instance_count: 1
subnet_name: subnet-1
vnet_name: vnet-1
subnet_rg_name: example1
resource_group: example
hostname: kubevela
writeConnectionSecretToRef:
name: vm1-conn
```
5. Apply it:
```
vela up -f first-vm.yaml
```
6. Wait untill all resource are healthy
7. vela delete vm-1
**Expected behavior**
<!--
A clear and concise description of what you expected to happen.
-->
As you can see in the components, the second component is a subnet under example1 resource group (the first component). The problem is: the virtual machine is in another resource group (example) and it has a NIC associate to it, which is linked to the subnet. So if you try to delete the first component created first, the resource group, it fails, because the vnet (the second resource) is under it and can't be deleted because the third resource (the vm) wasn't destroyed yet.
The idea is to have a way to refer it as `depends_on` (saying that the vm depends on the vnet and the vnet depends on the resource group) or simply delete them in the reverse order that they were created.
**Screenshots**
<!--
If applicable, add screenshots to help explain your problem.
-->
**KubeVela Version**
<!--
Describe your KubeVela controller or CLI version information.
-->
CLI Version: master
Core Version: 1.3.6
GitRevision: git-c8264b8c
GolangVersion: go1.18.2
**Cluster information**
<!--
Describe your kubernetes cluster information.
- Kubernetes Version [e.g. 1.16.9]
-->
Kubernetes version: 1.22.9
**Additional context**
<!--
Add any other context about the problem here.
-->
|
non_test
|
fail to delete applications due lack of dependency between components when using terraform describe the bug a clear and concise description of what the bug is when deleting a bunch of components in the same application using the terraform addon the controller decides to delete them in order of creation however depending of the provider in my case is azure it requires some order delete all resources under a resource group and after this deleting the resource group itself to reproduce steps to reproduce the behavior the yaml files of component trait i used the yaml file of application i applied other operations i have done see error configure the terraform and terraform azure addons vela addon install terraform vela addon install terraform azure vela provider add terraform azure arm client id arm client id arm client secret arm client secret arm subscription id arm subscription id arm tenant id arm tenant id name azure n vela system create a virtual machine componentdefinition apiversion core oam dev kind componentdefinition metadata annotations definition oam dev description terraform configuration for azure cloud virtual machine object labels type terraform name azure virtual machine namespace vela system spec workload definition apiversion terraform core oam dev kind configuration type configurations terraform core oam dev schematic terraform configuration provider azurerm features data azurerm subnet vnet name var subnet name virtual network name var vnet name resource group name var subnet rg name module server source azure compute azurerm resource group name var resource group nb instances var instance count vm os simple var os public ip dns vnet subnet id data azurerm subnet vnet id location var location vm hostname var hostname vm size var vm size ssh key ssh key values variable resource group description resource group name type string default example variable subnet name description subnet name type string variable vnet name description virtual network name type string variable subnet rg name description resource group of the subnet type string variable instance count description specify the number of vm instances type number default variable hostname description vm instance name type string default linux server variable location description azure location type string default westeurope variable os description specify ubuntuserver windowsserver rhel opensuse leap centos debian coreos or sles type string default ubuntuserver variable vm size description vm size type string default standard variable ssh key description ssh public key type string output vm public name value module server public ip dns name output vm public ip value module server public ip address deleteresource true providerref name azure namespace default edit the current azure subnet the current one doesn t work kubectl patch componentdefinitions azure subnet n vela system type merge p spec schematic terraform configuration create a file called first vm yaml apiversion core oam dev kind application metadata name vm spec components name example rg type azure resource group properties location westeurope name writeconnectionsecrettoref name example rg name example virtual network type azure subnet properties vnet name vnet create network true resource group name subnet prefixes subnet names subnet writeconnectionsecrettoref name example vnet name vm type azure virtual machine properties instance count subnet name subnet vnet name vnet subnet rg name resource group example hostname kubevela writeconnectionsecrettoref name conn apply it vela up f first vm yaml wait untill all resource are healthy vela delete vm expected behavior a clear and concise description of what you expected to happen as you can see in the components the second component is a subnet under resource group the first component the problem is the virtual machine is in another resource group example and it has a nic associate to it which is linked to the subnet so if you try to delete the first component created first the resource group it fails because the vnet the second resource is under it and can t be deleted because the third resource the vm wasn t destroyed yet the idea is to have a way to refer it as depends on saying that the vm depends on the vnet and the vnet depends on the resource group or simply delete them in the reverse order that they were created screenshots if applicable add screenshots to help explain your problem kubevela version describe your kubevela controller or cli version information cli version master core version gitrevision git golangversion cluster information describe your kubernetes cluster information kubernetes version kubernetes version additional context add any other context about the problem here
| 0
|
74,625
| 7,434,310,925
|
IssuesEvent
|
2018-03-26 10:34:02
|
Kademi/kademi-dev
|
https://api.github.com/repos/Kademi/kademi-dev
|
closed
|
Bulk find and select locations for organisations
|
Ready to Test - Dev
|
On the organisations map page (/organisations/map) do the following:
- show the number of organisations with locations (ie lattitude and longitude)
- show the number of organisations without locations
- if orgs without locations > 0 show a button to find and set them
- when pressed, for each organisation do a google maps search for the provided address (same as is currently done on the org details page), and if a matching location is found set its lat/lng as the organisation coords
Also, ensure orgs with no coords are not included in map search responses
|
1.0
|
Bulk find and select locations for organisations - On the organisations map page (/organisations/map) do the following:
- show the number of organisations with locations (ie lattitude and longitude)
- show the number of organisations without locations
- if orgs without locations > 0 show a button to find and set them
- when pressed, for each organisation do a google maps search for the provided address (same as is currently done on the org details page), and if a matching location is found set its lat/lng as the organisation coords
Also, ensure orgs with no coords are not included in map search responses
|
test
|
bulk find and select locations for organisations on the organisations map page organisations map do the following show the number of organisations with locations ie lattitude and longitude show the number of organisations without locations if orgs without locations show a button to find and set them when pressed for each organisation do a google maps search for the provided address same as is currently done on the org details page and if a matching location is found set its lat lng as the organisation coords also ensure orgs with no coords are not included in map search responses
| 1
|
226,718
| 18,043,955,427
|
IssuesEvent
|
2021-09-18 14:55:43
|
logicmoo/logicmoo_workspace
|
https://api.github.com/repos/logicmoo/logicmoo_workspace
|
opened
|
logicmoo.base.examples.fol.SANITY_EXISTS_03 JUnit
|
Test_9999 logicmoo.base.examples.fol unit_test SANITY_EXISTS_03
|
(cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif sanity_exists_03.pfc.pl)
GH_MASTER_ISSUE_FINFO=
ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ASANITY_EXISTS_03
GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/813ec17487381a026b83350c360d0c79a9e2d0ae
https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/
This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/65/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/
GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/813ec17487381a026b83350c360d0c79a9e2d0ae
https://github.com/logicmoo/logicmoo_workspace/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl
```
%
running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl'),
%~ this_test_might_need( :-( use_module( library(logicmoo_plarkc))))
:- module(t123).
:- dynamic(t123:ttExpressionType/1).
% :- process_this_script.
% :- process_this_script.
:- statistics.
/*~
% Started at Sat Sep 18 07:55:40 2021
% 0.875 seconds cpu time for 412,769 inferences
% 940,176 atoms, 31,018 functors, 29,482 predicates, 718 modules, 14,818,012 VM-codes
%
% Limit Allocated In use
% Local stack: - 52 Kb 4,216 b
% Global stack: - 64 Kb 49 Kb
% Trail stack: - 34 Kb 520 b
% Total: 1,024 Mb 150 Kb 53 Kb
%
% 5 garbage collections gained 305,208 bytes in 0.000 seconds.
% 1 atom garbage collections gained 1,137 atoms in 0.015 seconds.
% 4 clause garbage collections gained 118 clauses in 0.000 seconds.
% Stack shifts: 1 local, 0 global, 0 trail in 0.000 seconds
% 3 threads, 0 finished threads used 0.000 seconds
~*/
subtest_assert(I):-kif_assert(I).
subtest([subtest_assert(tAnimal(joe)),
mpred_test(isa(_,tHeart))]).
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl:18
%~ debugm( common_logic_loader,
%~ show_success( common_logic_loader,
%~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tAnimal(joe)),
%~ mpred_test( tHeart(Heart4))])))))
%~ kifi = subtest( [ subtest_assert( tAnimal(joe)),
%~ mpred_test( tHeart(Heart4))]).
%~ kifm = nesc( subtest( [ subtest_assert( tAnimal(joe)),
%~ mpred_test( tHeart(Heart4))])).
/*~
=======================================================
subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(tHeart('$VAR'('Heart4'))),[])))
============================================
?- kif_to_boxlog( subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))]) ).
% In English:
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ joe isa tAnimal isa subtest_assert ?Heart4 isa tHeart isa mpred_test isa subtest
%~
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ kifi=subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))])
%~ kifm=nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))]))
%~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(tHeart('$VAR'('Heart4'))),[]))))
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
% Results in the following 1 entailment(s):
nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))])).
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ it is necessarily true that joe isa tAnimal isa subtest_assert ?Heart4 isa tHeart isa mpred_test isa subtest
%~
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
nesc( subtest( [ subtest_assert( tAnimal(joe)),
mpred_test( tHeart(Heart4))])).
============================================
~*/
subtest([subtest_assert(tAnimal(joe)),
mpred_test(hasOrgan(joe,_))]).
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl:21
%~ debugm( common_logic_loader,
%~ show_success( common_logic_loader,
%~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tAnimal(joe)),
%~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))])))))
%~ kifi = subtest( [ subtest_assert( tAnimal(joe)),
%~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))]).
%~ kifm = nesc( subtest( [ subtest_assert( tAnimal(joe)),
%~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))])).
/*~
=======================================================
subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(hasOrgan(joe,'$VAR'('HasOrgan_Joe4'))),[])))
============================================
?- kif_to_boxlog( subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))]) ).
% In English:
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ joe isa tAnimal isa subtest_assert joe hasOrgan ?HasOrgan_Joe4 isa mpred_test isa subtest
%~
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ kifi=subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))])
%~ kifm=nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))]))
%~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(hasOrgan(joe,'$VAR'('HasOrgan_Joe4'))),[]))))
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
% Results in the following 1 entailment(s):
nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))])).
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ it is necessarily true that joe isa tAnimal isa subtest_assert joe hasOrgan ?HasOrgan_Joe4 isa mpred_test isa subtest
%~
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
nesc( subtest( [ subtest_assert( tAnimal(joe)),
mpred_test( hasOrgan(joe,HasOrgan_Joe4))])).
============================================
~*/
subtest([subtest_assert(tHeart(_)),
mpred_test(~hasOrgan(jack,_))]).
%~ debugm( common_logic_loader,
%~ show_success( common_logic_loader,
%~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tHeart(Heart7)),
%~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))])))))
%~ kifi = subtest( [ subtest_assert( tHeart(Heart7)),
%~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))]).
%~ kifm = nesc( subtest( [ subtest_assert( tHeart(Heart7)),
%~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))])).
/*~
=======================================================
subtest('[|]'(subtest_assert(tHeart('$VAR'('Heart7'))),'[|]'(mpred_test(~(hasOrgan(jack,'$VAR'('HasOrgan_Jack5')))),[])))
============================================
?- kif_to_boxlog( subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))]) ).
% In English:
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ ?Heart7 isa tHeart isa subtest_assert jack hasOrgan ?HasOrgan_Jack5 isa ~ isa mpred_test isa subtest
%~
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ kifi=subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))])
%~ kifm=nesc(subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))]))
%~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tHeart('$VAR'('Heart7'))),'[|]'(mpred_test(not(hasOrgan(jack,'$VAR'('HasOrgan_Jack5')))),[]))))
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
% Results in the following 1 entailment(s):
nesc(subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))])).
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ it is necessarily true that ?Heart7 isa tHeart isa subtest_assert jack hasOrgan ?HasOrgan_Jack5 isa ~ isa mpred_test isa subtest
%~
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
nesc( subtest( [ subtest_assert( tHeart(Heart7)),
mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))])).
============================================
~*/
dbanner:- nl,nl,dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%'),nl,nl.
%~ warn( db_expand_0( clause(unknown,cuz),
%~ dbanner :-
%~ nl ,
%~ nl ,
%~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') ,
%~ nl ,
%~ nl,
%~ dbanner :-
%~ nl ,
%~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') ,
%~ nl))
%~ message_hook_type(error)
%~ message_hook(
%~ initialization_exception( must_sanity : e( must_sanity : e( error(
%~ permission_error(modify,static_procedure,common_logic_sanity:dbanner/0),
%~ context(system:assertz/1,Context_Kw)),
%~ baseKB : assert_to_mu( dbanner :-
%~ nl ,
%~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') ,
%~ nl)),
%~ baseKB : ain_fast(
%~ baseKB : dbanner :-
%~ nl ,
%~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') ,
%~ nl,
%~ ( mfl4(BaseKB,baseKB,'/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl',27) ,
%~ ax)))),
%~ error,
%~ [ 'Prolog initialisation failed:',
%~ nl,
%~ 'Unknown message: ~p' - [ must_sanity : e( must_sanity : e( error(
%~ permission_error(modify,static_procedure,common_logic_sanity:dbanner/0),
%~ context(system:assertz/1,Context_Kw)),
%~ baseKB : assert_to_mu( dbanner :-
%~ nl ,
%~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') ,
%~ nl)),
%~ baseKB : ain_fast(
%~ baseKB : dbanner :-
%~ nl ,
%~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') ,
%~ nl,
%~ ( mfl4(BaseKB,baseKB,'/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl',27) ,
%~ ax)))]])
%~ unused(save_junit_results)
```
totalTime=1
ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ASANITY_EXISTS_03
GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/813ec17487381a026b83350c360d0c79a9e2d0ae
https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/
This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/65/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/
GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/813ec17487381a026b83350c360d0c79a9e2d0ae
https://github.com/logicmoo/logicmoo_workspace/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl
FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k sanity_exists_03.pfc.pl (returned 1)
|
2.0
|
logicmoo.base.examples.fol.SANITY_EXISTS_03 JUnit - (cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif sanity_exists_03.pfc.pl)
GH_MASTER_ISSUE_FINFO=
ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ASANITY_EXISTS_03
GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/813ec17487381a026b83350c360d0c79a9e2d0ae
https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/
This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/65/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/
GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/813ec17487381a026b83350c360d0c79a9e2d0ae
https://github.com/logicmoo/logicmoo_workspace/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl
```
%
running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl'),
%~ this_test_might_need( :-( use_module( library(logicmoo_plarkc))))
:- module(t123).
:- dynamic(t123:ttExpressionType/1).
% :- process_this_script.
% :- process_this_script.
:- statistics.
/*~
% Started at Sat Sep 18 07:55:40 2021
% 0.875 seconds cpu time for 412,769 inferences
% 940,176 atoms, 31,018 functors, 29,482 predicates, 718 modules, 14,818,012 VM-codes
%
% Limit Allocated In use
% Local stack: - 52 Kb 4,216 b
% Global stack: - 64 Kb 49 Kb
% Trail stack: - 34 Kb 520 b
% Total: 1,024 Mb 150 Kb 53 Kb
%
% 5 garbage collections gained 305,208 bytes in 0.000 seconds.
% 1 atom garbage collections gained 1,137 atoms in 0.015 seconds.
% 4 clause garbage collections gained 118 clauses in 0.000 seconds.
% Stack shifts: 1 local, 0 global, 0 trail in 0.000 seconds
% 3 threads, 0 finished threads used 0.000 seconds
~*/
subtest_assert(I):-kif_assert(I).
subtest([subtest_assert(tAnimal(joe)),
mpred_test(isa(_,tHeart))]).
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl:18
%~ debugm( common_logic_loader,
%~ show_success( common_logic_loader,
%~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tAnimal(joe)),
%~ mpred_test( tHeart(Heart4))])))))
%~ kifi = subtest( [ subtest_assert( tAnimal(joe)),
%~ mpred_test( tHeart(Heart4))]).
%~ kifm = nesc( subtest( [ subtest_assert( tAnimal(joe)),
%~ mpred_test( tHeart(Heart4))])).
/*~
=======================================================
subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(tHeart('$VAR'('Heart4'))),[])))
============================================
?- kif_to_boxlog( subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))]) ).
% In English:
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ joe isa tAnimal isa subtest_assert ?Heart4 isa tHeart isa mpred_test isa subtest
%~
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ kifi=subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))])
%~ kifm=nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))]))
%~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(tHeart('$VAR'('Heart4'))),[]))))
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
% Results in the following 1 entailment(s):
nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))])).
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ it is necessarily true that joe isa tAnimal isa subtest_assert ?Heart4 isa tHeart isa mpred_test isa subtest
%~
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
nesc( subtest( [ subtest_assert( tAnimal(joe)),
mpred_test( tHeart(Heart4))])).
============================================
~*/
subtest([subtest_assert(tAnimal(joe)),
mpred_test(hasOrgan(joe,_))]).
%~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl:21
%~ debugm( common_logic_loader,
%~ show_success( common_logic_loader,
%~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tAnimal(joe)),
%~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))])))))
%~ kifi = subtest( [ subtest_assert( tAnimal(joe)),
%~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))]).
%~ kifm = nesc( subtest( [ subtest_assert( tAnimal(joe)),
%~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))])).
/*~
=======================================================
subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(hasOrgan(joe,'$VAR'('HasOrgan_Joe4'))),[])))
============================================
?- kif_to_boxlog( subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))]) ).
% In English:
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ joe isa tAnimal isa subtest_assert joe hasOrgan ?HasOrgan_Joe4 isa mpred_test isa subtest
%~
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ kifi=subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))])
%~ kifm=nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))]))
%~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(hasOrgan(joe,'$VAR'('HasOrgan_Joe4'))),[]))))
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
% Results in the following 1 entailment(s):
nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))])).
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ it is necessarily true that joe isa tAnimal isa subtest_assert joe hasOrgan ?HasOrgan_Joe4 isa mpred_test isa subtest
%~
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
nesc( subtest( [ subtest_assert( tAnimal(joe)),
mpred_test( hasOrgan(joe,HasOrgan_Joe4))])).
============================================
~*/
subtest([subtest_assert(tHeart(_)),
mpred_test(~hasOrgan(jack,_))]).
%~ debugm( common_logic_loader,
%~ show_success( common_logic_loader,
%~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tHeart(Heart7)),
%~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))])))))
%~ kifi = subtest( [ subtest_assert( tHeart(Heart7)),
%~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))]).
%~ kifm = nesc( subtest( [ subtest_assert( tHeart(Heart7)),
%~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))])).
/*~
=======================================================
subtest('[|]'(subtest_assert(tHeart('$VAR'('Heart7'))),'[|]'(mpred_test(~(hasOrgan(jack,'$VAR'('HasOrgan_Jack5')))),[])))
============================================
?- kif_to_boxlog( subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))]) ).
% In English:
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ ?Heart7 isa tHeart isa subtest_assert jack hasOrgan ?HasOrgan_Jack5 isa ~ isa mpred_test isa subtest
%~
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ kifi=subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))])
%~ kifm=nesc(subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))]))
%~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tHeart('$VAR'('Heart7'))),'[|]'(mpred_test(not(hasOrgan(jack,'$VAR'('HasOrgan_Jack5')))),[]))))
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
% Results in the following 1 entailment(s):
nesc(subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))])).
%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
%~ it is necessarily true that ?Heart7 isa tHeart isa subtest_assert jack hasOrgan ?HasOrgan_Jack5 isa ~ isa mpred_test isa subtest
%~
%~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%
nesc( subtest( [ subtest_assert( tHeart(Heart7)),
mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))])).
============================================
~*/
dbanner:- nl,nl,dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%'),nl,nl.
%~ warn( db_expand_0( clause(unknown,cuz),
%~ dbanner :-
%~ nl ,
%~ nl ,
%~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') ,
%~ nl ,
%~ nl,
%~ dbanner :-
%~ nl ,
%~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') ,
%~ nl))
%~ message_hook_type(error)
%~ message_hook(
%~ initialization_exception( must_sanity : e( must_sanity : e( error(
%~ permission_error(modify,static_procedure,common_logic_sanity:dbanner/0),
%~ context(system:assertz/1,Context_Kw)),
%~ baseKB : assert_to_mu( dbanner :-
%~ nl ,
%~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') ,
%~ nl)),
%~ baseKB : ain_fast(
%~ baseKB : dbanner :-
%~ nl ,
%~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') ,
%~ nl,
%~ ( mfl4(BaseKB,baseKB,'/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl',27) ,
%~ ax)))),
%~ error,
%~ [ 'Prolog initialisation failed:',
%~ nl,
%~ 'Unknown message: ~p' - [ must_sanity : e( must_sanity : e( error(
%~ permission_error(modify,static_procedure,common_logic_sanity:dbanner/0),
%~ context(system:assertz/1,Context_Kw)),
%~ baseKB : assert_to_mu( dbanner :-
%~ nl ,
%~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') ,
%~ nl)),
%~ baseKB : ain_fast(
%~ baseKB : dbanner :-
%~ nl ,
%~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') ,
%~ nl,
%~ ( mfl4(BaseKB,baseKB,'/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl',27) ,
%~ ax)))]])
%~ unused(save_junit_results)
```
totalTime=1
ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ASANITY_EXISTS_03
GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/813ec17487381a026b83350c360d0c79a9e2d0ae
https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/
This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/65/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/
GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/813ec17487381a026b83350c360d0c79a9e2d0ae
https://github.com/logicmoo/logicmoo_workspace/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl
FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k sanity_exists_03.pfc.pl (returned 1)
|
test
|
logicmoo base examples fol sanity exists junit cd var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol timeout foreground preserve status s sigkill k lmoo clif sanity exists pfc pl gh master issue finfo issue search gitlab latest this build github running var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol sanity exists pfc pl this test might need use module library logicmoo plarkc module dynamic ttexpressiontype process this script process this script statistics started at sat sep seconds cpu time for inferences atoms functors predicates modules vm codes limit allocated in use local stack kb b global stack kb kb trail stack kb b total mb kb kb garbage collections gained bytes in seconds atom garbage collections gained atoms in seconds clause garbage collections gained clauses in seconds stack shifts local global trail in seconds threads finished threads used seconds subtest assert i kif assert i subtest subtest assert tanimal joe mpred test isa theart var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol sanity exists pfc pl debugm common logic loader show success common logic loader common logic loader ain clif subtest subtest assert tanimal joe mpred test theart kifi subtest subtest assert tanimal joe mpred test theart kifm nesc subtest subtest assert tanimal joe mpred test theart subtest subtest assert tanimal joe mpred test theart var kif to boxlog subtest in english joe isa tanimal isa subtest assert isa theart isa mpred test isa subtest kifi subtest kifm nesc subtest kif to boxlog necessary subtest subtest assert tanimal joe mpred test theart var results in the following entailment s nesc subtest it is necessarily true that joe isa tanimal isa subtest assert isa theart isa mpred test isa subtest nesc subtest subtest assert tanimal joe mpred test theart subtest subtest assert tanimal joe mpred test hasorgan joe var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol sanity exists pfc pl debugm common logic loader show success common logic loader common logic loader ain clif subtest subtest assert tanimal joe mpred test hasorgan joe hasorgan kifi subtest subtest assert tanimal joe mpred test hasorgan joe hasorgan kifm nesc subtest subtest assert tanimal joe mpred test hasorgan joe hasorgan subtest subtest assert tanimal joe mpred test hasorgan joe var hasorgan kif to boxlog subtest in english joe isa tanimal isa subtest assert joe hasorgan hasorgan isa mpred test isa subtest kifi subtest kifm nesc subtest kif to boxlog necessary subtest subtest assert tanimal joe mpred test hasorgan joe var hasorgan results in the following entailment s nesc subtest it is necessarily true that joe isa tanimal isa subtest assert joe hasorgan hasorgan isa mpred test isa subtest nesc subtest subtest assert tanimal joe mpred test hasorgan joe hasorgan subtest subtest assert theart mpred test hasorgan jack debugm common logic loader show success common logic loader common logic loader ain clif subtest subtest assert theart mpred test hasorgan jack hasorgan kifi subtest subtest assert theart mpred test hasorgan jack hasorgan kifm nesc subtest subtest assert theart mpred test hasorgan jack hasorgan subtest subtest assert theart var mpred test hasorgan jack var hasorgan kif to boxlog subtest in english isa theart isa subtest assert jack hasorgan hasorgan isa isa mpred test isa subtest kifi subtest kifm nesc subtest kif to boxlog necessary subtest subtest assert theart var mpred test not hasorgan jack var hasorgan results in the following entailment s nesc subtest it is necessarily true that isa theart isa subtest assert jack hasorgan hasorgan isa isa mpred test isa subtest nesc subtest subtest assert theart mpred test hasorgan jack hasorgan dbanner nl nl dmsg nl nl warn db expand clause unknown cuz dbanner nl nl dmsg nl nl dbanner nl dmsg nl message hook type error message hook initialization exception must sanity e must sanity e error permission error modify static procedure common logic sanity dbanner context system assertz context kw basekb assert to mu dbanner nl dmsg nl basekb ain fast basekb dbanner nl dmsg nl basekb basekb var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol sanity exists pfc pl ax error prolog initialisation failed nl unknown message p must sanity e must sanity e error permission error modify static procedure common logic sanity dbanner context system assertz context kw basekb assert to mu dbanner nl dmsg nl basekb ain fast basekb dbanner nl dmsg nl basekb basekb var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol sanity exists pfc pl ax unused save junit results totaltime issue search gitlab latest this build github failed var lib jenkins workspace logicmoo workspace bin lmoo junit minor k sanity exists pfc pl returned
| 1
|
50,615
| 13,187,624,433
|
IssuesEvent
|
2020-08-13 04:01:42
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
[photospline] photospline-pybindings fails during CMake's configure step if astropy is not present (Trac #1051)
|
Migrated from Trac combo simulation defect
|
I don't have astropy. So get this error for the current trunk when cmake configures itself:
Traceback (most recent call last):
File "/home/hdembinski/Code/icecube/icesim/trunk/photospline/resources/test/find_pyfits.py", line 6, in <module>
import astropy.io.fits as pyfits
ImportError: No module named astropy.io.fits
Maybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1051">https://code.icecube.wisc.edu/ticket/1051</a>, reported by hdembinski and owned by </em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-07-15T07:42:17",
"description": "I don't have astropy. So get this error for the current trunk when cmake configures itself:\n\nTraceback (most recent call last):\n File \"/home/hdembinski/Code/icecube/icesim/trunk/photospline/resources/test/find_pyfits.py\", line 6, in <module>\n import astropy.io.fits as pyfits\nImportError: No module named astropy.io.fits\n\nMaybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user.",
"reporter": "hdembinski",
"cc": "",
"resolution": "fixed",
"_ts": "1436946137322111",
"component": "combo simulation",
"summary": "[photospline] photospline-pybindings fails during CMake's configure step if astropy is not present",
"priority": "critical",
"keywords": "photospline astropy",
"time": "2015-07-15T00:12:23",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[photospline] photospline-pybindings fails during CMake's configure step if astropy is not present (Trac #1051) - I don't have astropy. So get this error for the current trunk when cmake configures itself:
Traceback (most recent call last):
File "/home/hdembinski/Code/icecube/icesim/trunk/photospline/resources/test/find_pyfits.py", line 6, in <module>
import astropy.io.fits as pyfits
ImportError: No module named astropy.io.fits
Maybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1051">https://code.icecube.wisc.edu/ticket/1051</a>, reported by hdembinski and owned by </em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-07-15T07:42:17",
"description": "I don't have astropy. So get this error for the current trunk when cmake configures itself:\n\nTraceback (most recent call last):\n File \"/home/hdembinski/Code/icecube/icesim/trunk/photospline/resources/test/find_pyfits.py\", line 6, in <module>\n import astropy.io.fits as pyfits\nImportError: No module named astropy.io.fits\n\nMaybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user.",
"reporter": "hdembinski",
"cc": "",
"resolution": "fixed",
"_ts": "1436946137322111",
"component": "combo simulation",
"summary": "[photospline] photospline-pybindings fails during CMake's configure step if astropy is not present",
"priority": "critical",
"keywords": "photospline astropy",
"time": "2015-07-15T00:12:23",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
|
non_test
|
photospline pybindings fails during cmake s configure step if astropy is not present trac i don t have astropy so get this error for the current trunk when cmake configures itself traceback most recent call last file home hdembinski code icecube icesim trunk photospline resources test find pyfits py line in import astropy io fits as pyfits importerror no module named astropy io fits maybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user migrated from json status closed changetime description i don t have astropy so get this error for the current trunk when cmake configures itself n ntraceback most recent call last n file home hdembinski code icecube icesim trunk photospline resources test find pyfits py line in n import astropy io fits as pyfits nimporterror no module named astropy io fits n nmaybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user reporter hdembinski cc resolution fixed ts component combo simulation summary photospline pybindings fails during cmake s configure step if astropy is not present priority critical keywords photospline astropy time milestone owner type defect
| 0
|
28,682
| 4,426,279,275
|
IssuesEvent
|
2016-08-16 17:51:43
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
closed
|
Update interop and stress tests to GA API
|
interop python security test
|
The only code in the repository that should call the Beta API should be the tests of the Beta API.
This is currently blocked on [issue 6799](https://github.com/grpc/grpc/issues/6799) because the [`oauth2_auth_token`](https://github.com/grpc/grpc/blob/master/doc/interop-test-descriptions.md#oauth2_auth_token), [`jwt_token_creds`](https://github.com/grpc/grpc/blob/master/doc/interop-test-descriptions.md#jwt_token_creds), and [`per_rpc_creds`](https://github.com/grpc/grpc/blob/master/doc/interop-test-descriptions.md#per_rpc_creds) interop tests are implemented in terms of the [`oauth2client`](https://github.com/google/oauth2client) library and the stress tests are implemented in terms of the interop tests.
|
1.0
|
Update interop and stress tests to GA API - The only code in the repository that should call the Beta API should be the tests of the Beta API.
This is currently blocked on [issue 6799](https://github.com/grpc/grpc/issues/6799) because the [`oauth2_auth_token`](https://github.com/grpc/grpc/blob/master/doc/interop-test-descriptions.md#oauth2_auth_token), [`jwt_token_creds`](https://github.com/grpc/grpc/blob/master/doc/interop-test-descriptions.md#jwt_token_creds), and [`per_rpc_creds`](https://github.com/grpc/grpc/blob/master/doc/interop-test-descriptions.md#per_rpc_creds) interop tests are implemented in terms of the [`oauth2client`](https://github.com/google/oauth2client) library and the stress tests are implemented in terms of the interop tests.
|
test
|
update interop and stress tests to ga api the only code in the repository that should call the beta api should be the tests of the beta api this is currently blocked on because the and interop tests are implemented in terms of the library and the stress tests are implemented in terms of the interop tests
| 1
|
8,446
| 11,614,671,832
|
IssuesEvent
|
2020-02-26 13:00:24
|
scikit-learn/scikit-learn
|
https://api.github.com/repos/scikit-learn/scikit-learn
|
closed
|
sklearn.preprocessing.StandardScaler gets NaN variance when partial_fit with sparse data
|
Bug module:preprocessing
|
#### Describe the bug
When I feed a specific dataset (which is sparse) to sklearn.preprocessing.StandardScaler.partial_fit in a specific order, I get variance which is NaN although data does **NOT** contains any NaNs and is very small.
When I convert the sparse arrays to dense, it works. When I change the order to feed the data, it works too.
#### Steps/Code to Reproduce
Please work with the data I attached. [sparse_data.tar.gz](https://github.com/scikit-learn/scikit-learn/files/4208684/sparse_data.tar.gz)
```python
import scipy.sparse as sp
from sklearn import preprocessing
s0 = sp.load_npz('0.npz')
s1 = sp.load_npz('1.npz')
# Buggy behavior
ss0 = preprocessing.StandardScaler(with_mean=False)
ss0.partial_fit(s0)
print(ss0.var_)
ss0.partial_fit(s1)
print(ss0.var_) # => gets NaN
# When use dence array, it works
ss1 = preprocessing.StandardScaler(with_mean=False)
ss1.partial_fit(s0.toarray())
print(ss1.var_)
ss1.partial_fit(s1.toarray())
print(ss1.var_)
# When change the order of data, it works
ss2 = preprocessing.StandardScaler(with_mean=False)
ss2.partial_fit(s1)
print(ss2.var_)
ss2.partial_fit(s0)
print(ss2.var_)
```
EDIT: Fix sample code around ss2
#### Expected Results
```python
ss0.var_ # => [0.15896542]
ss1.var_ # => [0.15896542]
ss2.var_ # => [0.15896542]
```
#### Actual Results
```python
ss0.var_ # => [nan]
ss1.var_ # => [0.15896542]
ss2.var_ # => [0.15896542]
```
#### Versions
I confirmed this issue in two different environments.
```
System:
python: 3.7.3 (default, Apr 22 2019, 02:40:09) [Clang 10.0.1 (clang-1001.0.46.4)]
executable: /usr/local/var/pyenv/versions/3.7.3/bin/python3
machine: Darwin-19.3.0-x86_64-i386-64bit
Python dependencies:
pip: 20.0.2
setuptools: 40.8.0
sklearn: 0.22
numpy: 1.18.0
scipy: 1.4.1
Cython: None
pandas: 0.25.3
matplotlib: 3.1.2
joblib: 0.14.1
Built with OpenMP: True
```
```
System:
python: 3.7.6 (default, Feb 14 2020, 16:41:52) [GCC 8.3.1 20190507 (Red Hat 8.3.1-4)]
executable: /home/***/ws/siml/.venv/bin/python3
machine: Linux-4.18.0-147.5.1.el8_1.x86_64-x86_64-with-centos-8.1.1911-Core
Python dependencies:
pip: 19.2.3
setuptools: 41.2.0
sklearn: 0.22.1
numpy: 1.18.1
scipy: 1.4.1
Cython: None
pandas: 0.25.3
matplotlib: 3.1.3
joblib: 0.14.1
Built with OpenMP: True
```
|
1.0
|
sklearn.preprocessing.StandardScaler gets NaN variance when partial_fit with sparse data - #### Describe the bug
When I feed a specific dataset (which is sparse) to sklearn.preprocessing.StandardScaler.partial_fit in a specific order, I get variance which is NaN although data does **NOT** contains any NaNs and is very small.
When I convert the sparse arrays to dense, it works. When I change the order to feed the data, it works too.
#### Steps/Code to Reproduce
Please work with the data I attached. [sparse_data.tar.gz](https://github.com/scikit-learn/scikit-learn/files/4208684/sparse_data.tar.gz)
```python
import scipy.sparse as sp
from sklearn import preprocessing
s0 = sp.load_npz('0.npz')
s1 = sp.load_npz('1.npz')
# Buggy behavior
ss0 = preprocessing.StandardScaler(with_mean=False)
ss0.partial_fit(s0)
print(ss0.var_)
ss0.partial_fit(s1)
print(ss0.var_) # => gets NaN
# When use dence array, it works
ss1 = preprocessing.StandardScaler(with_mean=False)
ss1.partial_fit(s0.toarray())
print(ss1.var_)
ss1.partial_fit(s1.toarray())
print(ss1.var_)
# When change the order of data, it works
ss2 = preprocessing.StandardScaler(with_mean=False)
ss2.partial_fit(s1)
print(ss2.var_)
ss2.partial_fit(s0)
print(ss2.var_)
```
EDIT: Fix sample code around ss2
#### Expected Results
```python
ss0.var_ # => [0.15896542]
ss1.var_ # => [0.15896542]
ss2.var_ # => [0.15896542]
```
#### Actual Results
```python
ss0.var_ # => [nan]
ss1.var_ # => [0.15896542]
ss2.var_ # => [0.15896542]
```
#### Versions
I confirmed this issue in two different environments.
```
System:
python: 3.7.3 (default, Apr 22 2019, 02:40:09) [Clang 10.0.1 (clang-1001.0.46.4)]
executable: /usr/local/var/pyenv/versions/3.7.3/bin/python3
machine: Darwin-19.3.0-x86_64-i386-64bit
Python dependencies:
pip: 20.0.2
setuptools: 40.8.0
sklearn: 0.22
numpy: 1.18.0
scipy: 1.4.1
Cython: None
pandas: 0.25.3
matplotlib: 3.1.2
joblib: 0.14.1
Built with OpenMP: True
```
```
System:
python: 3.7.6 (default, Feb 14 2020, 16:41:52) [GCC 8.3.1 20190507 (Red Hat 8.3.1-4)]
executable: /home/***/ws/siml/.venv/bin/python3
machine: Linux-4.18.0-147.5.1.el8_1.x86_64-x86_64-with-centos-8.1.1911-Core
Python dependencies:
pip: 19.2.3
setuptools: 41.2.0
sklearn: 0.22.1
numpy: 1.18.1
scipy: 1.4.1
Cython: None
pandas: 0.25.3
matplotlib: 3.1.3
joblib: 0.14.1
Built with OpenMP: True
```
|
non_test
|
sklearn preprocessing standardscaler gets nan variance when partial fit with sparse data describe the bug when i feed a specific dataset which is sparse to sklearn preprocessing standardscaler partial fit in a specific order i get variance which is nan although data does not contains any nans and is very small when i convert the sparse arrays to dense it works when i change the order to feed the data it works too steps code to reproduce please work with the data i attached python import scipy sparse as sp from sklearn import preprocessing sp load npz npz sp load npz npz buggy behavior preprocessing standardscaler with mean false partial fit print var partial fit print var gets nan when use dence array it works preprocessing standardscaler with mean false partial fit toarray print var partial fit toarray print var when change the order of data it works preprocessing standardscaler with mean false partial fit print var partial fit print var edit fix sample code around expected results python var var var actual results python var var var versions i confirmed this issue in two different environments system python default apr executable usr local var pyenv versions bin machine darwin python dependencies pip setuptools sklearn numpy scipy cython none pandas matplotlib joblib built with openmp true system python default feb executable home ws siml venv bin machine linux with centos core python dependencies pip setuptools sklearn numpy scipy cython none pandas matplotlib joblib built with openmp true
| 0
|
331,698
| 29,047,568,487
|
IssuesEvent
|
2023-05-13 19:15:26
|
T1rax/movies_ugc_service
|
https://api.github.com/repos/T1rax/movies_ugc_service
|
opened
|
Написать сквозные тесты e2e
|
test
|
Написать 1-2 сквозных теста всей системы. Можно взять за основу уже имебщийся e2e
|
1.0
|
Написать сквозные тесты e2e - Написать 1-2 сквозных теста всей системы. Можно взять за основу уже имебщийся e2e
|
test
|
написать сквозные тесты написать сквозных теста всей системы можно взять за основу уже имебщийся
| 1
|
27,331
| 11,471,757,098
|
IssuesEvent
|
2020-02-09 13:20:18
|
foretold-app/foretold
|
https://api.github.com/repos/foretold-app/foretold
|
opened
|
Figure out why FeedItem data was lost
|
security 🕷 bug
|
We seem to have lost some FeedItem data. Thankfully this data isn't particularly important to keep, but loosing data is quite bad.
If it would be relatively easy, it would be nice to recover it from the backups.
|
True
|
Figure out why FeedItem data was lost - We seem to have lost some FeedItem data. Thankfully this data isn't particularly important to keep, but loosing data is quite bad.
If it would be relatively easy, it would be nice to recover it from the backups.
|
non_test
|
figure out why feeditem data was lost we seem to have lost some feeditem data thankfully this data isn t particularly important to keep but loosing data is quite bad if it would be relatively easy it would be nice to recover it from the backups
| 0
|
163,545
| 12,734,481,878
|
IssuesEvent
|
2020-06-25 13:58:20
|
NLnetLabs/krill
|
https://api.github.com/repos/NLnetLabs/krill
|
closed
|
Add support for rsync URI in the TAL instead of HTTPS for testing purposes
|
interop testing
|
### Summary
When using `use_ta = true` the generated `.tal` file contains an HTTPS URI to the `.cer` file.
Routinator is able to consume this file but OctoRPKI and Fort Validator fail unless the URI is changed to use the RSYNC protocol instead.
### Versions tested:
- Krill Docker image `nlnetlabs/krill:v0.2.0` (built from the v0.2.0 GitHub tag).
- OctoRPKI Docker image `cloudflare/octorpki` (aka `cloudflare/octorpki:latest` with digest [93131096a0a9](https://hub.docker.com/layers/cloudflare/octorpki/latest/images/sha256-93131096a0a97cb1065660b046f6003f264e79193f2e04124980cb4e1829f6b5), where `/octorpki -version` reports `OctoRPKI v1.1.4 (2019-08-06T16:51:07-0700)`.
- Fort Validator Debian package https://github.com/NICMx/FORT-validator/releases/download/v1.1.0/fort_1.1.0-1_amd64.deb installed in a Docker container using image `debian` (`Debian GNU/Linux 10 \n \l` according to `/etc/issue`), where `fort -V` reports `fort 1.1.0`.
### OctoRPKI
Error behaviour when the `.cer` URI inside the TAL file uses the HTTPS protocol:
```
octorpki | time="2019-10-30T20:57:29Z" level=info msg="Got open https://ximontest.do.nlnetlabs.nl/ta/ta.cer: no such file or directory but repository not yet synchronized"
```
This is not a very helpful error message as the URI can definitely be fetched from inside the Docker container by WGet,
After replacing the URI inside the TAL file with an `rsync://` URI and placing the `.cer` file at that location in the RSYNC repo, OctoRPKI was then able to fetch the ROAs from Krill:
```
octorpki | {"metadata":{"counts":2,"generated":1572476435,"valid":1572480035,"signature":"3046022100992e43910ae22e24ff252527680d21fdd749a2a12975b729466c2e59649ccfe6022100b0ea9598ed5920202aafd8a19db51878a25c3235434afcfda53dd375f42ff285","signatureDate":"3044022028a5d2aad362ed58839b188cc1de8acd9e9892739fe833ad544760482987e22002206b81de1751f504a3b0f9643c70ee0bdfa4934fe5f3db6dc411d62fb1bb2e37a2"},"roas":[{"prefix":"10.0.0.0/24","maxLength":24,"asn":"AS64496","ta":"KRILL"},{"prefix":"10.0.1.0/24","maxLength":24,"asn":"AS64496","ta":"KRILL"}]}
```
### Fort Validator
Error behaviour when the `.cer` URI inside the TAL file uses the HTTPS protocol:
```
ERR: /tmp/ta.tal: Global URI 'https://ximontest.do.nlnetlabs.nl/ta/ta.cer' does not begin with 'rsync://'.
```
After replacing the URI inside the TAL file with an `rsync://` URI and placing the `.cer` file at that location in the RSYNC repo, Fort Validator then produced a different error message:
```
ERR: rsync://ximontest.do.nlnetlabs.nl/repo/ta/0/1169977D368819F3557C6B2353ED047D0CB84192.cer: Certificate validation failed: invalid or inconsistent certificate extension
```
Ximon
|
1.0
|
Add support for rsync URI in the TAL instead of HTTPS for testing purposes - ### Summary
When using `use_ta = true` the generated `.tal` file contains an HTTPS URI to the `.cer` file.
Routinator is able to consume this file but OctoRPKI and Fort Validator fail unless the URI is changed to use the RSYNC protocol instead.
### Versions tested:
- Krill Docker image `nlnetlabs/krill:v0.2.0` (built from the v0.2.0 GitHub tag).
- OctoRPKI Docker image `cloudflare/octorpki` (aka `cloudflare/octorpki:latest` with digest [93131096a0a9](https://hub.docker.com/layers/cloudflare/octorpki/latest/images/sha256-93131096a0a97cb1065660b046f6003f264e79193f2e04124980cb4e1829f6b5), where `/octorpki -version` reports `OctoRPKI v1.1.4 (2019-08-06T16:51:07-0700)`.
- Fort Validator Debian package https://github.com/NICMx/FORT-validator/releases/download/v1.1.0/fort_1.1.0-1_amd64.deb installed in a Docker container using image `debian` (`Debian GNU/Linux 10 \n \l` according to `/etc/issue`), where `fort -V` reports `fort 1.1.0`.
### OctoRPKI
Error behaviour when the `.cer` URI inside the TAL file uses the HTTPS protocol:
```
octorpki | time="2019-10-30T20:57:29Z" level=info msg="Got open https://ximontest.do.nlnetlabs.nl/ta/ta.cer: no such file or directory but repository not yet synchronized"
```
This is not a very helpful error message as the URI can definitely be fetched from inside the Docker container by WGet,
After replacing the URI inside the TAL file with an `rsync://` URI and placing the `.cer` file at that location in the RSYNC repo, OctoRPKI was then able to fetch the ROAs from Krill:
```
octorpki | {"metadata":{"counts":2,"generated":1572476435,"valid":1572480035,"signature":"3046022100992e43910ae22e24ff252527680d21fdd749a2a12975b729466c2e59649ccfe6022100b0ea9598ed5920202aafd8a19db51878a25c3235434afcfda53dd375f42ff285","signatureDate":"3044022028a5d2aad362ed58839b188cc1de8acd9e9892739fe833ad544760482987e22002206b81de1751f504a3b0f9643c70ee0bdfa4934fe5f3db6dc411d62fb1bb2e37a2"},"roas":[{"prefix":"10.0.0.0/24","maxLength":24,"asn":"AS64496","ta":"KRILL"},{"prefix":"10.0.1.0/24","maxLength":24,"asn":"AS64496","ta":"KRILL"}]}
```
### Fort Validator
Error behaviour when the `.cer` URI inside the TAL file uses the HTTPS protocol:
```
ERR: /tmp/ta.tal: Global URI 'https://ximontest.do.nlnetlabs.nl/ta/ta.cer' does not begin with 'rsync://'.
```
After replacing the URI inside the TAL file with an `rsync://` URI and placing the `.cer` file at that location in the RSYNC repo, Fort Validator then produced a different error message:
```
ERR: rsync://ximontest.do.nlnetlabs.nl/repo/ta/0/1169977D368819F3557C6B2353ED047D0CB84192.cer: Certificate validation failed: invalid or inconsistent certificate extension
```
Ximon
|
test
|
add support for rsync uri in the tal instead of https for testing purposes summary when using use ta true the generated tal file contains an https uri to the cer file routinator is able to consume this file but octorpki and fort validator fail unless the uri is changed to use the rsync protocol instead versions tested krill docker image nlnetlabs krill built from the github tag octorpki docker image cloudflare octorpki aka cloudflare octorpki latest with digest where octorpki version reports octorpki fort validator debian package installed in a docker container using image debian debian gnu linux n l according to etc issue where fort v reports fort octorpki error behaviour when the cer uri inside the tal file uses the https protocol octorpki time level info msg got open no such file or directory but repository not yet synchronized this is not a very helpful error message as the uri can definitely be fetched from inside the docker container by wget after replacing the uri inside the tal file with an rsync uri and placing the cer file at that location in the rsync repo octorpki was then able to fetch the roas from krill octorpki metadata counts generated valid signature signaturedate roas fort validator error behaviour when the cer uri inside the tal file uses the https protocol err tmp ta tal global uri does not begin with rsync after replacing the uri inside the tal file with an rsync uri and placing the cer file at that location in the rsync repo fort validator then produced a different error message err rsync ximontest do nlnetlabs nl repo ta cer certificate validation failed invalid or inconsistent certificate extension ximon
| 1
|
78,485
| 7,644,536,343
|
IssuesEvent
|
2018-05-08 15:44:31
|
Microsoft/AzureStorageExplorer
|
https://api.github.com/repos/Microsoft/AzureStorageExplorer
|
closed
|
The scrollbar still displays on Account Panel after maximizing Storage Explorer
|
resolved: won't fix testing
|
**Storage Explorer Version**: <!-- Enter the version of Storage Explorer in which you encountered your issue (see About dialog). -->1.1.0 RC
**Platform**: <!-- Enter the platform (see About dialog). -->Windows/Linux/Mac
**Architecture**: <!-- Enter the architecture of the app (see About dialog). -->ia32/x64
**Build Number**: <!-- Enter the build number (see About dialog). -->20180507.1
**Commit**: <!-- Enter the commit ID (see About dialog). -->bcc8990d
**Regression From**: <!-- Was this working on a previous version? If so, which one? --> Not a regression
#### Steps to Reproduce: ####
1. <!-- Describe the steps necessary to reproduce your issue. -->Launch Storage Explorer and sign in with your account.
2. <!-- Be as detailed as you can. -->Maximize Storage Explorer and check whether scrollbar displays on Account Panel or not.
#### Expected Experience: ####
<!-- Describe the outcome you expect when performing the above steps. -->
No scrollbar displays on Account Panel after maximizing Storage Explorer.
#### Actual Experience: ####
<!-- Describe the outcome that actually occurs when performing the above steps. -->
The scrollbar still displays on Account Panel after maximizing Storage Explorer.

|
1.0
|
The scrollbar still displays on Account Panel after maximizing Storage Explorer - **Storage Explorer Version**: <!-- Enter the version of Storage Explorer in which you encountered your issue (see About dialog). -->1.1.0 RC
**Platform**: <!-- Enter the platform (see About dialog). -->Windows/Linux/Mac
**Architecture**: <!-- Enter the architecture of the app (see About dialog). -->ia32/x64
**Build Number**: <!-- Enter the build number (see About dialog). -->20180507.1
**Commit**: <!-- Enter the commit ID (see About dialog). -->bcc8990d
**Regression From**: <!-- Was this working on a previous version? If so, which one? --> Not a regression
#### Steps to Reproduce: ####
1. <!-- Describe the steps necessary to reproduce your issue. -->Launch Storage Explorer and sign in with your account.
2. <!-- Be as detailed as you can. -->Maximize Storage Explorer and check whether scrollbar displays on Account Panel or not.
#### Expected Experience: ####
<!-- Describe the outcome you expect when performing the above steps. -->
No scrollbar displays on Account Panel after maximizing Storage Explorer.
#### Actual Experience: ####
<!-- Describe the outcome that actually occurs when performing the above steps. -->
The scrollbar still displays on Account Panel after maximizing Storage Explorer.

|
test
|
the scrollbar still displays on account panel after maximizing storage explorer storage explorer version rc platform windows linux mac architecture build number commit regression from not a regression steps to reproduce launch storage explorer and sign in with your account maximize storage explorer and check whether scrollbar displays on account panel or not expected experience no scrollbar displays on account panel after maximizing storage explorer actual experience the scrollbar still displays on account panel after maximizing storage explorer
| 1
|
125,974
| 10,372,065,467
|
IssuesEvent
|
2019-09-09 01:01:05
|
ssm-deepcove/discover_deep_cove
|
https://api.github.com/repos/ssm-deepcove/discover_deep_cove
|
closed
|
Add an image to photograph activity if possible
|
Client requested Investigate Prior: On-Site Testing enhancement
|
Clients have requested that they would like the ability to have an image included on the photograph activity i.e. so that they can show an image of what an item used to be for example.
|
1.0
|
Add an image to photograph activity if possible - Clients have requested that they would like the ability to have an image included on the photograph activity i.e. so that they can show an image of what an item used to be for example.
|
test
|
add an image to photograph activity if possible clients have requested that they would like the ability to have an image included on the photograph activity i e so that they can show an image of what an item used to be for example
| 1
|
38,763
| 2,850,254,072
|
IssuesEvent
|
2015-05-31 12:14:36
|
damonkohler/sl4a
|
https://api.github.com/repos/damonkohler/sl4a
|
opened
|
Ability to modify pitch of text-to-speech
|
auto-migrated Priority-Medium Type-Enhancement
|
_From @GoogleCodeExporter on May 31, 2015 11:31_
```
I am working on a project at the moment and I thought that being able to modify
the TTS pitch would be quite helpful. I decided to check out the source and add
this modification myself and it works (see attached). The only weird thing is
that, I had to make the parameter for setTtsPitch() a String rather than a
float or double (I get some "Argument is invalid" error, even after doing
something like setTtsPitch(float(5)) ).
It's not much but I hope a method like this does get included in future. :)
Cheers
```
Original issue reported on code.google.com by `chrispy...@gmail.com` on 3 Feb 2013 at 4:36
Attachments:
* [TextToSpeechFacade.java](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-677/comment-0/TextToSpeechFacade.java)
_Copied from original issue: damonkohler/android-scripting#677_
|
1.0
|
Ability to modify pitch of text-to-speech - _From @GoogleCodeExporter on May 31, 2015 11:31_
```
I am working on a project at the moment and I thought that being able to modify
the TTS pitch would be quite helpful. I decided to check out the source and add
this modification myself and it works (see attached). The only weird thing is
that, I had to make the parameter for setTtsPitch() a String rather than a
float or double (I get some "Argument is invalid" error, even after doing
something like setTtsPitch(float(5)) ).
It's not much but I hope a method like this does get included in future. :)
Cheers
```
Original issue reported on code.google.com by `chrispy...@gmail.com` on 3 Feb 2013 at 4:36
Attachments:
* [TextToSpeechFacade.java](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-677/comment-0/TextToSpeechFacade.java)
_Copied from original issue: damonkohler/android-scripting#677_
|
non_test
|
ability to modify pitch of text to speech from googlecodeexporter on may i am working on a project at the moment and i thought that being able to modify the tts pitch would be quite helpful i decided to check out the source and add this modification myself and it works see attached the only weird thing is that i had to make the parameter for setttspitch a string rather than a float or double i get some argument is invalid error even after doing something like setttspitch float it s not much but i hope a method like this does get included in future cheers original issue reported on code google com by chrispy gmail com on feb at attachments copied from original issue damonkohler android scripting
| 0
|
240,385
| 20,025,814,268
|
IssuesEvent
|
2022-02-01 21:10:37
|
microsoft/playwright
|
https://api.github.com/repos/microsoft/playwright
|
closed
|
[Feature]: Use source-maps when available in test-runner for reporting and errors
|
test-runner v1.19 inner-loop
|
See https://github.com/microsoft/playwright-vscode/issues/27
Users who are using a custom setup with the test-runner (feed transpiled files into the test-runner) currently can't use the extension. It should be possible that when there are source-map files, that they get used in the JSON reporter and in the error stack-traces.
|
1.0
|
[Feature]: Use source-maps when available in test-runner for reporting and errors - See https://github.com/microsoft/playwright-vscode/issues/27
Users who are using a custom setup with the test-runner (feed transpiled files into the test-runner) currently can't use the extension. It should be possible that when there are source-map files, that they get used in the JSON reporter and in the error stack-traces.
|
test
|
use source maps when available in test runner for reporting and errors see users who are using a custom setup with the test runner feed transpiled files into the test runner currently can t use the extension it should be possible that when there are source map files that they get used in the json reporter and in the error stack traces
| 1
|
127,690
| 17,353,964,643
|
IssuesEvent
|
2021-07-29 12:20:20
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Creating subspaces is very unintuitive
|
A-Spaces X-Needs-Design
|
You do so by hitting the 'add existing room' button. However, from the user perspective, spaces aren't rooms - they're spaces. So why would i select 'add existing room' when i want to find/add a space?
|
1.0
|
Creating subspaces is very unintuitive - You do so by hitting the 'add existing room' button. However, from the user perspective, spaces aren't rooms - they're spaces. So why would i select 'add existing room' when i want to find/add a space?
|
non_test
|
creating subspaces is very unintuitive you do so by hitting the add existing room button however from the user perspective spaces aren t rooms they re spaces so why would i select add existing room when i want to find add a space
| 0
|
226,581
| 18,040,860,140
|
IssuesEvent
|
2021-09-18 02:51:57
|
logicmoo/logicmoo_workspace
|
https://api.github.com/repos/logicmoo/logicmoo_workspace
|
opened
|
logicmoo.base.examples.fol.SANITY_FI_HUMAN_01@Test_9999_Line_9999__Exitcode_7
|
Test_9999 SANITY_FI_HUMAN_01 logicmoo.base.examples.fol
|
(cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif sanity_fi_human_01.pfc.pl)
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.examples.fol/SANITY_FI_HUMAN_01/logicmoo_base_examples_fol_SANITY_FI_HUMAN_01_Test_9999_Line_9999__Exitcode_7/
This: https://jenkins.logicmoo.org/job/logicmoo_workspace/41/testReport/logicmoo.base.examples.fol/SANITY_FI_HUMAN_01/logicmoo_base_examples_fol_SANITY_FI_HUMAN_01_Test_9999_Line_9999__Exitcode_7/
```
%
running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_fi_human_01.pfc.pl'),
%~ /var/lib/jenkins/.local/share/swi-prolog/pack/logicmoo_utils/prolog/logicmoo_test_header.pl:92
%~ this_test_might_need( :-( use_module( library(logicmoo_plarkc))))
:- include('sanity_fi_sk.pfc').
%~ message_hook_type(error)
%~ message_hook(
%~ initialization_exception( error(existence_error(source_sink,'sanity_fi_sk.pfc'),Existence_error)),
%~ error,
%~ [ 'Prolog initialisation failed:',
%~ nl,
%~ '~w `~p\' does not exist'-[source_sink,'sanity_fi_sk.pfc']])
%~ unused(save_junit_results)
```
totalTime=1
FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k sanity_fi_human_01.pfc.pl (returned 1)
|
1.0
|
logicmoo.base.examples.fol.SANITY_FI_HUMAN_01@Test_9999_Line_9999__Exitcode_7 - (cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif sanity_fi_human_01.pfc.pl)
Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.examples.fol/SANITY_FI_HUMAN_01/logicmoo_base_examples_fol_SANITY_FI_HUMAN_01_Test_9999_Line_9999__Exitcode_7/
This: https://jenkins.logicmoo.org/job/logicmoo_workspace/41/testReport/logicmoo.base.examples.fol/SANITY_FI_HUMAN_01/logicmoo_base_examples_fol_SANITY_FI_HUMAN_01_Test_9999_Line_9999__Exitcode_7/
```
%
running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_fi_human_01.pfc.pl'),
%~ /var/lib/jenkins/.local/share/swi-prolog/pack/logicmoo_utils/prolog/logicmoo_test_header.pl:92
%~ this_test_might_need( :-( use_module( library(logicmoo_plarkc))))
:- include('sanity_fi_sk.pfc').
%~ message_hook_type(error)
%~ message_hook(
%~ initialization_exception( error(existence_error(source_sink,'sanity_fi_sk.pfc'),Existence_error)),
%~ error,
%~ [ 'Prolog initialisation failed:',
%~ nl,
%~ '~w `~p\' does not exist'-[source_sink,'sanity_fi_sk.pfc']])
%~ unused(save_junit_results)
```
totalTime=1
FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k sanity_fi_human_01.pfc.pl (returned 1)
|
test
|
logicmoo base examples fol sanity fi human test line exitcode cd var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol timeout foreground preserve status s sigkill k lmoo clif sanity fi human pfc pl latest this running var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol sanity fi human pfc pl var lib jenkins local share swi prolog pack logicmoo utils prolog logicmoo test header pl this test might need use module library logicmoo plarkc include sanity fi sk pfc message hook type error message hook initialization exception error existence error source sink sanity fi sk pfc existence error error prolog initialisation failed nl w p does not exist unused save junit results totaltime failed var lib jenkins workspace logicmoo workspace bin lmoo junit minor k sanity fi human pfc pl returned
| 1
|
715,567
| 24,604,352,696
|
IssuesEvent
|
2022-10-14 14:56:56
|
hashicorp/terraform-cdk
|
https://api.github.com/repos/hashicorp/terraform-cdk
|
opened
|
cdktf debug should list prebuilt provider versions
|
enhancement cdktf-cli priority/important-soon ux/debugging size/medium pre-built providers cli/provider-subcommands
|
<!--- Please keep this note for the community --->
### Community Note
- Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
- Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
- If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please leave a helpful description of the feature request here. --->
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
And ideally also show the peer dependency of the provider in question so pre-built provider related dependency issues are easier to debug.
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation?
--->
- https://discuss.hashicorp.com/t/cdk-not-respecting-pom-xml/45418/7
|
1.0
|
cdktf debug should list prebuilt provider versions - <!--- Please keep this note for the community --->
### Community Note
- Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
- Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
- If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please leave a helpful description of the feature request here. --->
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
And ideally also show the peer dependency of the provider in question so pre-built provider related dependency issues are easier to debug.
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation?
--->
- https://discuss.hashicorp.com/t/cdk-not-respecting-pom-xml/45418/7
|
non_test
|
cdktf debug should list prebuilt provider versions community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description and ideally also show the peer dependency of the provider in question so pre built provider related dependency issues are easier to debug references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation
| 0
|
59,591
| 6,656,541,761
|
IssuesEvent
|
2017-09-29 21:03:22
|
pkrog/biodb
|
https://api.github.com/repos/pkrog/biodb
|
opened
|
Peakforest LCMS get-range issue
|
peakforest test
|
The request `https://peakforest-alpha.inra.fr/rest/spectra/lcms/peaks/get-range/321.047701453466/321.050911946534?token=...` returns a JSON of matching with the second element having its `source` field set to `null`. Check that it does not cause any issue in code.
|
1.0
|
Peakforest LCMS get-range issue - The request `https://peakforest-alpha.inra.fr/rest/spectra/lcms/peaks/get-range/321.047701453466/321.050911946534?token=...` returns a JSON of matching with the second element having its `source` field set to `null`. Check that it does not cause any issue in code.
|
test
|
peakforest lcms get range issue the request returns a json of matching with the second element having its source field set to null check that it does not cause any issue in code
| 1
|
77,515
| 7,575,437,406
|
IssuesEvent
|
2018-04-24 01:41:46
|
intesar/Fx-Test-Data
|
https://api.github.com/repos/intesar/Fx-Test-Data
|
reopened
|
Test-bvpoGI : enterprise_user_signup_valid
|
Test-bvpoGI
|
Project : Test-bvpoGI
Job : Stg
Env : Stg
Region : FxLabs/US_WEST_1
Result : fail
Status Code : 200
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Tue, 24 Apr 2018 00:53:27 GMT]}
Endpoint : http://stg1.fxlabs.io/api/v1/users/enterprise-sign-up
Request :
{
"name" : "first name",
"email" : "Iio@fxlabs.io",
"password" : "12345678",
"company" : "RK1rAe6hm"
}
Response :
{
"requestId" : "None",
"requestTime" : "2018-04-24T00:53:28.635+0000",
"errors" : false,
"messages" : [ {
"type" : "INFO",
"key" : "",
"value" : "Sign-up successful!"
} ],
"data" : true,
"totalPages" : 1,
"totalElements" : 1
}
Logs :
Assertion [@Response.errors == false] passed, expected [false] and found [false]Assertion [@StatusCode == 200] passed, expected [200] and found [200]
|
1.0
|
Test-bvpoGI : enterprise_user_signup_valid - Project : Test-bvpoGI
Job : Stg
Env : Stg
Region : FxLabs/US_WEST_1
Result : fail
Status Code : 200
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Tue, 24 Apr 2018 00:53:27 GMT]}
Endpoint : http://stg1.fxlabs.io/api/v1/users/enterprise-sign-up
Request :
{
"name" : "first name",
"email" : "Iio@fxlabs.io",
"password" : "12345678",
"company" : "RK1rAe6hm"
}
Response :
{
"requestId" : "None",
"requestTime" : "2018-04-24T00:53:28.635+0000",
"errors" : false,
"messages" : [ {
"type" : "INFO",
"key" : "",
"value" : "Sign-up successful!"
} ],
"data" : true,
"totalPages" : 1,
"totalElements" : 1
}
Logs :
Assertion [@Response.errors == false] passed, expected [false] and found [false]Assertion [@StatusCode == 200] passed, expected [200] and found [200]
|
test
|
test bvpogi enterprise user signup valid project test bvpogi job stg env stg region fxlabs us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options content type transfer encoding date endpoint request name first name email iio fxlabs io password company response requestid none requesttime errors false messages type info key value sign up successful data true totalpages totalelements logs assertion passed expected and found assertion passed expected and found
| 1
|
345,371
| 10,361,564,565
|
IssuesEvent
|
2019-09-06 10:20:00
|
diamm/diamm
|
https://api.github.com/repos/diamm/diamm
|
closed
|
Strange inventory note proliferation
|
Component: Metadata Priority: Medium Type: Bug
|
https://www.diamm.ac.uk/sources/635/#/inventory
If you click open the first piece in the inventory, you see a long list of notes (see screenshot). I don't know where these notes are coming from (they are obviously not to do with this work), and when I click on edit the only note that is in the record is not actually showing up in this list.
<img width="1488" alt="Screen Shot 2019-08-30 at 11 19 39" src="https://user-images.githubusercontent.com/10816169/64013704-53608980-cb18-11e9-92ef-f9a3099dbd3f.png">
|
1.0
|
Strange inventory note proliferation - https://www.diamm.ac.uk/sources/635/#/inventory
If you click open the first piece in the inventory, you see a long list of notes (see screenshot). I don't know where these notes are coming from (they are obviously not to do with this work), and when I click on edit the only note that is in the record is not actually showing up in this list.
<img width="1488" alt="Screen Shot 2019-08-30 at 11 19 39" src="https://user-images.githubusercontent.com/10816169/64013704-53608980-cb18-11e9-92ef-f9a3099dbd3f.png">
|
non_test
|
strange inventory note proliferation if you click open the first piece in the inventory you see a long list of notes see screenshot i don t know where these notes are coming from they are obviously not to do with this work and when i click on edit the only note that is in the record is not actually showing up in this list img width alt screen shot at src
| 0
|
383,436
| 11,356,173,645
|
IssuesEvent
|
2020-01-24 21:57:09
|
celo-org/celo-monorepo
|
https://api.github.com/repos/celo-org/celo-monorepo
|
closed
|
Users SBAT be prompted to switch to forno when experiencing connectivity issues
|
Priority: P1 applications feature investigate triaged wallet
|
### Expected Behavior
(Copied over from #1171, splitting into two issues- cc @nityas )
We propose a modal that gives users with connectivity issues the option to enable Forno mode.
NOTES:
If geth has restarted more than twice, or syncing has taken longer than 15s, but there is still some internet connection(eg type:unknown) show the modal below.
same modal template as the 'Are you sure?' Backup Key Modal, with enable being the primary cta:
a user should only see this modal after they have completed the 'Welcome to Celo' screen
once ‘Dismiss’ is pressed, we should not show the modal again, to avoid an overly intrusive or confusing experience.
(TODO: are there other signals we can use to fine tune this?)
_____________
**Switch Connection Mode?**
We’ve noticed you’re having some trouble connecting. We recommend enabling **Celo-Lite** mode to allow you to keep using the Celo Wallet with intermittent connection.
Celo-Lite allows you to communicate with the Celo Network through a trusted, hosted node, which can give faster performance than running a node on your device. You can always change this mode in Settings.
[Dismiss] [Enable Celo-Lite]
### Current Behavior
Working as toggle from settings screen
|
1.0
|
Users SBAT be prompted to switch to forno when experiencing connectivity issues - ### Expected Behavior
(Copied over from #1171, splitting into two issues- cc @nityas )
We propose a modal that gives users with connectivity issues the option to enable Forno mode.
NOTES:
If geth has restarted more than twice, or syncing has taken longer than 15s, but there is still some internet connection(eg type:unknown) show the modal below.
same modal template as the 'Are you sure?' Backup Key Modal, with enable being the primary cta:
a user should only see this modal after they have completed the 'Welcome to Celo' screen
once ‘Dismiss’ is pressed, we should not show the modal again, to avoid an overly intrusive or confusing experience.
(TODO: are there other signals we can use to fine tune this?)
_____________
**Switch Connection Mode?**
We’ve noticed you’re having some trouble connecting. We recommend enabling **Celo-Lite** mode to allow you to keep using the Celo Wallet with intermittent connection.
Celo-Lite allows you to communicate with the Celo Network through a trusted, hosted node, which can give faster performance than running a node on your device. You can always change this mode in Settings.
[Dismiss] [Enable Celo-Lite]
### Current Behavior
Working as toggle from settings screen
|
non_test
|
users sbat be prompted to switch to forno when experiencing connectivity issues expected behavior copied over from splitting into two issues cc nityas we propose a modal that gives users with connectivity issues the option to enable forno mode notes if geth has restarted more than twice or syncing has taken longer than but there is still some internet connection eg type unknown show the modal below same modal template as the are you sure backup key modal with enable being the primary cta a user should only see this modal after they have completed the welcome to celo screen once ‘dismiss’ is pressed we should not show the modal again to avoid an overly intrusive or confusing experience todo are there other signals we can use to fine tune this switch connection mode we’ve noticed you’re having some trouble connecting we recommend enabling celo lite mode to allow you to keep using the celo wallet with intermittent connection celo lite allows you to communicate with the celo network through a trusted hosted node which can give faster performance than running a node on your device you can always change this mode in settings current behavior working as toggle from settings screen
| 0
|
7,369
| 6,924,299,879
|
IssuesEvent
|
2017-11-30 12:13:28
|
aiidateam/aiida_core
|
https://api.github.com/repos/aiidateam/aiida_core
|
closed
|
debugging: more helpful error message for db-tests in case of import error
|
topic/TestInfrastructure type/enhancement type/proposal type/quality-of-life type/trivial
|
## Description
The following type of error message is common when running aiida tests in case of import errors (missing dependencies for instance).
```
File ".../aiida-core/aiida/backends/testbase.py", line 165, in run_aiida_db_tests
test_suite.addTest(test_loader.loadTestsFromName(modulename))
File ".../lib/python2.7/unittest/loader.py", line 100, in loadTestsFromName
parent, obj = obj, getattr(obj, part)
AttributeError: 'module' object has no attribute 'persistence'
```
The error message could easily be very helpful if only the name of the module was output too
## Possible solution
The AttributeError could be caught in `run_aiida_db_tests` in `aiida/backends/testase.py` and the more helpful message could be output containing the `modulename`.
|
1.0
|
debugging: more helpful error message for db-tests in case of import error - ## Description
The following type of error message is common when running aiida tests in case of import errors (missing dependencies for instance).
```
File ".../aiida-core/aiida/backends/testbase.py", line 165, in run_aiida_db_tests
test_suite.addTest(test_loader.loadTestsFromName(modulename))
File ".../lib/python2.7/unittest/loader.py", line 100, in loadTestsFromName
parent, obj = obj, getattr(obj, part)
AttributeError: 'module' object has no attribute 'persistence'
```
The error message could easily be very helpful if only the name of the module was output too
## Possible solution
The AttributeError could be caught in `run_aiida_db_tests` in `aiida/backends/testase.py` and the more helpful message could be output containing the `modulename`.
|
non_test
|
debugging more helpful error message for db tests in case of import error description the following type of error message is common when running aiida tests in case of import errors missing dependencies for instance file aiida core aiida backends testbase py line in run aiida db tests test suite addtest test loader loadtestsfromname modulename file lib unittest loader py line in loadtestsfromname parent obj obj getattr obj part attributeerror module object has no attribute persistence the error message could easily be very helpful if only the name of the module was output too possible solution the attributeerror could be caught in run aiida db tests in aiida backends testase py and the more helpful message could be output containing the modulename
| 0
|
95,655
| 12,032,463,673
|
IssuesEvent
|
2020-04-13 12:12:39
|
reach4help/reach4help
|
https://api.github.com/repos/reach4help/reach4help
|
closed
|
Design view: Name & address
|
UI Design
|
* [x] Initial design for desktop
* [x] Adapt design for mobile
<!--zenhub info: do not edit anything after this line, it will be automatically changed-->
--------
### [ZenHub Information](https://app.zenhub.com/workspaces/reach4help-5e8dcbfb14ac087f410cbabb/issues/reach4help/reach4help/344)
*This information is updated automatically. To modify it, please use ZenHub.*
**Belonging to Epics:**
* **[EPIC]** [#327 - View: Name & Address](https://github.com/reach4help/reach4help/issues/327)
**Blocking:**
* [BLOCKED BY: [#344](https://github.com/reach4help/reach4help/issues/344)] [#344 - Implement view: Name & Address](https://github.com/reach4help/reach4help/issues/100)
<!--zenhub info end-->
|
1.0
|
Design view: Name & address - * [x] Initial design for desktop
* [x] Adapt design for mobile
<!--zenhub info: do not edit anything after this line, it will be automatically changed-->
--------
### [ZenHub Information](https://app.zenhub.com/workspaces/reach4help-5e8dcbfb14ac087f410cbabb/issues/reach4help/reach4help/344)
*This information is updated automatically. To modify it, please use ZenHub.*
**Belonging to Epics:**
* **[EPIC]** [#327 - View: Name & Address](https://github.com/reach4help/reach4help/issues/327)
**Blocking:**
* [BLOCKED BY: [#344](https://github.com/reach4help/reach4help/issues/344)] [#344 - Implement view: Name & Address](https://github.com/reach4help/reach4help/issues/100)
<!--zenhub info end-->
|
non_test
|
design view name address initial design for desktop adapt design for mobile this information is updated automatically to modify it please use zenhub belonging to epics blocking
| 0
|
109,507
| 13,781,290,310
|
IssuesEvent
|
2020-10-08 15:58:54
|
brave/brave-ios
|
https://api.github.com/repos/brave/brave-ios
|
closed
|
Brave today extension
|
QA/Yes blocked: needs design enhancement release-notes/include
|
Search extension to display on today widgets screen. Search, new tab, private tab.
|
1.0
|
Brave today extension - Search extension to display on today widgets screen. Search, new tab, private tab.
|
non_test
|
brave today extension search extension to display on today widgets screen search new tab private tab
| 0
|
172,035
| 13,260,253,576
|
IssuesEvent
|
2020-08-20 17:55:08
|
ramp4-pcar4/ramp4-pcar4
|
https://api.github.com/repos/ramp4-pcar4/ramp4-pcar4
|
opened
|
Test for RampMap
|
effort: far away flavour: tests priority: nice type: preventative
|
Create tests for the RampMap class in GeoAPI. See `\packages\ramp-core\src\fixtures\mapnav\tests\` and `\packages\ramp-core\tests` for starter samples.
Suggest prioritizing `e2e` tests, as unit tests require mocking which can be hefty task that has the smell of redundancy.
Things to consider
- validation of loading promises
- validation of event wireup. e.g. mouse move, click, extent change
- suite of zoomies functions
- suite of `get` functions that return information about the map
- screen point to map point converter
will likely be more things added if this issue gets stale
|
1.0
|
Test for RampMap - Create tests for the RampMap class in GeoAPI. See `\packages\ramp-core\src\fixtures\mapnav\tests\` and `\packages\ramp-core\tests` for starter samples.
Suggest prioritizing `e2e` tests, as unit tests require mocking which can be hefty task that has the smell of redundancy.
Things to consider
- validation of loading promises
- validation of event wireup. e.g. mouse move, click, extent change
- suite of zoomies functions
- suite of `get` functions that return information about the map
- screen point to map point converter
will likely be more things added if this issue gets stale
|
test
|
test for rampmap create tests for the rampmap class in geoapi see packages ramp core src fixtures mapnav tests and packages ramp core tests for starter samples suggest prioritizing tests as unit tests require mocking which can be hefty task that has the smell of redundancy things to consider validation of loading promises validation of event wireup e g mouse move click extent change suite of zoomies functions suite of get functions that return information about the map screen point to map point converter will likely be more things added if this issue gets stale
| 1
|
294,866
| 25,410,850,009
|
IssuesEvent
|
2022-11-22 18:50:27
|
NuGet/Home
|
https://api.github.com/repos/NuGet/Home
|
closed
|
Dependencies tree unit tests
|
Priority:2 Functionality:VisualStudioUI Type:Engineering Type:Test Partner:Project-System
|
From https://github.com/NuGet/NuGet.Client/pull/3392
Tracking adding more tests to the code in NuGet.Client for populating transitive dependencies in Solution Explorer.
|
1.0
|
Dependencies tree unit tests - From https://github.com/NuGet/NuGet.Client/pull/3392
Tracking adding more tests to the code in NuGet.Client for populating transitive dependencies in Solution Explorer.
|
test
|
dependencies tree unit tests from tracking adding more tests to the code in nuget client for populating transitive dependencies in solution explorer
| 1
|
662,666
| 22,148,646,261
|
IssuesEvent
|
2022-06-03 14:32:58
|
kudobuilder/kudo
|
https://api.github.com/repos/kudobuilder/kudo
|
closed
|
Question: Deploy instance using CRD or API call instead of "kubectl kudo install"
|
component/cli kind/enhancement priority/long-term-important community
|
We're are currently evaluating kudo for deploying individual instances for larger tenants that are seperated into different namespaces. The idea is building a user interface that displays a form of parameters in order to deploy the infrastructure for each of the tenants.
Most operators found on operatorhub.io allow deploying application instances using CRDs. Our initial plan was creating CRDs for deployment instructions using a Java application.
I wasn't able to find anything specific in the documentation if that's officially supported. What I found is this Instance definition [01-upgrade.yml](https://github.com/kudobuilder/kudo/blob/master/test/integration/cli-install/01-upgrade.yaml)
``` yaml
apiVersion: kudo.dev/v1beta1
kind: Instance
metadata:
name: cli-install
labels:
kudo.dev/operator: cli-install-operator
spec:
operatorVersion:
name: cli-install-operator-0.1.0
parameters:
memory: "2Gi"
```
So my question is if a syntax like the snipped above or an official API can be used in order to fire up an instance using a standard kubernetes client library (e.g. [Java](https://github.com/kubernetes-client/java)).
Bottom line: Nice work! We believe declerative operators will be a huge step in the whole kubernetes ecosystem.
|
1.0
|
Question: Deploy instance using CRD or API call instead of "kubectl kudo install" - We're are currently evaluating kudo for deploying individual instances for larger tenants that are seperated into different namespaces. The idea is building a user interface that displays a form of parameters in order to deploy the infrastructure for each of the tenants.
Most operators found on operatorhub.io allow deploying application instances using CRDs. Our initial plan was creating CRDs for deployment instructions using a Java application.
I wasn't able to find anything specific in the documentation if that's officially supported. What I found is this Instance definition [01-upgrade.yml](https://github.com/kudobuilder/kudo/blob/master/test/integration/cli-install/01-upgrade.yaml)
``` yaml
apiVersion: kudo.dev/v1beta1
kind: Instance
metadata:
name: cli-install
labels:
kudo.dev/operator: cli-install-operator
spec:
operatorVersion:
name: cli-install-operator-0.1.0
parameters:
memory: "2Gi"
```
So my question is if a syntax like the snipped above or an official API can be used in order to fire up an instance using a standard kubernetes client library (e.g. [Java](https://github.com/kubernetes-client/java)).
Bottom line: Nice work! We believe declerative operators will be a huge step in the whole kubernetes ecosystem.
|
non_test
|
question deploy instance using crd or api call instead of kubectl kudo install we re are currently evaluating kudo for deploying individual instances for larger tenants that are seperated into different namespaces the idea is building a user interface that displays a form of parameters in order to deploy the infrastructure for each of the tenants most operators found on operatorhub io allow deploying application instances using crds our initial plan was creating crds for deployment instructions using a java application i wasn t able to find anything specific in the documentation if that s officially supported what i found is this instance definition yaml apiversion kudo dev kind instance metadata name cli install labels kudo dev operator cli install operator spec operatorversion name cli install operator parameters memory so my question is if a syntax like the snipped above or an official api can be used in order to fire up an instance using a standard kubernetes client library e g bottom line nice work we believe declerative operators will be a huge step in the whole kubernetes ecosystem
| 0
|
671,173
| 22,746,717,462
|
IssuesEvent
|
2022-07-07 09:48:13
|
aquasecurity/trivy-operator
|
https://api.github.com/repos/aquasecurity/trivy-operator
|
opened
|
ghcr is the default container registry in trivy-operator helm chart
|
good first issue kind/feature priority/important-soon kind/deployment
|
set ghcr to be the default container registry in trivy-operator helm chart
|
1.0
|
ghcr is the default container registry in trivy-operator helm chart - set ghcr to be the default container registry in trivy-operator helm chart
|
non_test
|
ghcr is the default container registry in trivy operator helm chart set ghcr to be the default container registry in trivy operator helm chart
| 0
|
184,128
| 14,271,902,287
|
IssuesEvent
|
2020-11-21 14:40:28
|
mui-org/material-ui
|
https://api.github.com/repos/mui-org/material-ui
|
closed
|
[Slider] Testing onChangeCommitted from click
|
component: Slider good first issue test
|
<!-- Provide a general summary of the issue in the Title above -->
<!--
Thank you very much for contributing to Material-UI by creating an issue! ❤️
To avoid duplicate issues we ask you to check off the following list.
-->
<!-- Checked checkbox should look like this: [x] -->
- [ ] The issue is present in the latest release.
- [ x] I have searched the [issues](https://github.com/mui-org/material-ui/issues) of this repository and believe that this is not a duplicate.
## Current Behavior 😯
When testing the slider functionality using rtl and jest no significant DOM update happens when clicking on labels (or doing mousedown or other events)
<!-- Describe what happens instead of the expected behavior. -->
## Expected Behavior 🤔
DOM should update with the selected mark as Slider value
<!-- Describe what should happen. -->
## Steps to Reproduce 🕹
In the codesadbox example i tried to test this component of a slider value change on label (slider mark) click. But i cant seem to get the onChangeCommitted to trigger when testing. The slider component dose not indicate any change
https://codesandbox.io/s/romantic-goldstine-gpjbj?file=/src/App.test.js
<!--
Provide a link to a live example (you can use codesandbox.io) and an unambiguous set of steps to reproduce this bug.
Include code to reproduce, if relevant (which it most likely is).
You should use the official codesandbox template as a starting point: https://material-ui.com/r/issue-template
If you have an issue concerning TypeScript please start from this TypeScript playground: https://material-ui.com/r/ts-issue-template
Issues without some form of live example have a longer response time.
-->
## Context 🔦
I am trying to test onChangeCommitted in the slider component. But ive tried up and down to test this event on the slider and its getting really frustrating as there dosent seem to be many threads or SO posts about so I might just be missing something trivial. Ive added a sample on codesandbox. Hope some one can help thanks :)
<!--
What are you trying to accomplish? How has this issue affected you?
Providing context helps us come up with a solution that is most useful in the real world.
-->
## Your Environment 🌎
<!--
Include as many relevant details about the environment with which you experienced the bug.
If you encounter issues with TypeScript please include version and tsconfig.
-->
| Tech | Version |
| ----------- | ------- |
| Material-UI | v4.3.1 |
| React | v16.8.4 |
| Browser | Chrome |
| TypeScript | Y (but not in example) |
| etc. | |
|
1.0
|
[Slider] Testing onChangeCommitted from click - <!-- Provide a general summary of the issue in the Title above -->
<!--
Thank you very much for contributing to Material-UI by creating an issue! ❤️
To avoid duplicate issues we ask you to check off the following list.
-->
<!-- Checked checkbox should look like this: [x] -->
- [ ] The issue is present in the latest release.
- [ x] I have searched the [issues](https://github.com/mui-org/material-ui/issues) of this repository and believe that this is not a duplicate.
## Current Behavior 😯
When testing the slider functionality using rtl and jest no significant DOM update happens when clicking on labels (or doing mousedown or other events)
<!-- Describe what happens instead of the expected behavior. -->
## Expected Behavior 🤔
DOM should update with the selected mark as Slider value
<!-- Describe what should happen. -->
## Steps to Reproduce 🕹
In the codesadbox example i tried to test this component of a slider value change on label (slider mark) click. But i cant seem to get the onChangeCommitted to trigger when testing. The slider component dose not indicate any change
https://codesandbox.io/s/romantic-goldstine-gpjbj?file=/src/App.test.js
<!--
Provide a link to a live example (you can use codesandbox.io) and an unambiguous set of steps to reproduce this bug.
Include code to reproduce, if relevant (which it most likely is).
You should use the official codesandbox template as a starting point: https://material-ui.com/r/issue-template
If you have an issue concerning TypeScript please start from this TypeScript playground: https://material-ui.com/r/ts-issue-template
Issues without some form of live example have a longer response time.
-->
## Context 🔦
I am trying to test onChangeCommitted in the slider component. But ive tried up and down to test this event on the slider and its getting really frustrating as there dosent seem to be many threads or SO posts about so I might just be missing something trivial. Ive added a sample on codesandbox. Hope some one can help thanks :)
<!--
What are you trying to accomplish? How has this issue affected you?
Providing context helps us come up with a solution that is most useful in the real world.
-->
## Your Environment 🌎
<!--
Include as many relevant details about the environment with which you experienced the bug.
If you encounter issues with TypeScript please include version and tsconfig.
-->
| Tech | Version |
| ----------- | ------- |
| Material-UI | v4.3.1 |
| React | v16.8.4 |
| Browser | Chrome |
| TypeScript | Y (but not in example) |
| etc. | |
|
test
|
testing onchangecommitted from click thank you very much for contributing to material ui by creating an issue ❤️ to avoid duplicate issues we ask you to check off the following list the issue is present in the latest release i have searched the of this repository and believe that this is not a duplicate current behavior 😯 when testing the slider functionality using rtl and jest no significant dom update happens when clicking on labels or doing mousedown or other events expected behavior 🤔 dom should update with the selected mark as slider value steps to reproduce 🕹 in the codesadbox example i tried to test this component of a slider value change on label slider mark click but i cant seem to get the onchangecommitted to trigger when testing the slider component dose not indicate any change provide a link to a live example you can use codesandbox io and an unambiguous set of steps to reproduce this bug include code to reproduce if relevant which it most likely is you should use the official codesandbox template as a starting point if you have an issue concerning typescript please start from this typescript playground issues without some form of live example have a longer response time context 🔦 i am trying to test onchangecommitted in the slider component but ive tried up and down to test this event on the slider and its getting really frustrating as there dosent seem to be many threads or so posts about so i might just be missing something trivial ive added a sample on codesandbox hope some one can help thanks what are you trying to accomplish how has this issue affected you providing context helps us come up with a solution that is most useful in the real world your environment 🌎 include as many relevant details about the environment with which you experienced the bug if you encounter issues with typescript please include version and tsconfig tech version material ui react browser chrome typescript y but not in example etc
| 1
|
49,760
| 6,039,394,262
|
IssuesEvent
|
2017-06-10 02:36:28
|
ClassicWoW/Nefarian_1.12.1_Bugtracker
|
https://api.github.com/repos/ClassicWoW/Nefarian_1.12.1_Bugtracker
|
closed
|
[Naxxramas] Outer Ring Mobs
|
Auf Testserver Datenbank Kreaturen Dev Behoben
|
Welches Verhalten wird beobachtet?
es fehlen dort 1 Trashpack direkt wenn man von Faerlina kommt
Wie sollte es sich verhalten?
es sollten 2 Ghulpacks mit jeweils 4 Ghulen sein
Schritte zur Reproduzierung
Outer ring aus der Richtung von Faerlina betreten
Zusätzliche Informationen (Screenshots, Videos, Klasse, Rasse, Level, etc.)
Giga liveraid von BC in Naxxramas man sieht 2 Ghulpacks
https://youtu.be/QtyEoJ6_nZI?t=2617
Alle Kreaturen, Items, Objekte, Quests, Zauber etc. müssen aus unserer Datenbank verlinkt sein.
https://datenbank.classic-wow.org/?npc=16447 nochmal dadrauf hingewiesen, besitzt wieder falsche Fähigkeiten als auf dem Testserver
|
1.0
|
[Naxxramas] Outer Ring Mobs - Welches Verhalten wird beobachtet?
es fehlen dort 1 Trashpack direkt wenn man von Faerlina kommt
Wie sollte es sich verhalten?
es sollten 2 Ghulpacks mit jeweils 4 Ghulen sein
Schritte zur Reproduzierung
Outer ring aus der Richtung von Faerlina betreten
Zusätzliche Informationen (Screenshots, Videos, Klasse, Rasse, Level, etc.)
Giga liveraid von BC in Naxxramas man sieht 2 Ghulpacks
https://youtu.be/QtyEoJ6_nZI?t=2617
Alle Kreaturen, Items, Objekte, Quests, Zauber etc. müssen aus unserer Datenbank verlinkt sein.
https://datenbank.classic-wow.org/?npc=16447 nochmal dadrauf hingewiesen, besitzt wieder falsche Fähigkeiten als auf dem Testserver
|
test
|
outer ring mobs welches verhalten wird beobachtet es fehlen dort trashpack direkt wenn man von faerlina kommt wie sollte es sich verhalten es sollten ghulpacks mit jeweils ghulen sein schritte zur reproduzierung outer ring aus der richtung von faerlina betreten zusätzliche informationen screenshots videos klasse rasse level etc giga liveraid von bc in naxxramas man sieht ghulpacks alle kreaturen items objekte quests zauber etc müssen aus unserer datenbank verlinkt sein nochmal dadrauf hingewiesen besitzt wieder falsche fähigkeiten als auf dem testserver
| 1
|
195,650
| 14,742,371,302
|
IssuesEvent
|
2021-01-07 12:10:53
|
pingcap/tidb
|
https://api.github.com/repos/pingcap/tidb
|
closed
|
Case failure[key]:CREATE TABLE t1 (\nname varchar(50) DEFAULT '' NOT NULL,\nauthor varchar(50) DEFAULT '' NOT NU...
|
component/test
|
Latest build: <a href="https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_integration_common_test/28045/display/redirect">https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_integration_common_test/28045/display/redirect</a>
|
1.0
|
Case failure[key]:CREATE TABLE t1 (\nname varchar(50) DEFAULT '' NOT NULL,\nauthor varchar(50) DEFAULT '' NOT NU... - Latest build: <a href="https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_integration_common_test/28045/display/redirect">https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_integration_common_test/28045/display/redirect</a>
|
test
|
case failure create table nname varchar default not null nauthor varchar default not nu latest build a href
| 1
|
73,015
| 7,320,336,937
|
IssuesEvent
|
2018-03-02 06:39:08
|
EyeSeeTea/malariapp
|
https://api.github.com/repos/EyeSeeTea/malariapp
|
closed
|
Change score classes
|
HNQIS complexity - med (1-5hr) priority - high testing type - feature
|
- Class A - green: >= 90%
- Class B - yellow: 80% >= and <= 89%
- Class C - red: <80%
|
1.0
|
Change score classes - - Class A - green: >= 90%
- Class B - yellow: 80% >= and <= 89%
- Class C - red: <80%
|
test
|
change score classes class a green class b yellow and class c red
| 1
|
101,447
| 8,788,090,232
|
IssuesEvent
|
2018-12-20 20:55:17
|
ONRR/doi-extractives-data
|
https://api.github.com/repos/ONRR/doi-extractives-data
|
closed
|
Write part two of our blog post series on homepage transition
|
blog workflow:testing
|
Part two will cover our transition to using GatsbyJS for new feature development, specifically our new homepage.
|
1.0
|
Write part two of our blog post series on homepage transition - Part two will cover our transition to using GatsbyJS for new feature development, specifically our new homepage.
|
test
|
write part two of our blog post series on homepage transition part two will cover our transition to using gatsbyjs for new feature development specifically our new homepage
| 1
|
331,156
| 28,537,327,852
|
IssuesEvent
|
2023-04-20 01:08:26
|
opera-adt/dolphin
|
https://api.github.com/repos/opera-adt/dolphin
|
closed
|
integration testing setup
|
enhancement testing
|
Setup a nightly or weekly integration test that runs through the primary workflow on some golden dataset.
The related parts to this are
1. Create a golden dataset: #15
2. Set up some way to store this data and pull it for the test (possibly using Zenodo or https://github.com/fatiando/pooch, depending on how big it is)
3. Figure out how to set up the CI system to have a longer-running test, and to trigger it on a schedule rather than on-push/on-PR
|
1.0
|
integration testing setup - Setup a nightly or weekly integration test that runs through the primary workflow on some golden dataset.
The related parts to this are
1. Create a golden dataset: #15
2. Set up some way to store this data and pull it for the test (possibly using Zenodo or https://github.com/fatiando/pooch, depending on how big it is)
3. Figure out how to set up the CI system to have a longer-running test, and to trigger it on a schedule rather than on-push/on-PR
|
test
|
integration testing setup setup a nightly or weekly integration test that runs through the primary workflow on some golden dataset the related parts to this are create a golden dataset set up some way to store this data and pull it for the test possibly using zenodo or depending on how big it is figure out how to set up the ci system to have a longer running test and to trigger it on a schedule rather than on push on pr
| 1
|
530,667
| 15,435,540,755
|
IssuesEvent
|
2021-03-07 09:19:51
|
woocommerce/woocommerce-gateway-stripe
|
https://api.github.com/repos/woocommerce/woocommerce-gateway-stripe
|
closed
|
!BROKEN! on AWS Cloudfront.
|
component: webhooks priority: high type: enhancement type: question
|
Update by @dwainm:
Add a filter is_valid_request return value . Pass along the function parameters.
https://github.com/woocommerce/woocommerce-gateway-stripe/blob/70640ab3a2dd3dfc4a700dee3dc2393fcaca95a7/includes/class-wc-stripe-webhook-handler.php#L93-L95
----
**Original Issue**
Plugin has conflict with Amazon Cloudfront CDN whereby the correct User Agent is forcefully rewritten by the CDN. On all requests, the User Agent is modified to "Amazon Cloudfront". This causes all webhook requests to return 400 even when correctly setup by failing on line 93 of /includes/class-wc-stripe-webhook-handler.php.
If I comment out lines 93-95 I get expected behaviour.
The Cloudfront specification can be found here: https://docs.aws.amazon.com/AmazonCloudFront/latest/DeveloperGuide/RequestAndResponseBehaviorCustomOrigin.html
Commenting out the three lines is sufficient but it’s not a long term fix. Please can you introduce an exception or a way of being able to disable this specific security measure.
Wordpress.org support ticket: https://wordpress.org/support/topic/cloudfront-user-agent/#post-12826607
**Environment (please complete the following information):**
- WordPress Version - all
- WooCommerce Version - 4.0.1
- Stripe Plugin Version - 4.3.3
- Browser [e.g. chrome, safari] and Version - all browsers
|
1.0
|
!BROKEN! on AWS Cloudfront. - Update by @dwainm:
Add a filter is_valid_request return value . Pass along the function parameters.
https://github.com/woocommerce/woocommerce-gateway-stripe/blob/70640ab3a2dd3dfc4a700dee3dc2393fcaca95a7/includes/class-wc-stripe-webhook-handler.php#L93-L95
----
**Original Issue**
Plugin has conflict with Amazon Cloudfront CDN whereby the correct User Agent is forcefully rewritten by the CDN. On all requests, the User Agent is modified to "Amazon Cloudfront". This causes all webhook requests to return 400 even when correctly setup by failing on line 93 of /includes/class-wc-stripe-webhook-handler.php.
If I comment out lines 93-95 I get expected behaviour.
The Cloudfront specification can be found here: https://docs.aws.amazon.com/AmazonCloudFront/latest/DeveloperGuide/RequestAndResponseBehaviorCustomOrigin.html
Commenting out the three lines is sufficient but it’s not a long term fix. Please can you introduce an exception or a way of being able to disable this specific security measure.
Wordpress.org support ticket: https://wordpress.org/support/topic/cloudfront-user-agent/#post-12826607
**Environment (please complete the following information):**
- WordPress Version - all
- WooCommerce Version - 4.0.1
- Stripe Plugin Version - 4.3.3
- Browser [e.g. chrome, safari] and Version - all browsers
|
non_test
|
broken on aws cloudfront update by dwainm add a filter is valid request return value pass along the function parameters original issue plugin has conflict with amazon cloudfront cdn whereby the correct user agent is forcefully rewritten by the cdn on all requests the user agent is modified to amazon cloudfront this causes all webhook requests to return even when correctly setup by failing on line of includes class wc stripe webhook handler php if i comment out lines i get expected behaviour the cloudfront specification can be found here commenting out the three lines is sufficient but it’s not a long term fix please can you introduce an exception or a way of being able to disable this specific security measure wordpress org support ticket environment please complete the following information wordpress version all woocommerce version stripe plugin version browser and version all browsers
| 0
|
214,303
| 16,580,643,019
|
IssuesEvent
|
2021-05-31 11:18:48
|
gardener/gardener
|
https://api.github.com/repos/gardener/gardener
|
closed
|
[Failing test] should add worker pool with containerd
|
area/testing kind/bug
|
/area testing
/kind bug
**What happened**:
Currently the `[Fail] Shoot container runtime testing [It] [DEFAULT] [SERIAL] [SHOOT] should add worker pool with containerd ` test fails with:
```
FAIL
FAIL github.com/gardener/gardener/test/suites/shoot 2295.248s
FAIL
--- FAIL: TestGardenerSuite (2295.18s)
FAIL! -- 7 Passed | 1 Failed | 0 Pending | 15 Skipped
Ran 8 of 23 Specs in 2295.167 seconds
/src/test/integration/shoots/operations/containerruntime.go:125
[Fail] Shoot container runtime testing [It] [DEFAULT] [SERIAL] [SHOOT] should add worker pool with containerd
Summarizing 1 Failure:
------------------------------
/src/test/integration/shoots/operations/containerruntime.go:125
<string>: found\n
to equal
<string>: Not found\n
Expected
/src/test/framework/gingko_utils.go:26
[DEFAULT] [SERIAL] [SHOOT] should add worker pool with containerd [It]
/src/test/integration/shoots/operations/containerruntime.go:30
Shoot container runtime testing
```
<img width="794" alt="Screenshot 2021-05-31 at 11 53 56" src="https://user-images.githubusercontent.com/9372594/120168155-3c782b00-c207-11eb-894b-53f2846918d6.png">
**Environment**:
- Gardener version: v1.24.0-dev-a12d3556f62c8a56e7364d1112b26a1d727eb604
- Kubernetes version (use `kubectl version`): v1.20.6
- Cloud provider or hardware configuration:
- Others:
|
1.0
|
[Failing test] should add worker pool with containerd - /area testing
/kind bug
**What happened**:
Currently the `[Fail] Shoot container runtime testing [It] [DEFAULT] [SERIAL] [SHOOT] should add worker pool with containerd ` test fails with:
```
FAIL
FAIL github.com/gardener/gardener/test/suites/shoot 2295.248s
FAIL
--- FAIL: TestGardenerSuite (2295.18s)
FAIL! -- 7 Passed | 1 Failed | 0 Pending | 15 Skipped
Ran 8 of 23 Specs in 2295.167 seconds
/src/test/integration/shoots/operations/containerruntime.go:125
[Fail] Shoot container runtime testing [It] [DEFAULT] [SERIAL] [SHOOT] should add worker pool with containerd
Summarizing 1 Failure:
------------------------------
/src/test/integration/shoots/operations/containerruntime.go:125
<string>: found\n
to equal
<string>: Not found\n
Expected
/src/test/framework/gingko_utils.go:26
[DEFAULT] [SERIAL] [SHOOT] should add worker pool with containerd [It]
/src/test/integration/shoots/operations/containerruntime.go:30
Shoot container runtime testing
```
<img width="794" alt="Screenshot 2021-05-31 at 11 53 56" src="https://user-images.githubusercontent.com/9372594/120168155-3c782b00-c207-11eb-894b-53f2846918d6.png">
**Environment**:
- Gardener version: v1.24.0-dev-a12d3556f62c8a56e7364d1112b26a1d727eb604
- Kubernetes version (use `kubectl version`): v1.20.6
- Cloud provider or hardware configuration:
- Others:
|
test
|
should add worker pool with containerd area testing kind bug what happened currently the shoot container runtime testing should add worker pool with containerd test fails with fail fail github com gardener gardener test suites shoot fail fail testgardenersuite fail passed failed pending skipped ran of specs in seconds src test integration shoots operations containerruntime go shoot container runtime testing should add worker pool with containerd summarizing failure src test integration shoots operations containerruntime go found n to equal not found n expected src test framework gingko utils go should add worker pool with containerd src test integration shoots operations containerruntime go shoot container runtime testing img width alt screenshot at src environment gardener version dev kubernetes version use kubectl version cloud provider or hardware configuration others
| 1
|
664,495
| 22,272,717,267
|
IssuesEvent
|
2022-06-10 13:48:43
|
opencrvs/opencrvs-core
|
https://api.github.com/repos/opencrvs/opencrvs-core
|
closed
|
In performance, no data is showing when location is an office
|
👹Bug Priority: high
|
**Bug description:**
In performance, no data is showing when location is an office. And for office Completeness bloc should be hidden
**Steps to reproduce:**
1. Login as any user other than field agent
2. Go to Performance
3. Select location as office and observe
**Actual result:**
1) all values are 0%
2) Completeness block is showing
**Expected result:**
1) should show actual values. If there is only 1 office in a district then all the values should be same as district
2) should hide completeness block
**Screenshot:**

**Tested on:**
https://login.farajaland-qa.opencrvs.org/
|
1.0
|
In performance, no data is showing when location is an office - **Bug description:**
In performance, no data is showing when location is an office. And for office Completeness bloc should be hidden
**Steps to reproduce:**
1. Login as any user other than field agent
2. Go to Performance
3. Select location as office and observe
**Actual result:**
1) all values are 0%
2) Completeness block is showing
**Expected result:**
1) should show actual values. If there is only 1 office in a district then all the values should be same as district
2) should hide completeness block
**Screenshot:**

**Tested on:**
https://login.farajaland-qa.opencrvs.org/
|
non_test
|
in performance no data is showing when location is an office bug description in performance no data is showing when location is an office and for office completeness bloc should be hidden steps to reproduce login as any user other than field agent go to performance select location as office and observe actual result all values are completeness block is showing expected result should show actual values if there is only office in a district then all the values should be same as district should hide completeness block screenshot tested on
| 0
|
50,417
| 10,500,534,066
|
IssuesEvent
|
2019-09-26 10:43:35
|
atomist/atomist-internal-sdm
|
https://api.github.com/repos/atomist/atomist-internal-sdm
|
closed
|
Code Inspection: npm audit on atomist/t29e48p34/npm-project-deps/master
|
bug code-inspection
|
### handlebars:<4.3.0
- _(error)_ [Prototype Pollution](https://npmjs.com/advisories/1164) _Upgrade to version 4.3.0 or later._
- `handlebars:4.2.0`:
- `@atomist/sdm-pack-aspect>express-handlebars>handlebars`
[atomist:code-inspection:atomist/t29e48p34/npm-project-deps/master=@atomist/atomist-sdm]
|
1.0
|
Code Inspection: npm audit on atomist/t29e48p34/npm-project-deps/master - ### handlebars:<4.3.0
- _(error)_ [Prototype Pollution](https://npmjs.com/advisories/1164) _Upgrade to version 4.3.0 or later._
- `handlebars:4.2.0`:
- `@atomist/sdm-pack-aspect>express-handlebars>handlebars`
[atomist:code-inspection:atomist/t29e48p34/npm-project-deps/master=@atomist/atomist-sdm]
|
non_test
|
code inspection npm audit on atomist npm project deps master handlebars error upgrade to version or later handlebars atomist sdm pack aspect express handlebars handlebars
| 0
|
343,665
| 30,682,389,137
|
IssuesEvent
|
2023-07-26 10:00:22
|
iho-ohi/S-101_Portrayal-Catalogue
|
https://api.github.com/repos/iho-ohi/S-101_Portrayal-Catalogue
|
closed
|
New symbol for Berth features with categoryOfCargo = 7 (Dangerous or Hazardous) - ENCWG7-5.3_2022 [PSWG #112]
|
enhancement test PC 1.1.0
|
SPEC at: https://github.com/S-101-Portrayal-subWG/Working-Documents/issues/112#issuecomment-1381193928
|
1.0
|
New symbol for Berth features with categoryOfCargo = 7 (Dangerous or Hazardous) - ENCWG7-5.3_2022 [PSWG #112] - SPEC at: https://github.com/S-101-Portrayal-subWG/Working-Documents/issues/112#issuecomment-1381193928
|
test
|
new symbol for berth features with categoryofcargo dangerous or hazardous spec at
| 1
|
75,027
| 7,458,753,560
|
IssuesEvent
|
2018-03-30 12:08:42
|
SunwellTracker/issues
|
https://api.github.com/repos/SunwellTracker/issues
|
closed
|
[Warrior]Deep Wounds
|
Works locally | Requires testing
|
Decription:
The damage from warriors talent "Deep Wounds" isnt affected by resilience or any other dmg reduce effect.
And the Base dmg is to High. 3,4k gs Warriors can stack that talent to 1300dmg per tick. On retail with Shadowmourne warrior i got that wounds on 1400dmg per tick in pve.
Quote:
PATCH 3.3.3: The damage from this talent is now affected by resilience as intended.
helping source:
http://wowwiki.wikia.com/wiki/Deep_Wounds
How it works:
Deep wounds arent effected by resilience or for expample painsub, dispersion, shadowform, shieldwall and all other things. So in Dispersion (90% dmg reduce) it does every 1 sec 600 dmg+.
How it should work:
That talent should do less dmg + should be affected by resilience and other dmg reduce effects.
Source (you should point out proofs of your report, please give us some source):
http://wowwiki.wikia.com/wiki/Deep_Wounds
There you can see the fact that it should be affected by resilience. Cant find a source about that it should do less dmg but there is something wrong when warriors do the most dmg with deepwounds in a whole bossfight and if the wounds do the same dmg as a 6.8k gs warrior with shadowmourne.
Pls analyse that talent with dev team
:)
|
1.0
|
[Warrior]Deep Wounds - Decription:
The damage from warriors talent "Deep Wounds" isnt affected by resilience or any other dmg reduce effect.
And the Base dmg is to High. 3,4k gs Warriors can stack that talent to 1300dmg per tick. On retail with Shadowmourne warrior i got that wounds on 1400dmg per tick in pve.
Quote:
PATCH 3.3.3: The damage from this talent is now affected by resilience as intended.
helping source:
http://wowwiki.wikia.com/wiki/Deep_Wounds
How it works:
Deep wounds arent effected by resilience or for expample painsub, dispersion, shadowform, shieldwall and all other things. So in Dispersion (90% dmg reduce) it does every 1 sec 600 dmg+.
How it should work:
That talent should do less dmg + should be affected by resilience and other dmg reduce effects.
Source (you should point out proofs of your report, please give us some source):
http://wowwiki.wikia.com/wiki/Deep_Wounds
There you can see the fact that it should be affected by resilience. Cant find a source about that it should do less dmg but there is something wrong when warriors do the most dmg with deepwounds in a whole bossfight and if the wounds do the same dmg as a 6.8k gs warrior with shadowmourne.
Pls analyse that talent with dev team
:)
|
test
|
deep wounds decription the damage from warriors talent deep wounds isnt affected by resilience or any other dmg reduce effect and the base dmg is to high gs warriors can stack that talent to per tick on retail with shadowmourne warrior i got that wounds on per tick in pve quote patch the damage from this talent is now affected by resilience as intended helping source how it works deep wounds arent effected by resilience or for expample painsub dispersion shadowform shieldwall and all other things so in dispersion dmg reduce it does every sec dmg how it should work that talent should do less dmg should be affected by resilience and other dmg reduce effects source you should point out proofs of your report please give us some source there you can see the fact that it should be affected by resilience cant find a source about that it should do less dmg but there is something wrong when warriors do the most dmg with deepwounds in a whole bossfight and if the wounds do the same dmg as a gs warrior with shadowmourne pls analyse that talent with dev team
| 1
|
122,505
| 10,225,332,902
|
IssuesEvent
|
2019-08-16 14:55:06
|
phetsims/energy-forms-and-changes
|
https://api.github.com/repos/phetsims/energy-forms-and-changes
|
opened
|
CT cannot read property name of undefined
|
type:automated-testing
|
```
energy-forms-and-changes : accessibility-fuzz : built : run
Query: fuzz&a11y&memoryLimit=1000
Uncaught TypeError: Cannot read property 'name' of undefined
TypeError: Cannot read property 'name' of undefined
at Object.getHeatTransferFactor (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:410992)
at A.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:1431513)
at e.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:1561001)
at e.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:1560656)
at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:3236413
at n.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:425381)
at t.stepSimulation (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:3247843)
at t.stepOneFrame (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:3247650)
at t.runAnimationLoop (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:3247511)
id: Bayes Chrome
Approximately 8/16/2019, 3:36:51 AM
energy-forms-and-changes : accessibility-fuzz : require.js : run
Query: brand=phet&ea&fuzz&a11y&memoryLimit=1000
Uncaught Error: Assertion failed: invalid category1: undefined
Error: Assertion failed: invalid category1: undefined
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/assert/js/assert.js:22:13)
at Object.getHeatTransferFactor (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/common/model/HeatTransferConstants.js?:77:17)
at BeakerHeater.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/BeakerHeater.js?:169:54)
at SystemsModel.stepModel (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/SystemsModel.js?:188:12)
at SystemsModel.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/SystemsModel.js?:169:14)
at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:217:22
at Action.execute (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/Action.js?:244:20)
at Sim.stepSimulation (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:933:33)
at Sim.stepOneFrame (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:914:14)
at Sim.runAnimationLoop (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:897:14)
id: Bayes Chrome
Approximately 8/16/2019, 3:36:51 AM
energy-forms-and-changes : accessibility-fuzzBoard : built : run
Query: fuzzBoard&a11y&memoryLimit=1000
Uncaught TypeError: Cannot read property 'name' of undefined
TypeError: Cannot read property 'name' of undefined
at Object.getHeatTransferFactor (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:410992)
at A.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:1431513)
at e.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:1561001)
at e.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:1560656)
at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:3236413
at n.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:425381)
at t.stepSimulation (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:3247843)
at t.stepOneFrame (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:3247650)
at t.runAnimationLoop (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:3247511)
id: Bayes Chrome
Approximately 8/16/2019, 3:36:51 AM
energy-forms-and-changes : fuzz : require.js : run
Query: brand=phet&ea&fuzz&memoryLimit=1000
Uncaught Error: Assertion failed: value failed isValidValue: -3.1164
Error: Assertion failed: value failed isValidValue: -3.1164
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/assert/js/assert.js:22:13)
at Object.isValueValid (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/ValidatorDef.js?:267:41)
at validate (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/validate.js?:28:20)
at NumberProperty.assertNumberPropertyValidateValue (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/NumberProperty.js?:67:26)
at NumberProperty.set (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/NumberProperty.js?:89:54)
at EFACIntroModel.fallToSurface (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/intro/model/EFACIntroModel.js?:743:47)
at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/intro/model/EFACIntroModel.js?:338:16
at Array.forEach (<anonymous>)
at EFACIntroModel.stepModel (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/intro/model/EFACIntroModel.js?:332:30)
at EFACIntroModel.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/intro/model/EFACIntroModel.js?:314:14)
id: Bayes Chrome
Approximately 8/16/2019, 3:36:51 AM
energy-forms-and-changes : xss-fuzz : run
Query: brand=phet&ea&fuzz&stringTest=xss&memoryLimit=1000
Uncaught Error: Assertion failed: invalid category1: undefined
Error: Assertion failed: invalid category1: undefined
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/assert/js/assert.js:22:13)
at Object.getHeatTransferFactor (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/common/model/HeatTransferConstants.js?:77:17)
at BeakerHeater.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/BeakerHeater.js?:169:54)
at SystemsModel.stepModel (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/SystemsModel.js?:188:12)
at SystemsModel.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/SystemsModel.js?:169:14)
at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:217:22
at Action.execute (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/Action.js?:244:20)
at Sim.stepSimulation (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:933:33)
at Sim.stepOneFrame (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:914:14)
at Sim.runAnimationLoop (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:897:14)
id: Bayes Chrome
Approximately 8/16/2019, 3:36:51 AM
```
|
1.0
|
CT cannot read property name of undefined - ```
energy-forms-and-changes : accessibility-fuzz : built : run
Query: fuzz&a11y&memoryLimit=1000
Uncaught TypeError: Cannot read property 'name' of undefined
TypeError: Cannot read property 'name' of undefined
at Object.getHeatTransferFactor (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:410992)
at A.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:1431513)
at e.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:1561001)
at e.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:1560656)
at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:3236413
at n.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:425381)
at t.stepSimulation (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:3247843)
at t.stepOneFrame (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:3247650)
at t.runAnimationLoop (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:3247511)
id: Bayes Chrome
Approximately 8/16/2019, 3:36:51 AM
energy-forms-and-changes : accessibility-fuzz : require.js : run
Query: brand=phet&ea&fuzz&a11y&memoryLimit=1000
Uncaught Error: Assertion failed: invalid category1: undefined
Error: Assertion failed: invalid category1: undefined
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/assert/js/assert.js:22:13)
at Object.getHeatTransferFactor (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/common/model/HeatTransferConstants.js?:77:17)
at BeakerHeater.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/BeakerHeater.js?:169:54)
at SystemsModel.stepModel (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/SystemsModel.js?:188:12)
at SystemsModel.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/SystemsModel.js?:169:14)
at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:217:22
at Action.execute (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/Action.js?:244:20)
at Sim.stepSimulation (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:933:33)
at Sim.stepOneFrame (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:914:14)
at Sim.runAnimationLoop (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:897:14)
id: Bayes Chrome
Approximately 8/16/2019, 3:36:51 AM
energy-forms-and-changes : accessibility-fuzzBoard : built : run
Query: fuzzBoard&a11y&memoryLimit=1000
Uncaught TypeError: Cannot read property 'name' of undefined
TypeError: Cannot read property 'name' of undefined
at Object.getHeatTransferFactor (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:410992)
at A.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:1431513)
at e.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:1561001)
at e.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:1560656)
at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:3236413
at n.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:425381)
at t.stepSimulation (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:3247843)
at t.stepOneFrame (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:3247650)
at t.runAnimationLoop (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:3247511)
id: Bayes Chrome
Approximately 8/16/2019, 3:36:51 AM
energy-forms-and-changes : fuzz : require.js : run
Query: brand=phet&ea&fuzz&memoryLimit=1000
Uncaught Error: Assertion failed: value failed isValidValue: -3.1164
Error: Assertion failed: value failed isValidValue: -3.1164
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/assert/js/assert.js:22:13)
at Object.isValueValid (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/ValidatorDef.js?:267:41)
at validate (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/validate.js?:28:20)
at NumberProperty.assertNumberPropertyValidateValue (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/NumberProperty.js?:67:26)
at NumberProperty.set (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/NumberProperty.js?:89:54)
at EFACIntroModel.fallToSurface (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/intro/model/EFACIntroModel.js?:743:47)
at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/intro/model/EFACIntroModel.js?:338:16
at Array.forEach (<anonymous>)
at EFACIntroModel.stepModel (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/intro/model/EFACIntroModel.js?:332:30)
at EFACIntroModel.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/intro/model/EFACIntroModel.js?:314:14)
id: Bayes Chrome
Approximately 8/16/2019, 3:36:51 AM
energy-forms-and-changes : xss-fuzz : run
Query: brand=phet&ea&fuzz&stringTest=xss&memoryLimit=1000
Uncaught Error: Assertion failed: invalid category1: undefined
Error: Assertion failed: invalid category1: undefined
at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/assert/js/assert.js:22:13)
at Object.getHeatTransferFactor (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/common/model/HeatTransferConstants.js?:77:17)
at BeakerHeater.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/BeakerHeater.js?:169:54)
at SystemsModel.stepModel (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/SystemsModel.js?:188:12)
at SystemsModel.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/SystemsModel.js?:169:14)
at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:217:22
at Action.execute (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/Action.js?:244:20)
at Sim.stepSimulation (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:933:33)
at Sim.stepOneFrame (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:914:14)
at Sim.runAnimationLoop (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:897:14)
id: Bayes Chrome
Approximately 8/16/2019, 3:36:51 AM
```
|
test
|
ct cannot read property name of undefined energy forms and changes accessibility fuzz built run query fuzz memorylimit uncaught typeerror cannot read property name of undefined typeerror cannot read property name of undefined at object getheattransferfactor at a value at e value at e value at at n value at t stepsimulation at t steponeframe at t runanimationloop id bayes chrome approximately am energy forms and changes accessibility fuzz require js run query brand phet ea fuzz memorylimit uncaught error assertion failed invalid undefined error assertion failed invalid undefined at window assertions assertfunction at object getheattransferfactor at beakerheater step at systemsmodel stepmodel at systemsmodel step at at action execute at sim stepsimulation at sim steponeframe at sim runanimationloop id bayes chrome approximately am energy forms and changes accessibility fuzzboard built run query fuzzboard memorylimit uncaught typeerror cannot read property name of undefined typeerror cannot read property name of undefined at object getheattransferfactor at a value at e value at e value at at n value at t stepsimulation at t steponeframe at t runanimationloop id bayes chrome approximately am energy forms and changes fuzz require js run query brand phet ea fuzz memorylimit uncaught error assertion failed value failed isvalidvalue error assertion failed value failed isvalidvalue at window assertions assertfunction at object isvaluevalid at validate at numberproperty assertnumberpropertyvalidatevalue at numberproperty set at efacintromodel falltosurface at at array foreach at efacintromodel stepmodel at efacintromodel step id bayes chrome approximately am energy forms and changes xss fuzz run query brand phet ea fuzz stringtest xss memorylimit uncaught error assertion failed invalid undefined error assertion failed invalid undefined at window assertions assertfunction at object getheattransferfactor at beakerheater step at systemsmodel stepmodel at systemsmodel step at at action execute at sim stepsimulation at sim steponeframe at sim runanimationloop id bayes chrome approximately am
| 1
|
194,831
| 14,689,666,811
|
IssuesEvent
|
2021-01-02 11:09:16
|
github-vet/rangeloop-pointer-findings
|
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
|
closed
|
kalyabin/vcsview: git_test.go; 53 LoC
|
fresh medium test
|
Found a possible issue in [kalyabin/vcsview](https://www.github.com/kalyabin/vcsview) at [git_test.go](https://github.com/kalyabin/vcsview/blob/cfe5e010e0206d6b98caad942e0322b2150ef7cd/git_test.go#L414-L466)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable key used in defer or goroutine at line 437
[Click here to see the code in its original context.](https://github.com/kalyabin/vcsview/blob/cfe5e010e0206d6b98caad942e0322b2150ef7cd/git_test.go#L414-L466)
<details>
<summary>Click here to show the 53 line(s) of Go which triggered the analyzer.</summary>
```go
for key, testCase := range cases {
var (
gotError error
gotCommit int
)
result := make(chan Commit, 1)
e := g.ReadHistory(testCase.repoPath, testCase.path, testCase.branch, testCase.offset, testCase.limit, result)
wg := sync.WaitGroup{}
wg.Add(1)
go func() {
defer wg.Done()
loop: for {
select {
case <-e.ctx.Done():
close(result)
break loop
case commit := <- result:
gotCommit++
if commit.Id() == "" {
t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty identifier", key, testCase)
}
if len(commit.Parents()) == 0 {
t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty parents", key, testCase)
}
if commit.Message() == "" {
t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty message", key, testCase)
}
if commit.Author().String() == "" {
t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty author", key, testCase)
}
if commit.Date().Unix() < 0 {
t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty date", key, testCase)
}
}
}
}()
gotError = e.Run()
wg.Wait()
if gotError != nil {
t.Fatalf("[%d] Git.ReadHistory(%v) has error: %v, want no errors", key, testCase, gotError)
}
if gotCommit != testCase.limit {
t.Fatalf("[%d] Git.ReadHistory(%v) got %v commits, want: %v", key, testCase, gotCommit, testCase.limit)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: cfe5e010e0206d6b98caad942e0322b2150ef7cd
|
1.0
|
kalyabin/vcsview: git_test.go; 53 LoC -
Found a possible issue in [kalyabin/vcsview](https://www.github.com/kalyabin/vcsview) at [git_test.go](https://github.com/kalyabin/vcsview/blob/cfe5e010e0206d6b98caad942e0322b2150ef7cd/git_test.go#L414-L466)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable key used in defer or goroutine at line 437
[Click here to see the code in its original context.](https://github.com/kalyabin/vcsview/blob/cfe5e010e0206d6b98caad942e0322b2150ef7cd/git_test.go#L414-L466)
<details>
<summary>Click here to show the 53 line(s) of Go which triggered the analyzer.</summary>
```go
for key, testCase := range cases {
var (
gotError error
gotCommit int
)
result := make(chan Commit, 1)
e := g.ReadHistory(testCase.repoPath, testCase.path, testCase.branch, testCase.offset, testCase.limit, result)
wg := sync.WaitGroup{}
wg.Add(1)
go func() {
defer wg.Done()
loop: for {
select {
case <-e.ctx.Done():
close(result)
break loop
case commit := <- result:
gotCommit++
if commit.Id() == "" {
t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty identifier", key, testCase)
}
if len(commit.Parents()) == 0 {
t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty parents", key, testCase)
}
if commit.Message() == "" {
t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty message", key, testCase)
}
if commit.Author().String() == "" {
t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty author", key, testCase)
}
if commit.Date().Unix() < 0 {
t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty date", key, testCase)
}
}
}
}()
gotError = e.Run()
wg.Wait()
if gotError != nil {
t.Fatalf("[%d] Git.ReadHistory(%v) has error: %v, want no errors", key, testCase, gotError)
}
if gotCommit != testCase.limit {
t.Fatalf("[%d] Git.ReadHistory(%v) got %v commits, want: %v", key, testCase, gotCommit, testCase.limit)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: cfe5e010e0206d6b98caad942e0322b2150ef7cd
|
test
|
kalyabin vcsview git test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable key used in defer or goroutine at line click here to show the line s of go which triggered the analyzer go for key testcase range cases var goterror error gotcommit int result make chan commit e g readhistory testcase repopath testcase path testcase branch testcase offset testcase limit result wg sync waitgroup wg add go func defer wg done loop for select case e ctx done close result break loop case commit result gotcommit if commit id t fatalf git readhistory v commit has empty identifier key testcase if len commit parents t fatalf git readhistory v commit has empty parents key testcase if commit message t fatalf git readhistory v commit has empty message key testcase if commit author string t fatalf git readhistory v commit has empty author key testcase if commit date unix t fatalf git readhistory v commit has empty date key testcase goterror e run wg wait if goterror nil t fatalf git readhistory v has error v want no errors key testcase goterror if gotcommit testcase limit t fatalf git readhistory v got v commits want v key testcase gotcommit testcase limit leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 1
|
288,055
| 24,882,269,637
|
IssuesEvent
|
2022-10-28 03:01:04
|
MPMG-DCC-UFMG/F01
|
https://api.github.com/repos/MPMG-DCC-UFMG/F01
|
closed
|
Teste de generalizacao para a tag Orçamento - Legislação - Jequitinhonha
|
generalization test development template - Memory (66) tag - Orçamento subtag - Legislação
|
DoD: Realizar o teste de Generalização do validador da tag Orçamento - Legislação para o Município de Jequitinhonha.
|
1.0
|
Teste de generalizacao para a tag Orçamento - Legislação - Jequitinhonha - DoD: Realizar o teste de Generalização do validador da tag Orçamento - Legislação para o Município de Jequitinhonha.
|
test
|
teste de generalizacao para a tag orçamento legislação jequitinhonha dod realizar o teste de generalização do validador da tag orçamento legislação para o município de jequitinhonha
| 1
|
131,939
| 18,262,075,889
|
IssuesEvent
|
2021-10-04 00:55:31
|
ghc-dev/Steve-Chambers-MD
|
https://api.github.com/repos/ghc-dev/Steve-Chambers-MD
|
opened
|
CVE-2020-13956 (Medium) detected in httpclient-4.5.10.jar
|
security vulnerability
|
## CVE-2020-13956 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>httpclient-4.5.10.jar</b></p></summary>
<p>Apache HttpComponents Client</p>
<p>Path to dependency file: Steve-Chambers-MD/build.gradle</p>
<p>Path to vulnerable library: hes/modules-2/files-2.1/org.apache.httpcomponents/httpclient/4.5.10/7ca2e4276f4ef95e4db725a8cd4a1d1e7585b9e5/httpclient-4.5.10.jar</p>
<p>
Dependency Hierarchy:
- :x: **httpclient-4.5.10.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Steve-Chambers-MD/commit/a94641a73803572678190ca83e1504bf0ea37d1f">a94641a73803572678190ca83e1504bf0ea37d1f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution.
<p>Publish Date: 2020-12-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956>CVE-2020-13956</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.httpcomponents","packageName":"httpclient","packageVersion":"4.5.10","packageFilePaths":["/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"org.apache.httpcomponents:httpclient:4.5.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-13956","vulnerabilityDetails":"Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-13956 (Medium) detected in httpclient-4.5.10.jar - ## CVE-2020-13956 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>httpclient-4.5.10.jar</b></p></summary>
<p>Apache HttpComponents Client</p>
<p>Path to dependency file: Steve-Chambers-MD/build.gradle</p>
<p>Path to vulnerable library: hes/modules-2/files-2.1/org.apache.httpcomponents/httpclient/4.5.10/7ca2e4276f4ef95e4db725a8cd4a1d1e7585b9e5/httpclient-4.5.10.jar</p>
<p>
Dependency Hierarchy:
- :x: **httpclient-4.5.10.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Steve-Chambers-MD/commit/a94641a73803572678190ca83e1504bf0ea37d1f">a94641a73803572678190ca83e1504bf0ea37d1f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution.
<p>Publish Date: 2020-12-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956>CVE-2020-13956</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.httpcomponents","packageName":"httpclient","packageVersion":"4.5.10","packageFilePaths":["/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"org.apache.httpcomponents:httpclient:4.5.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-13956","vulnerabilityDetails":"Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
non_test
|
cve medium detected in httpclient jar cve medium severity vulnerability vulnerable library httpclient jar apache httpcomponents client path to dependency file steve chambers md build gradle path to vulnerable library hes modules files org apache httpcomponents httpclient httpclient jar dependency hierarchy x httpclient jar vulnerable library found in head commit a href found in base branch master vulnerability details apache httpclient versions prior to version and can misinterpret malformed authority component in request uris passed to the library as java net uri object and pick the wrong target host for request execution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache httpcomponents httpclient org apache httpcomponents httpclient osgi org apache httpcomponents org apache httpcomponents osgi isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org apache httpcomponents httpclient isminimumfixversionavailable true minimumfixversion org apache httpcomponents httpclient org apache httpcomponents httpclient osgi org apache httpcomponents org apache httpcomponents osgi basebranches vulnerabilityidentifier cve vulnerabilitydetails apache httpclient versions prior to version and can misinterpret malformed authority component in request uris passed to the library as java net uri object and pick the wrong target host for request execution vulnerabilityurl
| 0
|
311,451
| 26,792,744,537
|
IssuesEvent
|
2023-02-01 09:42:47
|
akademia-envelo-3/MELO-front
|
https://api.github.com/repos/akademia-envelo-3/MELO-front
|
closed
|
[MKT-05] Uszczegółowienie makiety admina odnośnie kategorii
|
frontend UI sp1 test-ok
|
### Dodatkowe informacje
Należy stworzyć makietę szczegółową dla widoku kategorii dla admina. Makieta ta składać się będzie z widoku wszystkich kategorii, widoku edycji kategorii oraz widoku dodania nowej kategorii. Makieta ma przedstawiać zarówno wersję widoku dla desktop oraz mobile.
### Makiety
[Makieta 01](https://www.figma.com/file/ybxNkntmTWg7yLBzHhBx4R/MELO?node-id=333%3A8921&t=wFubEuu6vVmKgmdq-0)
[Makieta 02](https://www.figma.com/file/ybxNkntmTWg7yLBzHhBx4R/MELO?node-id=333%3A8316&t=wFubEuu6vVmKgmdq-0)
[Makieta 03](https://www.figma.com/file/ybxNkntmTWg7yLBzHhBx4R/MELO?node-id=394%3A8841&t=wFubEuu6vVmKgmdq-0)
### Kryteria akceptacji
1. Została utworzona makieta szczegółowa dla widoku wszystkich kategorii
2. Została utworzona makieta szczegółowa dla dodania nowej kategorii
3. Została utworzona makieta szczegółowa dla widoku edycji kategorii
|
1.0
|
[MKT-05] Uszczegółowienie makiety admina odnośnie kategorii - ### Dodatkowe informacje
Należy stworzyć makietę szczegółową dla widoku kategorii dla admina. Makieta ta składać się będzie z widoku wszystkich kategorii, widoku edycji kategorii oraz widoku dodania nowej kategorii. Makieta ma przedstawiać zarówno wersję widoku dla desktop oraz mobile.
### Makiety
[Makieta 01](https://www.figma.com/file/ybxNkntmTWg7yLBzHhBx4R/MELO?node-id=333%3A8921&t=wFubEuu6vVmKgmdq-0)
[Makieta 02](https://www.figma.com/file/ybxNkntmTWg7yLBzHhBx4R/MELO?node-id=333%3A8316&t=wFubEuu6vVmKgmdq-0)
[Makieta 03](https://www.figma.com/file/ybxNkntmTWg7yLBzHhBx4R/MELO?node-id=394%3A8841&t=wFubEuu6vVmKgmdq-0)
### Kryteria akceptacji
1. Została utworzona makieta szczegółowa dla widoku wszystkich kategorii
2. Została utworzona makieta szczegółowa dla dodania nowej kategorii
3. Została utworzona makieta szczegółowa dla widoku edycji kategorii
|
test
|
uszczegółowienie makiety admina odnośnie kategorii dodatkowe informacje należy stworzyć makietę szczegółową dla widoku kategorii dla admina makieta ta składać się będzie z widoku wszystkich kategorii widoku edycji kategorii oraz widoku dodania nowej kategorii makieta ma przedstawiać zarówno wersję widoku dla desktop oraz mobile makiety kryteria akceptacji została utworzona makieta szczegółowa dla widoku wszystkich kategorii została utworzona makieta szczegółowa dla dodania nowej kategorii została utworzona makieta szczegółowa dla widoku edycji kategorii
| 1
|
129,308
| 10,569,527,896
|
IssuesEvent
|
2019-10-06 20:07:59
|
mgba-emu/mgba
|
https://api.github.com/repos/mgba-emu/mgba
|
closed
|
[GBA] Volume of audio channels changes frequently
|
category:needs retest platform:GBA severity:regression
|
Compared to mGBA 0.6.3, the development build's volume for GBA games seems to frequently change in an odd way, causing some of the audio channels to sound louder than usual in comparison for brief moments. This can be noticed in games like the Sonic Advance games.
Also, the following is probably unrelated to this issue, but the fps counter in the development build indicates that the fps is hovering around 65.6 fps even though I have the target fps set to the native amount.
|
1.0
|
[GBA] Volume of audio channels changes frequently - Compared to mGBA 0.6.3, the development build's volume for GBA games seems to frequently change in an odd way, causing some of the audio channels to sound louder than usual in comparison for brief moments. This can be noticed in games like the Sonic Advance games.
Also, the following is probably unrelated to this issue, but the fps counter in the development build indicates that the fps is hovering around 65.6 fps even though I have the target fps set to the native amount.
|
test
|
volume of audio channels changes frequently compared to mgba the development build s volume for gba games seems to frequently change in an odd way causing some of the audio channels to sound louder than usual in comparison for brief moments this can be noticed in games like the sonic advance games also the following is probably unrelated to this issue but the fps counter in the development build indicates that the fps is hovering around fps even though i have the target fps set to the native amount
| 1
|
5,903
| 2,798,041,191
|
IssuesEvent
|
2015-05-12 16:34:37
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
Test failure in CI build 2339
|
PTAL test-failure
|
The following test appears to have failed:
[#2339](https://circleci.com/gh/cockroachdb/cockroach/2339):
```
level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="770" msg="node 100000001: group 5 raft ready"
level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="775" msg="HardState updated: {Term:6 Vote:4294967297 Commit:13 XXX_unrecognized:[]}"
level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="778" msg="New Entry[0]: 6/13 EntryNormal 00000000000000006b4ee14837a7cb41: raft_id:5 origin_node_id:4294967297 cmd:<internal_resolve_intent:<header:<timestamp:<wall_time:0 logical:933 > cmd_id:<wall_time:0 random:0 > key:\"\\0"
level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="781" msg="Committed Entry[0]: 6/13 EntryNormal 00000000000000006b4ee14837a7cb41: raft_id:5 origin_node_id:4294967297 cmd:<internal_resolve_intent:<header:<timestamp:<wall_time:0 logical:933 > cmd_id:<wall_time:0 random:0 > key:\"\\0"
==================
WARNING: DATA RACE
Read by goroutine 79:
github.com/cockroachdb/cockroach/storage.(*rangeStats).GetSize()
/go/src/github.com/cockroachdb/cockroach/storage/stats.go:68 +0x96
github.com/cockroachdb/cockroach/storage.(*splitQueue).process()
/go/src/github.com/cockroachdb/cockroach/storage/split_queue.go:103 +0x81d
github.com/cockroachdb/cockroach/storage.(*baseQueue).processOne()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:261 +0x5ce
github.com/cockroachdb/cockroach/storage.func·011()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:214 +0x1b7
github.com/cockroachdb/cockroach/util.func·001()
/go/src/github.com/cockroachdb/cockroach/util/stopper.go:73 +0x7f
Previous write by goroutine 87:
github.com/cockroachdb/cockroach/proto.(*MVCCStats).Add()
/go/src/github.com/cockroachdb/cockroach/proto/data.go:630 +0xa6
github.com/cockroachdb/cockroach/storage.(*rangeStats).MergeMVCCStats()
/go/src/github.com/cockroachdb/cockroach/storage/stats.go:82 +0x25c
github.com/cockroachdb/cockroach/storage.(*Range).applyRaftCommand()
/go/src/github.com/cockroachdb/cockroach/storage/range.go:845 +0xca3
github.com/cockroachdb/cockroach/storage.(*Range).processRaftCommand()
/go/src/github.com/cockroachdb/cockroach/storage/range.go:767 +0x3b5
github.com/cockroachdb/cockroach/storage.func·036()
/go/src/github.com/cockroachdb/cockroach/storage/store.go:1339 +0xce8
--
/go/src/github.com/cockroachdb/cockroach/kv/split_test.go:137 +0x7d
testing.tRunner()
/usr/src/go/src/testing/testing.go:447 +0x133
==================
==================
WARNING: DATA RACE
Read by goroutine 79:
github.com/cockroachdb/cockroach/storage.(*rangeStats).GetSize()
/go/src/github.com/cockroachdb/cockroach/storage/stats.go:68 +0xae
github.com/cockroachdb/cockroach/storage.(*splitQueue).process()
/go/src/github.com/cockroachdb/cockroach/storage/split_queue.go:103 +0x81d
github.com/cockroachdb/cockroach/storage.(*baseQueue).processOne()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:261 +0x5ce
github.com/cockroachdb/cockroach/storage.func·011()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:214 +0x1b7
github.com/cockroachdb/cockroach/util.func·001()
/go/src/github.com/cockroachdb/cockroach/util/stopper.go:73 +0x7f
Previous write by goroutine 87:
github.com/cockroachdb/cockroach/proto.(*MVCCStats).Add()
/go/src/github.com/cockroachdb/cockroach/proto/data.go:631 +0x107
github.com/cockroachdb/cockroach/storage.(*rangeStats).MergeMVCCStats()
/go/src/github.com/cockroachdb/cockroach/storage/stats.go:82 +0x25c
github.com/cockroachdb/cockroach/storage.(*Range).applyRaftCommand()
/go/src/github.com/cockroachdb/cockroach/storage/range.go:845 +0xca3
github.com/cockroachdb/cockroach/storage.(*Range).processRaftCommand()
/go/src/github.com/cockroachdb/cockroach/storage/range.go:767 +0x3b5
github.com/cockroachdb/cockroach/storage.func·036()
/go/src/github.com/cockroachdb/cockroach/storage/store.go:1339 +0xce8
level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="770" msg="node 100000001: group 5 raft ready"
level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="775" msg="HardState updated: {Term:6 Vote:4294967297 Commit:13 XXX_unrecognized:[]}"
level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="778" msg="New Entry[0]: 6/13 EntryNormal 00000000000000006b4ee14837a7cb41: raft_id:5 origin_node_id:4294967297 cmd:<internal_resolve_intent:<header:<timestamp:<wall_time:0 logical:933 > cmd_id:<wall_time:0 random:0 > key:\"\\0"
level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="781" msg="Committed Entry[0]: 6/13 EntryNormal 00000000000000006b4ee14837a7cb41: raft_id:5 origin_node_id:4294967297 cmd:<internal_resolve_intent:<header:<timestamp:<wall_time:0 logical:933 > cmd_id:<wall_time:0 random:0 > key:\"\\0"
==================
WARNING: DATA RACE
Read by goroutine 79:
github.com/cockroachdb/cockroach/storage.(*rangeStats).GetSize()
/go/src/github.com/cockroachdb/cockroach/storage/stats.go:68 +0x96
github.com/cockroachdb/cockroach/storage.(*splitQueue).process()
/go/src/github.com/cockroachdb/cockroach/storage/split_queue.go:103 +0x81d
github.com/cockroachdb/cockroach/storage.(*baseQueue).processOne()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:261 +0x5ce
github.com/cockroachdb/cockroach/storage.func·011()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:214 +0x1b7
github.com/cockroachdb/cockroach/util.func·001()
/go/src/github.com/cockroachdb/cockroach/util/stopper.go:73 +0x7f
Previous write by goroutine 87:
github.com/cockroachdb/cockroach/proto.(*MVCCStats).Add()
/go/src/github.com/cockroachdb/cockroach/proto/data.go:630 +0xa6
github.com/cockroachdb/cockroach/storage.(*rangeStats).MergeMVCCStats()
/go/src/github.com/cockroachdb/cockroach/storage/stats.go:82 +0x25c
github.com/cockroachdb/cockroach/storage.(*Range).applyRaftCommand()
/go/src/github.com/cockroachdb/cockroach/storage/range.go:845 +0xca3
github.com/cockroachdb/cockroach/storage.(*Range).processRaftCommand()
/go/src/github.com/cockroachdb/cockroach/storage/range.go:767 +0x3b5
github.com/cockroachdb/cockroach/storage.func·036()
/go/src/github.com/cockroachdb/cockroach/storage/store.go:1339 +0xce8
--
/go/src/github.com/cockroachdb/cockroach/kv/split_test.go:137 +0x7d
testing.tRunner()
/usr/src/go/src/testing/testing.go:447 +0x133
==================
==================
WARNING: DATA RACE
Read by goroutine 79:
github.com/cockroachdb/cockroach/storage.(*rangeStats).GetSize()
/go/src/github.com/cockroachdb/cockroach/storage/stats.go:68 +0xae
github.com/cockroachdb/cockroach/storage.(*splitQueue).process()
/go/src/github.com/cockroachdb/cockroach/storage/split_queue.go:103 +0x81d
github.com/cockroachdb/cockroach/storage.(*baseQueue).processOne()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:261 +0x5ce
github.com/cockroachdb/cockroach/storage.func·011()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:214 +0x1b7
github.com/cockroachdb/cockroach/util.func·001()
/go/src/github.com/cockroachdb/cockroach/util/stopper.go:73 +0x7f
Previous write by goroutine 87:
github.com/cockroachdb/cockroach/proto.(*MVCCStats).Add()
/go/src/github.com/cockroachdb/cockroach/proto/data.go:631 +0x107
github.com/cockroachdb/cockroach/storage.(*rangeStats).MergeMVCCStats()
/go/src/github.com/cockroachdb/cockroach/storage/stats.go:82 +0x25c
github.com/cockroachdb/cockroach/storage.(*Range).applyRaftCommand()
/go/src/github.com/cockroachdb/cockroach/storage/range.go:845 +0xca3
github.com/cockroachdb/cockroach/storage.(*Range).processRaftCommand()
/go/src/github.com/cockroachdb/cockroach/storage/range.go:767 +0x3b5
github.com/cockroachdb/cockroach/storage.func·036()
/go/src/github.com/cockroachdb/cockroach/storage/store.go:1339 +0xce8
```
Please assign, take a look and update the issue accordingly.
|
1.0
|
Test failure in CI build 2339 - The following test appears to have failed:
[#2339](https://circleci.com/gh/cockroachdb/cockroach/2339):
```
level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="770" msg="node 100000001: group 5 raft ready"
level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="775" msg="HardState updated: {Term:6 Vote:4294967297 Commit:13 XXX_unrecognized:[]}"
level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="778" msg="New Entry[0]: 6/13 EntryNormal 00000000000000006b4ee14837a7cb41: raft_id:5 origin_node_id:4294967297 cmd:<internal_resolve_intent:<header:<timestamp:<wall_time:0 logical:933 > cmd_id:<wall_time:0 random:0 > key:\"\\0"
level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="781" msg="Committed Entry[0]: 6/13 EntryNormal 00000000000000006b4ee14837a7cb41: raft_id:5 origin_node_id:4294967297 cmd:<internal_resolve_intent:<header:<timestamp:<wall_time:0 logical:933 > cmd_id:<wall_time:0 random:0 > key:\"\\0"
==================
WARNING: DATA RACE
Read by goroutine 79:
github.com/cockroachdb/cockroach/storage.(*rangeStats).GetSize()
/go/src/github.com/cockroachdb/cockroach/storage/stats.go:68 +0x96
github.com/cockroachdb/cockroach/storage.(*splitQueue).process()
/go/src/github.com/cockroachdb/cockroach/storage/split_queue.go:103 +0x81d
github.com/cockroachdb/cockroach/storage.(*baseQueue).processOne()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:261 +0x5ce
github.com/cockroachdb/cockroach/storage.func·011()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:214 +0x1b7
github.com/cockroachdb/cockroach/util.func·001()
/go/src/github.com/cockroachdb/cockroach/util/stopper.go:73 +0x7f
Previous write by goroutine 87:
github.com/cockroachdb/cockroach/proto.(*MVCCStats).Add()
/go/src/github.com/cockroachdb/cockroach/proto/data.go:630 +0xa6
github.com/cockroachdb/cockroach/storage.(*rangeStats).MergeMVCCStats()
/go/src/github.com/cockroachdb/cockroach/storage/stats.go:82 +0x25c
github.com/cockroachdb/cockroach/storage.(*Range).applyRaftCommand()
/go/src/github.com/cockroachdb/cockroach/storage/range.go:845 +0xca3
github.com/cockroachdb/cockroach/storage.(*Range).processRaftCommand()
/go/src/github.com/cockroachdb/cockroach/storage/range.go:767 +0x3b5
github.com/cockroachdb/cockroach/storage.func·036()
/go/src/github.com/cockroachdb/cockroach/storage/store.go:1339 +0xce8
--
/go/src/github.com/cockroachdb/cockroach/kv/split_test.go:137 +0x7d
testing.tRunner()
/usr/src/go/src/testing/testing.go:447 +0x133
==================
==================
WARNING: DATA RACE
Read by goroutine 79:
github.com/cockroachdb/cockroach/storage.(*rangeStats).GetSize()
/go/src/github.com/cockroachdb/cockroach/storage/stats.go:68 +0xae
github.com/cockroachdb/cockroach/storage.(*splitQueue).process()
/go/src/github.com/cockroachdb/cockroach/storage/split_queue.go:103 +0x81d
github.com/cockroachdb/cockroach/storage.(*baseQueue).processOne()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:261 +0x5ce
github.com/cockroachdb/cockroach/storage.func·011()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:214 +0x1b7
github.com/cockroachdb/cockroach/util.func·001()
/go/src/github.com/cockroachdb/cockroach/util/stopper.go:73 +0x7f
Previous write by goroutine 87:
github.com/cockroachdb/cockroach/proto.(*MVCCStats).Add()
/go/src/github.com/cockroachdb/cockroach/proto/data.go:631 +0x107
github.com/cockroachdb/cockroach/storage.(*rangeStats).MergeMVCCStats()
/go/src/github.com/cockroachdb/cockroach/storage/stats.go:82 +0x25c
github.com/cockroachdb/cockroach/storage.(*Range).applyRaftCommand()
/go/src/github.com/cockroachdb/cockroach/storage/range.go:845 +0xca3
github.com/cockroachdb/cockroach/storage.(*Range).processRaftCommand()
/go/src/github.com/cockroachdb/cockroach/storage/range.go:767 +0x3b5
github.com/cockroachdb/cockroach/storage.func·036()
/go/src/github.com/cockroachdb/cockroach/storage/store.go:1339 +0xce8
level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="770" msg="node 100000001: group 5 raft ready"
level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="775" msg="HardState updated: {Term:6 Vote:4294967297 Commit:13 XXX_unrecognized:[]}"
level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="778" msg="New Entry[0]: 6/13 EntryNormal 00000000000000006b4ee14837a7cb41: raft_id:5 origin_node_id:4294967297 cmd:<internal_resolve_intent:<header:<timestamp:<wall_time:0 logical:933 > cmd_id:<wall_time:0 random:0 > key:\"\\0"
level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="781" msg="Committed Entry[0]: 6/13 EntryNormal 00000000000000006b4ee14837a7cb41: raft_id:5 origin_node_id:4294967297 cmd:<internal_resolve_intent:<header:<timestamp:<wall_time:0 logical:933 > cmd_id:<wall_time:0 random:0 > key:\"\\0"
==================
WARNING: DATA RACE
Read by goroutine 79:
github.com/cockroachdb/cockroach/storage.(*rangeStats).GetSize()
/go/src/github.com/cockroachdb/cockroach/storage/stats.go:68 +0x96
github.com/cockroachdb/cockroach/storage.(*splitQueue).process()
/go/src/github.com/cockroachdb/cockroach/storage/split_queue.go:103 +0x81d
github.com/cockroachdb/cockroach/storage.(*baseQueue).processOne()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:261 +0x5ce
github.com/cockroachdb/cockroach/storage.func·011()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:214 +0x1b7
github.com/cockroachdb/cockroach/util.func·001()
/go/src/github.com/cockroachdb/cockroach/util/stopper.go:73 +0x7f
Previous write by goroutine 87:
github.com/cockroachdb/cockroach/proto.(*MVCCStats).Add()
/go/src/github.com/cockroachdb/cockroach/proto/data.go:630 +0xa6
github.com/cockroachdb/cockroach/storage.(*rangeStats).MergeMVCCStats()
/go/src/github.com/cockroachdb/cockroach/storage/stats.go:82 +0x25c
github.com/cockroachdb/cockroach/storage.(*Range).applyRaftCommand()
/go/src/github.com/cockroachdb/cockroach/storage/range.go:845 +0xca3
github.com/cockroachdb/cockroach/storage.(*Range).processRaftCommand()
/go/src/github.com/cockroachdb/cockroach/storage/range.go:767 +0x3b5
github.com/cockroachdb/cockroach/storage.func·036()
/go/src/github.com/cockroachdb/cockroach/storage/store.go:1339 +0xce8
--
/go/src/github.com/cockroachdb/cockroach/kv/split_test.go:137 +0x7d
testing.tRunner()
/usr/src/go/src/testing/testing.go:447 +0x133
==================
==================
WARNING: DATA RACE
Read by goroutine 79:
github.com/cockroachdb/cockroach/storage.(*rangeStats).GetSize()
/go/src/github.com/cockroachdb/cockroach/storage/stats.go:68 +0xae
github.com/cockroachdb/cockroach/storage.(*splitQueue).process()
/go/src/github.com/cockroachdb/cockroach/storage/split_queue.go:103 +0x81d
github.com/cockroachdb/cockroach/storage.(*baseQueue).processOne()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:261 +0x5ce
github.com/cockroachdb/cockroach/storage.func·011()
/go/src/github.com/cockroachdb/cockroach/storage/queue.go:214 +0x1b7
github.com/cockroachdb/cockroach/util.func·001()
/go/src/github.com/cockroachdb/cockroach/util/stopper.go:73 +0x7f
Previous write by goroutine 87:
github.com/cockroachdb/cockroach/proto.(*MVCCStats).Add()
/go/src/github.com/cockroachdb/cockroach/proto/data.go:631 +0x107
github.com/cockroachdb/cockroach/storage.(*rangeStats).MergeMVCCStats()
/go/src/github.com/cockroachdb/cockroach/storage/stats.go:82 +0x25c
github.com/cockroachdb/cockroach/storage.(*Range).applyRaftCommand()
/go/src/github.com/cockroachdb/cockroach/storage/range.go:845 +0xca3
github.com/cockroachdb/cockroach/storage.(*Range).processRaftCommand()
/go/src/github.com/cockroachdb/cockroach/storage/range.go:767 +0x3b5
github.com/cockroachdb/cockroach/storage.func·036()
/go/src/github.com/cockroachdb/cockroach/storage/store.go:1339 +0xce8
```
Please assign, take a look and update the issue accordingly.
|
test
|
test failure in ci build the following test appears to have failed level info time file multiraft go line msg node group raft ready level info time file multiraft go line msg hardstate updated term vote commit xxx unrecognized level info time file multiraft go line msg new entry entrynormal raft id origin node id cmd cmd id key level info time file multiraft go line msg committed entry entrynormal raft id origin node id cmd cmd id key warning data race read by goroutine github com cockroachdb cockroach storage rangestats getsize go src github com cockroachdb cockroach storage stats go github com cockroachdb cockroach storage splitqueue process go src github com cockroachdb cockroach storage split queue go github com cockroachdb cockroach storage basequeue processone go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach storage func· go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util func· go src github com cockroachdb cockroach util stopper go previous write by goroutine github com cockroachdb cockroach proto mvccstats add go src github com cockroachdb cockroach proto data go github com cockroachdb cockroach storage rangestats mergemvccstats go src github com cockroachdb cockroach storage stats go github com cockroachdb cockroach storage range applyraftcommand go src github com cockroachdb cockroach storage range go github com cockroachdb cockroach storage range processraftcommand go src github com cockroachdb cockroach storage range go github com cockroachdb cockroach storage func· go src github com cockroachdb cockroach storage store go go src github com cockroachdb cockroach kv split test go testing trunner usr src go src testing testing go warning data race read by goroutine github com cockroachdb cockroach storage rangestats getsize go src github com cockroachdb cockroach storage stats go github com cockroachdb cockroach storage splitqueue process go src github com cockroachdb cockroach storage split queue go github com cockroachdb cockroach storage basequeue processone go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach storage func· go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util func· go src github com cockroachdb cockroach util stopper go previous write by goroutine github com cockroachdb cockroach proto mvccstats add go src github com cockroachdb cockroach proto data go github com cockroachdb cockroach storage rangestats mergemvccstats go src github com cockroachdb cockroach storage stats go github com cockroachdb cockroach storage range applyraftcommand go src github com cockroachdb cockroach storage range go github com cockroachdb cockroach storage range processraftcommand go src github com cockroachdb cockroach storage range go github com cockroachdb cockroach storage func· go src github com cockroachdb cockroach storage store go level info time file multiraft go line msg node group raft ready level info time file multiraft go line msg hardstate updated term vote commit xxx unrecognized level info time file multiraft go line msg new entry entrynormal raft id origin node id cmd cmd id key level info time file multiraft go line msg committed entry entrynormal raft id origin node id cmd cmd id key warning data race read by goroutine github com cockroachdb cockroach storage rangestats getsize go src github com cockroachdb cockroach storage stats go github com cockroachdb cockroach storage splitqueue process go src github com cockroachdb cockroach storage split queue go github com cockroachdb cockroach storage basequeue processone go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach storage func· go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util func· go src github com cockroachdb cockroach util stopper go previous write by goroutine github com cockroachdb cockroach proto mvccstats add go src github com cockroachdb cockroach proto data go github com cockroachdb cockroach storage rangestats mergemvccstats go src github com cockroachdb cockroach storage stats go github com cockroachdb cockroach storage range applyraftcommand go src github com cockroachdb cockroach storage range go github com cockroachdb cockroach storage range processraftcommand go src github com cockroachdb cockroach storage range go github com cockroachdb cockroach storage func· go src github com cockroachdb cockroach storage store go go src github com cockroachdb cockroach kv split test go testing trunner usr src go src testing testing go warning data race read by goroutine github com cockroachdb cockroach storage rangestats getsize go src github com cockroachdb cockroach storage stats go github com cockroachdb cockroach storage splitqueue process go src github com cockroachdb cockroach storage split queue go github com cockroachdb cockroach storage basequeue processone go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach storage func· go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util func· go src github com cockroachdb cockroach util stopper go previous write by goroutine github com cockroachdb cockroach proto mvccstats add go src github com cockroachdb cockroach proto data go github com cockroachdb cockroach storage rangestats mergemvccstats go src github com cockroachdb cockroach storage stats go github com cockroachdb cockroach storage range applyraftcommand go src github com cockroachdb cockroach storage range go github com cockroachdb cockroach storage range processraftcommand go src github com cockroachdb cockroach storage range go github com cockroachdb cockroach storage func· go src github com cockroachdb cockroach storage store go please assign take a look and update the issue accordingly
| 1
|
169,626
| 13,155,092,586
|
IssuesEvent
|
2020-08-10 08:12:59
|
ekaterinailin/AltaiPony
|
https://api.github.com/repos/ekaterinailin/AltaiPony
|
closed
|
Test the preservation of units throughout the pipeline
|
testing
|
<!-- Fill in the information below before opening an issue. -->
#### What needs to be created or improved?
<!-- Provide a clear and concise description of the issue. -->
Write tests that explicitly check units of outputs of functions if the input units are known.
**Question**: Would this be a good use case for decorators? (@gully)
#### Can you provide an example?
<!-- Provide a link or minimal code snippet that demonstrates the issue. -->
```python
def test_units():
lc = get_k2sc_lc()
flares = lc.find_flares(*args, **kwargs)
test_find_flares(lc, flares)
#OR
@testunits
def find_flares(*args, **kwargs):
[...]
return flares
```
#### What is the goal / expected behaviour?
<!-- Describe the behavior you expected and how it differs from the behavior observed in the example. -->
Time is given in days, but ED is typically measured in seconds, frequencies can be _per hour_ or _per year_, amplitued can be relative or e-/s.
|
1.0
|
Test the preservation of units throughout the pipeline - <!-- Fill in the information below before opening an issue. -->
#### What needs to be created or improved?
<!-- Provide a clear and concise description of the issue. -->
Write tests that explicitly check units of outputs of functions if the input units are known.
**Question**: Would this be a good use case for decorators? (@gully)
#### Can you provide an example?
<!-- Provide a link or minimal code snippet that demonstrates the issue. -->
```python
def test_units():
lc = get_k2sc_lc()
flares = lc.find_flares(*args, **kwargs)
test_find_flares(lc, flares)
#OR
@testunits
def find_flares(*args, **kwargs):
[...]
return flares
```
#### What is the goal / expected behaviour?
<!-- Describe the behavior you expected and how it differs from the behavior observed in the example. -->
Time is given in days, but ED is typically measured in seconds, frequencies can be _per hour_ or _per year_, amplitued can be relative or e-/s.
|
test
|
test the preservation of units throughout the pipeline what needs to be created or improved write tests that explicitly check units of outputs of functions if the input units are known question would this be a good use case for decorators gully can you provide an example python def test units lc get lc flares lc find flares args kwargs test find flares lc flares or testunits def find flares args kwargs return flares what is the goal expected behaviour time is given in days but ed is typically measured in seconds frequencies can be per hour or per year amplitued can be relative or e s
| 1
|
46,417
| 11,840,195,642
|
IssuesEvent
|
2020-03-23 18:25:42
|
apache/incubator-mxnet
|
https://api.github.com/repos/apache/incubator-mxnet
|
closed
|
USE_TVM_OP=1 build broken with DMLC_LOG_FATAL_THROW=0
|
Bug Build TVM OP
|
## Description
Setting `-DDMLC_LOG_FATAL_THROW=0` is useful for debugging with `gdb` or `lldb`. Unfortunately we can't compile MXNet with `USE_TVM_OP=1` in that case.
### Error Message
```
[658/658] Linking CXX shared library libmxnet.so
FAILED: libmxnet.so
: && /usr/bin/c++ -fPIC -Wall -Wno-sign-compare -O3 -std=c++11 -fopenmp -shared -Wl,-soname,libmxnet.so -o libmxnet.so CMakeFiles/mxnet.dir/dummy.c.o -L/home/ubuntu/src/mxnet-master/build/3rdparty/tvm -Wl,-rpath,/usr/local/lib:/home/ubuntu/src/mxnet-master/build/3rdparty/tvm: -Wl,--whole-archive libmxnet.a -Wl,--no-whole-archive libmxnet.a 3rdparty/dmlc-core/libdmlc.a -ldl 3rdparty/mkldnn/src/libdnnl.a -lpthread -Wl,--start-group /opt/intel/mkl/lib/intel64/libmkl_intel_lp64.a /opt/intel/mkl/lib/intel64/libmkl_intel_thread.a /opt/intel/mkl/lib/intel64/libmkl_core.a /usr/local/lib/libiomp5.so -Wl,--end-group /usr/lib/x86_64-linux-gnu/librt.so /usr/lib/x86_64-linux-gnu/libopencv_highgui.so.3.2.0 /usr/lib/x86_64-linux-gnu/libopencv_imgcodecs.so.3.2.0 /usr/lib/x86_64-linux-gnu/libopencv_imgproc.so.3.2.0 /usr/lib/x86_64-linux-gnu/libopencv_core.so.3.2.0 -llapack -ltvm_runtime /usr/lib/gcc/x86_64-linux-gnu/7/libgomp.so -lpthread -lrt && cd /home/ubuntu/src/mxnet-master/build && /home/ubuntu/.pyenv/versions/3.8.2-debug/lib/python3.8/site-packages/cmake/data/bin/cmake -E copy /home/ubuntu/src/mxnet-master/build/3rdparty/mkldnn/include/dnnl_config.h /home/ubuntu/src/mxnet-master/include/mkldnn/ && /home/ubuntu/.pyenv/versions/3.8.2-debug/lib/python3.8/site-packages/cmake/data/bin/cmake -E copy /home/ubuntu/src/mxnet-master/build/3rdparty/mkldnn/include/dnnl_version.h /home/ubuntu/src/mxnet-master/include/mkldnn/ && cd /home/ubuntu/src/mxnet-master/build && /home/ubuntu/.pyenv/versions/3.8.2-debug/lib/python3.8/site-packages/cmake/data/bin/cmake -E env PYTHONPATH="/home/ubuntu/src/mxnet-master/3rdparty/tvm/python:/home/ubuntu/src/mxnet-master/3rdparty/tvm/topi/python:/home/ubuntu/src/mxnet-master/contrib" LD_LIBRARY_PATH=/home/ubuntu/src/mxnet-master/build:/home/ubuntu/src/mxnet-master/build/3rdparty/tvm:/usr/local/cuda/lib64:/usr/local/lib:/usr/lib:/usr/local/cuda/extras/CUPTI/lib64:/usr/local/mpi/lib:/home/ubuntu/local/lib:/home/ubuntu/local/lib:/usr/local/cuda/lib64:/usr/local/lib:/usr/lib:/usr/local/cuda/extras/CUPTI/lib64:/usr/local/mpi/lib:/home/ubuntu/local/lib /home/ubuntu/.pyenv/versions/3.8.2-debug/bin/python3.8 /home/ubuntu/src/mxnet-master/contrib/tvmop/compile.py -o/home/ubuntu/src/mxnet-master/build --config /home/ubuntu/src/mxnet-master/build/tvmop.conf -L /home/ubuntu/src/mxnet-master/build/3rdparty/tvm
[19:56:21] ../3rdparty/tvm/src/node/reflection.cc:113: AttributeError: Variable object has no attributed shape
Child aborted
ninja: build stopped: subcommand failed.
```
## To Reproduce
### Steps to reproduce
1. Apply patch below
2. Compile with `cmake -GNinja -DUSE_CUDA=0 -DUSE_TVM_OP=ON ..; ninja`
``` diff
diff --git a/CMakeLists.txt b/CMakeLists.txt
index 365da2d90..b4c9d267b 100644
--- a/CMakeLists.txt
+++ b/CMakeLists.txt
@@ -151,7 +151,7 @@ endif()
add_definitions(-DDMLC_MODERN_THREAD_LOCAL=0)
# disable stack trace in exception by default.
add_definitions(-DDMLC_LOG_STACK_TRACE_SIZE=0)
-add_definitions(-DDMLC_LOG_FATAL_THROW=1)
+add_definitions(-DDMLC_LOG_FATAL_THROW=0)
if(MSVC)
add_definitions(-DWIN32_LEAN_AND_MEAN)
```
|
1.0
|
USE_TVM_OP=1 build broken with DMLC_LOG_FATAL_THROW=0 - ## Description
Setting `-DDMLC_LOG_FATAL_THROW=0` is useful for debugging with `gdb` or `lldb`. Unfortunately we can't compile MXNet with `USE_TVM_OP=1` in that case.
### Error Message
```
[658/658] Linking CXX shared library libmxnet.so
FAILED: libmxnet.so
: && /usr/bin/c++ -fPIC -Wall -Wno-sign-compare -O3 -std=c++11 -fopenmp -shared -Wl,-soname,libmxnet.so -o libmxnet.so CMakeFiles/mxnet.dir/dummy.c.o -L/home/ubuntu/src/mxnet-master/build/3rdparty/tvm -Wl,-rpath,/usr/local/lib:/home/ubuntu/src/mxnet-master/build/3rdparty/tvm: -Wl,--whole-archive libmxnet.a -Wl,--no-whole-archive libmxnet.a 3rdparty/dmlc-core/libdmlc.a -ldl 3rdparty/mkldnn/src/libdnnl.a -lpthread -Wl,--start-group /opt/intel/mkl/lib/intel64/libmkl_intel_lp64.a /opt/intel/mkl/lib/intel64/libmkl_intel_thread.a /opt/intel/mkl/lib/intel64/libmkl_core.a /usr/local/lib/libiomp5.so -Wl,--end-group /usr/lib/x86_64-linux-gnu/librt.so /usr/lib/x86_64-linux-gnu/libopencv_highgui.so.3.2.0 /usr/lib/x86_64-linux-gnu/libopencv_imgcodecs.so.3.2.0 /usr/lib/x86_64-linux-gnu/libopencv_imgproc.so.3.2.0 /usr/lib/x86_64-linux-gnu/libopencv_core.so.3.2.0 -llapack -ltvm_runtime /usr/lib/gcc/x86_64-linux-gnu/7/libgomp.so -lpthread -lrt && cd /home/ubuntu/src/mxnet-master/build && /home/ubuntu/.pyenv/versions/3.8.2-debug/lib/python3.8/site-packages/cmake/data/bin/cmake -E copy /home/ubuntu/src/mxnet-master/build/3rdparty/mkldnn/include/dnnl_config.h /home/ubuntu/src/mxnet-master/include/mkldnn/ && /home/ubuntu/.pyenv/versions/3.8.2-debug/lib/python3.8/site-packages/cmake/data/bin/cmake -E copy /home/ubuntu/src/mxnet-master/build/3rdparty/mkldnn/include/dnnl_version.h /home/ubuntu/src/mxnet-master/include/mkldnn/ && cd /home/ubuntu/src/mxnet-master/build && /home/ubuntu/.pyenv/versions/3.8.2-debug/lib/python3.8/site-packages/cmake/data/bin/cmake -E env PYTHONPATH="/home/ubuntu/src/mxnet-master/3rdparty/tvm/python:/home/ubuntu/src/mxnet-master/3rdparty/tvm/topi/python:/home/ubuntu/src/mxnet-master/contrib" LD_LIBRARY_PATH=/home/ubuntu/src/mxnet-master/build:/home/ubuntu/src/mxnet-master/build/3rdparty/tvm:/usr/local/cuda/lib64:/usr/local/lib:/usr/lib:/usr/local/cuda/extras/CUPTI/lib64:/usr/local/mpi/lib:/home/ubuntu/local/lib:/home/ubuntu/local/lib:/usr/local/cuda/lib64:/usr/local/lib:/usr/lib:/usr/local/cuda/extras/CUPTI/lib64:/usr/local/mpi/lib:/home/ubuntu/local/lib /home/ubuntu/.pyenv/versions/3.8.2-debug/bin/python3.8 /home/ubuntu/src/mxnet-master/contrib/tvmop/compile.py -o/home/ubuntu/src/mxnet-master/build --config /home/ubuntu/src/mxnet-master/build/tvmop.conf -L /home/ubuntu/src/mxnet-master/build/3rdparty/tvm
[19:56:21] ../3rdparty/tvm/src/node/reflection.cc:113: AttributeError: Variable object has no attributed shape
Child aborted
ninja: build stopped: subcommand failed.
```
## To Reproduce
### Steps to reproduce
1. Apply patch below
2. Compile with `cmake -GNinja -DUSE_CUDA=0 -DUSE_TVM_OP=ON ..; ninja`
``` diff
diff --git a/CMakeLists.txt b/CMakeLists.txt
index 365da2d90..b4c9d267b 100644
--- a/CMakeLists.txt
+++ b/CMakeLists.txt
@@ -151,7 +151,7 @@ endif()
add_definitions(-DDMLC_MODERN_THREAD_LOCAL=0)
# disable stack trace in exception by default.
add_definitions(-DDMLC_LOG_STACK_TRACE_SIZE=0)
-add_definitions(-DDMLC_LOG_FATAL_THROW=1)
+add_definitions(-DDMLC_LOG_FATAL_THROW=0)
if(MSVC)
add_definitions(-DWIN32_LEAN_AND_MEAN)
```
|
non_test
|
use tvm op build broken with dmlc log fatal throw description setting ddmlc log fatal throw is useful for debugging with gdb or lldb unfortunately we can t compile mxnet with use tvm op in that case error message linking cxx shared library libmxnet so failed libmxnet so usr bin c fpic wall wno sign compare std c fopenmp shared wl soname libmxnet so o libmxnet so cmakefiles mxnet dir dummy c o l home ubuntu src mxnet master build tvm wl rpath usr local lib home ubuntu src mxnet master build tvm wl whole archive libmxnet a wl no whole archive libmxnet a dmlc core libdmlc a ldl mkldnn src libdnnl a lpthread wl start group opt intel mkl lib libmkl intel a opt intel mkl lib libmkl intel thread a opt intel mkl lib libmkl core a usr local lib so wl end group usr lib linux gnu librt so usr lib linux gnu libopencv highgui so usr lib linux gnu libopencv imgcodecs so usr lib linux gnu libopencv imgproc so usr lib linux gnu libopencv core so llapack ltvm runtime usr lib gcc linux gnu libgomp so lpthread lrt cd home ubuntu src mxnet master build home ubuntu pyenv versions debug lib site packages cmake data bin cmake e copy home ubuntu src mxnet master build mkldnn include dnnl config h home ubuntu src mxnet master include mkldnn home ubuntu pyenv versions debug lib site packages cmake data bin cmake e copy home ubuntu src mxnet master build mkldnn include dnnl version h home ubuntu src mxnet master include mkldnn cd home ubuntu src mxnet master build home ubuntu pyenv versions debug lib site packages cmake data bin cmake e env pythonpath home ubuntu src mxnet master tvm python home ubuntu src mxnet master tvm topi python home ubuntu src mxnet master contrib ld library path home ubuntu src mxnet master build home ubuntu src mxnet master build tvm usr local cuda usr local lib usr lib usr local cuda extras cupti usr local mpi lib home ubuntu local lib home ubuntu local lib usr local cuda usr local lib usr lib usr local cuda extras cupti usr local mpi lib home ubuntu local lib home ubuntu pyenv versions debug bin home ubuntu src mxnet master contrib tvmop compile py o home ubuntu src mxnet master build config home ubuntu src mxnet master build tvmop conf l home ubuntu src mxnet master build tvm tvm src node reflection cc attributeerror variable object has no attributed shape child aborted ninja build stopped subcommand failed to reproduce steps to reproduce apply patch below compile with cmake gninja duse cuda duse tvm op on ninja diff diff git a cmakelists txt b cmakelists txt index a cmakelists txt b cmakelists txt endif add definitions ddmlc modern thread local disable stack trace in exception by default add definitions ddmlc log stack trace size add definitions ddmlc log fatal throw add definitions ddmlc log fatal throw if msvc add definitions lean and mean
| 0
|
363,805
| 10,755,416,028
|
IssuesEvent
|
2019-10-31 09:05:36
|
acidanthera/bugtracker
|
https://api.github.com/repos/acidanthera/bugtracker
|
closed
|
VoodooPS2, Thinkpad T440s, 3 physical button not working
|
enhancement priority:normal project:ps2
|
hi,
hava an t440s with the trackpad from a t450( extra 3physical buttons).
all gestures are working fine except the three button.
in consol.log if I press on elf the three button it produce
standard 17:20:16.553291 +0200 kernel synaptics_parse_hw_state: finger 0 pressure 0 width 8.
thx
[debug_2105.zip](https://github.com/acidanthera/bugtracker/files/3108267/debug_2105.zip)
|
1.0
|
VoodooPS2, Thinkpad T440s, 3 physical button not working - hi,
hava an t440s with the trackpad from a t450( extra 3physical buttons).
all gestures are working fine except the three button.
in consol.log if I press on elf the three button it produce
standard 17:20:16.553291 +0200 kernel synaptics_parse_hw_state: finger 0 pressure 0 width 8.
thx
[debug_2105.zip](https://github.com/acidanthera/bugtracker/files/3108267/debug_2105.zip)
|
non_test
|
thinkpad physical button not working hi hava an with the trackpad from a extra buttons all gestures are working fine except the three button in consol log if i press on elf the three button it produce standard kernel synaptics parse hw state finger pressure width thx
| 0
|
433,197
| 12,503,507,124
|
IssuesEvent
|
2020-06-02 07:22:24
|
metal3-io/metal3-dev-env
|
https://api.github.com/repos/metal3-io/metal3-dev-env
|
closed
|
Add cloud init log out put
|
kind/feature priority/important-longterm
|
In the CI, some jobs are failing and the output does not reveal much information.
This [PR](281) is a case in point. It was failing due to missing binaries and the outputs did not show that. We were required to login to the provisioned machines and get the logs.
It would be good the content of ```/var/log/cloud-init-output.log``` is also shown at the end as it would make pin pointing errors easy.
|
1.0
|
Add cloud init log out put - In the CI, some jobs are failing and the output does not reveal much information.
This [PR](281) is a case in point. It was failing due to missing binaries and the outputs did not show that. We were required to login to the provisioned machines and get the logs.
It would be good the content of ```/var/log/cloud-init-output.log``` is also shown at the end as it would make pin pointing errors easy.
|
non_test
|
add cloud init log out put in the ci some jobs are failing and the output does not reveal much information this is a case in point it was failing due to missing binaries and the outputs did not show that we were required to login to the provisioned machines and get the logs it would be good the content of var log cloud init output log is also shown at the end as it would make pin pointing errors easy
| 0
|
119,536
| 10,057,783,106
|
IssuesEvent
|
2019-07-22 12:31:18
|
ICIJ/datashare
|
https://api.github.com/repos/ICIJ/datashare
|
closed
|
Put 'Tags' above 'Details'
|
bug front need testing
|
In order to avoid horizontal scroll bars and a blank space behing 'Tags', we can put 'Tags' above 'Details' as it should not take too much room.
<img width="835" alt="Screenshot 2019-07-15 at 13 56 32" src="https://user-images.githubusercontent.com/17233829/61214545-74a41c80-a708-11e9-8256-65bd7bd463cd.png">
|
1.0
|
Put 'Tags' above 'Details' - In order to avoid horizontal scroll bars and a blank space behing 'Tags', we can put 'Tags' above 'Details' as it should not take too much room.
<img width="835" alt="Screenshot 2019-07-15 at 13 56 32" src="https://user-images.githubusercontent.com/17233829/61214545-74a41c80-a708-11e9-8256-65bd7bd463cd.png">
|
test
|
put tags above details in order to avoid horizontal scroll bars and a blank space behing tags we can put tags above details as it should not take too much room img width alt screenshot at src
| 1
|
46,469
| 2,957,910,188
|
IssuesEvent
|
2015-07-08 18:41:49
|
CenterForOpenScience/osf.io
|
https://api.github.com/repos/CenterForOpenScience/osf.io
|
closed
|
ID requested for linkedin profile does not lead to public profile
|
2 - ready bug: production intern priority - high
|
From the community repo: https://github.com/CenterForOpenScience/community/issues/269
## Current behavior
When adding a Linkedin profile, the user is shown the beginning of a URL that they are expected to fill out with their "profile ID" (https://www.linkedin.com/profile/view?id=)

My Linkedin user ID is davidtmellor, which does not successfully complete the expected URL, it results in a 404 error on Linkedin. The correct URL that I think would complete it is: ttps://www.linkedin.com/profile/view?id=77582777 but that URL only works if the user is signed into Linkedin, a visitor to my OSF profile who clicks on that link would be prompted to sign into Linkedin before seeing my Linkefin profile. The suggest change below allows a user to access my Linkedin profile without signing up for Linkedin.

The public profile also shows a number (the ID) instead of their user ID, which would be slightly more welcoming:

## Suggested behavior
Allow the user to enter in their Linkedin ID to complete this URL: https://www.linkedin.com/in/USERID example: https://www.linkedin.com/in/davidtmellor. Show this on the public profile
OR
Allow user to enter in any URL they want for their Linkedin profile.
|
1.0
|
ID requested for linkedin profile does not lead to public profile - From the community repo: https://github.com/CenterForOpenScience/community/issues/269
## Current behavior
When adding a Linkedin profile, the user is shown the beginning of a URL that they are expected to fill out with their "profile ID" (https://www.linkedin.com/profile/view?id=)

My Linkedin user ID is davidtmellor, which does not successfully complete the expected URL, it results in a 404 error on Linkedin. The correct URL that I think would complete it is: ttps://www.linkedin.com/profile/view?id=77582777 but that URL only works if the user is signed into Linkedin, a visitor to my OSF profile who clicks on that link would be prompted to sign into Linkedin before seeing my Linkefin profile. The suggest change below allows a user to access my Linkedin profile without signing up for Linkedin.

The public profile also shows a number (the ID) instead of their user ID, which would be slightly more welcoming:

## Suggested behavior
Allow the user to enter in their Linkedin ID to complete this URL: https://www.linkedin.com/in/USERID example: https://www.linkedin.com/in/davidtmellor. Show this on the public profile
OR
Allow user to enter in any URL they want for their Linkedin profile.
|
non_test
|
id requested for linkedin profile does not lead to public profile from the community repo current behavior when adding a linkedin profile the user is shown the beginning of a url that they are expected to fill out with their profile id my linkedin user id is davidtmellor which does not successfully complete the expected url it results in a error on linkedin the correct url that i think would complete it is ttps but that url only works if the user is signed into linkedin a visitor to my osf profile who clicks on that link would be prompted to sign into linkedin before seeing my linkefin profile the suggest change below allows a user to access my linkedin profile without signing up for linkedin the public profile also shows a number the id instead of their user id which would be slightly more welcoming suggested behavior allow the user to enter in their linkedin id to complete this url example show this on the public profile or allow user to enter in any url they want for their linkedin profile
| 0
|
50,481
| 6,097,900,444
|
IssuesEvent
|
2017-06-20 05:33:36
|
c2corg/v6_ui
|
https://api.github.com/repos/c2corg/v6_ui
|
closed
|
Unsupported L# tags in v6.1.10
|
ready for testing
|
For some reason L# are not correctly supported in release v6.1.10 (v6.1.9 was OK but with a simpler support I think).
It can be observed in the demo. For instance http://www.demov6.camptocamp.org/routes/368130/fr/vallee-d-ailefroide-poire-d-ailefroide-bonne-poire
I have switched the redis cache off on the demo to make sure rendered documents are not retrieved from the cache.
The weird thing seems that sometimes it works, sometimes it doesn't. Well it seems that it works until the dcument is edited, than it's broken.
@olaurendeau Would you have an idea?
|
1.0
|
Unsupported L# tags in v6.1.10 - For some reason L# are not correctly supported in release v6.1.10 (v6.1.9 was OK but with a simpler support I think).
It can be observed in the demo. For instance http://www.demov6.camptocamp.org/routes/368130/fr/vallee-d-ailefroide-poire-d-ailefroide-bonne-poire
I have switched the redis cache off on the demo to make sure rendered documents are not retrieved from the cache.
The weird thing seems that sometimes it works, sometimes it doesn't. Well it seems that it works until the dcument is edited, than it's broken.
@olaurendeau Would you have an idea?
|
test
|
unsupported l tags in for some reason l are not correctly supported in release was ok but with a simpler support i think it can be observed in the demo for instance i have switched the redis cache off on the demo to make sure rendered documents are not retrieved from the cache the weird thing seems that sometimes it works sometimes it doesn t well it seems that it works until the dcument is edited than it s broken olaurendeau would you have an idea
| 1
|
118,100
| 17,576,352,460
|
IssuesEvent
|
2021-08-15 17:32:48
|
turkdevops/brackets
|
https://api.github.com/repos/turkdevops/brackets
|
reopened
|
CVE-2017-16026 (Medium) detected in request-2.42.0.tgz, request-2.9.153.tgz
|
security vulnerability
|
## CVE-2017-16026 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>request-2.42.0.tgz</b>, <b>request-2.9.153.tgz</b></p></summary>
<p>
<details><summary><b>request-2.42.0.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.42.0.tgz">https://registry.npmjs.org/request/-/request-2.42.0.tgz</a></p>
<p>Path to dependency file: brackets/package.json</p>
<p>Path to vulnerable library: brackets/node_modules/phantomjs/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- phantomjs-1.9.18.tgz (Root Library)
- :x: **request-2.42.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>request-2.9.153.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.9.153.tgz">https://registry.npmjs.org/request/-/request-2.9.153.tgz</a></p>
<p>Path to dependency file: brackets/src/extensions/default/JavaScriptQuickEdit/unittest-files/jquery-ui/package.json</p>
<p>Path to vulnerable library: /src/extensions/default/JavaScriptQuickEdit/unittest-files/jquery-ui/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- :x: **request-2.9.153.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/brackets/commit/38b389e26019954346ecebf98a10c2f5fa9a0488">38b389e26019954346ecebf98a10c2f5fa9a0488</a></p>
<p>Found in base branch: <b>checkTravis</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Request is an http client. If a request is made using ```multipart```, and the body type is a ```number```, then the specified number of non-zero memory is passed in the body. This affects Request >=2.2.6 <2.47.0 || >2.51.0 <=2.67.0.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16026>CVE-2017-16026</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-16026">https://nvd.nist.gov/vuln/detail/CVE-2017-16026</a></p>
<p>Release Date: 2018-06-04</p>
<p>Fix Resolution: 2.47.1,2.67.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2017-16026 (Medium) detected in request-2.42.0.tgz, request-2.9.153.tgz - ## CVE-2017-16026 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>request-2.42.0.tgz</b>, <b>request-2.9.153.tgz</b></p></summary>
<p>
<details><summary><b>request-2.42.0.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.42.0.tgz">https://registry.npmjs.org/request/-/request-2.42.0.tgz</a></p>
<p>Path to dependency file: brackets/package.json</p>
<p>Path to vulnerable library: brackets/node_modules/phantomjs/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- phantomjs-1.9.18.tgz (Root Library)
- :x: **request-2.42.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>request-2.9.153.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.9.153.tgz">https://registry.npmjs.org/request/-/request-2.9.153.tgz</a></p>
<p>Path to dependency file: brackets/src/extensions/default/JavaScriptQuickEdit/unittest-files/jquery-ui/package.json</p>
<p>Path to vulnerable library: /src/extensions/default/JavaScriptQuickEdit/unittest-files/jquery-ui/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- :x: **request-2.9.153.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/brackets/commit/38b389e26019954346ecebf98a10c2f5fa9a0488">38b389e26019954346ecebf98a10c2f5fa9a0488</a></p>
<p>Found in base branch: <b>checkTravis</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Request is an http client. If a request is made using ```multipart```, and the body type is a ```number```, then the specified number of non-zero memory is passed in the body. This affects Request >=2.2.6 <2.47.0 || >2.51.0 <=2.67.0.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16026>CVE-2017-16026</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-16026">https://nvd.nist.gov/vuln/detail/CVE-2017-16026</a></p>
<p>Release Date: 2018-06-04</p>
<p>Fix Resolution: 2.47.1,2.67.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in request tgz request tgz cve medium severity vulnerability vulnerable libraries request tgz request tgz request tgz simplified http request client library home page a href path to dependency file brackets package json path to vulnerable library brackets node modules phantomjs node modules request package json dependency hierarchy phantomjs tgz root library x request tgz vulnerable library request tgz simplified http request client library home page a href path to dependency file brackets src extensions default javascriptquickedit unittest files jquery ui package json path to vulnerable library src extensions default javascriptquickedit unittest files jquery ui node modules request package json dependency hierarchy x request tgz vulnerable library found in head commit a href found in base branch checktravis vulnerability details request is an http client if a request is made using multipart and the body type is a number then the specified number of non zero memory is passed in the body this affects request publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
287,288
| 24,822,078,303
|
IssuesEvent
|
2022-10-25 17:18:05
|
godotengine/godot
|
https://api.github.com/repos/godotengine/godot
|
closed
|
Shader visually differs from playing scene and in-editor
|
topic:rendering needs testing topic:3d
|
### Godot version
4.0beta3
### System information
PopOS 22.04, 5.19.0-76051900-generic kernel, Ryzen 6800h, Radeon 6700m, amdgpu, Vulkan
### Issue description
In editor:


When playing scene:

### Steps to reproduce
Modifying shaders update in editor but don't show the same in game. It's not clear which is visually correct. Far and near of camera are set to the same as in editor.
### Minimal reproduction project
Reproduction Repo: https://github.com/Mikeysax/example-project-for-godot-bug-67807
It appears like the noise being used for the foam is never applied:

Shader Code: https://gist.github.com/Mikeysax/a97f14e8f1d00a53c3ba83d964eeb350
|
1.0
|
Shader visually differs from playing scene and in-editor - ### Godot version
4.0beta3
### System information
PopOS 22.04, 5.19.0-76051900-generic kernel, Ryzen 6800h, Radeon 6700m, amdgpu, Vulkan
### Issue description
In editor:


When playing scene:

### Steps to reproduce
Modifying shaders update in editor but don't show the same in game. It's not clear which is visually correct. Far and near of camera are set to the same as in editor.
### Minimal reproduction project
Reproduction Repo: https://github.com/Mikeysax/example-project-for-godot-bug-67807
It appears like the noise being used for the foam is never applied:

Shader Code: https://gist.github.com/Mikeysax/a97f14e8f1d00a53c3ba83d964eeb350
|
test
|
shader visually differs from playing scene and in editor godot version system information popos generic kernel ryzen radeon amdgpu vulkan issue description in editor when playing scene steps to reproduce modifying shaders update in editor but don t show the same in game it s not clear which is visually correct far and near of camera are set to the same as in editor minimal reproduction project reproduction repo it appears like the noise being used for the foam is never applied shader code
| 1
|
133,197
| 12,534,458,856
|
IssuesEvent
|
2020-06-04 19:28:29
|
spring-projects/spring-boot
|
https://api.github.com/repos/spring-projects/spring-boot
|
closed
|
Use highlightjs for syntax highlighting in HTML documentation
|
type: documentation
|
It looks like it got lost when the build was converted to Gradle.
|
1.0
|
Use highlightjs for syntax highlighting in HTML documentation - It looks like it got lost when the build was converted to Gradle.
|
non_test
|
use highlightjs for syntax highlighting in html documentation it looks like it got lost when the build was converted to gradle
| 0
|
312,816
| 23,444,530,534
|
IssuesEvent
|
2022-08-15 18:12:13
|
edgexfoundry/edgex-docs
|
https://api.github.com/repos/edgexfoundry/edgex-docs
|
closed
|
Document Device System Events
|
documentation enhancement
|
See implementation PRs for reference:
https://github.com/edgexfoundry/edgex-go/pull/4101
https://github.com/edgexfoundry/edgex-go/pull/4121
And the ADR here:
https://docs.edgexfoundry.org/2.3/design/adr/0024-system-events/
|
1.0
|
Document Device System Events - See implementation PRs for reference:
https://github.com/edgexfoundry/edgex-go/pull/4101
https://github.com/edgexfoundry/edgex-go/pull/4121
And the ADR here:
https://docs.edgexfoundry.org/2.3/design/adr/0024-system-events/
|
non_test
|
document device system events see implementation prs for reference and the adr here
| 0
|
320,076
| 27,420,002,996
|
IssuesEvent
|
2023-03-01 16:06:32
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
opened
|
Fix math.test_tensorflow_squared_difference
|
TensorFlow Frontend Sub Task Failing Test
|
| | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4301928602/jobs/7499787609" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|None
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4301928602/jobs/7499798636" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|None
<details>
<summary>Not found</summary>
Not found
</details>
<details>
<summary>Not found</summary>
Not found
</details>
|
1.0
|
Fix math.test_tensorflow_squared_difference - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4301928602/jobs/7499787609" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|None
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4301928602/jobs/7499798636" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|None
<details>
<summary>Not found</summary>
Not found
</details>
<details>
<summary>Not found</summary>
Not found
</details>
|
test
|
fix math test tensorflow squared difference tensorflow img src torch none numpy img src jax none not found not found not found not found
| 1
|
20,289
| 2,622,438,446
|
IssuesEvent
|
2015-03-04 02:44:47
|
calblueprint/PHC
|
https://api.github.com/repos/calblueprint/PHC
|
opened
|
standardize all text and views
|
medium priority UI/UX
|
standardize text size, font, format, style.
standardize button sizes and other view sizes.
|
1.0
|
standardize all text and views - standardize text size, font, format, style.
standardize button sizes and other view sizes.
|
non_test
|
standardize all text and views standardize text size font format style standardize button sizes and other view sizes
| 0
|
68,139
| 8,221,632,785
|
IssuesEvent
|
2018-09-06 03:05:37
|
rbastic/dyndao
|
https://api.github.com/repos/rbastic/dyndao
|
closed
|
Consider adding column Length (and Precision, for floats) as a parameter to the Is*Type() functions
|
design
|
There are situations where these things are important (due to ambiguity in type affinity.)
|
1.0
|
Consider adding column Length (and Precision, for floats) as a parameter to the Is*Type() functions - There are situations where these things are important (due to ambiguity in type affinity.)
|
non_test
|
consider adding column length and precision for floats as a parameter to the is type functions there are situations where these things are important due to ambiguity in type affinity
| 0
|
276,291
| 8,596,906,768
|
IssuesEvent
|
2018-11-15 17:07:28
|
InfiniteFlightAirportEditing/Airports
|
https://api.github.com/repos/InfiniteFlightAirportEditing/Airports
|
opened
|
OAHR-Herat-HERAT-AFGHANISTAN
|
Being Redone Low Priority
|
# Airport Name
Herat
# Country?
Afghanistan
# Improvements that need to be made?
from scratch
# Are you working on this airport?
Yes
# Airport Priority? (A380, 10000ft+ Runway)
9880ft Rwy
|
1.0
|
OAHR-Herat-HERAT-AFGHANISTAN - # Airport Name
Herat
# Country?
Afghanistan
# Improvements that need to be made?
from scratch
# Are you working on this airport?
Yes
# Airport Priority? (A380, 10000ft+ Runway)
9880ft Rwy
|
non_test
|
oahr herat herat afghanistan airport name herat country afghanistan improvements that need to be made from scratch are you working on this airport yes airport priority runway rwy
| 0
|
101,802
| 11,258,338,685
|
IssuesEvent
|
2020-01-13 04:08:57
|
KisaraBlue/ManaClash
|
https://api.github.com/repos/KisaraBlue/ManaClash
|
opened
|
Swagger UI documentation
|
documentation
|
Documentation can be generated in about an hour of work. I will do it on January 13th, after the deadline passes (I have an exam at 10:45 AM and I need to try to get a few hours of sleep before that, so I'll add documentation same day at night).
|
1.0
|
Swagger UI documentation - Documentation can be generated in about an hour of work. I will do it on January 13th, after the deadline passes (I have an exam at 10:45 AM and I need to try to get a few hours of sleep before that, so I'll add documentation same day at night).
|
non_test
|
swagger ui documentation documentation can be generated in about an hour of work i will do it on january after the deadline passes i have an exam at am and i need to try to get a few hours of sleep before that so i ll add documentation same day at night
| 0
|
226,473
| 17,352,981,655
|
IssuesEvent
|
2021-07-29 11:05:48
|
FreshRSS/FreshRSS
|
https://api.github.com/repos/FreshRSS/FreshRSS
|
closed
|
[Feature] Support for Raspberry Pi
|
Documentation :books:
|
**Is your feature request related to a problem? Please describe.**
Hi, there. I ran FreshRSS on my cloud VPS for a couple of months and everything worked fine. After some time I decided to buy Raspberry Pi and move all my cloud self-hosted services there. I had some issues even with PostgreSQL but found [a solution](https://blog.samcater.com/fix-workaround-rpi4-docker-libseccomp2-docker-20/) that fixed that issue. Regardless of it FreshRSS app still can't start (I am running it in a container, `docker-compose.yml` is taken from [here](https://github.com/FreshRSS/FreshRSS/blob/edge/Docker/docker-compose.yml) without additional proxies or other stuff)...
```
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
33ab4f55a320 freshrss/freshrss:latest "./Docker/entrypoint…" 25 seconds ago Restarting (1) 6 seconds ago freshrss-app
f5dafa34d425 postgres:12-alpine "docker-entrypoint.s…" 27 seconds ago Up 25 seconds 0.0.0.0:5432->5432/tcp freshrss-db
```
Here are logs of FreshRSS app:
```
Attaching to freshrss-app
freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error
freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error
freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error
freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error
freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error
freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error
freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error
```
**Describe the solution you'd like**
It would be nice to have FreshRSS app able to run on Raspberry Pi since it is common platform for serving different services at home.
**Additional context**
Since I'm pretty sure it is not problem with FreshRSS app itself, but some underlying stuff with linux and so on (which I'm not so familiar with) I considered this GitHub issue to be a feature request and not bug. In the end it would be useful to have the solution written in the documentation. So it is like cry for a help plus documentation feature request😅
|
1.0
|
[Feature] Support for Raspberry Pi - **Is your feature request related to a problem? Please describe.**
Hi, there. I ran FreshRSS on my cloud VPS for a couple of months and everything worked fine. After some time I decided to buy Raspberry Pi and move all my cloud self-hosted services there. I had some issues even with PostgreSQL but found [a solution](https://blog.samcater.com/fix-workaround-rpi4-docker-libseccomp2-docker-20/) that fixed that issue. Regardless of it FreshRSS app still can't start (I am running it in a container, `docker-compose.yml` is taken from [here](https://github.com/FreshRSS/FreshRSS/blob/edge/Docker/docker-compose.yml) without additional proxies or other stuff)...
```
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
33ab4f55a320 freshrss/freshrss:latest "./Docker/entrypoint…" 25 seconds ago Restarting (1) 6 seconds ago freshrss-app
f5dafa34d425 postgres:12-alpine "docker-entrypoint.s…" 27 seconds ago Up 25 seconds 0.0.0.0:5432->5432/tcp freshrss-db
```
Here are logs of FreshRSS app:
```
Attaching to freshrss-app
freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error
freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error
freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error
freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error
freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error
freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error
freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error
```
**Describe the solution you'd like**
It would be nice to have FreshRSS app able to run on Raspberry Pi since it is common platform for serving different services at home.
**Additional context**
Since I'm pretty sure it is not problem with FreshRSS app itself, but some underlying stuff with linux and so on (which I'm not so familiar with) I considered this GitHub issue to be a feature request and not bug. In the end it would be useful to have the solution written in the documentation. So it is like cry for a help plus documentation feature request😅
|
non_test
|
support for raspberry pi is your feature request related to a problem please describe hi there i ran freshrss on my cloud vps for a couple of months and everything worked fine after some time i decided to buy raspberry pi and move all my cloud self hosted services there i had some issues even with postgresql but found that fixed that issue regardless of it freshrss app still can t start i am running it in a container docker compose yml is taken from without additional proxies or other stuff container id image command created status ports names freshrss freshrss latest docker entrypoint… seconds ago restarting seconds ago freshrss app postgres alpine docker entrypoint s… seconds ago up seconds tcp freshrss db here are logs of freshrss app attaching to freshrss app freshrss app standard init linux go exec user process caused exec format error freshrss app standard init linux go exec user process caused exec format error freshrss app standard init linux go exec user process caused exec format error freshrss app standard init linux go exec user process caused exec format error freshrss app standard init linux go exec user process caused exec format error freshrss app standard init linux go exec user process caused exec format error freshrss app standard init linux go exec user process caused exec format error describe the solution you d like it would be nice to have freshrss app able to run on raspberry pi since it is common platform for serving different services at home additional context since i m pretty sure it is not problem with freshrss app itself but some underlying stuff with linux and so on which i m not so familiar with i considered this github issue to be a feature request and not bug in the end it would be useful to have the solution written in the documentation so it is like cry for a help plus documentation feature request😅
| 0
|
3,532
| 2,678,027,248
|
IssuesEvent
|
2015-03-26 07:12:09
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
test failure #914
|
test-failure
|
The following test appears to have failed:
[#914](https://circleci.com/gh/cockroachdb/cockroach/914):
```
I0326 05:26:29.194389 81 multiraft.go:449] node 1: group 1 got message 8->1 MsgHeartbeatResp Term:6 Log:0/0
I0326 05:26:29.194408 81 multiraft.go:449] node 1: group 1 got message 4->1 MsgHeartbeatResp Term:6 Log:0/0
I0326 05:26:29.194422 81 multiraft.go:449] node 1: group 1 got message 5->1 MsgAppResp Term:6 Log:0/11
I0326 05:26:29.194440 81 multiraft.go:449] node 1: group 1 got message 6->1 MsgAppResp Term:6 Log:0/11
E0326 05:26:29.687754 81 heartbeat_test.go:51] timeout when reading from intercept channel
panic: test timed out after 30s
goroutine 384 [running]:
testing.func·008()
/usr/src/go/src/testing/testing.go:681 +0x12f
created by time.goFunc
/usr/src/go/src/time/sleep.go:129 +0x4b
goroutine 1 [chan receive]:
testing.RunTests(0x9a5d20, 0xb0d340, 0x7, 0x7, 0xc207ffc301)
/usr/src/go/src/testing/testing.go:556 +0xad6
--
goroutine 345 [select]:
github.com/cockroachdb/cockroach/multiraft.func·009()
/go/src/github.com/cockroachdb/cockroach/multiraft/events_test.go:51 +0x3fc
created by github.com/cockroachdb/cockroach/multiraft.(*eventDemux).start
/go/src/github.com/cockroachdb/cockroach/multiraft/events_test.go:72 +0x8c
FAIL github.com/cockroachdb/cockroach/multiraft 30.023s
=== RUN TestMemoryStorage
--- PASS: TestMemoryStorage (0.00s)
PASS
ok github.com/cockroachdb/cockroach/multiraft/storagetest 0.004s
=== RUN TestClientCmdIDIsEmpty
--- PASS: TestClientCmdIDIsEmpty (0.00s)
=== RUN TestResponseHeaderSetGoError
--- PASS: TestResponseHeaderSetGoError (0.00s)
=== RUN TestResponseHeaderNilError
--- PASS: TestResponseHeaderNilError (0.00s)
I0326 05:26:29.194389 81 multiraft.go:449] node 1: group 1 got message 8->1 MsgHeartbeatResp Term:6 Log:0/0
I0326 05:26:29.194408 81 multiraft.go:449] node 1: group 1 got message 4->1 MsgHeartbeatResp Term:6 Log:0/0
I0326 05:26:29.194422 81 multiraft.go:449] node 1: group 1 got message 5->1 MsgAppResp Term:6 Log:0/11
I0326 05:26:29.194440 81 multiraft.go:449] node 1: group 1 got message 6->1 MsgAppResp Term:6 Log:0/11
E0326 05:26:29.687754 81 heartbeat_test.go:51] timeout when reading from intercept channel
panic: test timed out after 30s
goroutine 384 [running]:
testing.func·008()
/usr/src/go/src/testing/testing.go:681 +0x12f
created by time.goFunc
/usr/src/go/src/time/sleep.go:129 +0x4b
goroutine 1 [chan receive]:
testing.RunTests(0x9a5d20, 0xb0d340, 0x7, 0x7, 0xc207ffc301)
/usr/src/go/src/testing/testing.go:556 +0xad6
--
goroutine 345 [select]:
github.com/cockroachdb/cockroach/multiraft.func·009()
/go/src/github.com/cockroachdb/cockroach/multiraft/events_test.go:51 +0x3fc
created by github.com/cockroachdb/cockroach/multiraft.(*eventDemux).start
/go/src/github.com/cockroachdb/cockroach/multiraft/events_test.go:72 +0x8c
FAIL github.com/cockroachdb/cockroach/multiraft 30.023s
=== RUN TestMemoryStorage
--- PASS: TestMemoryStorage (0.00s)
PASS
ok github.com/cockroachdb/cockroach/multiraft/storagetest 0.004s
=== RUN TestClientCmdIDIsEmpty
--- PASS: TestClientCmdIDIsEmpty (0.00s)
=== RUN TestResponseHeaderSetGoError
--- PASS: TestResponseHeaderSetGoError (0.00s)
=== RUN TestResponseHeaderNilError
--- PASS: TestResponseHeaderNilError (0.00s)
```
Please assign, take a look and update the issue accordingly.
|
1.0
|
test failure #914 - The following test appears to have failed:
[#914](https://circleci.com/gh/cockroachdb/cockroach/914):
```
I0326 05:26:29.194389 81 multiraft.go:449] node 1: group 1 got message 8->1 MsgHeartbeatResp Term:6 Log:0/0
I0326 05:26:29.194408 81 multiraft.go:449] node 1: group 1 got message 4->1 MsgHeartbeatResp Term:6 Log:0/0
I0326 05:26:29.194422 81 multiraft.go:449] node 1: group 1 got message 5->1 MsgAppResp Term:6 Log:0/11
I0326 05:26:29.194440 81 multiraft.go:449] node 1: group 1 got message 6->1 MsgAppResp Term:6 Log:0/11
E0326 05:26:29.687754 81 heartbeat_test.go:51] timeout when reading from intercept channel
panic: test timed out after 30s
goroutine 384 [running]:
testing.func·008()
/usr/src/go/src/testing/testing.go:681 +0x12f
created by time.goFunc
/usr/src/go/src/time/sleep.go:129 +0x4b
goroutine 1 [chan receive]:
testing.RunTests(0x9a5d20, 0xb0d340, 0x7, 0x7, 0xc207ffc301)
/usr/src/go/src/testing/testing.go:556 +0xad6
--
goroutine 345 [select]:
github.com/cockroachdb/cockroach/multiraft.func·009()
/go/src/github.com/cockroachdb/cockroach/multiraft/events_test.go:51 +0x3fc
created by github.com/cockroachdb/cockroach/multiraft.(*eventDemux).start
/go/src/github.com/cockroachdb/cockroach/multiraft/events_test.go:72 +0x8c
FAIL github.com/cockroachdb/cockroach/multiraft 30.023s
=== RUN TestMemoryStorage
--- PASS: TestMemoryStorage (0.00s)
PASS
ok github.com/cockroachdb/cockroach/multiraft/storagetest 0.004s
=== RUN TestClientCmdIDIsEmpty
--- PASS: TestClientCmdIDIsEmpty (0.00s)
=== RUN TestResponseHeaderSetGoError
--- PASS: TestResponseHeaderSetGoError (0.00s)
=== RUN TestResponseHeaderNilError
--- PASS: TestResponseHeaderNilError (0.00s)
I0326 05:26:29.194389 81 multiraft.go:449] node 1: group 1 got message 8->1 MsgHeartbeatResp Term:6 Log:0/0
I0326 05:26:29.194408 81 multiraft.go:449] node 1: group 1 got message 4->1 MsgHeartbeatResp Term:6 Log:0/0
I0326 05:26:29.194422 81 multiraft.go:449] node 1: group 1 got message 5->1 MsgAppResp Term:6 Log:0/11
I0326 05:26:29.194440 81 multiraft.go:449] node 1: group 1 got message 6->1 MsgAppResp Term:6 Log:0/11
E0326 05:26:29.687754 81 heartbeat_test.go:51] timeout when reading from intercept channel
panic: test timed out after 30s
goroutine 384 [running]:
testing.func·008()
/usr/src/go/src/testing/testing.go:681 +0x12f
created by time.goFunc
/usr/src/go/src/time/sleep.go:129 +0x4b
goroutine 1 [chan receive]:
testing.RunTests(0x9a5d20, 0xb0d340, 0x7, 0x7, 0xc207ffc301)
/usr/src/go/src/testing/testing.go:556 +0xad6
--
goroutine 345 [select]:
github.com/cockroachdb/cockroach/multiraft.func·009()
/go/src/github.com/cockroachdb/cockroach/multiraft/events_test.go:51 +0x3fc
created by github.com/cockroachdb/cockroach/multiraft.(*eventDemux).start
/go/src/github.com/cockroachdb/cockroach/multiraft/events_test.go:72 +0x8c
FAIL github.com/cockroachdb/cockroach/multiraft 30.023s
=== RUN TestMemoryStorage
--- PASS: TestMemoryStorage (0.00s)
PASS
ok github.com/cockroachdb/cockroach/multiraft/storagetest 0.004s
=== RUN TestClientCmdIDIsEmpty
--- PASS: TestClientCmdIDIsEmpty (0.00s)
=== RUN TestResponseHeaderSetGoError
--- PASS: TestResponseHeaderSetGoError (0.00s)
=== RUN TestResponseHeaderNilError
--- PASS: TestResponseHeaderNilError (0.00s)
```
Please assign, take a look and update the issue accordingly.
|
test
|
test failure the following test appears to have failed multiraft go node group got message msgheartbeatresp term log multiraft go node group got message msgheartbeatresp term log multiraft go node group got message msgappresp term log multiraft go node group got message msgappresp term log heartbeat test go timeout when reading from intercept channel panic test timed out after goroutine testing func· usr src go src testing testing go created by time gofunc usr src go src time sleep go goroutine testing runtests usr src go src testing testing go goroutine github com cockroachdb cockroach multiraft func· go src github com cockroachdb cockroach multiraft events test go created by github com cockroachdb cockroach multiraft eventdemux start go src github com cockroachdb cockroach multiraft events test go fail github com cockroachdb cockroach multiraft run testmemorystorage pass testmemorystorage pass ok github com cockroachdb cockroach multiraft storagetest run testclientcmdidisempty pass testclientcmdidisempty run testresponseheadersetgoerror pass testresponseheadersetgoerror run testresponseheadernilerror pass testresponseheadernilerror multiraft go node group got message msgheartbeatresp term log multiraft go node group got message msgheartbeatresp term log multiraft go node group got message msgappresp term log multiraft go node group got message msgappresp term log heartbeat test go timeout when reading from intercept channel panic test timed out after goroutine testing func· usr src go src testing testing go created by time gofunc usr src go src time sleep go goroutine testing runtests usr src go src testing testing go goroutine github com cockroachdb cockroach multiraft func· go src github com cockroachdb cockroach multiraft events test go created by github com cockroachdb cockroach multiraft eventdemux start go src github com cockroachdb cockroach multiraft events test go fail github com cockroachdb cockroach multiraft run testmemorystorage pass testmemorystorage pass ok github com cockroachdb cockroach multiraft storagetest run testclientcmdidisempty pass testclientcmdidisempty run testresponseheadersetgoerror pass testresponseheadersetgoerror run testresponseheadernilerror pass testresponseheadernilerror please assign take a look and update the issue accordingly
| 1
|
126,949
| 17,146,277,094
|
IssuesEvent
|
2021-07-13 14:54:29
|
wordpress-mobile/gutenberg-mobile
|
https://api.github.com/repos/wordpress-mobile/gutenberg-mobile
|
closed
|
Add Layout Grid - 1st iteration: Read only version
|
Blocks [Status] Needs Design
|
What we are aiming on this iteration is rendering the grid correctly on mobile in different content size options but not yet allowing the users to change settings.
There will be a [separate effort](https://github.com/wordpress-mobile/gutenberg-mobile/issues/2516) to decide for which sites we should enable this block on mobile.
@enejb After making an initial investigation please drop a comment explaining how you'll approach developing this, what your sub-tasks will look like. Thanks!
@iamthomasbishop We need some directions about UX, we are planning to start working on this ~Aug 10-13. So the sooner we can clarify somethings the more efficient our work will be. Thanks!
|
1.0
|
Add Layout Grid - 1st iteration: Read only version - What we are aiming on this iteration is rendering the grid correctly on mobile in different content size options but not yet allowing the users to change settings.
There will be a [separate effort](https://github.com/wordpress-mobile/gutenberg-mobile/issues/2516) to decide for which sites we should enable this block on mobile.
@enejb After making an initial investigation please drop a comment explaining how you'll approach developing this, what your sub-tasks will look like. Thanks!
@iamthomasbishop We need some directions about UX, we are planning to start working on this ~Aug 10-13. So the sooner we can clarify somethings the more efficient our work will be. Thanks!
|
non_test
|
add layout grid iteration read only version what we are aiming on this iteration is rendering the grid correctly on mobile in different content size options but not yet allowing the users to change settings there will be a to decide for which sites we should enable this block on mobile enejb after making an initial investigation please drop a comment explaining how you ll approach developing this what your sub tasks will look like thanks iamthomasbishop we need some directions about ux we are planning to start working on this aug so the sooner we can clarify somethings the more efficient our work will be thanks
| 0
|
26,719
| 6,796,334,460
|
IssuesEvent
|
2017-11-01 18:39:06
|
Dijji/RepairTasks
|
https://api.github.com/repos/Dijji/RepairTasks
|
closed
|
wind 10 to win 7 still errors
|
CodePlexMigrationInitiated Impact: Unassigned
|
I couldn't make the unzip and remove function work. Still orig 5 errors. After scan, comment was not formatted.
I am a klutx, so maybe its me.
thankks
#### This work item was migrated from CodePlex
CodePlex work item ID: '3'
Vote count: '1'
|
1.0
|
wind 10 to win 7 still errors - I couldn't make the unzip and remove function work. Still orig 5 errors. After scan, comment was not formatted.
I am a klutx, so maybe its me.
thankks
#### This work item was migrated from CodePlex
CodePlex work item ID: '3'
Vote count: '1'
|
non_test
|
wind to win still errors i couldn t make the unzip and remove function work still orig errors after scan comment was not formatted i am a klutx so maybe its me thankks this work item was migrated from codeplex codeplex work item id vote count
| 0
|
510,759
| 14,815,778,583
|
IssuesEvent
|
2021-01-14 07:57:28
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
lms.autozone.com - site is not usable
|
browser-firefox engine-gecko os-linux priority-normal type-flash
|
<!-- @browser: Firefox 68.0 -->
<!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Firefox/68.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/65511 -->
**URL**: https://lms.autozone.com/Kview/CustomCodeBehind/base/courseware/scorm/scorm12courseframe.aspx
**Browser / Version**: Firefox 68.0
**Operating System**: Linux
**Tested Another Browser**: Yes Other
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/1/dc4c0dd8-565a-496f-a6fb-728343d627c9.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200409090751</li><li>channel: default</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/1/db35845e-e7e9-464d-bf24-f51f2cda37a1)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
lms.autozone.com - site is not usable - <!-- @browser: Firefox 68.0 -->
<!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Firefox/68.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/65511 -->
**URL**: https://lms.autozone.com/Kview/CustomCodeBehind/base/courseware/scorm/scorm12courseframe.aspx
**Browser / Version**: Firefox 68.0
**Operating System**: Linux
**Tested Another Browser**: Yes Other
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/1/dc4c0dd8-565a-496f-a6fb-728343d627c9.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200409090751</li><li>channel: default</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2021/1/db35845e-e7e9-464d-bf24-f51f2cda37a1)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_test
|
lms autozone com site is not usable url browser version firefox operating system linux tested another browser yes other problem type site is not usable description page not loading correctly steps to reproduce view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel default hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
18,572
| 3,072,603,030
|
IssuesEvent
|
2015-08-19 17:45:34
|
CocoaPods/CocoaPods
|
https://api.github.com/repos/CocoaPods/CocoaPods
|
closed
|
Missing Required Modules
|
s2:confirmed t2:defect
|
I amusing the 0.36.0 beta 2 version of the plugin and working to create an app using an Extension. I want my app framework to be the only part that uses certain cocoa pods. However when I try to import my AppKit into my main project (using swift) the complier complains that it is missing the required modules, which happen to be the cocoapods that I am not including in my main app.


|
1.0
|
Missing Required Modules - I amusing the 0.36.0 beta 2 version of the plugin and working to create an app using an Extension. I want my app framework to be the only part that uses certain cocoa pods. However when I try to import my AppKit into my main project (using swift) the complier complains that it is missing the required modules, which happen to be the cocoapods that I am not including in my main app.


|
non_test
|
missing required modules i amusing the beta version of the plugin and working to create an app using an extension i want my app framework to be the only part that uses certain cocoa pods however when i try to import my appkit into my main project using swift the complier complains that it is missing the required modules which happen to be the cocoapods that i am not including in my main app
| 0
|
76,604
| 7,541,418,562
|
IssuesEvent
|
2018-04-17 09:43:04
|
ISISScientificComputing/autoreduce
|
https://api.github.com/repos/ISISScientificComputing/autoreduce
|
closed
|
Static analysis
|
Maintenance Testing
|
This should be added after we have integrated Travis to the Repo. I have already tested that this works, but there are a notable amount of checks to fix - however they are mostly white space at a glance. Pylint or Flake8 can be used for these they both use the same standard (PEP8) so should produce roughly the same results.
As a result of this issue, the warnings will also have to be fixed!
|
1.0
|
Static analysis - This should be added after we have integrated Travis to the Repo. I have already tested that this works, but there are a notable amount of checks to fix - however they are mostly white space at a glance. Pylint or Flake8 can be used for these they both use the same standard (PEP8) so should produce roughly the same results.
As a result of this issue, the warnings will also have to be fixed!
|
test
|
static analysis this should be added after we have integrated travis to the repo i have already tested that this works but there are a notable amount of checks to fix however they are mostly white space at a glance pylint or can be used for these they both use the same standard so should produce roughly the same results as a result of this issue the warnings will also have to be fixed
| 1
|
318,445
| 27,304,568,627
|
IssuesEvent
|
2023-02-24 06:53:01
|
lightningnetwork/lnd
|
https://api.github.com/repos/lightningnetwork/lnd
|
closed
|
itest-flake: co-op channel fail during update chan policy test teardown
|
test flake itests btcd
|
At the very end of the test, for `btcd` backend (so far):
```
test_harness.go:88: Failed: (update channel policy): exited with error:
*errors.errorString unable to close channel: error while waiting for broadcast tx: wanted aedef3c01bb9e4c046acb0dddd6aaf65476a9a5c2cd7500ab41b06909ccc7e49, found 0 txs in mempool: []
/home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:242 (0xf32df8)
closeChannelAndAssertType: t.Fatalf("unable to close channel: %v", err)
/home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:217 (0xf3ae7f)
testUpdateChannelPolicy: return closeChannelAndAssertType(ctx, t, net, node, fundingChanPoint, false, force)
/home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:2058 (0xf3ae31)
testUpdateChannelPolicy: closeChannelAndAssert(ctxt, t, net, net.Alice, chanPoint, false)
/home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/test_harness.go:112 (0xeeb1ee)
(*harnessTest).RunTestCase: testCase.test(h.lndHarness, h)
/home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:11702 (0xf9ceaf)
TestLightningNetworkDaemon.func4: ht.RunTestCase(testCase)
/home/travis/.gimme/versions/go1.16.3.linux.amd64/src/testing/testing.go:1193 (0x51c72f)
tRunner: fn(t)
/home/travis/.gimme/versions/go1.16.3.linux.amd64/src/runtime/asm_amd64.s:1371 (0x471c01)
goexit: BYTE $0x90 // NOP
=== CONT TestLightningNetworkDaemon
lnd_test.go:11710: Failure time: 2021-07-03 13:30:47.057
--- FAIL: TestLightningNetworkDaemon (806.43s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc (596.71s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy (232.50s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/local_force_close_immediate_expiry (27.72s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/receiver_chain_claim (30.49s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/local_force_close_on-chain_htlc_timeout (26.10s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/remote_force_close_on-chain_htlc_timeout (25.80s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/local_chain_claim (45.21s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/remote_chain_claim (34.75s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/local_htlc_aggregation (37.98s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors (343.77s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/local_force_close_immediate_expiry (44.10s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/receiver_chain_claim (38.92s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/local_force_close_on-chain_htlc_timeout (50.20s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/remote_force_close_on-chain_htlc_timeout (33.90s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/local_chain_claim (68.85s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/remote_chain_claim (57.60s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/local_htlc_aggregation (45.52s)
--- PASS: TestLightningNetworkDaemon/02-of-82/btcd/sweep_coins (6.09s)
--- PASS: TestLightningNetworkDaemon/03-of-82/btcd/recovery_info (14.40s)
--- PASS: TestLightningNetworkDaemon/04-of-82/btcd/onchain_fund_recovery (29.89s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow (66.18s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=legacy,dave_commit=legacy (7.02s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=legacy,dave_commit=tweakless (7.02s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=legacy,dave_commit=anchors (6.99s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=tweakless,dave_commit=legacy (6.99s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=tweakless,dave_commit=tweakless (6.98s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=tweakless,dave_commit=anchors (6.99s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=anchors,dave_commit=legacy (7.04s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=anchors,dave_commit=tweakless (6.99s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=anchors,dave_commit=anchors (6.97s)
--- PASS: TestLightningNetworkDaemon/06-of-82/btcd/unconfirmed_channel_funding (7.32s)
--- FAIL: TestLightningNetworkDaemon/07-of-82/btcd/update_channel_policy (74.45s)
```
|
2.0
|
itest-flake: co-op channel fail during update chan policy test teardown - At the very end of the test, for `btcd` backend (so far):
```
test_harness.go:88: Failed: (update channel policy): exited with error:
*errors.errorString unable to close channel: error while waiting for broadcast tx: wanted aedef3c01bb9e4c046acb0dddd6aaf65476a9a5c2cd7500ab41b06909ccc7e49, found 0 txs in mempool: []
/home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:242 (0xf32df8)
closeChannelAndAssertType: t.Fatalf("unable to close channel: %v", err)
/home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:217 (0xf3ae7f)
testUpdateChannelPolicy: return closeChannelAndAssertType(ctx, t, net, node, fundingChanPoint, false, force)
/home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:2058 (0xf3ae31)
testUpdateChannelPolicy: closeChannelAndAssert(ctxt, t, net, net.Alice, chanPoint, false)
/home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/test_harness.go:112 (0xeeb1ee)
(*harnessTest).RunTestCase: testCase.test(h.lndHarness, h)
/home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:11702 (0xf9ceaf)
TestLightningNetworkDaemon.func4: ht.RunTestCase(testCase)
/home/travis/.gimme/versions/go1.16.3.linux.amd64/src/testing/testing.go:1193 (0x51c72f)
tRunner: fn(t)
/home/travis/.gimme/versions/go1.16.3.linux.amd64/src/runtime/asm_amd64.s:1371 (0x471c01)
goexit: BYTE $0x90 // NOP
=== CONT TestLightningNetworkDaemon
lnd_test.go:11710: Failure time: 2021-07-03 13:30:47.057
--- FAIL: TestLightningNetworkDaemon (806.43s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc (596.71s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy (232.50s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/local_force_close_immediate_expiry (27.72s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/receiver_chain_claim (30.49s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/local_force_close_on-chain_htlc_timeout (26.10s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/remote_force_close_on-chain_htlc_timeout (25.80s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/local_chain_claim (45.21s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/remote_chain_claim (34.75s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/local_htlc_aggregation (37.98s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors (343.77s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/local_force_close_immediate_expiry (44.10s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/receiver_chain_claim (38.92s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/local_force_close_on-chain_htlc_timeout (50.20s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/remote_force_close_on-chain_htlc_timeout (33.90s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/local_chain_claim (68.85s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/remote_chain_claim (57.60s)
--- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/local_htlc_aggregation (45.52s)
--- PASS: TestLightningNetworkDaemon/02-of-82/btcd/sweep_coins (6.09s)
--- PASS: TestLightningNetworkDaemon/03-of-82/btcd/recovery_info (14.40s)
--- PASS: TestLightningNetworkDaemon/04-of-82/btcd/onchain_fund_recovery (29.89s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow (66.18s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=legacy,dave_commit=legacy (7.02s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=legacy,dave_commit=tweakless (7.02s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=legacy,dave_commit=anchors (6.99s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=tweakless,dave_commit=legacy (6.99s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=tweakless,dave_commit=tweakless (6.98s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=tweakless,dave_commit=anchors (6.99s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=anchors,dave_commit=legacy (7.04s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=anchors,dave_commit=tweakless (6.99s)
--- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=anchors,dave_commit=anchors (6.97s)
--- PASS: TestLightningNetworkDaemon/06-of-82/btcd/unconfirmed_channel_funding (7.32s)
--- FAIL: TestLightningNetworkDaemon/07-of-82/btcd/update_channel_policy (74.45s)
```
|
test
|
itest flake co op channel fail during update chan policy test teardown at the very end of the test for btcd backend so far test harness go failed update channel policy exited with error errors errorstring unable to close channel error while waiting for broadcast tx wanted found txs in mempool home travis gopath src github com lightningnetwork lnd lntest itest lnd test go closechannelandasserttype t fatalf unable to close channel v err home travis gopath src github com lightningnetwork lnd lntest itest lnd test go testupdatechannelpolicy return closechannelandasserttype ctx t net node fundingchanpoint false force home travis gopath src github com lightningnetwork lnd lntest itest lnd test go testupdatechannelpolicy closechannelandassert ctxt t net net alice chanpoint false home travis gopath src github com lightningnetwork lnd lntest itest test harness go harnesstest runtestcase testcase test h lndharness h home travis gopath src github com lightningnetwork lnd lntest itest lnd test go testlightningnetworkdaemon ht runtestcase testcase home travis gimme versions linux src testing testing go trunner fn t home travis gimme versions linux src runtime asm s goexit byte nop cont testlightningnetworkdaemon lnd test go failure time fail testlightningnetworkdaemon pass testlightningnetworkdaemon of btcd test multi hop htlc pass testlightningnetworkdaemon of btcd test multi hop htlc committype legacy pass testlightningnetworkdaemon of btcd test multi hop htlc committype legacy local force close immediate expiry pass testlightningnetworkdaemon of btcd test multi hop htlc committype legacy receiver chain claim pass testlightningnetworkdaemon of btcd test multi hop htlc committype legacy local force close on chain htlc timeout pass testlightningnetworkdaemon of btcd test multi hop htlc committype legacy remote force close on chain htlc timeout pass testlightningnetworkdaemon of btcd test multi hop htlc committype legacy local chain claim pass testlightningnetworkdaemon of btcd test multi hop htlc committype legacy remote chain claim pass testlightningnetworkdaemon of btcd test multi hop htlc committype legacy local htlc aggregation pass testlightningnetworkdaemon of btcd test multi hop htlc committype anchors pass testlightningnetworkdaemon of btcd test multi hop htlc committype anchors local force close immediate expiry pass testlightningnetworkdaemon of btcd test multi hop htlc committype anchors receiver chain claim pass testlightningnetworkdaemon of btcd test multi hop htlc committype anchors local force close on chain htlc timeout pass testlightningnetworkdaemon of btcd test multi hop htlc committype anchors remote force close on chain htlc timeout pass testlightningnetworkdaemon of btcd test multi hop htlc committype anchors local chain claim pass testlightningnetworkdaemon of btcd test multi hop htlc committype anchors remote chain claim pass testlightningnetworkdaemon of btcd test multi hop htlc committype anchors local htlc aggregation pass testlightningnetworkdaemon of btcd sweep coins pass testlightningnetworkdaemon of btcd recovery info pass testlightningnetworkdaemon of btcd onchain fund recovery pass testlightningnetworkdaemon of btcd basic funding flow pass testlightningnetworkdaemon of btcd basic funding flow carol commit legacy dave commit legacy pass testlightningnetworkdaemon of btcd basic funding flow carol commit legacy dave commit tweakless pass testlightningnetworkdaemon of btcd basic funding flow carol commit legacy dave commit anchors pass testlightningnetworkdaemon of btcd basic funding flow carol commit tweakless dave commit legacy pass testlightningnetworkdaemon of btcd basic funding flow carol commit tweakless dave commit tweakless pass testlightningnetworkdaemon of btcd basic funding flow carol commit tweakless dave commit anchors pass testlightningnetworkdaemon of btcd basic funding flow carol commit anchors dave commit legacy pass testlightningnetworkdaemon of btcd basic funding flow carol commit anchors dave commit tweakless pass testlightningnetworkdaemon of btcd basic funding flow carol commit anchors dave commit anchors pass testlightningnetworkdaemon of btcd unconfirmed channel funding fail testlightningnetworkdaemon of btcd update channel policy
| 1
|
349,699
| 31,822,190,424
|
IssuesEvent
|
2023-09-14 03:54:37
|
longhorn/longhorn
|
https://api.github.com/repos/longhorn/longhorn
|
opened
|
[TEST][FEATURE] SPDK volumes support volume backup/restore
|
kind/test
|
adding/updating auto e2e test cases for https://github.com/longhorn/longhorn/issues/6138 if they can be automated
cc @longhorn/qa
|
1.0
|
[TEST][FEATURE] SPDK volumes support volume backup/restore - adding/updating auto e2e test cases for https://github.com/longhorn/longhorn/issues/6138 if they can be automated
cc @longhorn/qa
|
test
|
spdk volumes support volume backup restore adding updating auto test cases for if they can be automated cc longhorn qa
| 1
|
1,898
| 2,578,422,325
|
IssuesEvent
|
2015-02-12 23:08:06
|
Elgg/Elgg
|
https://api.github.com/repos/Elgg/Elgg
|
closed
|
Running unit tests using Developer Tools
|
tests
|
It looks like we are shipping the distribution without dev dependencies included, so unit tests won't run unless dependencies are installed using composer. Perhaps worth noting this somewhere.
|
1.0
|
Running unit tests using Developer Tools - It looks like we are shipping the distribution without dev dependencies included, so unit tests won't run unless dependencies are installed using composer. Perhaps worth noting this somewhere.
|
test
|
running unit tests using developer tools it looks like we are shipping the distribution without dev dependencies included so unit tests won t run unless dependencies are installed using composer perhaps worth noting this somewhere
| 1
|
230,854
| 18,719,852,325
|
IssuesEvent
|
2021-11-03 10:30:33
|
WordPress/gutenberg
|
https://api.github.com/repos/WordPress/gutenberg
|
closed
|
Cover Block min height unit "vh" missing / only "vw" selectable
|
[Type] Bug Needs Testing [Block] Cover
|
### Description
I have noticed that you cannot select "vh" as unit for the min-height block setting of cover blocks, only "vh" (besides px etc.)


### Step-by-step reproduction instructions
- Editor view
- Insert Cover Block
- Select unit list for min height
### Screenshots, screen recording, code snippet
_No response_
### Environment info
WordPress 5.8.1
Gutenberg Version 11.8.0 (Plugin)
### Please confirm that you have searched existing issues in the repo.
Yes
### Please confirm that you have tested with all plugins deactivated except Gutenberg.
Yes
|
1.0
|
Cover Block min height unit "vh" missing / only "vw" selectable - ### Description
I have noticed that you cannot select "vh" as unit for the min-height block setting of cover blocks, only "vh" (besides px etc.)


### Step-by-step reproduction instructions
- Editor view
- Insert Cover Block
- Select unit list for min height
### Screenshots, screen recording, code snippet
_No response_
### Environment info
WordPress 5.8.1
Gutenberg Version 11.8.0 (Plugin)
### Please confirm that you have searched existing issues in the repo.
Yes
### Please confirm that you have tested with all plugins deactivated except Gutenberg.
Yes
|
test
|
cover block min height unit vh missing only vw selectable description i have noticed that you cannot select vh as unit for the min height block setting of cover blocks only vh besides px etc step by step reproduction instructions editor view insert cover block select unit list for min height screenshots screen recording code snippet no response environment info wordpress gutenberg version plugin please confirm that you have searched existing issues in the repo yes please confirm that you have tested with all plugins deactivated except gutenberg yes
| 1
|
265,225
| 20,074,203,099
|
IssuesEvent
|
2022-02-04 10:47:41
|
HWR-Berlin-SWE-II-Gruppe-2-Team-3-2022/Dokumentation
|
https://api.github.com/repos/HWR-Berlin-SWE-II-Gruppe-2-Team-3-2022/Dokumentation
|
closed
|
[Documentation] Textueller Projekt-Teaser
|
Documentation
|
## Detailed Description
Projekt-Teaser in Form einer textuellen Kurzbeschreibung des Projektthemas (200-250 Wörter),
Abgabe eines Dokuments im PDF-Format.
(5 / 100 Punkte)
### Tasks
- [x] Wortanzahl erreichen
- [x] "Corporate Design" etablieren
- [x] Logo anpassen
|
1.0
|
[Documentation] Textueller Projekt-Teaser - ## Detailed Description
Projekt-Teaser in Form einer textuellen Kurzbeschreibung des Projektthemas (200-250 Wörter),
Abgabe eines Dokuments im PDF-Format.
(5 / 100 Punkte)
### Tasks
- [x] Wortanzahl erreichen
- [x] "Corporate Design" etablieren
- [x] Logo anpassen
|
non_test
|
textueller projekt teaser detailed description projekt teaser in form einer textuellen kurzbeschreibung des projektthemas wörter abgabe eines dokuments im pdf format punkte tasks wortanzahl erreichen corporate design etablieren logo anpassen
| 0
|
163,204
| 12,707,279,524
|
IssuesEvent
|
2020-06-23 08:39:40
|
wazuh/wazuh-kibana-app
|
https://api.github.com/repos/wazuh/wazuh-kibana-app
|
closed
|
Testing WUI 3.13.0 - 7.7.0/7.7.1
|
testing
|
All the tests must be done on every common browser and using the Open Distro plugin.
- [x] Chrome
- [x] Firefox
- [x] Safari
|
1.0
|
Testing WUI 3.13.0 - 7.7.0/7.7.1 - All the tests must be done on every common browser and using the Open Distro plugin.
- [x] Chrome
- [x] Firefox
- [x] Safari
|
test
|
testing wui all the tests must be done on every common browser and using the open distro plugin chrome firefox safari
| 1
|
230,662
| 18,681,839,874
|
IssuesEvent
|
2021-11-01 07:09:12
|
RasaHQ/rasa
|
https://api.github.com/repos/RasaHQ/rasa
|
opened
|
Scheduled Model Regression Test Failed
|
type:bug :bug: tool:model-regression-tests
|
*This PR is automatically created by the Scheduled Model Regression Test workflow. Checkout the Github Action Run [here](https://github.com/RasaHQ/rasa/actions/runs/1405766870).* <br> --- <br> **Description of Problem:** <br> Scheduled Model Regression Test failed. <br> **Configuration**: `Rules + TED` <br> **Dataset**: `retail-demo`
|
1.0
|
Scheduled Model Regression Test Failed - *This PR is automatically created by the Scheduled Model Regression Test workflow. Checkout the Github Action Run [here](https://github.com/RasaHQ/rasa/actions/runs/1405766870).* <br> --- <br> **Description of Problem:** <br> Scheduled Model Regression Test failed. <br> **Configuration**: `Rules + TED` <br> **Dataset**: `retail-demo`
|
test
|
scheduled model regression test failed this pr is automatically created by the scheduled model regression test workflow checkout the github action run description of problem scheduled model regression test failed configuration rules ted dataset retail demo
| 1
|
642,257
| 20,871,923,513
|
IssuesEvent
|
2022-03-22 12:46:49
|
datavisyn/tdp_core
|
https://api.github.com/repos/datavisyn/tdp_core
|
closed
|
Add checksum to score states
|
type: feature priority: low
|
In case a score is restored using the score parameters stored in the provenance state, it would be useful to store a checksum that tells the user if the re-calculated scores are the same as the original ones. When the data in the database changes, this will not the case.
|
1.0
|
Add checksum to score states - In case a score is restored using the score parameters stored in the provenance state, it would be useful to store a checksum that tells the user if the re-calculated scores are the same as the original ones. When the data in the database changes, this will not the case.
|
non_test
|
add checksum to score states in case a score is restored using the score parameters stored in the provenance state it would be useful to store a checksum that tells the user if the re calculated scores are the same as the original ones when the data in the database changes this will not the case
| 0
|
19,063
| 3,749,579,549
|
IssuesEvent
|
2016-03-11 00:40:14
|
Microsoft/RTVS
|
https://api.github.com/repos/Microsoft/RTVS
|
closed
|
Need better output for unit-tests in case of RHost start failure
|
area:REPL-Backend type:enhancement type:test issue
|
Currently it isn't visible in test output and require debugging
|
1.0
|
Need better output for unit-tests in case of RHost start failure - Currently it isn't visible in test output and require debugging
|
test
|
need better output for unit tests in case of rhost start failure currently it isn t visible in test output and require debugging
| 1
|
14,074
| 3,374,045,686
|
IssuesEvent
|
2015-11-24 10:59:57
|
mantidproject/mantid
|
https://api.github.com/repos/mantidproject/mantid
|
opened
|
ISISPowderDiffractionPolarisTest: add a test case with ExistingV=yes
|
Component: Diffraction Group: Powder Quality: System Tests
|
The test added in #14534 checks the longer workflow (with ExistingV=no) with simple calibration files. We need to check with Ron what would be good calibration files from the last or a recent cycle to use for a test with `ExistingV=no` (that is, assuming a first calibration has been done at the beginning of the cycle and its results are ready to be re-used).
|
1.0
|
ISISPowderDiffractionPolarisTest: add a test case with ExistingV=yes - The test added in #14534 checks the longer workflow (with ExistingV=no) with simple calibration files. We need to check with Ron what would be good calibration files from the last or a recent cycle to use for a test with `ExistingV=no` (that is, assuming a first calibration has been done at the beginning of the cycle and its results are ready to be re-used).
|
test
|
isispowderdiffractionpolaristest add a test case with existingv yes the test added in checks the longer workflow with existingv no with simple calibration files we need to check with ron what would be good calibration files from the last or a recent cycle to use for a test with existingv no that is assuming a first calibration has been done at the beginning of the cycle and its results are ready to be re used
| 1
|
176,350
| 13,637,642,812
|
IssuesEvent
|
2020-09-25 08:10:11
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
opened
|
tests: arch: arm: arm_thread_swap: fails with bus fault
|
area: Tests bug
|
**Describe the bug**
tests/arch/arm/arm_thread_swap/ fails with bus fault on nucleo_l552ze_q
**To Reproduce**
Steps to reproduce the behavior:
1. sanitycheck -N --device-testing --hardware-map ../map.yaml -c -p nucleo_l552ze_q -T tests/arch/arm/arm_thread_swap/
2. See error
**Expected behavior**
test passed
**Logs and console output**
*** Booting Zephyr OS build v2.4.0-rc2-80-g1c3659519f69 ***
```
Running test suite arm_thread_swap
===================================================================
START - test_arm_thread_swap
E: ***** BUS FAULT *****
E: Precise data bus error
E: BFAR Address: 0x0
E: r0/a1: 0x20000334 r1/a2: 0x00000000 r2/a3: 0x00000040
E: r3/a4: 0x20000334 r12/ip: 0x00000000 r14/lr: 0x08001705
E: xpsr: 0x81000200
E: s[ 0]: 0x00000000 s[ 1]: 0x00000000 s[ 2]: 0x00000000 s[ 3]: 0x00000000
E: s[ 4]: 0x00000000 s[ 5]: 0x00000000 s[ 6]: 0x00000000 s[ 7]: 0x00000000
E: s[ 8]: 0x00000000 s[ 9]: 0x00000000 s[10]: 0x00000000 s[11]: 0x00000000
E: s[12]: 0x00000000 s[13]: 0x00000000 s[14]: 0x00000000 s[15]: 0x00000000
E: fpscr: 0x00000000
E: Faulting instruction address (r15/pc): 0x0800c104
E: >>> ZEPHYR FATAL ERROR 0: CPU exception on CPU 0
E: Current thread: 0x20000280 (unknown)
E: Halting system
```
**Environment (please complete the following information):**
- OS: Linux
- Toolchain Zephyr SDK
- Commit SHA 1c3659519f69e5f121b9f90b23b89a6cbbd26f8e
|
1.0
|
tests: arch: arm: arm_thread_swap: fails with bus fault - **Describe the bug**
tests/arch/arm/arm_thread_swap/ fails with bus fault on nucleo_l552ze_q
**To Reproduce**
Steps to reproduce the behavior:
1. sanitycheck -N --device-testing --hardware-map ../map.yaml -c -p nucleo_l552ze_q -T tests/arch/arm/arm_thread_swap/
2. See error
**Expected behavior**
test passed
**Logs and console output**
*** Booting Zephyr OS build v2.4.0-rc2-80-g1c3659519f69 ***
```
Running test suite arm_thread_swap
===================================================================
START - test_arm_thread_swap
E: ***** BUS FAULT *****
E: Precise data bus error
E: BFAR Address: 0x0
E: r0/a1: 0x20000334 r1/a2: 0x00000000 r2/a3: 0x00000040
E: r3/a4: 0x20000334 r12/ip: 0x00000000 r14/lr: 0x08001705
E: xpsr: 0x81000200
E: s[ 0]: 0x00000000 s[ 1]: 0x00000000 s[ 2]: 0x00000000 s[ 3]: 0x00000000
E: s[ 4]: 0x00000000 s[ 5]: 0x00000000 s[ 6]: 0x00000000 s[ 7]: 0x00000000
E: s[ 8]: 0x00000000 s[ 9]: 0x00000000 s[10]: 0x00000000 s[11]: 0x00000000
E: s[12]: 0x00000000 s[13]: 0x00000000 s[14]: 0x00000000 s[15]: 0x00000000
E: fpscr: 0x00000000
E: Faulting instruction address (r15/pc): 0x0800c104
E: >>> ZEPHYR FATAL ERROR 0: CPU exception on CPU 0
E: Current thread: 0x20000280 (unknown)
E: Halting system
```
**Environment (please complete the following information):**
- OS: Linux
- Toolchain Zephyr SDK
- Commit SHA 1c3659519f69e5f121b9f90b23b89a6cbbd26f8e
|
test
|
tests arch arm arm thread swap fails with bus fault describe the bug tests arch arm arm thread swap fails with bus fault on nucleo q to reproduce steps to reproduce the behavior sanitycheck n device testing hardware map map yaml c p nucleo q t tests arch arm arm thread swap see error expected behavior test passed logs and console output booting zephyr os build running test suite arm thread swap start test arm thread swap e bus fault e precise data bus error e bfar address e e ip lr e xpsr e s s s s e s s s s e s s s s e s s s s e fpscr e faulting instruction address pc e zephyr fatal error cpu exception on cpu e current thread unknown e halting system environment please complete the following information os linux toolchain zephyr sdk commit sha
| 1
|
762,138
| 26,709,895,996
|
IssuesEvent
|
2023-01-27 22:13:49
|
UTDNebula/planner
|
https://api.github.com/repos/UTDNebula/planner
|
closed
|
[Feature Request] Taking into Account Course Difficulty When Generating Degree Plans
|
Type: Feature Request Category: User Experience Status: On Hold Product Priority: Low Size: 8
|
# Overview
In cooperation with Trends, Planner could create degree plans that take into account/visualize course difficulty. A star rating could be shown under each course to show relative difficulty in order to assist the user further when deciding which classes to take at certain times. Further down the line, it would be nice if when Planner generates a degree plan, it adds courses in a way that the average difficulty of the semester is relatively close to each other. This would allow users who want to add a minor/extra courses later on to see the best opportunity to take them.
# Proposal
First would be working with trends to track a relative difficulty for a course (using grade distribution, professor grades, etc.) Then assigning a value to each course would be necessary. Following that, displaying the difficulty would be a start and then taking an average difficulty for a semester and displaying to the user would also be helpful. Later implementations can begin messing around with the generation of degree plans in a way that the remaining courses for someone is organized such that the difficulty is not outrageously different from semester to semester.
## Supplementary Information
Feel free to respond to this for more context.
|
1.0
|
[Feature Request] Taking into Account Course Difficulty When Generating Degree Plans - # Overview
In cooperation with Trends, Planner could create degree plans that take into account/visualize course difficulty. A star rating could be shown under each course to show relative difficulty in order to assist the user further when deciding which classes to take at certain times. Further down the line, it would be nice if when Planner generates a degree plan, it adds courses in a way that the average difficulty of the semester is relatively close to each other. This would allow users who want to add a minor/extra courses later on to see the best opportunity to take them.
# Proposal
First would be working with trends to track a relative difficulty for a course (using grade distribution, professor grades, etc.) Then assigning a value to each course would be necessary. Following that, displaying the difficulty would be a start and then taking an average difficulty for a semester and displaying to the user would also be helpful. Later implementations can begin messing around with the generation of degree plans in a way that the remaining courses for someone is organized such that the difficulty is not outrageously different from semester to semester.
## Supplementary Information
Feel free to respond to this for more context.
|
non_test
|
taking into account course difficulty when generating degree plans overview in cooperation with trends planner could create degree plans that take into account visualize course difficulty a star rating could be shown under each course to show relative difficulty in order to assist the user further when deciding which classes to take at certain times further down the line it would be nice if when planner generates a degree plan it adds courses in a way that the average difficulty of the semester is relatively close to each other this would allow users who want to add a minor extra courses later on to see the best opportunity to take them proposal first would be working with trends to track a relative difficulty for a course using grade distribution professor grades etc then assigning a value to each course would be necessary following that displaying the difficulty would be a start and then taking an average difficulty for a semester and displaying to the user would also be helpful later implementations can begin messing around with the generation of degree plans in a way that the remaining courses for someone is organized such that the difficulty is not outrageously different from semester to semester supplementary information feel free to respond to this for more context
| 0
|
336,470
| 30,195,477,313
|
IssuesEvent
|
2023-07-04 20:22:35
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: copy/bank/rows=100000,nodes=5,txn=true failed
|
C-test-failure O-robot A-testing O-roachtest T-kv branch-release-23.1
|
roachtest.copy/bank/rows=100000,nodes=5,txn=true [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/10756610?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/10756610?buildTab=artifacts#/copy/bank/rows=100000,nodes=5,txn=true) on release-23.1 @ [e12e85479312972b551677203849d29aeb38ad5f](https://github.com/cockroachdb/cockroach/commits/e12e85479312972b551677203849d29aeb38ad5f):
```
(cluster.go:2247).Run: output in run_063348.848455572_n1_workload-fixtures-lo: ./workload fixtures load bank --rows=100000 --payload-bytes=100 --seed 1 {pgurl:1} returned: COMMAND_PROBLEM: exit status 1
(monitor.go:137).Wait: monitor failure: monitor task failed: t.Fatal() was called
test artifacts and logs in: /artifacts/copy/bank/rows=100000_nodes=5_txn=true/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=aws</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*copy/bank/rows=100000,nodes=5,txn=true.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-29307
|
3.0
|
roachtest: copy/bank/rows=100000,nodes=5,txn=true failed - roachtest.copy/bank/rows=100000,nodes=5,txn=true [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/10756610?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/10756610?buildTab=artifacts#/copy/bank/rows=100000,nodes=5,txn=true) on release-23.1 @ [e12e85479312972b551677203849d29aeb38ad5f](https://github.com/cockroachdb/cockroach/commits/e12e85479312972b551677203849d29aeb38ad5f):
```
(cluster.go:2247).Run: output in run_063348.848455572_n1_workload-fixtures-lo: ./workload fixtures load bank --rows=100000 --payload-bytes=100 --seed 1 {pgurl:1} returned: COMMAND_PROBLEM: exit status 1
(monitor.go:137).Wait: monitor failure: monitor task failed: t.Fatal() was called
test artifacts and logs in: /artifacts/copy/bank/rows=100000_nodes=5_txn=true/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=aws</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*copy/bank/rows=100000,nodes=5,txn=true.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-29307
|
test
|
roachtest copy bank rows nodes txn true failed roachtest copy bank rows nodes txn true with on release cluster go run output in run workload fixtures lo workload fixtures load bank rows payload bytes seed pgurl returned command problem exit status monitor go wait monitor failure monitor task failed t fatal was called test artifacts and logs in artifacts copy bank rows nodes txn true run parameters roachtest arch roachtest cloud aws roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb kv triage jira issue crdb
| 1
|
236,986
| 7,754,586,259
|
IssuesEvent
|
2018-05-31 07:19:10
|
molgenis/molgenis
|
https://api.github.com/repos/molgenis/molgenis
|
closed
|
Internal server error when idAttribute of refEntity is visible == false and no other labelAttribute specified
|
2.1.0 4.1 5.2 7.0.0-RC bug mod:data-importer mod:dataexplorer priority-first
|
### How to Reproduce
Create EMX model with one entity referencing to another, where idAttribute of the reference entity is set on visible == false (and no other attribute is specified as labelAttribute)
### Expected behavior
The importer will give a warning or error
### Observed behavior
The EMX is correctly uploaded, but you get an internal server error 500 when selecting the entity
-Marieke has an example if this is not clear enough ;-)
|
1.0
|
Internal server error when idAttribute of refEntity is visible == false and no other labelAttribute specified - ### How to Reproduce
Create EMX model with one entity referencing to another, where idAttribute of the reference entity is set on visible == false (and no other attribute is specified as labelAttribute)
### Expected behavior
The importer will give a warning or error
### Observed behavior
The EMX is correctly uploaded, but you get an internal server error 500 when selecting the entity
-Marieke has an example if this is not clear enough ;-)
|
non_test
|
internal server error when idattribute of refentity is visible false and no other labelattribute specified how to reproduce create emx model with one entity referencing to another where idattribute of the reference entity is set on visible false and no other attribute is specified as labelattribute expected behavior the importer will give a warning or error observed behavior the emx is correctly uploaded but you get an internal server error when selecting the entity marieke has an example if this is not clear enough
| 0
|
11,714
| 14,546,541,572
|
IssuesEvent
|
2020-12-15 21:24:17
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
No Support for Generic Environment Resources
|
Pri2 devops-cicd-process/tech devops/prod product-feedback
|
I see that Azure DevOps yaml supports Kubernetes and Virtual Machines for deployment tracability in the form of "resources". Is there a reason there doesnt just exists a "generic" resource that is created if the resource is referenced through yml but doesnt exists yet? The same as environments are managed. That way jobs and commits will be recorded by resource based on the name of the resources.
Example:
`
environment: environment_name.resource_name
`
And the environment **environment_name** is created and the "generic" resource is created **resource_name**
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 77d95db6-9983-7346-d0eb-4b7443e4e252
* Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087
* Content: [Environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops)
* Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
No Support for Generic Environment Resources -
I see that Azure DevOps yaml supports Kubernetes and Virtual Machines for deployment tracability in the form of "resources". Is there a reason there doesnt just exists a "generic" resource that is created if the resource is referenced through yml but doesnt exists yet? The same as environments are managed. That way jobs and commits will be recorded by resource based on the name of the resources.
Example:
`
environment: environment_name.resource_name
`
And the environment **environment_name** is created and the "generic" resource is created **resource_name**
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 77d95db6-9983-7346-d0eb-4b7443e4e252
* Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087
* Content: [Environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops)
* Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
non_test
|
no support for generic environment resources i see that azure devops yaml supports kubernetes and virtual machines for deployment tracability in the form of resources is there a reason there doesnt just exists a generic resource that is created if the resource is referenced through yml but doesnt exists yet the same as environments are managed that way jobs and commits will be recorded by resource based on the name of the resources example environment environment name resource name and the environment environment name is created and the generic resource is created resource name document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 0
|
249,000
| 21,093,293,186
|
IssuesEvent
|
2022-04-04 07:57:44
|
blumilksoftware/meetup
|
https://api.github.com/repos/blumilksoftware/meetup
|
opened
|
Initialize Behat
|
environment feature tests
|
With Behat tests initially provided in #45, it would be nice to have a way to run these scenarios.
Behat should be added with newest version. Contexts should be properly configured and ready to be extended. The aim of this task is to setup Behat and is not to have all tests green.
|
1.0
|
Initialize Behat - With Behat tests initially provided in #45, it would be nice to have a way to run these scenarios.
Behat should be added with newest version. Contexts should be properly configured and ready to be extended. The aim of this task is to setup Behat and is not to have all tests green.
|
test
|
initialize behat with behat tests initially provided in it would be nice to have a way to run these scenarios behat should be added with newest version contexts should be properly configured and ready to be extended the aim of this task is to setup behat and is not to have all tests green
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.