Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
268,248
8,405,323,677
IssuesEvent
2018-10-11 14:59:29
CS2103-AY1819S1-F10-2/main
https://api.github.com/repos/CS2103-AY1819S1-F10-2/main
opened
I can register my bicycles within the app
priority.High type.Story
- Implement the Bike class. - Let the AddressBook class contain a set of Bikes. - Implement an addBike command.
1.0
I can register my bicycles within the app - - Implement the Bike class. - Let the AddressBook class contain a set of Bikes. - Implement an addBike command.
non_test
i can register my bicycles within the app implement the bike class let the addressbook class contain a set of bikes implement an addbike command
0
338,247
30,288,514,465
IssuesEvent
2023-07-09 01:21:11
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix creation.test_frombuffer
Sub Task Ivy API Experimental Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4902117212/jobs/8753770307"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4902117212/jobs/8753770307"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4902117212/jobs/8753770307"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4902117212/jobs/8753770307"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5497189625/jobs/10017714381"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix creation.test_frombuffer - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4902117212/jobs/8753770307"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4902117212/jobs/8753770307"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4902117212/jobs/8753770307"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4902117212/jobs/8753770307"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5497189625/jobs/10017714381"><img src=https://img.shields.io/badge/-success-success></a>
test
fix creation test frombuffer tensorflow a href src torch a href src jax a href src numpy a href src paddle a href src
1
247,000
20,955,364,259
IssuesEvent
2022-03-27 02:54:28
azuline/repertoire
https://api.github.com/repos/azuline/repertoire
closed
Storybook for components
frontend testing
For the `components/` directory. Consider also testing the components with Storybook.
1.0
Storybook for components - For the `components/` directory. Consider also testing the components with Storybook.
test
storybook for components for the components directory consider also testing the components with storybook
1
192,464
14,617,547,397
IssuesEvent
2020-12-22 14:56:36
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
praparn/kubernetes_201907: WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go; 5 LoC
fresh test tiny
Found a possible issue in [praparn/kubernetes_201907](https://www.github.com/praparn/kubernetes_201907) at [WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go](https://github.com/praparn/kubernetes_201907/blob/309486c01dfaa48464c7dce2ac5ec6742df92eef/WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go#L56-L60) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to m at line 58 may start a goroutine [Click here to see the code in its original context.](https://github.com/praparn/kubernetes_201907/blob/309486c01dfaa48464c7dce2ac5ec6742df92eef/WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go#L56-L60) <details> <summary>Click here to show the 5 line(s) of Go which triggered the analyzer.</summary> ```go for _, m := range metas[5:] { var buf bytes.Buffer testutil.Ok(t, json.NewEncoder(&buf).Encode(&m)) testutil.Ok(t, bkt.Upload(ctx, path.Join(m.ULID.String(), block.MetaFilename), &buf)) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 309486c01dfaa48464c7dce2ac5ec6742df92eef
1.0
praparn/kubernetes_201907: WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go; 5 LoC - Found a possible issue in [praparn/kubernetes_201907](https://www.github.com/praparn/kubernetes_201907) at [WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go](https://github.com/praparn/kubernetes_201907/blob/309486c01dfaa48464c7dce2ac5ec6742df92eef/WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go#L56-L60) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to m at line 58 may start a goroutine [Click here to see the code in its original context.](https://github.com/praparn/kubernetes_201907/blob/309486c01dfaa48464c7dce2ac5ec6742df92eef/WorkShop_2.10_Monitor/prometheus-operator/vendor/github.com/improbable-eng/thanos/pkg/compact/compact_e2e_test.go#L56-L60) <details> <summary>Click here to show the 5 line(s) of Go which triggered the analyzer.</summary> ```go for _, m := range metas[5:] { var buf bytes.Buffer testutil.Ok(t, json.NewEncoder(&buf).Encode(&m)) testutil.Ok(t, bkt.Upload(ctx, path.Join(m.ULID.String(), block.MetaFilename), &buf)) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 309486c01dfaa48464c7dce2ac5ec6742df92eef
test
praparn kubernetes workshop monitor prometheus operator vendor github com improbable eng thanos pkg compact compact test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to m at line may start a goroutine click here to show the line s of go which triggered the analyzer go for m range metas var buf bytes buffer testutil ok t json newencoder buf encode m testutil ok t bkt upload ctx path join m ulid string block metafilename buf leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
17,491
3,619,229,431
IssuesEvent
2016-02-08 15:17:18
Microsoft/vscode
https://api.github.com/repos/Microsoft/vscode
opened
Test working files picker
testplan-item
You can now quick open working files. * verify the picker is well integrated into our quick open story and either shows you all working files or the ones matching your search if searching * verify the order matches the order in the explorer unless you search, where we then use fuzzy matching/scoring * verify you see untitled files as well as dirty indicators for dirty files
1.0
Test working files picker - You can now quick open working files. * verify the picker is well integrated into our quick open story and either shows you all working files or the ones matching your search if searching * verify the order matches the order in the explorer unless you search, where we then use fuzzy matching/scoring * verify you see untitled files as well as dirty indicators for dirty files
test
test working files picker you can now quick open working files verify the picker is well integrated into our quick open story and either shows you all working files or the ones matching your search if searching verify the order matches the order in the explorer unless you search where we then use fuzzy matching scoring verify you see untitled files as well as dirty indicators for dirty files
1
7,682
9,935,667,963
IssuesEvent
2019-07-02 17:07:44
Direwolf20-MC/BuildingGadgets
https://api.github.com/repos/Direwolf20-MC/BuildingGadgets
closed
Gadget's drop Items, when Dim with bound Inv is deleted
1.12 bug incompatibility
A user reported this on Direcord: - bound to a chest in an rftools dim - delete said dim - Gadget's will now drop Items instead of putting them into the player inventory Notice that this happened on a dedicated Server - I don't know whether this can be reproduced using SP-mode. I presume he was using 2.6.8 but I didn't ask either.
True
Gadget's drop Items, when Dim with bound Inv is deleted - A user reported this on Direcord: - bound to a chest in an rftools dim - delete said dim - Gadget's will now drop Items instead of putting them into the player inventory Notice that this happened on a dedicated Server - I don't know whether this can be reproduced using SP-mode. I presume he was using 2.6.8 but I didn't ask either.
non_test
gadget s drop items when dim with bound inv is deleted a user reported this on direcord bound to a chest in an rftools dim delete said dim gadget s will now drop items instead of putting them into the player inventory notice that this happened on a dedicated server i don t know whether this can be reproduced using sp mode i presume he was using but i didn t ask either
0
90,492
11,405,758,977
IssuesEvent
2020-01-31 12:53:32
fecgov/fec-cms
https://api.github.com/repos/fecgov/fec-cms
opened
Design homepage change to link to map
High priority Work: UX/Design
**What we're after:** We're rebuilding the presidential data map on .gov and we need to create a pathway from the homepage to that map. ### Completion criteria - [ ] Mockup design(s) for the homepage to link to the map visualization - [ ] Review with stakeholders - [ ] Make necessary changes and move to an implementation issue
1.0
Design homepage change to link to map - **What we're after:** We're rebuilding the presidential data map on .gov and we need to create a pathway from the homepage to that map. ### Completion criteria - [ ] Mockup design(s) for the homepage to link to the map visualization - [ ] Review with stakeholders - [ ] Make necessary changes and move to an implementation issue
non_test
design homepage change to link to map what we re after we re rebuilding the presidential data map on gov and we need to create a pathway from the homepage to that map completion criteria mockup design s for the homepage to link to the map visualization review with stakeholders make necessary changes and move to an implementation issue
0
137,539
11,140,404,631
IssuesEvent
2019-12-21 14:08:27
dexpenses/dexpenses-extract
https://api.github.com/repos/dexpenses/dexpenses-extract
closed
Implement test receipt ec/bs-l-osteria-credit
enhancement test-data
Receipt to implement: ![ec/bs-l-osteria-credit](https://firebasestorage.googleapis.com/v0/b/dexpenses-207219-test-images/o/ec%2Fbs-l-osteria-credit.JPEG?alt=media "ec/bs-l-osteria-credit")
1.0
Implement test receipt ec/bs-l-osteria-credit - Receipt to implement: ![ec/bs-l-osteria-credit](https://firebasestorage.googleapis.com/v0/b/dexpenses-207219-test-images/o/ec%2Fbs-l-osteria-credit.JPEG?alt=media "ec/bs-l-osteria-credit")
test
implement test receipt ec bs l osteria credit receipt to implement ec bs l osteria credit
1
284,167
24,581,251,578
IssuesEvent
2022-10-13 15:46:31
IntellectualSites/PlotSquared
https://api.github.com/repos/IntellectualSites/PlotSquared
closed
Road schematic only paste´s top blocks not blocks under the top layer
Requires Testing
### Server Implementation Paper ### Server Version 1.19.2 ### Describe the bug Hey there, I´m trying to build a proper road schematic. I built it around the plot but not only did I modify the top layer. There are blocks beneath the top layer because I want to put beacons into the intersections with diamonds beneath. When I create the road schematic and try it out, only the top layer is put onto roads. Is this intended? Sincerely cynicalmax ### To Reproduce 1. Create plotworld 2. Modify road with blocks beneath toplayer 3. Create road schematic 4. Debugpaste the road schematic 5. done ### Expected behaviour Paste the whole road from Y:1 between the plots ### Screenshots / Videos _No response_ ### Error log (if applicable) _No response_ ### Plot Debugpaste https://athion.net/ISPaster/paste/view/21a47b95c5764a2589496e861cd28564 ### PlotSquared Version 6.9.4 Premium ### Checklist - [X] I have included a Plot debugpaste. - [X] I am using the newest build from https://www.spigotmc.org/resources/77506/ and the issue still persists. ### Anything else? _No response_
1.0
Road schematic only paste´s top blocks not blocks under the top layer - ### Server Implementation Paper ### Server Version 1.19.2 ### Describe the bug Hey there, I´m trying to build a proper road schematic. I built it around the plot but not only did I modify the top layer. There are blocks beneath the top layer because I want to put beacons into the intersections with diamonds beneath. When I create the road schematic and try it out, only the top layer is put onto roads. Is this intended? Sincerely cynicalmax ### To Reproduce 1. Create plotworld 2. Modify road with blocks beneath toplayer 3. Create road schematic 4. Debugpaste the road schematic 5. done ### Expected behaviour Paste the whole road from Y:1 between the plots ### Screenshots / Videos _No response_ ### Error log (if applicable) _No response_ ### Plot Debugpaste https://athion.net/ISPaster/paste/view/21a47b95c5764a2589496e861cd28564 ### PlotSquared Version 6.9.4 Premium ### Checklist - [X] I have included a Plot debugpaste. - [X] I am using the newest build from https://www.spigotmc.org/resources/77506/ and the issue still persists. ### Anything else? _No response_
test
road schematic only paste´s top blocks not blocks under the top layer server implementation paper server version describe the bug hey there i´m trying to build a proper road schematic i built it around the plot but not only did i modify the top layer there are blocks beneath the top layer because i want to put beacons into the intersections with diamonds beneath when i create the road schematic and try it out only the top layer is put onto roads is this intended sincerely cynicalmax to reproduce create plotworld modify road with blocks beneath toplayer create road schematic debugpaste the road schematic done expected behaviour paste the whole road from y between the plots screenshots videos no response error log if applicable no response plot debugpaste plotsquared version premium checklist i have included a plot debugpaste i am using the newest build from and the issue still persists anything else no response
1
114,903
9,764,285,076
IssuesEvent
2019-06-05 15:28:01
IBM/kui
https://api.github.com/repos/IBM/kui
closed
npm test with all layers miss k8s1 and k8s2
bug tests
LAYERS is hardcoded for non-openwhisk developers running tests.
1.0
npm test with all layers miss k8s1 and k8s2 - LAYERS is hardcoded for non-openwhisk developers running tests.
test
npm test with all layers miss and layers is hardcoded for non openwhisk developers running tests
1
72,500
13,879,638,912
IssuesEvent
2020-10-17 15:19:23
z3t0/Arduino-IRremote
https://api.github.com/repos/z3t0/Arduino-IRremote
closed
Mitsubishi AC Protocol Creation
Write Code
**Board:** Arduino Uno **Library Version:** 2.1.0 **Protocol:** Creation of Mitsubishi AC Protocol If anyone is up to it, would you be so kind as to create a Mitsubishi **Electric** AC protocol? The current Mitsubishi protocol is for a Mitsubishi TV (16 bits) and, thus, doesn't work for an AC (32 bits) Or, could you kindly guide me on what I would have to do create my own protocol?
1.0
Mitsubishi AC Protocol Creation - **Board:** Arduino Uno **Library Version:** 2.1.0 **Protocol:** Creation of Mitsubishi AC Protocol If anyone is up to it, would you be so kind as to create a Mitsubishi **Electric** AC protocol? The current Mitsubishi protocol is for a Mitsubishi TV (16 bits) and, thus, doesn't work for an AC (32 bits) Or, could you kindly guide me on what I would have to do create my own protocol?
non_test
mitsubishi ac protocol creation board arduino uno library version protocol creation of mitsubishi ac protocol if anyone is up to it would you be so kind as to create a mitsubishi electric ac protocol the current mitsubishi protocol is for a mitsubishi tv bits and thus doesn t work for an ac bits or could you kindly guide me on what i would have to do create my own protocol
0
13,451
9,975,252,459
IssuesEvent
2019-07-09 12:40:47
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Gpu Sharing on Azure Kubernetes
azure-container-service/svc container-service/svc cxp product-question triaged
Although there is currently no official support on k8s to share GPU resources among pods, there is an open source solution (https://github.com/AliyunContainerService/gpushare-scheduler-extender). The problem is that in order to apply the solution there is the need of direct access to the master node, and from what I understood the master node is controlled entirely by Azure. So my question is: is there already a solution to do gpu sharing on Azure Kubernetes? If not is it possible to do modification on the master node? I feel that the only solution I have for now is to create the VMs and set up both the master and worker nodes myself. Thanks in advance for any help/clarification. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: deaac79e-bd40-2d95-7288-e15371726e91 * Version Independent ID: 9fd80908-20f1-7970-3bda-5a4283e4af21 * Content: [Use GPUs on Azure Kubernetes Service (AKS)](https://docs.microsoft.com/en-us/azure/aks/gpu-cluster) * Content Source: [articles/aks/gpu-cluster.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/gpu-cluster.md) * Service: **container-service** * GitHub Login: @zr-msft * Microsoft Alias: **zarhoads**
2.0
Gpu Sharing on Azure Kubernetes - Although there is currently no official support on k8s to share GPU resources among pods, there is an open source solution (https://github.com/AliyunContainerService/gpushare-scheduler-extender). The problem is that in order to apply the solution there is the need of direct access to the master node, and from what I understood the master node is controlled entirely by Azure. So my question is: is there already a solution to do gpu sharing on Azure Kubernetes? If not is it possible to do modification on the master node? I feel that the only solution I have for now is to create the VMs and set up both the master and worker nodes myself. Thanks in advance for any help/clarification. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: deaac79e-bd40-2d95-7288-e15371726e91 * Version Independent ID: 9fd80908-20f1-7970-3bda-5a4283e4af21 * Content: [Use GPUs on Azure Kubernetes Service (AKS)](https://docs.microsoft.com/en-us/azure/aks/gpu-cluster) * Content Source: [articles/aks/gpu-cluster.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/gpu-cluster.md) * Service: **container-service** * GitHub Login: @zr-msft * Microsoft Alias: **zarhoads**
non_test
gpu sharing on azure kubernetes although there is currently no official support on to share gpu resources among pods there is an open source solution the problem is that in order to apply the solution there is the need of direct access to the master node and from what i understood the master node is controlled entirely by azure so my question is is there already a solution to do gpu sharing on azure kubernetes if not is it possible to do modification on the master node i feel that the only solution i have for now is to create the vms and set up both the master and worker nodes myself thanks in advance for any help clarification document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login zr msft microsoft alias zarhoads
0
57,880
16,114,546,169
IssuesEvent
2021-04-28 05:01:30
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Can't package dmg electron app on MacOS
A-Electron A-Packaging P2 S-Minor T-Defect Z-Platform-Specific
### Description Can't package electron app to dmg on MacOS High Sierra v10.13.2, process hangs. ### Steps to reproduce - git clone https://github.com/vector-im/riot-web.git && cd riot-web - npm install - npm run build - npm install electron - node_modules/.bin/build -m --x64 Describe how what happens differs from what you expected. The process just hangs. But even before trying to build the app open with ```npm run electron``` but get only a white window. ![screen shot 2017-12-25 at 3 15 54 pm](https://user-images.githubusercontent.com/13173679/34336078-afcf822c-e986-11e7-9fee-9e88802f17aa.png) ### Version information 0.13.3 <!-- IMPORTANT: please answer the following questions, to help us narrow down the problem --> - **Platform**: web (in-browser) or desktop? Desktop app, MacOS High Sierra 10.13.2 P.S: I'm pretty much a noob so it is absolutely possible that I'm doing something wrong, but following the instructions provided in the README.
1.0
Can't package dmg electron app on MacOS - ### Description Can't package electron app to dmg on MacOS High Sierra v10.13.2, process hangs. ### Steps to reproduce - git clone https://github.com/vector-im/riot-web.git && cd riot-web - npm install - npm run build - npm install electron - node_modules/.bin/build -m --x64 Describe how what happens differs from what you expected. The process just hangs. But even before trying to build the app open with ```npm run electron``` but get only a white window. ![screen shot 2017-12-25 at 3 15 54 pm](https://user-images.githubusercontent.com/13173679/34336078-afcf822c-e986-11e7-9fee-9e88802f17aa.png) ### Version information 0.13.3 <!-- IMPORTANT: please answer the following questions, to help us narrow down the problem --> - **Platform**: web (in-browser) or desktop? Desktop app, MacOS High Sierra 10.13.2 P.S: I'm pretty much a noob so it is absolutely possible that I'm doing something wrong, but following the instructions provided in the README.
non_test
can t package dmg electron app on macos description can t package electron app to dmg on macos high sierra process hangs steps to reproduce git clone cd riot web npm install npm run build npm install electron node modules bin build m describe how what happens differs from what you expected the process just hangs but even before trying to build the app open with npm run electron but get only a white window version information platform web in browser or desktop desktop app macos high sierra p s i m pretty much a noob so it is absolutely possible that i m doing something wrong but following the instructions provided in the readme
0
335,411
30,028,728,556
IssuesEvent
2023-06-27 08:11:09
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Cypress: Notifications toast cypress tests failing
A-Developer-Experience Z-Flaky-Test
### Steps to reproduce https://github.com/matrix-org/matrix-react-sdk/actions/runs/5386706635/jobs/9777756363#step:8:10104 https://cloud.cypress.io/projects/ppvnzg/runs/15772/test-results?actions=%5B%5D&browsers=%5B%5D&groups=%5B%5D&isFlaky=%5B%5D&modificationDateRange=%7B%22startDate%22%3A%221970-01-01%22%2C%22endDate%22%3A%222038-01-19%22%7D&orderBy=EXECUTION_ORDER&oses=%5B%5D&specs=%5B%5D&statuses=%5B%7B%22value%22%3A%22FAILED%22%2C%22label%22%3A%22FAILED%22%7D%5D&testingTypesEnum=%5B%5D&utm_source=github ### Outcome x ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp _No response_ ### Application version _No response_ ### Homeserver _No response_ ### Will you send logs? No
1.0
Cypress: Notifications toast cypress tests failing - ### Steps to reproduce https://github.com/matrix-org/matrix-react-sdk/actions/runs/5386706635/jobs/9777756363#step:8:10104 https://cloud.cypress.io/projects/ppvnzg/runs/15772/test-results?actions=%5B%5D&browsers=%5B%5D&groups=%5B%5D&isFlaky=%5B%5D&modificationDateRange=%7B%22startDate%22%3A%221970-01-01%22%2C%22endDate%22%3A%222038-01-19%22%7D&orderBy=EXECUTION_ORDER&oses=%5B%5D&specs=%5B%5D&statuses=%5B%7B%22value%22%3A%22FAILED%22%2C%22label%22%3A%22FAILED%22%7D%5D&testingTypesEnum=%5B%5D&utm_source=github ### Outcome x ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp _No response_ ### Application version _No response_ ### Homeserver _No response_ ### Will you send logs? No
test
cypress notifications toast cypress tests failing steps to reproduce outcome x operating system no response browser information no response url for webapp no response application version no response homeserver no response will you send logs no
1
278,284
24,142,940,858
IssuesEvent
2022-09-21 16:10:10
TahmidU/mosaic
https://api.github.com/repos/TahmidU/mosaic
closed
Jest RTL tests
test
# TODO ## Components - [x] Atom/Button (https://github.com/TahmidU/mosaic/pull/177) - [x] Atom/Card (https://github.com/TahmidU/mosaic/pull/178) - [x] Atom/Checkbox (https://github.com/TahmidU/mosaic/pull/179) - [x] Atom/DropdownList (https://github.com/TahmidU/mosaic/pull/180) - [x] Atom/ExternalLink (https://github.com/TahmidU/mosaic/pull/181) - [x] Atom/Input (https://github.com/TahmidU/mosaic/pull/182, https://github.com/TahmidU/mosaic/pull/184) - [x] Atom/ProgressiveCircle (https://github.com/TahmidU/mosaic/pull/183) - [x] Atom/SelectTitleList (https://github.com/TahmidU/mosaic/pull/185) - [x] Molecules/CastCard (https://github.com/TahmidU/mosaic/pull/186) - [x] Molecules/TypeFilter (https://github.com/TahmidU/mosaic/pull/187) - [x] Molecules/Filter (https://github.com/TahmidU/mosaic/pull/190) - [x] Molecules/HorizonalList (https://github.com/TahmidU/mosaic/pull/191, https://github.com/TahmidU/mosaic/pull/201) - [x] Molecules/MediaCard (https://github.com/TahmidU/mosaic/pull/193) - [x] Molecules/MovieCard (https://github.com/TahmidU/mosaic/pull/194) - [x] Molecules/ReviewStat (Already tested) - [x] Molecules/ScrollDownBtn (No point testing this) - [x] Molecules/SearchCard (https://github.com/TahmidU/mosaic/pull/195) - [x] Molecules/SlideButton (https://github.com/TahmidU/mosaic/pull/196) - [x] Molecules/Socials (https://github.com/TahmidU/mosaic/pull/197) - [x] Molecules/StarRating (https://github.com/TahmidU/mosaic/pull/198) - [x] Molecules/SubList (No point testing it, it's just atom/SelectTitleList and atom/DropdownList) - [x] Molecules/VideoModal (https://github.com/TahmidU/mosaic/pull/199) - [x] Organisms/Carousel (https://github.com/TahmidU/mosaic/pull/200) - [x] Organisms/CastList (No point testing this, it's basically testing HorizontalList again) - [x] Organisms/DetailsCard (https://github.com/TahmidU/mosaic/pull/202) - [x] Organisms/Footer (No point testing this) - [x] Organisms/Layout (No point testing this) - [x] Organisms/MediaList (No point testing this, it's basically testing HorizontalList again) - [x] Organisms/MovieList (No point testing this, it's basically testing HorizontalList again) - [x] Organisms/Navbar (No point testing this) - [x] Organisms/SearchInput (https://github.com/TahmidU/mosaic/pull/204) ## Utils - [x] MathUtils (https://github.com/TahmidU/mosaic/pull/209) - [x] RoutingUtils (https://github.com/TahmidU/mosaic/pull/207) - [x] TestUtils (https://github.com/TahmidU/mosaic/pull/210) - [x] TextUtils (https://github.com/TahmidU/mosaic/pull/206)
1.0
Jest RTL tests - # TODO ## Components - [x] Atom/Button (https://github.com/TahmidU/mosaic/pull/177) - [x] Atom/Card (https://github.com/TahmidU/mosaic/pull/178) - [x] Atom/Checkbox (https://github.com/TahmidU/mosaic/pull/179) - [x] Atom/DropdownList (https://github.com/TahmidU/mosaic/pull/180) - [x] Atom/ExternalLink (https://github.com/TahmidU/mosaic/pull/181) - [x] Atom/Input (https://github.com/TahmidU/mosaic/pull/182, https://github.com/TahmidU/mosaic/pull/184) - [x] Atom/ProgressiveCircle (https://github.com/TahmidU/mosaic/pull/183) - [x] Atom/SelectTitleList (https://github.com/TahmidU/mosaic/pull/185) - [x] Molecules/CastCard (https://github.com/TahmidU/mosaic/pull/186) - [x] Molecules/TypeFilter (https://github.com/TahmidU/mosaic/pull/187) - [x] Molecules/Filter (https://github.com/TahmidU/mosaic/pull/190) - [x] Molecules/HorizonalList (https://github.com/TahmidU/mosaic/pull/191, https://github.com/TahmidU/mosaic/pull/201) - [x] Molecules/MediaCard (https://github.com/TahmidU/mosaic/pull/193) - [x] Molecules/MovieCard (https://github.com/TahmidU/mosaic/pull/194) - [x] Molecules/ReviewStat (Already tested) - [x] Molecules/ScrollDownBtn (No point testing this) - [x] Molecules/SearchCard (https://github.com/TahmidU/mosaic/pull/195) - [x] Molecules/SlideButton (https://github.com/TahmidU/mosaic/pull/196) - [x] Molecules/Socials (https://github.com/TahmidU/mosaic/pull/197) - [x] Molecules/StarRating (https://github.com/TahmidU/mosaic/pull/198) - [x] Molecules/SubList (No point testing it, it's just atom/SelectTitleList and atom/DropdownList) - [x] Molecules/VideoModal (https://github.com/TahmidU/mosaic/pull/199) - [x] Organisms/Carousel (https://github.com/TahmidU/mosaic/pull/200) - [x] Organisms/CastList (No point testing this, it's basically testing HorizontalList again) - [x] Organisms/DetailsCard (https://github.com/TahmidU/mosaic/pull/202) - [x] Organisms/Footer (No point testing this) - [x] Organisms/Layout (No point testing this) - [x] Organisms/MediaList (No point testing this, it's basically testing HorizontalList again) - [x] Organisms/MovieList (No point testing this, it's basically testing HorizontalList again) - [x] Organisms/Navbar (No point testing this) - [x] Organisms/SearchInput (https://github.com/TahmidU/mosaic/pull/204) ## Utils - [x] MathUtils (https://github.com/TahmidU/mosaic/pull/209) - [x] RoutingUtils (https://github.com/TahmidU/mosaic/pull/207) - [x] TestUtils (https://github.com/TahmidU/mosaic/pull/210) - [x] TextUtils (https://github.com/TahmidU/mosaic/pull/206)
test
jest rtl tests todo components atom button atom card atom checkbox atom dropdownlist atom externallink atom input atom progressivecircle atom selecttitlelist molecules castcard molecules typefilter molecules filter molecules horizonallist molecules mediacard molecules moviecard molecules reviewstat already tested molecules scrolldownbtn no point testing this molecules searchcard molecules slidebutton molecules socials molecules starrating molecules sublist no point testing it it s just atom selecttitlelist and atom dropdownlist molecules videomodal organisms carousel organisms castlist no point testing this it s basically testing horizontallist again organisms detailscard organisms footer no point testing this organisms layout no point testing this organisms medialist no point testing this it s basically testing horizontallist again organisms movielist no point testing this it s basically testing horizontallist again organisms navbar no point testing this organisms searchinput utils mathutils routingutils testutils textutils
1
661,865
22,092,675,306
IssuesEvent
2022-06-01 07:30:51
kubevela/kubevela
https://api.github.com/repos/kubevela/kubevela
closed
Fail to delete applications due lack of dependency between components when using terraform
priority/important-soon cloud-resource area/addon
**Describe the bug** <!-- A clear and concise description of what the bug is. --> When deleting a bunch of components in the same application using the terraform addon, the controller decides to delete them in order of creation, however, depending of the provider (in my case is Azure) it requires some order (delete all resources under a resource group and after this deleting the resource group itself). **To Reproduce** <!-- Steps to reproduce the behavior: 1. The YAML files of Component/Trait I used. 2. The YAML file of Application I applied. 3. Other operations I have done. 4. See error --> 1. Configure the terraform and terraform-azure addons ``` vela addon install terraform vela addon install terraform-azure vela provider add terraform-azure --ARM_CLIENT_ID $ARM_CLIENT_ID --ARM_CLIENT_SECRET $ARM_CLIENT_SECRET --ARM_SUBSCRIPTION_ID $ARM_SUBSCRIPTION_ID --ARM_TENANT_ID $ARM_TENANT_ID --name azure -n vela-system ``` 2. Create a Virtual Machine ComponentDefinition ``` apiVersion: core.oam.dev/v1beta1 kind: ComponentDefinition metadata: annotations: definition.oam.dev/description: Terraform configuration for Azure Cloud Virtual Machine object labels: type: terraform name: azure-virtual-machine namespace: vela-system spec: workload: definition: apiVersion: terraform.core.oam.dev/v1beta1 kind: Configuration type: configurations.terraform.core.oam.dev schematic: terraform: configuration: | provider "azurerm" { features {} } data "azurerm_subnet" "vnet" { name = var.subnet_name virtual_network_name = var.vnet_name resource_group_name = var.subnet_rg_name } module "server" { source = "Azure/compute/azurerm" resource_group_name = var.resource_group nb_instances = var.instance_count vm_os_simple = var.os public_ip_dns = [var.hostname] vnet_subnet_id = data.azurerm_subnet.vnet.id location = var.location vm_hostname = var.hostname vm_size = var.vm_size ssh_key = "" ssh_key_values = [var.ssh_key] } variable "resource_group" { description = "Resource group name" type = string default = "example" } variable "subnet_name" { description = "Subnet Name" type = string } variable "vnet_name" { description = "Virtual Network Name" type = string } variable "subnet_rg_name" { description = "Resource group of the Subnet" type = string } variable "instance_count" { description = "Specify the number of vm instances." type = number default = 1 } variable "hostname" { description = "VM instance name" type = string default = "linux-server" } variable "location" { description = "Azure Location" type = string default = "westeurope" } variable "os" { description = "Specify UbuntuServer, WindowsServer, RHEL, openSUSE-Leap, CentOS, Debian, CoreOS or SLES" type = string default = "UbuntuServer" } variable "vm_size" { description = "VM size" type = string default = "Standard_B1s" } variable "ssh_key" { description = "SSH Public Key" type = string } output "vm_public_name" { value = module.server.public_ip_dns_name } output "vm_public_ip" { value = module.server.public_ip_address } deleteResource: true providerRef: name: azure namespace: default ``` 3. Edit the current azure-subnet (the current one doesn't work) ``` kubectl patch componentDefinitions azure-subnet -n vela-system --type merge -p '{"spec":{"schematic":{"terraform": {"configuration": "https://github.com/alisson276/terraform-modules.git"}}}}' ``` 4. Create a file called first-vm.yaml: ``` apiVersion: core.oam.dev/v1beta1 kind: Application metadata: name: vm-1 spec: components: - name: example-rg type: azure-resource-group properties: location: westeurope name: example1 writeConnectionSecretToRef: name: example-rg - name: example-virtual-network type: azure-subnet properties: vnet_name: vnet-1 create_network: true resource_group_name: example1 subnet_prefixes: - 10.0.0.0/24 subnet_names: - subnet-1 writeConnectionSecretToRef: name: example-vnet - name: vm-1-1 type: azure-virtual-machine properties: instance_count: 1 subnet_name: subnet-1 vnet_name: vnet-1 subnet_rg_name: example1 resource_group: example hostname: kubevela writeConnectionSecretToRef: name: vm1-conn ``` 5. Apply it: ``` vela up -f first-vm.yaml ``` 6. Wait untill all resource are healthy 7. vela delete vm-1 **Expected behavior** <!-- A clear and concise description of what you expected to happen. --> As you can see in the components, the second component is a subnet under example1 resource group (the first component). The problem is: the virtual machine is in another resource group (example) and it has a NIC associate to it, which is linked to the subnet. So if you try to delete the first component created first, the resource group, it fails, because the vnet (the second resource) is under it and can't be deleted because the third resource (the vm) wasn't destroyed yet. The idea is to have a way to refer it as `depends_on` (saying that the vm depends on the vnet and the vnet depends on the resource group) or simply delete them in the reverse order that they were created. **Screenshots** <!-- If applicable, add screenshots to help explain your problem. --> **KubeVela Version** <!-- Describe your KubeVela controller or CLI version information. --> CLI Version: master Core Version: 1.3.6 GitRevision: git-c8264b8c GolangVersion: go1.18.2 **Cluster information** <!-- Describe your kubernetes cluster information. - Kubernetes Version [e.g. 1.16.9] --> Kubernetes version: 1.22.9 **Additional context** <!-- Add any other context about the problem here. -->
1.0
Fail to delete applications due lack of dependency between components when using terraform - **Describe the bug** <!-- A clear and concise description of what the bug is. --> When deleting a bunch of components in the same application using the terraform addon, the controller decides to delete them in order of creation, however, depending of the provider (in my case is Azure) it requires some order (delete all resources under a resource group and after this deleting the resource group itself). **To Reproduce** <!-- Steps to reproduce the behavior: 1. The YAML files of Component/Trait I used. 2. The YAML file of Application I applied. 3. Other operations I have done. 4. See error --> 1. Configure the terraform and terraform-azure addons ``` vela addon install terraform vela addon install terraform-azure vela provider add terraform-azure --ARM_CLIENT_ID $ARM_CLIENT_ID --ARM_CLIENT_SECRET $ARM_CLIENT_SECRET --ARM_SUBSCRIPTION_ID $ARM_SUBSCRIPTION_ID --ARM_TENANT_ID $ARM_TENANT_ID --name azure -n vela-system ``` 2. Create a Virtual Machine ComponentDefinition ``` apiVersion: core.oam.dev/v1beta1 kind: ComponentDefinition metadata: annotations: definition.oam.dev/description: Terraform configuration for Azure Cloud Virtual Machine object labels: type: terraform name: azure-virtual-machine namespace: vela-system spec: workload: definition: apiVersion: terraform.core.oam.dev/v1beta1 kind: Configuration type: configurations.terraform.core.oam.dev schematic: terraform: configuration: | provider "azurerm" { features {} } data "azurerm_subnet" "vnet" { name = var.subnet_name virtual_network_name = var.vnet_name resource_group_name = var.subnet_rg_name } module "server" { source = "Azure/compute/azurerm" resource_group_name = var.resource_group nb_instances = var.instance_count vm_os_simple = var.os public_ip_dns = [var.hostname] vnet_subnet_id = data.azurerm_subnet.vnet.id location = var.location vm_hostname = var.hostname vm_size = var.vm_size ssh_key = "" ssh_key_values = [var.ssh_key] } variable "resource_group" { description = "Resource group name" type = string default = "example" } variable "subnet_name" { description = "Subnet Name" type = string } variable "vnet_name" { description = "Virtual Network Name" type = string } variable "subnet_rg_name" { description = "Resource group of the Subnet" type = string } variable "instance_count" { description = "Specify the number of vm instances." type = number default = 1 } variable "hostname" { description = "VM instance name" type = string default = "linux-server" } variable "location" { description = "Azure Location" type = string default = "westeurope" } variable "os" { description = "Specify UbuntuServer, WindowsServer, RHEL, openSUSE-Leap, CentOS, Debian, CoreOS or SLES" type = string default = "UbuntuServer" } variable "vm_size" { description = "VM size" type = string default = "Standard_B1s" } variable "ssh_key" { description = "SSH Public Key" type = string } output "vm_public_name" { value = module.server.public_ip_dns_name } output "vm_public_ip" { value = module.server.public_ip_address } deleteResource: true providerRef: name: azure namespace: default ``` 3. Edit the current azure-subnet (the current one doesn't work) ``` kubectl patch componentDefinitions azure-subnet -n vela-system --type merge -p '{"spec":{"schematic":{"terraform": {"configuration": "https://github.com/alisson276/terraform-modules.git"}}}}' ``` 4. Create a file called first-vm.yaml: ``` apiVersion: core.oam.dev/v1beta1 kind: Application metadata: name: vm-1 spec: components: - name: example-rg type: azure-resource-group properties: location: westeurope name: example1 writeConnectionSecretToRef: name: example-rg - name: example-virtual-network type: azure-subnet properties: vnet_name: vnet-1 create_network: true resource_group_name: example1 subnet_prefixes: - 10.0.0.0/24 subnet_names: - subnet-1 writeConnectionSecretToRef: name: example-vnet - name: vm-1-1 type: azure-virtual-machine properties: instance_count: 1 subnet_name: subnet-1 vnet_name: vnet-1 subnet_rg_name: example1 resource_group: example hostname: kubevela writeConnectionSecretToRef: name: vm1-conn ``` 5. Apply it: ``` vela up -f first-vm.yaml ``` 6. Wait untill all resource are healthy 7. vela delete vm-1 **Expected behavior** <!-- A clear and concise description of what you expected to happen. --> As you can see in the components, the second component is a subnet under example1 resource group (the first component). The problem is: the virtual machine is in another resource group (example) and it has a NIC associate to it, which is linked to the subnet. So if you try to delete the first component created first, the resource group, it fails, because the vnet (the second resource) is under it and can't be deleted because the third resource (the vm) wasn't destroyed yet. The idea is to have a way to refer it as `depends_on` (saying that the vm depends on the vnet and the vnet depends on the resource group) or simply delete them in the reverse order that they were created. **Screenshots** <!-- If applicable, add screenshots to help explain your problem. --> **KubeVela Version** <!-- Describe your KubeVela controller or CLI version information. --> CLI Version: master Core Version: 1.3.6 GitRevision: git-c8264b8c GolangVersion: go1.18.2 **Cluster information** <!-- Describe your kubernetes cluster information. - Kubernetes Version [e.g. 1.16.9] --> Kubernetes version: 1.22.9 **Additional context** <!-- Add any other context about the problem here. -->
non_test
fail to delete applications due lack of dependency between components when using terraform describe the bug a clear and concise description of what the bug is when deleting a bunch of components in the same application using the terraform addon the controller decides to delete them in order of creation however depending of the provider in my case is azure it requires some order delete all resources under a resource group and after this deleting the resource group itself to reproduce steps to reproduce the behavior the yaml files of component trait i used the yaml file of application i applied other operations i have done see error configure the terraform and terraform azure addons vela addon install terraform vela addon install terraform azure vela provider add terraform azure arm client id arm client id arm client secret arm client secret arm subscription id arm subscription id arm tenant id arm tenant id name azure n vela system create a virtual machine componentdefinition apiversion core oam dev kind componentdefinition metadata annotations definition oam dev description terraform configuration for azure cloud virtual machine object labels type terraform name azure virtual machine namespace vela system spec workload definition apiversion terraform core oam dev kind configuration type configurations terraform core oam dev schematic terraform configuration provider azurerm features data azurerm subnet vnet name var subnet name virtual network name var vnet name resource group name var subnet rg name module server source azure compute azurerm resource group name var resource group nb instances var instance count vm os simple var os public ip dns vnet subnet id data azurerm subnet vnet id location var location vm hostname var hostname vm size var vm size ssh key ssh key values variable resource group description resource group name type string default example variable subnet name description subnet name type string variable vnet name description virtual network name type string variable subnet rg name description resource group of the subnet type string variable instance count description specify the number of vm instances type number default variable hostname description vm instance name type string default linux server variable location description azure location type string default westeurope variable os description specify ubuntuserver windowsserver rhel opensuse leap centos debian coreos or sles type string default ubuntuserver variable vm size description vm size type string default standard variable ssh key description ssh public key type string output vm public name value module server public ip dns name output vm public ip value module server public ip address deleteresource true providerref name azure namespace default edit the current azure subnet the current one doesn t work kubectl patch componentdefinitions azure subnet n vela system type merge p spec schematic terraform configuration create a file called first vm yaml apiversion core oam dev kind application metadata name vm spec components name example rg type azure resource group properties location westeurope name writeconnectionsecrettoref name example rg name example virtual network type azure subnet properties vnet name vnet create network true resource group name subnet prefixes subnet names subnet writeconnectionsecrettoref name example vnet name vm type azure virtual machine properties instance count subnet name subnet vnet name vnet subnet rg name resource group example hostname kubevela writeconnectionsecrettoref name conn apply it vela up f first vm yaml wait untill all resource are healthy vela delete vm expected behavior a clear and concise description of what you expected to happen as you can see in the components the second component is a subnet under resource group the first component the problem is the virtual machine is in another resource group example and it has a nic associate to it which is linked to the subnet so if you try to delete the first component created first the resource group it fails because the vnet the second resource is under it and can t be deleted because the third resource the vm wasn t destroyed yet the idea is to have a way to refer it as depends on saying that the vm depends on the vnet and the vnet depends on the resource group or simply delete them in the reverse order that they were created screenshots if applicable add screenshots to help explain your problem kubevela version describe your kubevela controller or cli version information cli version master core version gitrevision git golangversion cluster information describe your kubernetes cluster information kubernetes version kubernetes version additional context add any other context about the problem here
0
74,625
7,434,310,925
IssuesEvent
2018-03-26 10:34:02
Kademi/kademi-dev
https://api.github.com/repos/Kademi/kademi-dev
closed
Bulk find and select locations for organisations
Ready to Test - Dev
On the organisations map page (/organisations/map) do the following: - show the number of organisations with locations (ie lattitude and longitude) - show the number of organisations without locations - if orgs without locations > 0 show a button to find and set them - when pressed, for each organisation do a google maps search for the provided address (same as is currently done on the org details page), and if a matching location is found set its lat/lng as the organisation coords Also, ensure orgs with no coords are not included in map search responses
1.0
Bulk find and select locations for organisations - On the organisations map page (/organisations/map) do the following: - show the number of organisations with locations (ie lattitude and longitude) - show the number of organisations without locations - if orgs without locations > 0 show a button to find and set them - when pressed, for each organisation do a google maps search for the provided address (same as is currently done on the org details page), and if a matching location is found set its lat/lng as the organisation coords Also, ensure orgs with no coords are not included in map search responses
test
bulk find and select locations for organisations on the organisations map page organisations map do the following show the number of organisations with locations ie lattitude and longitude show the number of organisations without locations if orgs without locations show a button to find and set them when pressed for each organisation do a google maps search for the provided address same as is currently done on the org details page and if a matching location is found set its lat lng as the organisation coords also ensure orgs with no coords are not included in map search responses
1
226,718
18,043,955,427
IssuesEvent
2021-09-18 14:55:43
logicmoo/logicmoo_workspace
https://api.github.com/repos/logicmoo/logicmoo_workspace
opened
logicmoo.base.examples.fol.SANITY_EXISTS_03 JUnit
Test_9999 logicmoo.base.examples.fol unit_test SANITY_EXISTS_03
(cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif sanity_exists_03.pfc.pl) GH_MASTER_ISSUE_FINFO= ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ASANITY_EXISTS_03 GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/813ec17487381a026b83350c360d0c79a9e2d0ae https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/ This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/65/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/ GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/813ec17487381a026b83350c360d0c79a9e2d0ae https://github.com/logicmoo/logicmoo_workspace/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl ``` % running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl'), %~ this_test_might_need( :-( use_module( library(logicmoo_plarkc)))) :- module(t123). :- dynamic(t123:ttExpressionType/1). % :- process_this_script. % :- process_this_script. :- statistics. /*~ % Started at Sat Sep 18 07:55:40 2021 % 0.875 seconds cpu time for 412,769 inferences % 940,176 atoms, 31,018 functors, 29,482 predicates, 718 modules, 14,818,012 VM-codes % % Limit Allocated In use % Local stack: - 52 Kb 4,216 b % Global stack: - 64 Kb 49 Kb % Trail stack: - 34 Kb 520 b % Total: 1,024 Mb 150 Kb 53 Kb % % 5 garbage collections gained 305,208 bytes in 0.000 seconds. % 1 atom garbage collections gained 1,137 atoms in 0.015 seconds. % 4 clause garbage collections gained 118 clauses in 0.000 seconds. % Stack shifts: 1 local, 0 global, 0 trail in 0.000 seconds % 3 threads, 0 finished threads used 0.000 seconds ~*/ subtest_assert(I):-kif_assert(I). subtest([subtest_assert(tAnimal(joe)), mpred_test(isa(_,tHeart))]). %~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl:18 %~ debugm( common_logic_loader, %~ show_success( common_logic_loader, %~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( tHeart(Heart4))]))))) %~ kifi = subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( tHeart(Heart4))]). %~ kifm = nesc( subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( tHeart(Heart4))])). /*~ ======================================================= subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(tHeart('$VAR'('Heart4'))),[]))) ============================================ ?- kif_to_boxlog( subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))]) ). % In English: %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ joe isa tAnimal isa subtest_assert ?Heart4 isa tHeart isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ kifi=subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))]) %~ kifm=nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))])) %~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(tHeart('$VAR'('Heart4'))),[])))) %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % Results in the following 1 entailment(s): nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))])). %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ it is necessarily true that joe isa tAnimal isa subtest_assert ?Heart4 isa tHeart isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% nesc( subtest( [ subtest_assert( tAnimal(joe)), mpred_test( tHeart(Heart4))])). ============================================ ~*/ subtest([subtest_assert(tAnimal(joe)), mpred_test(hasOrgan(joe,_))]). %~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl:21 %~ debugm( common_logic_loader, %~ show_success( common_logic_loader, %~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))]))))) %~ kifi = subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))]). %~ kifm = nesc( subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))])). /*~ ======================================================= subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(hasOrgan(joe,'$VAR'('HasOrgan_Joe4'))),[]))) ============================================ ?- kif_to_boxlog( subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))]) ). % In English: %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ joe isa tAnimal isa subtest_assert joe hasOrgan ?HasOrgan_Joe4 isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ kifi=subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))]) %~ kifm=nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))])) %~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(hasOrgan(joe,'$VAR'('HasOrgan_Joe4'))),[])))) %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % Results in the following 1 entailment(s): nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))])). %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ it is necessarily true that joe isa tAnimal isa subtest_assert joe hasOrgan ?HasOrgan_Joe4 isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% nesc( subtest( [ subtest_assert( tAnimal(joe)), mpred_test( hasOrgan(joe,HasOrgan_Joe4))])). ============================================ ~*/ subtest([subtest_assert(tHeart(_)), mpred_test(~hasOrgan(jack,_))]). %~ debugm( common_logic_loader, %~ show_success( common_logic_loader, %~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tHeart(Heart7)), %~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))]))))) %~ kifi = subtest( [ subtest_assert( tHeart(Heart7)), %~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))]). %~ kifm = nesc( subtest( [ subtest_assert( tHeart(Heart7)), %~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))])). /*~ ======================================================= subtest('[|]'(subtest_assert(tHeart('$VAR'('Heart7'))),'[|]'(mpred_test(~(hasOrgan(jack,'$VAR'('HasOrgan_Jack5')))),[]))) ============================================ ?- kif_to_boxlog( subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))]) ). % In English: %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ ?Heart7 isa tHeart isa subtest_assert jack hasOrgan ?HasOrgan_Jack5 isa ~ isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ kifi=subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))]) %~ kifm=nesc(subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))])) %~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tHeart('$VAR'('Heart7'))),'[|]'(mpred_test(not(hasOrgan(jack,'$VAR'('HasOrgan_Jack5')))),[])))) %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % Results in the following 1 entailment(s): nesc(subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))])). %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ it is necessarily true that ?Heart7 isa tHeart isa subtest_assert jack hasOrgan ?HasOrgan_Jack5 isa ~ isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% nesc( subtest( [ subtest_assert( tHeart(Heart7)), mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))])). ============================================ ~*/ dbanner:- nl,nl,dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%'),nl,nl. %~ warn( db_expand_0( clause(unknown,cuz), %~ dbanner :- %~ nl , %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl , %~ nl, %~ dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl)) %~ message_hook_type(error) %~ message_hook( %~ initialization_exception( must_sanity : e( must_sanity : e( error( %~ permission_error(modify,static_procedure,common_logic_sanity:dbanner/0), %~ context(system:assertz/1,Context_Kw)), %~ baseKB : assert_to_mu( dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl)), %~ baseKB : ain_fast( %~ baseKB : dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl, %~ ( mfl4(BaseKB,baseKB,'/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl',27) , %~ ax)))), %~ error, %~ [ 'Prolog initialisation failed:', %~ nl, %~ 'Unknown message: ~p' - [ must_sanity : e( must_sanity : e( error( %~ permission_error(modify,static_procedure,common_logic_sanity:dbanner/0), %~ context(system:assertz/1,Context_Kw)), %~ baseKB : assert_to_mu( dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl)), %~ baseKB : ain_fast( %~ baseKB : dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl, %~ ( mfl4(BaseKB,baseKB,'/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl',27) , %~ ax)))]]) %~ unused(save_junit_results) ``` totalTime=1 ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ASANITY_EXISTS_03 GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/813ec17487381a026b83350c360d0c79a9e2d0ae https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/ This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/65/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/ GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/813ec17487381a026b83350c360d0c79a9e2d0ae https://github.com/logicmoo/logicmoo_workspace/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k sanity_exists_03.pfc.pl (returned 1)
2.0
logicmoo.base.examples.fol.SANITY_EXISTS_03 JUnit - (cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif sanity_exists_03.pfc.pl) GH_MASTER_ISSUE_FINFO= ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ASANITY_EXISTS_03 GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/813ec17487381a026b83350c360d0c79a9e2d0ae https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/ This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/65/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/ GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/813ec17487381a026b83350c360d0c79a9e2d0ae https://github.com/logicmoo/logicmoo_workspace/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl ``` % running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl'), %~ this_test_might_need( :-( use_module( library(logicmoo_plarkc)))) :- module(t123). :- dynamic(t123:ttExpressionType/1). % :- process_this_script. % :- process_this_script. :- statistics. /*~ % Started at Sat Sep 18 07:55:40 2021 % 0.875 seconds cpu time for 412,769 inferences % 940,176 atoms, 31,018 functors, 29,482 predicates, 718 modules, 14,818,012 VM-codes % % Limit Allocated In use % Local stack: - 52 Kb 4,216 b % Global stack: - 64 Kb 49 Kb % Trail stack: - 34 Kb 520 b % Total: 1,024 Mb 150 Kb 53 Kb % % 5 garbage collections gained 305,208 bytes in 0.000 seconds. % 1 atom garbage collections gained 1,137 atoms in 0.015 seconds. % 4 clause garbage collections gained 118 clauses in 0.000 seconds. % Stack shifts: 1 local, 0 global, 0 trail in 0.000 seconds % 3 threads, 0 finished threads used 0.000 seconds ~*/ subtest_assert(I):-kif_assert(I). subtest([subtest_assert(tAnimal(joe)), mpred_test(isa(_,tHeart))]). %~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl:18 %~ debugm( common_logic_loader, %~ show_success( common_logic_loader, %~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( tHeart(Heart4))]))))) %~ kifi = subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( tHeart(Heart4))]). %~ kifm = nesc( subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( tHeart(Heart4))])). /*~ ======================================================= subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(tHeart('$VAR'('Heart4'))),[]))) ============================================ ?- kif_to_boxlog( subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))]) ). % In English: %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ joe isa tAnimal isa subtest_assert ?Heart4 isa tHeart isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ kifi=subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))]) %~ kifm=nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))])) %~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(tHeart('$VAR'('Heart4'))),[])))) %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % Results in the following 1 entailment(s): nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(tHeart(Heart4))])). %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ it is necessarily true that joe isa tAnimal isa subtest_assert ?Heart4 isa tHeart isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% nesc( subtest( [ subtest_assert( tAnimal(joe)), mpred_test( tHeart(Heart4))])). ============================================ ~*/ subtest([subtest_assert(tAnimal(joe)), mpred_test(hasOrgan(joe,_))]). %~ /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl:21 %~ debugm( common_logic_loader, %~ show_success( common_logic_loader, %~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))]))))) %~ kifi = subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))]). %~ kifm = nesc( subtest( [ subtest_assert( tAnimal(joe)), %~ mpred_test( hasOrgan(joe,HasOrgan_Joe4))])). /*~ ======================================================= subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(hasOrgan(joe,'$VAR'('HasOrgan_Joe4'))),[]))) ============================================ ?- kif_to_boxlog( subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))]) ). % In English: %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ joe isa tAnimal isa subtest_assert joe hasOrgan ?HasOrgan_Joe4 isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ kifi=subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))]) %~ kifm=nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))])) %~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tAnimal(joe)),'[|]'(mpred_test(hasOrgan(joe,'$VAR'('HasOrgan_Joe4'))),[])))) %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % Results in the following 1 entailment(s): nesc(subtest([subtest_assert(tAnimal(joe)),mpred_test(hasOrgan(joe,HasOrgan_Joe4))])). %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ it is necessarily true that joe isa tAnimal isa subtest_assert joe hasOrgan ?HasOrgan_Joe4 isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% nesc( subtest( [ subtest_assert( tAnimal(joe)), mpred_test( hasOrgan(joe,HasOrgan_Joe4))])). ============================================ ~*/ subtest([subtest_assert(tHeart(_)), mpred_test(~hasOrgan(jack,_))]). %~ debugm( common_logic_loader, %~ show_success( common_logic_loader, %~ common_logic_loader : ain( clif( subtest( [ subtest_assert( tHeart(Heart7)), %~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))]))))) %~ kifi = subtest( [ subtest_assert( tHeart(Heart7)), %~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))]). %~ kifm = nesc( subtest( [ subtest_assert( tHeart(Heart7)), %~ mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))])). /*~ ======================================================= subtest('[|]'(subtest_assert(tHeart('$VAR'('Heart7'))),'[|]'(mpred_test(~(hasOrgan(jack,'$VAR'('HasOrgan_Jack5')))),[]))) ============================================ ?- kif_to_boxlog( subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))]) ). % In English: %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ ?Heart7 isa tHeart isa subtest_assert jack hasOrgan ?HasOrgan_Jack5 isa ~ isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ kifi=subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))]) %~ kifm=nesc(subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))])) %~ kif_to_boxlog_attvars2 = necessary(subtest('[|]'(subtest_assert(tHeart('$VAR'('Heart7'))),'[|]'(mpred_test(not(hasOrgan(jack,'$VAR'('HasOrgan_Jack5')))),[])))) %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% % Results in the following 1 entailment(s): nesc(subtest([subtest_assert(tHeart(Heart7)),mpred_test(~hasOrgan(jack,HasOrgan_Jack5))])). %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% %~ it is necessarily true that ?Heart7 isa tHeart isa subtest_assert jack hasOrgan ?HasOrgan_Jack5 isa ~ isa mpred_test isa subtest %~ %~ %%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%% nesc( subtest( [ subtest_assert( tHeart(Heart7)), mpred_test( ~( hasOrgan(jack,HasOrgan_Jack5)))])). ============================================ ~*/ dbanner:- nl,nl,dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%'),nl,nl. %~ warn( db_expand_0( clause(unknown,cuz), %~ dbanner :- %~ nl , %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl , %~ nl, %~ dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl)) %~ message_hook_type(error) %~ message_hook( %~ initialization_exception( must_sanity : e( must_sanity : e( error( %~ permission_error(modify,static_procedure,common_logic_sanity:dbanner/0), %~ context(system:assertz/1,Context_Kw)), %~ baseKB : assert_to_mu( dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl)), %~ baseKB : ain_fast( %~ baseKB : dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl, %~ ( mfl4(BaseKB,baseKB,'/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl',27) , %~ ax)))), %~ error, %~ [ 'Prolog initialisation failed:', %~ nl, %~ 'Unknown message: ~p' - [ must_sanity : e( must_sanity : e( error( %~ permission_error(modify,static_procedure,common_logic_sanity:dbanner/0), %~ context(system:assertz/1,Context_Kw)), %~ baseKB : assert_to_mu( dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl)), %~ baseKB : ain_fast( %~ baseKB : dbanner :- %~ nl , %~ dmsg('%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%') , %~ nl, %~ ( mfl4(BaseKB,baseKB,'/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl',27) , %~ ax)))]]) %~ unused(save_junit_results) ``` totalTime=1 ISSUE_SEARCH: https://github.com/logicmoo/logicmoo_workspace/issues?q=is%3Aissue+label%3ASANITY_EXISTS_03 GITLAB: https://logicmoo.org:2082/gitlab/logicmoo/logicmoo_workspace/-/commit/813ec17487381a026b83350c360d0c79a9e2d0ae https://gitlab.logicmoo.org/gitlab/logicmoo/logicmoo_workspace/-/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/ This Build: https://jenkins.logicmoo.org/job/logicmoo_workspace/65/testReport/logicmoo.base.examples.fol/SANITY_EXISTS_03/logicmoo_base_examples_fol_SANITY_EXISTS_03_JUnit/ GITHUB: https://github.com/logicmoo/logicmoo_workspace/commit/813ec17487381a026b83350c360d0c79a9e2d0ae https://github.com/logicmoo/logicmoo_workspace/blob/813ec17487381a026b83350c360d0c79a9e2d0ae/packs_sys/logicmoo_base/t/examples/fol/sanity_exists_03.pfc.pl FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k sanity_exists_03.pfc.pl (returned 1)
test
logicmoo base examples fol sanity exists junit cd var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol timeout foreground preserve status s sigkill k lmoo clif sanity exists pfc pl gh master issue finfo issue search gitlab latest this build github running var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol sanity exists pfc pl this test might need use module library logicmoo plarkc module dynamic ttexpressiontype process this script process this script statistics started at sat sep seconds cpu time for inferences atoms functors predicates modules vm codes limit allocated in use local stack kb b global stack kb kb trail stack kb b total mb kb kb garbage collections gained bytes in seconds atom garbage collections gained atoms in seconds clause garbage collections gained clauses in seconds stack shifts local global trail in seconds threads finished threads used seconds subtest assert i kif assert i subtest subtest assert tanimal joe mpred test isa theart var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol sanity exists pfc pl debugm common logic loader show success common logic loader common logic loader ain clif subtest subtest assert tanimal joe mpred test theart kifi subtest subtest assert tanimal joe mpred test theart kifm nesc subtest subtest assert tanimal joe mpred test theart subtest subtest assert tanimal joe mpred test theart var kif to boxlog subtest in english joe isa tanimal isa subtest assert isa theart isa mpred test isa subtest kifi subtest kifm nesc subtest kif to boxlog necessary subtest subtest assert tanimal joe mpred test theart var results in the following entailment s nesc subtest it is necessarily true that joe isa tanimal isa subtest assert isa theart isa mpred test isa subtest nesc subtest subtest assert tanimal joe mpred test theart subtest subtest assert tanimal joe mpred test hasorgan joe var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol sanity exists pfc pl debugm common logic loader show success common logic loader common logic loader ain clif subtest subtest assert tanimal joe mpred test hasorgan joe hasorgan kifi subtest subtest assert tanimal joe mpred test hasorgan joe hasorgan kifm nesc subtest subtest assert tanimal joe mpred test hasorgan joe hasorgan subtest subtest assert tanimal joe mpred test hasorgan joe var hasorgan kif to boxlog subtest in english joe isa tanimal isa subtest assert joe hasorgan hasorgan isa mpred test isa subtest kifi subtest kifm nesc subtest kif to boxlog necessary subtest subtest assert tanimal joe mpred test hasorgan joe var hasorgan results in the following entailment s nesc subtest it is necessarily true that joe isa tanimal isa subtest assert joe hasorgan hasorgan isa mpred test isa subtest nesc subtest subtest assert tanimal joe mpred test hasorgan joe hasorgan subtest subtest assert theart mpred test hasorgan jack debugm common logic loader show success common logic loader common logic loader ain clif subtest subtest assert theart mpred test hasorgan jack hasorgan kifi subtest subtest assert theart mpred test hasorgan jack hasorgan kifm nesc subtest subtest assert theart mpred test hasorgan jack hasorgan subtest subtest assert theart var mpred test hasorgan jack var hasorgan kif to boxlog subtest in english isa theart isa subtest assert jack hasorgan hasorgan isa isa mpred test isa subtest kifi subtest kifm nesc subtest kif to boxlog necessary subtest subtest assert theart var mpred test not hasorgan jack var hasorgan results in the following entailment s nesc subtest it is necessarily true that isa theart isa subtest assert jack hasorgan hasorgan isa isa mpred test isa subtest nesc subtest subtest assert theart mpred test hasorgan jack hasorgan dbanner nl nl dmsg nl nl warn db expand clause unknown cuz dbanner nl nl dmsg nl nl dbanner nl dmsg nl message hook type error message hook initialization exception must sanity e must sanity e error permission error modify static procedure common logic sanity dbanner context system assertz context kw basekb assert to mu dbanner nl dmsg nl basekb ain fast basekb dbanner nl dmsg nl basekb basekb var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol sanity exists pfc pl ax error prolog initialisation failed nl unknown message p must sanity e must sanity e error permission error modify static procedure common logic sanity dbanner context system assertz context kw basekb assert to mu dbanner nl dmsg nl basekb ain fast basekb dbanner nl dmsg nl basekb basekb var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol sanity exists pfc pl ax unused save junit results totaltime issue search gitlab latest this build github failed var lib jenkins workspace logicmoo workspace bin lmoo junit minor k sanity exists pfc pl returned
1
50,615
13,187,624,433
IssuesEvent
2020-08-13 04:01:42
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
[photospline] photospline-pybindings fails during CMake's configure step if astropy is not present (Trac #1051)
Migrated from Trac combo simulation defect
I don't have astropy. So get this error for the current trunk when cmake configures itself: Traceback (most recent call last): File "/home/hdembinski/Code/icecube/icesim/trunk/photospline/resources/test/find_pyfits.py", line 6, in <module> import astropy.io.fits as pyfits ImportError: No module named astropy.io.fits Maybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1051">https://code.icecube.wisc.edu/ticket/1051</a>, reported by hdembinski and owned by </em></summary> <p> ```json { "status": "closed", "changetime": "2015-07-15T07:42:17", "description": "I don't have astropy. So get this error for the current trunk when cmake configures itself:\n\nTraceback (most recent call last):\n File \"/home/hdembinski/Code/icecube/icesim/trunk/photospline/resources/test/find_pyfits.py\", line 6, in <module>\n import astropy.io.fits as pyfits\nImportError: No module named astropy.io.fits\n\nMaybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user.", "reporter": "hdembinski", "cc": "", "resolution": "fixed", "_ts": "1436946137322111", "component": "combo simulation", "summary": "[photospline] photospline-pybindings fails during CMake's configure step if astropy is not present", "priority": "critical", "keywords": "photospline astropy", "time": "2015-07-15T00:12:23", "milestone": "", "owner": "", "type": "defect" } ``` </p> </details>
1.0
[photospline] photospline-pybindings fails during CMake's configure step if astropy is not present (Trac #1051) - I don't have astropy. So get this error for the current trunk when cmake configures itself: Traceback (most recent call last): File "/home/hdembinski/Code/icecube/icesim/trunk/photospline/resources/test/find_pyfits.py", line 6, in <module> import astropy.io.fits as pyfits ImportError: No module named astropy.io.fits Maybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1051">https://code.icecube.wisc.edu/ticket/1051</a>, reported by hdembinski and owned by </em></summary> <p> ```json { "status": "closed", "changetime": "2015-07-15T07:42:17", "description": "I don't have astropy. So get this error for the current trunk when cmake configures itself:\n\nTraceback (most recent call last):\n File \"/home/hdembinski/Code/icecube/icesim/trunk/photospline/resources/test/find_pyfits.py\", line 6, in <module>\n import astropy.io.fits as pyfits\nImportError: No module named astropy.io.fits\n\nMaybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user.", "reporter": "hdembinski", "cc": "", "resolution": "fixed", "_ts": "1436946137322111", "component": "combo simulation", "summary": "[photospline] photospline-pybindings fails during CMake's configure step if astropy is not present", "priority": "critical", "keywords": "photospline astropy", "time": "2015-07-15T00:12:23", "milestone": "", "owner": "", "type": "defect" } ``` </p> </details>
non_test
photospline pybindings fails during cmake s configure step if astropy is not present trac i don t have astropy so get this error for the current trunk when cmake configures itself traceback most recent call last file home hdembinski code icecube icesim trunk photospline resources test find pyfits py line in import astropy io fits as pyfits importerror no module named astropy io fits maybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user migrated from json status closed changetime description i don t have astropy so get this error for the current trunk when cmake configures itself n ntraceback most recent call last n file home hdembinski code icecube icesim trunk photospline resources test find pyfits py line in n import astropy io fits as pyfits nimporterror no module named astropy io fits n nmaybe the compilation of photospline should be turned off in this case or a more helpful error message should be given to the user reporter hdembinski cc resolution fixed ts component combo simulation summary photospline pybindings fails during cmake s configure step if astropy is not present priority critical keywords photospline astropy time milestone owner type defect
0
28,682
4,426,279,275
IssuesEvent
2016-08-16 17:51:43
grpc/grpc
https://api.github.com/repos/grpc/grpc
closed
Update interop and stress tests to GA API
interop python security test
The only code in the repository that should call the Beta API should be the tests of the Beta API. This is currently blocked on [issue 6799](https://github.com/grpc/grpc/issues/6799) because the [`oauth2_auth_token`](https://github.com/grpc/grpc/blob/master/doc/interop-test-descriptions.md#oauth2_auth_token), [`jwt_token_creds`](https://github.com/grpc/grpc/blob/master/doc/interop-test-descriptions.md#jwt_token_creds), and [`per_rpc_creds`](https://github.com/grpc/grpc/blob/master/doc/interop-test-descriptions.md#per_rpc_creds) interop tests are implemented in terms of the [`oauth2client`](https://github.com/google/oauth2client) library and the stress tests are implemented in terms of the interop tests.
1.0
Update interop and stress tests to GA API - The only code in the repository that should call the Beta API should be the tests of the Beta API. This is currently blocked on [issue 6799](https://github.com/grpc/grpc/issues/6799) because the [`oauth2_auth_token`](https://github.com/grpc/grpc/blob/master/doc/interop-test-descriptions.md#oauth2_auth_token), [`jwt_token_creds`](https://github.com/grpc/grpc/blob/master/doc/interop-test-descriptions.md#jwt_token_creds), and [`per_rpc_creds`](https://github.com/grpc/grpc/blob/master/doc/interop-test-descriptions.md#per_rpc_creds) interop tests are implemented in terms of the [`oauth2client`](https://github.com/google/oauth2client) library and the stress tests are implemented in terms of the interop tests.
test
update interop and stress tests to ga api the only code in the repository that should call the beta api should be the tests of the beta api this is currently blocked on because the and interop tests are implemented in terms of the library and the stress tests are implemented in terms of the interop tests
1
8,446
11,614,671,832
IssuesEvent
2020-02-26 13:00:24
scikit-learn/scikit-learn
https://api.github.com/repos/scikit-learn/scikit-learn
closed
sklearn.preprocessing.StandardScaler gets NaN variance when partial_fit with sparse data
Bug module:preprocessing
#### Describe the bug When I feed a specific dataset (which is sparse) to sklearn.preprocessing.StandardScaler.partial_fit in a specific order, I get variance which is NaN although data does **NOT** contains any NaNs and is very small. When I convert the sparse arrays to dense, it works. When I change the order to feed the data, it works too. #### Steps/Code to Reproduce Please work with the data I attached. [sparse_data.tar.gz](https://github.com/scikit-learn/scikit-learn/files/4208684/sparse_data.tar.gz) ```python import scipy.sparse as sp from sklearn import preprocessing s0 = sp.load_npz('0.npz') s1 = sp.load_npz('1.npz') # Buggy behavior ss0 = preprocessing.StandardScaler(with_mean=False) ss0.partial_fit(s0) print(ss0.var_) ss0.partial_fit(s1) print(ss0.var_) # => gets NaN # When use dence array, it works ss1 = preprocessing.StandardScaler(with_mean=False) ss1.partial_fit(s0.toarray()) print(ss1.var_) ss1.partial_fit(s1.toarray()) print(ss1.var_) # When change the order of data, it works ss2 = preprocessing.StandardScaler(with_mean=False) ss2.partial_fit(s1) print(ss2.var_) ss2.partial_fit(s0) print(ss2.var_) ``` EDIT: Fix sample code around ss2 #### Expected Results ```python ss0.var_ # => [0.15896542] ss1.var_ # => [0.15896542] ss2.var_ # => [0.15896542] ``` #### Actual Results ```python ss0.var_ # => [nan] ss1.var_ # => [0.15896542] ss2.var_ # => [0.15896542] ``` #### Versions I confirmed this issue in two different environments. ``` System: python: 3.7.3 (default, Apr 22 2019, 02:40:09) [Clang 10.0.1 (clang-1001.0.46.4)] executable: /usr/local/var/pyenv/versions/3.7.3/bin/python3 machine: Darwin-19.3.0-x86_64-i386-64bit Python dependencies: pip: 20.0.2 setuptools: 40.8.0 sklearn: 0.22 numpy: 1.18.0 scipy: 1.4.1 Cython: None pandas: 0.25.3 matplotlib: 3.1.2 joblib: 0.14.1 Built with OpenMP: True ``` ``` System: python: 3.7.6 (default, Feb 14 2020, 16:41:52) [GCC 8.3.1 20190507 (Red Hat 8.3.1-4)] executable: /home/***/ws/siml/.venv/bin/python3 machine: Linux-4.18.0-147.5.1.el8_1.x86_64-x86_64-with-centos-8.1.1911-Core Python dependencies: pip: 19.2.3 setuptools: 41.2.0 sklearn: 0.22.1 numpy: 1.18.1 scipy: 1.4.1 Cython: None pandas: 0.25.3 matplotlib: 3.1.3 joblib: 0.14.1 Built with OpenMP: True ```
1.0
sklearn.preprocessing.StandardScaler gets NaN variance when partial_fit with sparse data - #### Describe the bug When I feed a specific dataset (which is sparse) to sklearn.preprocessing.StandardScaler.partial_fit in a specific order, I get variance which is NaN although data does **NOT** contains any NaNs and is very small. When I convert the sparse arrays to dense, it works. When I change the order to feed the data, it works too. #### Steps/Code to Reproduce Please work with the data I attached. [sparse_data.tar.gz](https://github.com/scikit-learn/scikit-learn/files/4208684/sparse_data.tar.gz) ```python import scipy.sparse as sp from sklearn import preprocessing s0 = sp.load_npz('0.npz') s1 = sp.load_npz('1.npz') # Buggy behavior ss0 = preprocessing.StandardScaler(with_mean=False) ss0.partial_fit(s0) print(ss0.var_) ss0.partial_fit(s1) print(ss0.var_) # => gets NaN # When use dence array, it works ss1 = preprocessing.StandardScaler(with_mean=False) ss1.partial_fit(s0.toarray()) print(ss1.var_) ss1.partial_fit(s1.toarray()) print(ss1.var_) # When change the order of data, it works ss2 = preprocessing.StandardScaler(with_mean=False) ss2.partial_fit(s1) print(ss2.var_) ss2.partial_fit(s0) print(ss2.var_) ``` EDIT: Fix sample code around ss2 #### Expected Results ```python ss0.var_ # => [0.15896542] ss1.var_ # => [0.15896542] ss2.var_ # => [0.15896542] ``` #### Actual Results ```python ss0.var_ # => [nan] ss1.var_ # => [0.15896542] ss2.var_ # => [0.15896542] ``` #### Versions I confirmed this issue in two different environments. ``` System: python: 3.7.3 (default, Apr 22 2019, 02:40:09) [Clang 10.0.1 (clang-1001.0.46.4)] executable: /usr/local/var/pyenv/versions/3.7.3/bin/python3 machine: Darwin-19.3.0-x86_64-i386-64bit Python dependencies: pip: 20.0.2 setuptools: 40.8.0 sklearn: 0.22 numpy: 1.18.0 scipy: 1.4.1 Cython: None pandas: 0.25.3 matplotlib: 3.1.2 joblib: 0.14.1 Built with OpenMP: True ``` ``` System: python: 3.7.6 (default, Feb 14 2020, 16:41:52) [GCC 8.3.1 20190507 (Red Hat 8.3.1-4)] executable: /home/***/ws/siml/.venv/bin/python3 machine: Linux-4.18.0-147.5.1.el8_1.x86_64-x86_64-with-centos-8.1.1911-Core Python dependencies: pip: 19.2.3 setuptools: 41.2.0 sklearn: 0.22.1 numpy: 1.18.1 scipy: 1.4.1 Cython: None pandas: 0.25.3 matplotlib: 3.1.3 joblib: 0.14.1 Built with OpenMP: True ```
non_test
sklearn preprocessing standardscaler gets nan variance when partial fit with sparse data describe the bug when i feed a specific dataset which is sparse to sklearn preprocessing standardscaler partial fit in a specific order i get variance which is nan although data does not contains any nans and is very small when i convert the sparse arrays to dense it works when i change the order to feed the data it works too steps code to reproduce please work with the data i attached python import scipy sparse as sp from sklearn import preprocessing sp load npz npz sp load npz npz buggy behavior preprocessing standardscaler with mean false partial fit print var partial fit print var gets nan when use dence array it works preprocessing standardscaler with mean false partial fit toarray print var partial fit toarray print var when change the order of data it works preprocessing standardscaler with mean false partial fit print var partial fit print var edit fix sample code around expected results python var var var actual results python var var var versions i confirmed this issue in two different environments system python default apr executable usr local var pyenv versions bin machine darwin python dependencies pip setuptools sklearn numpy scipy cython none pandas matplotlib joblib built with openmp true system python default feb executable home ws siml venv bin machine linux with centos core python dependencies pip setuptools sklearn numpy scipy cython none pandas matplotlib joblib built with openmp true
0
331,698
29,047,568,487
IssuesEvent
2023-05-13 19:15:26
T1rax/movies_ugc_service
https://api.github.com/repos/T1rax/movies_ugc_service
opened
Написать сквозные тесты e2e
test
Написать 1-2 сквозных теста всей системы. Можно взять за основу уже имебщийся e2e
1.0
Написать сквозные тесты e2e - Написать 1-2 сквозных теста всей системы. Можно взять за основу уже имебщийся e2e
test
написать сквозные тесты написать сквозных теста всей системы можно взять за основу уже имебщийся
1
27,331
11,471,757,098
IssuesEvent
2020-02-09 13:20:18
foretold-app/foretold
https://api.github.com/repos/foretold-app/foretold
opened
Figure out why FeedItem data was lost
security 🕷 bug
We seem to have lost some FeedItem data. Thankfully this data isn't particularly important to keep, but loosing data is quite bad. If it would be relatively easy, it would be nice to recover it from the backups.
True
Figure out why FeedItem data was lost - We seem to have lost some FeedItem data. Thankfully this data isn't particularly important to keep, but loosing data is quite bad. If it would be relatively easy, it would be nice to recover it from the backups.
non_test
figure out why feeditem data was lost we seem to have lost some feeditem data thankfully this data isn t particularly important to keep but loosing data is quite bad if it would be relatively easy it would be nice to recover it from the backups
0
163,545
12,734,481,878
IssuesEvent
2020-06-25 13:58:20
NLnetLabs/krill
https://api.github.com/repos/NLnetLabs/krill
closed
Add support for rsync URI in the TAL instead of HTTPS for testing purposes
interop testing
### Summary When using `use_ta = true` the generated `.tal` file contains an HTTPS URI to the `.cer` file. Routinator is able to consume this file but OctoRPKI and Fort Validator fail unless the URI is changed to use the RSYNC protocol instead. ### Versions tested: - Krill Docker image `nlnetlabs/krill:v0.2.0` (built from the v0.2.0 GitHub tag). - OctoRPKI Docker image `cloudflare/octorpki` (aka `cloudflare/octorpki:latest` with digest [93131096a0a9](https://hub.docker.com/layers/cloudflare/octorpki/latest/images/sha256-93131096a0a97cb1065660b046f6003f264e79193f2e04124980cb4e1829f6b5), where `/octorpki -version` reports `OctoRPKI v1.1.4 (2019-08-06T16:51:07-0700)`. - Fort Validator Debian package https://github.com/NICMx/FORT-validator/releases/download/v1.1.0/fort_1.1.0-1_amd64.deb installed in a Docker container using image `debian` (`Debian GNU/Linux 10 \n \l` according to `/etc/issue`), where `fort -V` reports `fort 1.1.0`. ### OctoRPKI Error behaviour when the `.cer` URI inside the TAL file uses the HTTPS protocol: ``` octorpki | time="2019-10-30T20:57:29Z" level=info msg="Got open https://ximontest.do.nlnetlabs.nl/ta/ta.cer: no such file or directory but repository not yet synchronized" ``` This is not a very helpful error message as the URI can definitely be fetched from inside the Docker container by WGet, After replacing the URI inside the TAL file with an `rsync://` URI and placing the `.cer` file at that location in the RSYNC repo, OctoRPKI was then able to fetch the ROAs from Krill: ``` octorpki | {"metadata":{"counts":2,"generated":1572476435,"valid":1572480035,"signature":"3046022100992e43910ae22e24ff252527680d21fdd749a2a12975b729466c2e59649ccfe6022100b0ea9598ed5920202aafd8a19db51878a25c3235434afcfda53dd375f42ff285","signatureDate":"3044022028a5d2aad362ed58839b188cc1de8acd9e9892739fe833ad544760482987e22002206b81de1751f504a3b0f9643c70ee0bdfa4934fe5f3db6dc411d62fb1bb2e37a2"},"roas":[{"prefix":"10.0.0.0/24","maxLength":24,"asn":"AS64496","ta":"KRILL"},{"prefix":"10.0.1.0/24","maxLength":24,"asn":"AS64496","ta":"KRILL"}]} ``` ### Fort Validator Error behaviour when the `.cer` URI inside the TAL file uses the HTTPS protocol: ``` ERR: /tmp/ta.tal: Global URI 'https://ximontest.do.nlnetlabs.nl/ta/ta.cer' does not begin with 'rsync://'. ``` After replacing the URI inside the TAL file with an `rsync://` URI and placing the `.cer` file at that location in the RSYNC repo, Fort Validator then produced a different error message: ``` ERR: rsync://ximontest.do.nlnetlabs.nl/repo/ta/0/1169977D368819F3557C6B2353ED047D0CB84192.cer: Certificate validation failed: invalid or inconsistent certificate extension ``` Ximon
1.0
Add support for rsync URI in the TAL instead of HTTPS for testing purposes - ### Summary When using `use_ta = true` the generated `.tal` file contains an HTTPS URI to the `.cer` file. Routinator is able to consume this file but OctoRPKI and Fort Validator fail unless the URI is changed to use the RSYNC protocol instead. ### Versions tested: - Krill Docker image `nlnetlabs/krill:v0.2.0` (built from the v0.2.0 GitHub tag). - OctoRPKI Docker image `cloudflare/octorpki` (aka `cloudflare/octorpki:latest` with digest [93131096a0a9](https://hub.docker.com/layers/cloudflare/octorpki/latest/images/sha256-93131096a0a97cb1065660b046f6003f264e79193f2e04124980cb4e1829f6b5), where `/octorpki -version` reports `OctoRPKI v1.1.4 (2019-08-06T16:51:07-0700)`. - Fort Validator Debian package https://github.com/NICMx/FORT-validator/releases/download/v1.1.0/fort_1.1.0-1_amd64.deb installed in a Docker container using image `debian` (`Debian GNU/Linux 10 \n \l` according to `/etc/issue`), where `fort -V` reports `fort 1.1.0`. ### OctoRPKI Error behaviour when the `.cer` URI inside the TAL file uses the HTTPS protocol: ``` octorpki | time="2019-10-30T20:57:29Z" level=info msg="Got open https://ximontest.do.nlnetlabs.nl/ta/ta.cer: no such file or directory but repository not yet synchronized" ``` This is not a very helpful error message as the URI can definitely be fetched from inside the Docker container by WGet, After replacing the URI inside the TAL file with an `rsync://` URI and placing the `.cer` file at that location in the RSYNC repo, OctoRPKI was then able to fetch the ROAs from Krill: ``` octorpki | {"metadata":{"counts":2,"generated":1572476435,"valid":1572480035,"signature":"3046022100992e43910ae22e24ff252527680d21fdd749a2a12975b729466c2e59649ccfe6022100b0ea9598ed5920202aafd8a19db51878a25c3235434afcfda53dd375f42ff285","signatureDate":"3044022028a5d2aad362ed58839b188cc1de8acd9e9892739fe833ad544760482987e22002206b81de1751f504a3b0f9643c70ee0bdfa4934fe5f3db6dc411d62fb1bb2e37a2"},"roas":[{"prefix":"10.0.0.0/24","maxLength":24,"asn":"AS64496","ta":"KRILL"},{"prefix":"10.0.1.0/24","maxLength":24,"asn":"AS64496","ta":"KRILL"}]} ``` ### Fort Validator Error behaviour when the `.cer` URI inside the TAL file uses the HTTPS protocol: ``` ERR: /tmp/ta.tal: Global URI 'https://ximontest.do.nlnetlabs.nl/ta/ta.cer' does not begin with 'rsync://'. ``` After replacing the URI inside the TAL file with an `rsync://` URI and placing the `.cer` file at that location in the RSYNC repo, Fort Validator then produced a different error message: ``` ERR: rsync://ximontest.do.nlnetlabs.nl/repo/ta/0/1169977D368819F3557C6B2353ED047D0CB84192.cer: Certificate validation failed: invalid or inconsistent certificate extension ``` Ximon
test
add support for rsync uri in the tal instead of https for testing purposes summary when using use ta true the generated tal file contains an https uri to the cer file routinator is able to consume this file but octorpki and fort validator fail unless the uri is changed to use the rsync protocol instead versions tested krill docker image nlnetlabs krill built from the github tag octorpki docker image cloudflare octorpki aka cloudflare octorpki latest with digest where octorpki version reports octorpki fort validator debian package installed in a docker container using image debian debian gnu linux n l according to etc issue where fort v reports fort octorpki error behaviour when the cer uri inside the tal file uses the https protocol octorpki time level info msg got open no such file or directory but repository not yet synchronized this is not a very helpful error message as the uri can definitely be fetched from inside the docker container by wget after replacing the uri inside the tal file with an rsync uri and placing the cer file at that location in the rsync repo octorpki was then able to fetch the roas from krill octorpki metadata counts generated valid signature signaturedate roas fort validator error behaviour when the cer uri inside the tal file uses the https protocol err tmp ta tal global uri does not begin with rsync after replacing the uri inside the tal file with an rsync uri and placing the cer file at that location in the rsync repo fort validator then produced a different error message err rsync ximontest do nlnetlabs nl repo ta cer certificate validation failed invalid or inconsistent certificate extension ximon
1
78,485
7,644,536,343
IssuesEvent
2018-05-08 15:44:31
Microsoft/AzureStorageExplorer
https://api.github.com/repos/Microsoft/AzureStorageExplorer
closed
The scrollbar still displays on Account Panel after maximizing Storage Explorer
resolved: won't fix testing
**Storage Explorer Version**: <!-- Enter the version of Storage Explorer in which you encountered your issue (see About dialog). -->1.1.0 RC **Platform**: <!-- Enter the platform (see About dialog). -->Windows/Linux/Mac **Architecture**: <!-- Enter the architecture of the app (see About dialog). -->ia32/x64 **Build Number**: <!-- Enter the build number (see About dialog). -->20180507.1 **Commit**: <!-- Enter the commit ID (see About dialog). -->bcc8990d **Regression From**: <!-- Was this working on a previous version? If so, which one? --> Not a regression #### Steps to Reproduce: #### 1. <!-- Describe the steps necessary to reproduce your issue. -->Launch Storage Explorer and sign in with your account. 2. <!-- Be as detailed as you can. -->Maximize Storage Explorer and check whether scrollbar displays on Account Panel or not. #### Expected Experience: #### <!-- Describe the outcome you expect when performing the above steps. --> No scrollbar displays on Account Panel after maximizing Storage Explorer. #### Actual Experience: #### <!-- Describe the outcome that actually occurs when performing the above steps. --> The scrollbar still displays on Account Panel after maximizing Storage Explorer. ![scrollbar](https://user-images.githubusercontent.com/34729022/39752029-b840c40c-52a9-11e8-9330-813d5c6f29d0.png)
1.0
The scrollbar still displays on Account Panel after maximizing Storage Explorer - **Storage Explorer Version**: <!-- Enter the version of Storage Explorer in which you encountered your issue (see About dialog). -->1.1.0 RC **Platform**: <!-- Enter the platform (see About dialog). -->Windows/Linux/Mac **Architecture**: <!-- Enter the architecture of the app (see About dialog). -->ia32/x64 **Build Number**: <!-- Enter the build number (see About dialog). -->20180507.1 **Commit**: <!-- Enter the commit ID (see About dialog). -->bcc8990d **Regression From**: <!-- Was this working on a previous version? If so, which one? --> Not a regression #### Steps to Reproduce: #### 1. <!-- Describe the steps necessary to reproduce your issue. -->Launch Storage Explorer and sign in with your account. 2. <!-- Be as detailed as you can. -->Maximize Storage Explorer and check whether scrollbar displays on Account Panel or not. #### Expected Experience: #### <!-- Describe the outcome you expect when performing the above steps. --> No scrollbar displays on Account Panel after maximizing Storage Explorer. #### Actual Experience: #### <!-- Describe the outcome that actually occurs when performing the above steps. --> The scrollbar still displays on Account Panel after maximizing Storage Explorer. ![scrollbar](https://user-images.githubusercontent.com/34729022/39752029-b840c40c-52a9-11e8-9330-813d5c6f29d0.png)
test
the scrollbar still displays on account panel after maximizing storage explorer storage explorer version rc platform windows linux mac architecture build number commit regression from not a regression steps to reproduce launch storage explorer and sign in with your account maximize storage explorer and check whether scrollbar displays on account panel or not expected experience no scrollbar displays on account panel after maximizing storage explorer actual experience the scrollbar still displays on account panel after maximizing storage explorer
1
125,974
10,372,065,467
IssuesEvent
2019-09-09 01:01:05
ssm-deepcove/discover_deep_cove
https://api.github.com/repos/ssm-deepcove/discover_deep_cove
closed
Add an image to photograph activity if possible
Client requested Investigate Prior: On-Site Testing enhancement
Clients have requested that they would like the ability to have an image included on the photograph activity i.e. so that they can show an image of what an item used to be for example.
1.0
Add an image to photograph activity if possible - Clients have requested that they would like the ability to have an image included on the photograph activity i.e. so that they can show an image of what an item used to be for example.
test
add an image to photograph activity if possible clients have requested that they would like the ability to have an image included on the photograph activity i e so that they can show an image of what an item used to be for example
1
38,763
2,850,254,072
IssuesEvent
2015-05-31 12:14:36
damonkohler/sl4a
https://api.github.com/repos/damonkohler/sl4a
opened
Ability to modify pitch of text-to-speech
auto-migrated Priority-Medium Type-Enhancement
_From @GoogleCodeExporter on May 31, 2015 11:31_ ``` I am working on a project at the moment and I thought that being able to modify the TTS pitch would be quite helpful. I decided to check out the source and add this modification myself and it works (see attached). The only weird thing is that, I had to make the parameter for setTtsPitch() a String rather than a float or double (I get some "Argument is invalid" error, even after doing something like setTtsPitch(float(5)) ). It's not much but I hope a method like this does get included in future. :) Cheers ``` Original issue reported on code.google.com by `chrispy...@gmail.com` on 3 Feb 2013 at 4:36 Attachments: * [TextToSpeechFacade.java](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-677/comment-0/TextToSpeechFacade.java) _Copied from original issue: damonkohler/android-scripting#677_
1.0
Ability to modify pitch of text-to-speech - _From @GoogleCodeExporter on May 31, 2015 11:31_ ``` I am working on a project at the moment and I thought that being able to modify the TTS pitch would be quite helpful. I decided to check out the source and add this modification myself and it works (see attached). The only weird thing is that, I had to make the parameter for setTtsPitch() a String rather than a float or double (I get some "Argument is invalid" error, even after doing something like setTtsPitch(float(5)) ). It's not much but I hope a method like this does get included in future. :) Cheers ``` Original issue reported on code.google.com by `chrispy...@gmail.com` on 3 Feb 2013 at 4:36 Attachments: * [TextToSpeechFacade.java](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-677/comment-0/TextToSpeechFacade.java) _Copied from original issue: damonkohler/android-scripting#677_
non_test
ability to modify pitch of text to speech from googlecodeexporter on may i am working on a project at the moment and i thought that being able to modify the tts pitch would be quite helpful i decided to check out the source and add this modification myself and it works see attached the only weird thing is that i had to make the parameter for setttspitch a string rather than a float or double i get some argument is invalid error even after doing something like setttspitch float it s not much but i hope a method like this does get included in future cheers original issue reported on code google com by chrispy gmail com on feb at attachments copied from original issue damonkohler android scripting
0
240,385
20,025,814,268
IssuesEvent
2022-02-01 21:10:37
microsoft/playwright
https://api.github.com/repos/microsoft/playwright
closed
[Feature]: Use source-maps when available in test-runner for reporting and errors
test-runner v1.19 inner-loop
See https://github.com/microsoft/playwright-vscode/issues/27 Users who are using a custom setup with the test-runner (feed transpiled files into the test-runner) currently can't use the extension. It should be possible that when there are source-map files, that they get used in the JSON reporter and in the error stack-traces.
1.0
[Feature]: Use source-maps when available in test-runner for reporting and errors - See https://github.com/microsoft/playwright-vscode/issues/27 Users who are using a custom setup with the test-runner (feed transpiled files into the test-runner) currently can't use the extension. It should be possible that when there are source-map files, that they get used in the JSON reporter and in the error stack-traces.
test
use source maps when available in test runner for reporting and errors see users who are using a custom setup with the test runner feed transpiled files into the test runner currently can t use the extension it should be possible that when there are source map files that they get used in the json reporter and in the error stack traces
1
127,690
17,353,964,643
IssuesEvent
2021-07-29 12:20:20
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Creating subspaces is very unintuitive
A-Spaces X-Needs-Design
You do so by hitting the 'add existing room' button. However, from the user perspective, spaces aren't rooms - they're spaces. So why would i select 'add existing room' when i want to find/add a space?
1.0
Creating subspaces is very unintuitive - You do so by hitting the 'add existing room' button. However, from the user perspective, spaces aren't rooms - they're spaces. So why would i select 'add existing room' when i want to find/add a space?
non_test
creating subspaces is very unintuitive you do so by hitting the add existing room button however from the user perspective spaces aren t rooms they re spaces so why would i select add existing room when i want to find add a space
0
226,581
18,040,860,140
IssuesEvent
2021-09-18 02:51:57
logicmoo/logicmoo_workspace
https://api.github.com/repos/logicmoo/logicmoo_workspace
opened
logicmoo.base.examples.fol.SANITY_FI_HUMAN_01@Test_9999_Line_9999__Exitcode_7
Test_9999 SANITY_FI_HUMAN_01 logicmoo.base.examples.fol
(cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif sanity_fi_human_01.pfc.pl) Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.examples.fol/SANITY_FI_HUMAN_01/logicmoo_base_examples_fol_SANITY_FI_HUMAN_01_Test_9999_Line_9999__Exitcode_7/ This: https://jenkins.logicmoo.org/job/logicmoo_workspace/41/testReport/logicmoo.base.examples.fol/SANITY_FI_HUMAN_01/logicmoo_base_examples_fol_SANITY_FI_HUMAN_01_Test_9999_Line_9999__Exitcode_7/ ``` % running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_fi_human_01.pfc.pl'), %~ /var/lib/jenkins/.local/share/swi-prolog/pack/logicmoo_utils/prolog/logicmoo_test_header.pl:92 %~ this_test_might_need( :-( use_module( library(logicmoo_plarkc)))) :- include('sanity_fi_sk.pfc'). %~ message_hook_type(error) %~ message_hook( %~ initialization_exception( error(existence_error(source_sink,'sanity_fi_sk.pfc'),Existence_error)), %~ error, %~ [ 'Prolog initialisation failed:', %~ nl, %~ '~w `~p\' does not exist'-[source_sink,'sanity_fi_sk.pfc']]) %~ unused(save_junit_results) ``` totalTime=1 FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k sanity_fi_human_01.pfc.pl (returned 1)
1.0
logicmoo.base.examples.fol.SANITY_FI_HUMAN_01@Test_9999_Line_9999__Exitcode_7 - (cd /var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol ; timeout --foreground --preserve-status -s SIGKILL -k 10s 10s lmoo-clif sanity_fi_human_01.pfc.pl) Latest: https://jenkins.logicmoo.org/job/logicmoo_workspace/lastBuild/testReport/logicmoo.base.examples.fol/SANITY_FI_HUMAN_01/logicmoo_base_examples_fol_SANITY_FI_HUMAN_01_Test_9999_Line_9999__Exitcode_7/ This: https://jenkins.logicmoo.org/job/logicmoo_workspace/41/testReport/logicmoo.base.examples.fol/SANITY_FI_HUMAN_01/logicmoo_base_examples_fol_SANITY_FI_HUMAN_01_Test_9999_Line_9999__Exitcode_7/ ``` % running('/var/lib/jenkins/workspace/logicmoo_workspace/packs_sys/logicmoo_base/t/examples/fol/sanity_fi_human_01.pfc.pl'), %~ /var/lib/jenkins/.local/share/swi-prolog/pack/logicmoo_utils/prolog/logicmoo_test_header.pl:92 %~ this_test_might_need( :-( use_module( library(logicmoo_plarkc)))) :- include('sanity_fi_sk.pfc'). %~ message_hook_type(error) %~ message_hook( %~ initialization_exception( error(existence_error(source_sink,'sanity_fi_sk.pfc'),Existence_error)), %~ error, %~ [ 'Prolog initialisation failed:', %~ nl, %~ '~w `~p\' does not exist'-[source_sink,'sanity_fi_sk.pfc']]) %~ unused(save_junit_results) ``` totalTime=1 FAILED: /var/lib/jenkins/workspace/logicmoo_workspace/bin/lmoo-junit-minor -k sanity_fi_human_01.pfc.pl (returned 1)
test
logicmoo base examples fol sanity fi human test line exitcode cd var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol timeout foreground preserve status s sigkill k lmoo clif sanity fi human pfc pl latest this running var lib jenkins workspace logicmoo workspace packs sys logicmoo base t examples fol sanity fi human pfc pl var lib jenkins local share swi prolog pack logicmoo utils prolog logicmoo test header pl this test might need use module library logicmoo plarkc include sanity fi sk pfc message hook type error message hook initialization exception error existence error source sink sanity fi sk pfc existence error error prolog initialisation failed nl w p does not exist unused save junit results totaltime failed var lib jenkins workspace logicmoo workspace bin lmoo junit minor k sanity fi human pfc pl returned
1
715,567
24,604,352,696
IssuesEvent
2022-10-14 14:56:56
hashicorp/terraform-cdk
https://api.github.com/repos/hashicorp/terraform-cdk
opened
cdktf debug should list prebuilt provider versions
enhancement cdktf-cli priority/important-soon ux/debugging size/medium pre-built providers cli/provider-subcommands
<!--- Please keep this note for the community ---> ### Community Note - Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request - Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request - If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description <!--- Please leave a helpful description of the feature request here. ---> <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> And ideally also show the peer dependency of the provider in question so pre-built provider related dependency issues are easier to debug. ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? ---> - https://discuss.hashicorp.com/t/cdk-not-respecting-pom-xml/45418/7
1.0
cdktf debug should list prebuilt provider versions - <!--- Please keep this note for the community ---> ### Community Note - Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request - Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request - If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description <!--- Please leave a helpful description of the feature request here. ---> <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> And ideally also show the peer dependency of the provider in question so pre-built provider related dependency issues are easier to debug. ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? ---> - https://discuss.hashicorp.com/t/cdk-not-respecting-pom-xml/45418/7
non_test
cdktf debug should list prebuilt provider versions community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description and ideally also show the peer dependency of the provider in question so pre built provider related dependency issues are easier to debug references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation
0
59,591
6,656,541,761
IssuesEvent
2017-09-29 21:03:22
pkrog/biodb
https://api.github.com/repos/pkrog/biodb
opened
Peakforest LCMS get-range issue
peakforest test
The request `https://peakforest-alpha.inra.fr/rest/spectra/lcms/peaks/get-range/321.047701453466/321.050911946534?token=...` returns a JSON of matching with the second element having its `source` field set to `null`. Check that it does not cause any issue in code.
1.0
Peakforest LCMS get-range issue - The request `https://peakforest-alpha.inra.fr/rest/spectra/lcms/peaks/get-range/321.047701453466/321.050911946534?token=...` returns a JSON of matching with the second element having its `source` field set to `null`. Check that it does not cause any issue in code.
test
peakforest lcms get range issue the request returns a json of matching with the second element having its source field set to null check that it does not cause any issue in code
1
77,515
7,575,437,406
IssuesEvent
2018-04-24 01:41:46
intesar/Fx-Test-Data
https://api.github.com/repos/intesar/Fx-Test-Data
reopened
Test-bvpoGI : enterprise_user_signup_valid
Test-bvpoGI
Project : Test-bvpoGI Job : Stg Env : Stg Region : FxLabs/US_WEST_1 Result : fail Status Code : 200 Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Tue, 24 Apr 2018 00:53:27 GMT]} Endpoint : http://stg1.fxlabs.io/api/v1/users/enterprise-sign-up Request : { "name" : "first name", "email" : "Iio@fxlabs.io", "password" : "12345678", "company" : "RK1rAe6hm" } Response : { "requestId" : "None", "requestTime" : "2018-04-24T00:53:28.635+0000", "errors" : false, "messages" : [ { "type" : "INFO", "key" : "", "value" : "Sign-up successful!" } ], "data" : true, "totalPages" : 1, "totalElements" : 1 } Logs : Assertion [@Response.errors == false] passed, expected [false] and found [false]Assertion [@StatusCode == 200] passed, expected [200] and found [200]
1.0
Test-bvpoGI : enterprise_user_signup_valid - Project : Test-bvpoGI Job : Stg Env : Stg Region : FxLabs/US_WEST_1 Result : fail Status Code : 200 Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Tue, 24 Apr 2018 00:53:27 GMT]} Endpoint : http://stg1.fxlabs.io/api/v1/users/enterprise-sign-up Request : { "name" : "first name", "email" : "Iio@fxlabs.io", "password" : "12345678", "company" : "RK1rAe6hm" } Response : { "requestId" : "None", "requestTime" : "2018-04-24T00:53:28.635+0000", "errors" : false, "messages" : [ { "type" : "INFO", "key" : "", "value" : "Sign-up successful!" } ], "data" : true, "totalPages" : 1, "totalElements" : 1 } Logs : Assertion [@Response.errors == false] passed, expected [false] and found [false]Assertion [@StatusCode == 200] passed, expected [200] and found [200]
test
test bvpogi enterprise user signup valid project test bvpogi job stg env stg region fxlabs us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options content type transfer encoding date endpoint request name first name email iio fxlabs io password company response requestid none requesttime errors false messages type info key value sign up successful data true totalpages totalelements logs assertion passed expected and found assertion passed expected and found
1
345,371
10,361,564,565
IssuesEvent
2019-09-06 10:20:00
diamm/diamm
https://api.github.com/repos/diamm/diamm
closed
Strange inventory note proliferation
Component: Metadata Priority: Medium Type: Bug
https://www.diamm.ac.uk/sources/635/#/inventory If you click open the first piece in the inventory, you see a long list of notes (see screenshot). I don't know where these notes are coming from (they are obviously not to do with this work), and when I click on edit the only note that is in the record is not actually showing up in this list. <img width="1488" alt="Screen Shot 2019-08-30 at 11 19 39" src="https://user-images.githubusercontent.com/10816169/64013704-53608980-cb18-11e9-92ef-f9a3099dbd3f.png">
1.0
Strange inventory note proliferation - https://www.diamm.ac.uk/sources/635/#/inventory If you click open the first piece in the inventory, you see a long list of notes (see screenshot). I don't know where these notes are coming from (they are obviously not to do with this work), and when I click on edit the only note that is in the record is not actually showing up in this list. <img width="1488" alt="Screen Shot 2019-08-30 at 11 19 39" src="https://user-images.githubusercontent.com/10816169/64013704-53608980-cb18-11e9-92ef-f9a3099dbd3f.png">
non_test
strange inventory note proliferation if you click open the first piece in the inventory you see a long list of notes see screenshot i don t know where these notes are coming from they are obviously not to do with this work and when i click on edit the only note that is in the record is not actually showing up in this list img width alt screen shot at src
0
383,436
11,356,173,645
IssuesEvent
2020-01-24 21:57:09
celo-org/celo-monorepo
https://api.github.com/repos/celo-org/celo-monorepo
closed
Users SBAT be prompted to switch to forno when experiencing connectivity issues
Priority: P1 applications feature investigate triaged wallet
### Expected Behavior (Copied over from #1171, splitting into two issues- cc @nityas ) We propose a modal that gives users with connectivity issues the option to enable Forno mode. NOTES: If geth has restarted more than twice, or syncing has taken longer than 15s, but there is still some internet connection(eg type:unknown) show the modal below. same modal template as the 'Are you sure?' Backup Key Modal, with enable being the primary cta: a user should only see this modal after they have completed the 'Welcome to Celo' screen once ‘Dismiss’ is pressed, we should not show the modal again, to avoid an overly intrusive or confusing experience. (TODO: are there other signals we can use to fine tune this?) _____________ **Switch Connection Mode?** We’ve noticed you’re having some trouble connecting. We recommend enabling **Celo-Lite** mode to allow you to keep using the Celo Wallet with intermittent connection. Celo-Lite allows you to communicate with the Celo Network through a trusted, hosted node, which can give faster performance than running a node on your device. You can always change this mode in Settings. [Dismiss] [Enable Celo-Lite] ### Current Behavior Working as toggle from settings screen
1.0
Users SBAT be prompted to switch to forno when experiencing connectivity issues - ### Expected Behavior (Copied over from #1171, splitting into two issues- cc @nityas ) We propose a modal that gives users with connectivity issues the option to enable Forno mode. NOTES: If geth has restarted more than twice, or syncing has taken longer than 15s, but there is still some internet connection(eg type:unknown) show the modal below. same modal template as the 'Are you sure?' Backup Key Modal, with enable being the primary cta: a user should only see this modal after they have completed the 'Welcome to Celo' screen once ‘Dismiss’ is pressed, we should not show the modal again, to avoid an overly intrusive or confusing experience. (TODO: are there other signals we can use to fine tune this?) _____________ **Switch Connection Mode?** We’ve noticed you’re having some trouble connecting. We recommend enabling **Celo-Lite** mode to allow you to keep using the Celo Wallet with intermittent connection. Celo-Lite allows you to communicate with the Celo Network through a trusted, hosted node, which can give faster performance than running a node on your device. You can always change this mode in Settings. [Dismiss] [Enable Celo-Lite] ### Current Behavior Working as toggle from settings screen
non_test
users sbat be prompted to switch to forno when experiencing connectivity issues expected behavior copied over from splitting into two issues cc nityas we propose a modal that gives users with connectivity issues the option to enable forno mode notes if geth has restarted more than twice or syncing has taken longer than but there is still some internet connection eg type unknown show the modal below same modal template as the are you sure backup key modal with enable being the primary cta a user should only see this modal after they have completed the welcome to celo screen once ‘dismiss’ is pressed we should not show the modal again to avoid an overly intrusive or confusing experience todo are there other signals we can use to fine tune this switch connection mode we’ve noticed you’re having some trouble connecting we recommend enabling celo lite mode to allow you to keep using the celo wallet with intermittent connection celo lite allows you to communicate with the celo network through a trusted hosted node which can give faster performance than running a node on your device you can always change this mode in settings current behavior working as toggle from settings screen
0
7,369
6,924,299,879
IssuesEvent
2017-11-30 12:13:28
aiidateam/aiida_core
https://api.github.com/repos/aiidateam/aiida_core
closed
debugging: more helpful error message for db-tests in case of import error
topic/TestInfrastructure type/enhancement type/proposal type/quality-of-life type/trivial
## Description The following type of error message is common when running aiida tests in case of import errors (missing dependencies for instance). ``` File ".../aiida-core/aiida/backends/testbase.py", line 165, in run_aiida_db_tests test_suite.addTest(test_loader.loadTestsFromName(modulename)) File ".../lib/python2.7/unittest/loader.py", line 100, in loadTestsFromName parent, obj = obj, getattr(obj, part) AttributeError: 'module' object has no attribute 'persistence' ``` The error message could easily be very helpful if only the name of the module was output too ## Possible solution The AttributeError could be caught in `run_aiida_db_tests` in `aiida/backends/testase.py` and the more helpful message could be output containing the `modulename`.
1.0
debugging: more helpful error message for db-tests in case of import error - ## Description The following type of error message is common when running aiida tests in case of import errors (missing dependencies for instance). ``` File ".../aiida-core/aiida/backends/testbase.py", line 165, in run_aiida_db_tests test_suite.addTest(test_loader.loadTestsFromName(modulename)) File ".../lib/python2.7/unittest/loader.py", line 100, in loadTestsFromName parent, obj = obj, getattr(obj, part) AttributeError: 'module' object has no attribute 'persistence' ``` The error message could easily be very helpful if only the name of the module was output too ## Possible solution The AttributeError could be caught in `run_aiida_db_tests` in `aiida/backends/testase.py` and the more helpful message could be output containing the `modulename`.
non_test
debugging more helpful error message for db tests in case of import error description the following type of error message is common when running aiida tests in case of import errors missing dependencies for instance file aiida core aiida backends testbase py line in run aiida db tests test suite addtest test loader loadtestsfromname modulename file lib unittest loader py line in loadtestsfromname parent obj obj getattr obj part attributeerror module object has no attribute persistence the error message could easily be very helpful if only the name of the module was output too possible solution the attributeerror could be caught in run aiida db tests in aiida backends testase py and the more helpful message could be output containing the modulename
0
95,655
12,032,463,673
IssuesEvent
2020-04-13 12:12:39
reach4help/reach4help
https://api.github.com/repos/reach4help/reach4help
closed
Design view: Name & address
UI Design
* [x] Initial design for desktop * [x] Adapt design for mobile <!--zenhub info: do not edit anything after this line, it will be automatically changed--> -------- ### [ZenHub Information](https://app.zenhub.com/workspaces/reach4help-5e8dcbfb14ac087f410cbabb/issues/reach4help/reach4help/344) *This information is updated automatically. To modify it, please use ZenHub.* **Belonging to Epics:** * **[EPIC]** [#327 - View: Name & Address](https://github.com/reach4help/reach4help/issues/327) **Blocking:** * [BLOCKED BY: [#344](https://github.com/reach4help/reach4help/issues/344)] [#344 - Implement view: Name & Address](https://github.com/reach4help/reach4help/issues/100) <!--zenhub info end-->
1.0
Design view: Name & address - * [x] Initial design for desktop * [x] Adapt design for mobile <!--zenhub info: do not edit anything after this line, it will be automatically changed--> -------- ### [ZenHub Information](https://app.zenhub.com/workspaces/reach4help-5e8dcbfb14ac087f410cbabb/issues/reach4help/reach4help/344) *This information is updated automatically. To modify it, please use ZenHub.* **Belonging to Epics:** * **[EPIC]** [#327 - View: Name & Address](https://github.com/reach4help/reach4help/issues/327) **Blocking:** * [BLOCKED BY: [#344](https://github.com/reach4help/reach4help/issues/344)] [#344 - Implement view: Name & Address](https://github.com/reach4help/reach4help/issues/100) <!--zenhub info end-->
non_test
design view name address initial design for desktop adapt design for mobile this information is updated automatically to modify it please use zenhub belonging to epics blocking
0
109,507
13,781,290,310
IssuesEvent
2020-10-08 15:58:54
brave/brave-ios
https://api.github.com/repos/brave/brave-ios
closed
Brave today extension
QA/Yes blocked: needs design enhancement release-notes/include
Search extension to display on today widgets screen. Search, new tab, private tab.
1.0
Brave today extension - Search extension to display on today widgets screen. Search, new tab, private tab.
non_test
brave today extension search extension to display on today widgets screen search new tab private tab
0
172,035
13,260,253,576
IssuesEvent
2020-08-20 17:55:08
ramp4-pcar4/ramp4-pcar4
https://api.github.com/repos/ramp4-pcar4/ramp4-pcar4
opened
Test for RampMap
effort: far away flavour: tests priority: nice type: preventative
Create tests for the RampMap class in GeoAPI. See `\packages\ramp-core\src\fixtures\mapnav\tests\` and `\packages\ramp-core\tests` for starter samples. Suggest prioritizing `e2e` tests, as unit tests require mocking which can be hefty task that has the smell of redundancy. Things to consider - validation of loading promises - validation of event wireup. e.g. mouse move, click, extent change - suite of zoomies functions - suite of `get` functions that return information about the map - screen point to map point converter will likely be more things added if this issue gets stale
1.0
Test for RampMap - Create tests for the RampMap class in GeoAPI. See `\packages\ramp-core\src\fixtures\mapnav\tests\` and `\packages\ramp-core\tests` for starter samples. Suggest prioritizing `e2e` tests, as unit tests require mocking which can be hefty task that has the smell of redundancy. Things to consider - validation of loading promises - validation of event wireup. e.g. mouse move, click, extent change - suite of zoomies functions - suite of `get` functions that return information about the map - screen point to map point converter will likely be more things added if this issue gets stale
test
test for rampmap create tests for the rampmap class in geoapi see packages ramp core src fixtures mapnav tests and packages ramp core tests for starter samples suggest prioritizing tests as unit tests require mocking which can be hefty task that has the smell of redundancy things to consider validation of loading promises validation of event wireup e g mouse move click extent change suite of zoomies functions suite of get functions that return information about the map screen point to map point converter will likely be more things added if this issue gets stale
1
294,866
25,410,850,009
IssuesEvent
2022-11-22 18:50:27
NuGet/Home
https://api.github.com/repos/NuGet/Home
closed
Dependencies tree unit tests
Priority:2 Functionality:VisualStudioUI Type:Engineering Type:Test Partner:Project-System
From https://github.com/NuGet/NuGet.Client/pull/3392 Tracking adding more tests to the code in NuGet.Client for populating transitive dependencies in Solution Explorer.
1.0
Dependencies tree unit tests - From https://github.com/NuGet/NuGet.Client/pull/3392 Tracking adding more tests to the code in NuGet.Client for populating transitive dependencies in Solution Explorer.
test
dependencies tree unit tests from tracking adding more tests to the code in nuget client for populating transitive dependencies in solution explorer
1
662,666
22,148,646,261
IssuesEvent
2022-06-03 14:32:58
kudobuilder/kudo
https://api.github.com/repos/kudobuilder/kudo
closed
Question: Deploy instance using CRD or API call instead of "kubectl kudo install"
component/cli kind/enhancement priority/long-term-important community
We're are currently evaluating kudo for deploying individual instances for larger tenants that are seperated into different namespaces. The idea is building a user interface that displays a form of parameters in order to deploy the infrastructure for each of the tenants. Most operators found on operatorhub.io allow deploying application instances using CRDs. Our initial plan was creating CRDs for deployment instructions using a Java application. I wasn't able to find anything specific in the documentation if that's officially supported. What I found is this Instance definition [01-upgrade.yml](https://github.com/kudobuilder/kudo/blob/master/test/integration/cli-install/01-upgrade.yaml) ``` yaml apiVersion: kudo.dev/v1beta1 kind: Instance metadata: name: cli-install labels: kudo.dev/operator: cli-install-operator spec: operatorVersion: name: cli-install-operator-0.1.0 parameters: memory: "2Gi" ``` So my question is if a syntax like the snipped above or an official API can be used in order to fire up an instance using a standard kubernetes client library (e.g. [Java](https://github.com/kubernetes-client/java)). Bottom line: Nice work! We believe declerative operators will be a huge step in the whole kubernetes ecosystem.
1.0
Question: Deploy instance using CRD or API call instead of "kubectl kudo install" - We're are currently evaluating kudo for deploying individual instances for larger tenants that are seperated into different namespaces. The idea is building a user interface that displays a form of parameters in order to deploy the infrastructure for each of the tenants. Most operators found on operatorhub.io allow deploying application instances using CRDs. Our initial plan was creating CRDs for deployment instructions using a Java application. I wasn't able to find anything specific in the documentation if that's officially supported. What I found is this Instance definition [01-upgrade.yml](https://github.com/kudobuilder/kudo/blob/master/test/integration/cli-install/01-upgrade.yaml) ``` yaml apiVersion: kudo.dev/v1beta1 kind: Instance metadata: name: cli-install labels: kudo.dev/operator: cli-install-operator spec: operatorVersion: name: cli-install-operator-0.1.0 parameters: memory: "2Gi" ``` So my question is if a syntax like the snipped above or an official API can be used in order to fire up an instance using a standard kubernetes client library (e.g. [Java](https://github.com/kubernetes-client/java)). Bottom line: Nice work! We believe declerative operators will be a huge step in the whole kubernetes ecosystem.
non_test
question deploy instance using crd or api call instead of kubectl kudo install we re are currently evaluating kudo for deploying individual instances for larger tenants that are seperated into different namespaces the idea is building a user interface that displays a form of parameters in order to deploy the infrastructure for each of the tenants most operators found on operatorhub io allow deploying application instances using crds our initial plan was creating crds for deployment instructions using a java application i wasn t able to find anything specific in the documentation if that s officially supported what i found is this instance definition yaml apiversion kudo dev kind instance metadata name cli install labels kudo dev operator cli install operator spec operatorversion name cli install operator parameters memory so my question is if a syntax like the snipped above or an official api can be used in order to fire up an instance using a standard kubernetes client library e g bottom line nice work we believe declerative operators will be a huge step in the whole kubernetes ecosystem
0
671,173
22,746,717,462
IssuesEvent
2022-07-07 09:48:13
aquasecurity/trivy-operator
https://api.github.com/repos/aquasecurity/trivy-operator
opened
ghcr is the default container registry in trivy-operator helm chart
good first issue kind/feature priority/important-soon kind/deployment
set ghcr to be the default container registry in trivy-operator helm chart
1.0
ghcr is the default container registry in trivy-operator helm chart - set ghcr to be the default container registry in trivy-operator helm chart
non_test
ghcr is the default container registry in trivy operator helm chart set ghcr to be the default container registry in trivy operator helm chart
0
184,128
14,271,902,287
IssuesEvent
2020-11-21 14:40:28
mui-org/material-ui
https://api.github.com/repos/mui-org/material-ui
closed
[Slider] Testing onChangeCommitted from click
component: Slider good first issue test
<!-- Provide a general summary of the issue in the Title above --> <!-- Thank you very much for contributing to Material-UI by creating an issue! ❤️ To avoid duplicate issues we ask you to check off the following list. --> <!-- Checked checkbox should look like this: [x] --> - [ ] The issue is present in the latest release. - [ x] I have searched the [issues](https://github.com/mui-org/material-ui/issues) of this repository and believe that this is not a duplicate. ## Current Behavior 😯 When testing the slider functionality using rtl and jest no significant DOM update happens when clicking on labels (or doing mousedown or other events) <!-- Describe what happens instead of the expected behavior. --> ## Expected Behavior 🤔 DOM should update with the selected mark as Slider value <!-- Describe what should happen. --> ## Steps to Reproduce 🕹 In the codesadbox example i tried to test this component of a slider value change on label (slider mark) click. But i cant seem to get the onChangeCommitted to trigger when testing. The slider component dose not indicate any change https://codesandbox.io/s/romantic-goldstine-gpjbj?file=/src/App.test.js <!-- Provide a link to a live example (you can use codesandbox.io) and an unambiguous set of steps to reproduce this bug. Include code to reproduce, if relevant (which it most likely is). You should use the official codesandbox template as a starting point: https://material-ui.com/r/issue-template If you have an issue concerning TypeScript please start from this TypeScript playground: https://material-ui.com/r/ts-issue-template Issues without some form of live example have a longer response time. --> ## Context 🔦 I am trying to test onChangeCommitted in the slider component. But ive tried up and down to test this event on the slider and its getting really frustrating as there dosent seem to be many threads or SO posts about so I might just be missing something trivial. Ive added a sample on codesandbox. Hope some one can help thanks :) <!-- What are you trying to accomplish? How has this issue affected you? Providing context helps us come up with a solution that is most useful in the real world. --> ## Your Environment 🌎 <!-- Include as many relevant details about the environment with which you experienced the bug. If you encounter issues with TypeScript please include version and tsconfig. --> | Tech | Version | | ----------- | ------- | | Material-UI | v4.3.1 | | React | v16.8.4 | | Browser | Chrome | | TypeScript | Y (but not in example) | | etc. | |
1.0
[Slider] Testing onChangeCommitted from click - <!-- Provide a general summary of the issue in the Title above --> <!-- Thank you very much for contributing to Material-UI by creating an issue! ❤️ To avoid duplicate issues we ask you to check off the following list. --> <!-- Checked checkbox should look like this: [x] --> - [ ] The issue is present in the latest release. - [ x] I have searched the [issues](https://github.com/mui-org/material-ui/issues) of this repository and believe that this is not a duplicate. ## Current Behavior 😯 When testing the slider functionality using rtl and jest no significant DOM update happens when clicking on labels (or doing mousedown or other events) <!-- Describe what happens instead of the expected behavior. --> ## Expected Behavior 🤔 DOM should update with the selected mark as Slider value <!-- Describe what should happen. --> ## Steps to Reproduce 🕹 In the codesadbox example i tried to test this component of a slider value change on label (slider mark) click. But i cant seem to get the onChangeCommitted to trigger when testing. The slider component dose not indicate any change https://codesandbox.io/s/romantic-goldstine-gpjbj?file=/src/App.test.js <!-- Provide a link to a live example (you can use codesandbox.io) and an unambiguous set of steps to reproduce this bug. Include code to reproduce, if relevant (which it most likely is). You should use the official codesandbox template as a starting point: https://material-ui.com/r/issue-template If you have an issue concerning TypeScript please start from this TypeScript playground: https://material-ui.com/r/ts-issue-template Issues without some form of live example have a longer response time. --> ## Context 🔦 I am trying to test onChangeCommitted in the slider component. But ive tried up and down to test this event on the slider and its getting really frustrating as there dosent seem to be many threads or SO posts about so I might just be missing something trivial. Ive added a sample on codesandbox. Hope some one can help thanks :) <!-- What are you trying to accomplish? How has this issue affected you? Providing context helps us come up with a solution that is most useful in the real world. --> ## Your Environment 🌎 <!-- Include as many relevant details about the environment with which you experienced the bug. If you encounter issues with TypeScript please include version and tsconfig. --> | Tech | Version | | ----------- | ------- | | Material-UI | v4.3.1 | | React | v16.8.4 | | Browser | Chrome | | TypeScript | Y (but not in example) | | etc. | |
test
testing onchangecommitted from click thank you very much for contributing to material ui by creating an issue ❤️ to avoid duplicate issues we ask you to check off the following list the issue is present in the latest release i have searched the of this repository and believe that this is not a duplicate current behavior 😯 when testing the slider functionality using rtl and jest no significant dom update happens when clicking on labels or doing mousedown or other events expected behavior 🤔 dom should update with the selected mark as slider value steps to reproduce 🕹 in the codesadbox example i tried to test this component of a slider value change on label slider mark click but i cant seem to get the onchangecommitted to trigger when testing the slider component dose not indicate any change provide a link to a live example you can use codesandbox io and an unambiguous set of steps to reproduce this bug include code to reproduce if relevant which it most likely is you should use the official codesandbox template as a starting point if you have an issue concerning typescript please start from this typescript playground issues without some form of live example have a longer response time context 🔦 i am trying to test onchangecommitted in the slider component but ive tried up and down to test this event on the slider and its getting really frustrating as there dosent seem to be many threads or so posts about so i might just be missing something trivial ive added a sample on codesandbox hope some one can help thanks what are you trying to accomplish how has this issue affected you providing context helps us come up with a solution that is most useful in the real world your environment 🌎 include as many relevant details about the environment with which you experienced the bug if you encounter issues with typescript please include version and tsconfig tech version material ui react browser chrome typescript y but not in example etc
1
49,760
6,039,394,262
IssuesEvent
2017-06-10 02:36:28
ClassicWoW/Nefarian_1.12.1_Bugtracker
https://api.github.com/repos/ClassicWoW/Nefarian_1.12.1_Bugtracker
closed
[Naxxramas] Outer Ring Mobs
Auf Testserver Datenbank Kreaturen Dev Behoben
Welches Verhalten wird beobachtet? es fehlen dort 1 Trashpack direkt wenn man von Faerlina kommt Wie sollte es sich verhalten? es sollten 2 Ghulpacks mit jeweils 4 Ghulen sein Schritte zur Reproduzierung Outer ring aus der Richtung von Faerlina betreten Zusätzliche Informationen (Screenshots, Videos, Klasse, Rasse, Level, etc.) Giga liveraid von BC in Naxxramas man sieht 2 Ghulpacks https://youtu.be/QtyEoJ6_nZI?t=2617 Alle Kreaturen, Items, Objekte, Quests, Zauber etc. müssen aus unserer Datenbank verlinkt sein. https://datenbank.classic-wow.org/?npc=16447 nochmal dadrauf hingewiesen, besitzt wieder falsche Fähigkeiten als auf dem Testserver
1.0
[Naxxramas] Outer Ring Mobs - Welches Verhalten wird beobachtet? es fehlen dort 1 Trashpack direkt wenn man von Faerlina kommt Wie sollte es sich verhalten? es sollten 2 Ghulpacks mit jeweils 4 Ghulen sein Schritte zur Reproduzierung Outer ring aus der Richtung von Faerlina betreten Zusätzliche Informationen (Screenshots, Videos, Klasse, Rasse, Level, etc.) Giga liveraid von BC in Naxxramas man sieht 2 Ghulpacks https://youtu.be/QtyEoJ6_nZI?t=2617 Alle Kreaturen, Items, Objekte, Quests, Zauber etc. müssen aus unserer Datenbank verlinkt sein. https://datenbank.classic-wow.org/?npc=16447 nochmal dadrauf hingewiesen, besitzt wieder falsche Fähigkeiten als auf dem Testserver
test
outer ring mobs welches verhalten wird beobachtet es fehlen dort trashpack direkt wenn man von faerlina kommt wie sollte es sich verhalten es sollten ghulpacks mit jeweils ghulen sein schritte zur reproduzierung outer ring aus der richtung von faerlina betreten zusätzliche informationen screenshots videos klasse rasse level etc giga liveraid von bc in naxxramas man sieht ghulpacks alle kreaturen items objekte quests zauber etc müssen aus unserer datenbank verlinkt sein nochmal dadrauf hingewiesen besitzt wieder falsche fähigkeiten als auf dem testserver
1
195,650
14,742,371,302
IssuesEvent
2021-01-07 12:10:53
pingcap/tidb
https://api.github.com/repos/pingcap/tidb
closed
Case failure[key]:CREATE TABLE t1 (\nname varchar(50) DEFAULT '' NOT NULL,\nauthor varchar(50) DEFAULT '' NOT NU...
component/test
Latest build: <a href="https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_integration_common_test/28045/display/redirect">https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_integration_common_test/28045/display/redirect</a>
1.0
Case failure[key]:CREATE TABLE t1 (\nname varchar(50) DEFAULT '' NOT NULL,\nauthor varchar(50) DEFAULT '' NOT NU... - Latest build: <a href="https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_integration_common_test/28045/display/redirect">https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_integration_common_test/28045/display/redirect</a>
test
case failure create table nname varchar default not null nauthor varchar default not nu latest build a href
1
73,015
7,320,336,937
IssuesEvent
2018-03-02 06:39:08
EyeSeeTea/malariapp
https://api.github.com/repos/EyeSeeTea/malariapp
closed
Change score classes
HNQIS complexity - med (1-5hr) priority - high testing type - feature
- Class A - green: >= 90% - Class B - yellow: 80% >= and <= 89% - Class C - red: <80%
1.0
Change score classes - - Class A - green: >= 90% - Class B - yellow: 80% >= and <= 89% - Class C - red: <80%
test
change score classes class a green class b yellow and class c red
1
101,447
8,788,090,232
IssuesEvent
2018-12-20 20:55:17
ONRR/doi-extractives-data
https://api.github.com/repos/ONRR/doi-extractives-data
closed
Write part two of our blog post series on homepage transition
blog workflow:testing
Part two will cover our transition to using GatsbyJS for new feature development, specifically our new homepage.
1.0
Write part two of our blog post series on homepage transition - Part two will cover our transition to using GatsbyJS for new feature development, specifically our new homepage.
test
write part two of our blog post series on homepage transition part two will cover our transition to using gatsbyjs for new feature development specifically our new homepage
1
331,156
28,537,327,852
IssuesEvent
2023-04-20 01:08:26
opera-adt/dolphin
https://api.github.com/repos/opera-adt/dolphin
closed
integration testing setup
enhancement testing
Setup a nightly or weekly integration test that runs through the primary workflow on some golden dataset. The related parts to this are 1. Create a golden dataset: #15 2. Set up some way to store this data and pull it for the test (possibly using Zenodo or https://github.com/fatiando/pooch, depending on how big it is) 3. Figure out how to set up the CI system to have a longer-running test, and to trigger it on a schedule rather than on-push/on-PR
1.0
integration testing setup - Setup a nightly or weekly integration test that runs through the primary workflow on some golden dataset. The related parts to this are 1. Create a golden dataset: #15 2. Set up some way to store this data and pull it for the test (possibly using Zenodo or https://github.com/fatiando/pooch, depending on how big it is) 3. Figure out how to set up the CI system to have a longer-running test, and to trigger it on a schedule rather than on-push/on-PR
test
integration testing setup setup a nightly or weekly integration test that runs through the primary workflow on some golden dataset the related parts to this are create a golden dataset set up some way to store this data and pull it for the test possibly using zenodo or depending on how big it is figure out how to set up the ci system to have a longer running test and to trigger it on a schedule rather than on push on pr
1
530,667
15,435,540,755
IssuesEvent
2021-03-07 09:19:51
woocommerce/woocommerce-gateway-stripe
https://api.github.com/repos/woocommerce/woocommerce-gateway-stripe
closed
!BROKEN! on AWS Cloudfront.
component: webhooks priority: high type: enhancement type: question
Update by @dwainm: Add a filter is_valid_request return value . Pass along the function parameters. https://github.com/woocommerce/woocommerce-gateway-stripe/blob/70640ab3a2dd3dfc4a700dee3dc2393fcaca95a7/includes/class-wc-stripe-webhook-handler.php#L93-L95 ---- **Original Issue** Plugin has conflict with Amazon Cloudfront CDN whereby the correct User Agent is forcefully rewritten by the CDN. On all requests, the User Agent is modified to "Amazon Cloudfront". This causes all webhook requests to return 400 even when correctly setup by failing on line 93 of /includes/class-wc-stripe-webhook-handler.php. If I comment out lines 93-95 I get expected behaviour. The Cloudfront specification can be found here: https://docs.aws.amazon.com/AmazonCloudFront/latest/DeveloperGuide/RequestAndResponseBehaviorCustomOrigin.html Commenting out the three lines is sufficient but it’s not a long term fix. Please can you introduce an exception or a way of being able to disable this specific security measure. Wordpress.org support ticket: https://wordpress.org/support/topic/cloudfront-user-agent/#post-12826607 **Environment (please complete the following information):** - WordPress Version - all - WooCommerce Version - 4.0.1 - Stripe Plugin Version - 4.3.3 - Browser [e.g. chrome, safari] and Version - all browsers
1.0
!BROKEN! on AWS Cloudfront. - Update by @dwainm: Add a filter is_valid_request return value . Pass along the function parameters. https://github.com/woocommerce/woocommerce-gateway-stripe/blob/70640ab3a2dd3dfc4a700dee3dc2393fcaca95a7/includes/class-wc-stripe-webhook-handler.php#L93-L95 ---- **Original Issue** Plugin has conflict with Amazon Cloudfront CDN whereby the correct User Agent is forcefully rewritten by the CDN. On all requests, the User Agent is modified to "Amazon Cloudfront". This causes all webhook requests to return 400 even when correctly setup by failing on line 93 of /includes/class-wc-stripe-webhook-handler.php. If I comment out lines 93-95 I get expected behaviour. The Cloudfront specification can be found here: https://docs.aws.amazon.com/AmazonCloudFront/latest/DeveloperGuide/RequestAndResponseBehaviorCustomOrigin.html Commenting out the three lines is sufficient but it’s not a long term fix. Please can you introduce an exception or a way of being able to disable this specific security measure. Wordpress.org support ticket: https://wordpress.org/support/topic/cloudfront-user-agent/#post-12826607 **Environment (please complete the following information):** - WordPress Version - all - WooCommerce Version - 4.0.1 - Stripe Plugin Version - 4.3.3 - Browser [e.g. chrome, safari] and Version - all browsers
non_test
broken on aws cloudfront update by dwainm add a filter is valid request return value pass along the function parameters original issue plugin has conflict with amazon cloudfront cdn whereby the correct user agent is forcefully rewritten by the cdn on all requests the user agent is modified to amazon cloudfront this causes all webhook requests to return even when correctly setup by failing on line of includes class wc stripe webhook handler php if i comment out lines i get expected behaviour the cloudfront specification can be found here commenting out the three lines is sufficient but it’s not a long term fix please can you introduce an exception or a way of being able to disable this specific security measure wordpress org support ticket environment please complete the following information wordpress version all woocommerce version stripe plugin version browser and version all browsers
0
214,303
16,580,643,019
IssuesEvent
2021-05-31 11:18:48
gardener/gardener
https://api.github.com/repos/gardener/gardener
closed
[Failing test] should add worker pool with containerd
area/testing kind/bug
/area testing /kind bug **What happened**: Currently the `[Fail] Shoot container runtime testing [It] [DEFAULT] [SERIAL] [SHOOT] should add worker pool with containerd ` test fails with: ``` FAIL FAIL github.com/gardener/gardener/test/suites/shoot 2295.248s FAIL --- FAIL: TestGardenerSuite (2295.18s) FAIL! -- 7 Passed | 1 Failed | 0 Pending | 15 Skipped Ran 8 of 23 Specs in 2295.167 seconds /src/test/integration/shoots/operations/containerruntime.go:125 [Fail] Shoot container runtime testing [It] [DEFAULT] [SERIAL] [SHOOT] should add worker pool with containerd Summarizing 1 Failure: ------------------------------ /src/test/integration/shoots/operations/containerruntime.go:125 <string>: found\n to equal <string>: Not found\n Expected /src/test/framework/gingko_utils.go:26 [DEFAULT] [SERIAL] [SHOOT] should add worker pool with containerd [It] /src/test/integration/shoots/operations/containerruntime.go:30 Shoot container runtime testing ``` <img width="794" alt="Screenshot 2021-05-31 at 11 53 56" src="https://user-images.githubusercontent.com/9372594/120168155-3c782b00-c207-11eb-894b-53f2846918d6.png"> **Environment**: - Gardener version: v1.24.0-dev-a12d3556f62c8a56e7364d1112b26a1d727eb604 - Kubernetes version (use `kubectl version`): v1.20.6 - Cloud provider or hardware configuration: - Others:
1.0
[Failing test] should add worker pool with containerd - /area testing /kind bug **What happened**: Currently the `[Fail] Shoot container runtime testing [It] [DEFAULT] [SERIAL] [SHOOT] should add worker pool with containerd ` test fails with: ``` FAIL FAIL github.com/gardener/gardener/test/suites/shoot 2295.248s FAIL --- FAIL: TestGardenerSuite (2295.18s) FAIL! -- 7 Passed | 1 Failed | 0 Pending | 15 Skipped Ran 8 of 23 Specs in 2295.167 seconds /src/test/integration/shoots/operations/containerruntime.go:125 [Fail] Shoot container runtime testing [It] [DEFAULT] [SERIAL] [SHOOT] should add worker pool with containerd Summarizing 1 Failure: ------------------------------ /src/test/integration/shoots/operations/containerruntime.go:125 <string>: found\n to equal <string>: Not found\n Expected /src/test/framework/gingko_utils.go:26 [DEFAULT] [SERIAL] [SHOOT] should add worker pool with containerd [It] /src/test/integration/shoots/operations/containerruntime.go:30 Shoot container runtime testing ``` <img width="794" alt="Screenshot 2021-05-31 at 11 53 56" src="https://user-images.githubusercontent.com/9372594/120168155-3c782b00-c207-11eb-894b-53f2846918d6.png"> **Environment**: - Gardener version: v1.24.0-dev-a12d3556f62c8a56e7364d1112b26a1d727eb604 - Kubernetes version (use `kubectl version`): v1.20.6 - Cloud provider or hardware configuration: - Others:
test
should add worker pool with containerd area testing kind bug what happened currently the shoot container runtime testing should add worker pool with containerd test fails with fail fail github com gardener gardener test suites shoot fail fail testgardenersuite fail passed failed pending skipped ran of specs in seconds src test integration shoots operations containerruntime go shoot container runtime testing should add worker pool with containerd summarizing failure src test integration shoots operations containerruntime go found n to equal not found n expected src test framework gingko utils go should add worker pool with containerd src test integration shoots operations containerruntime go shoot container runtime testing img width alt screenshot at src environment gardener version dev kubernetes version use kubectl version cloud provider or hardware configuration others
1
664,495
22,272,717,267
IssuesEvent
2022-06-10 13:48:43
opencrvs/opencrvs-core
https://api.github.com/repos/opencrvs/opencrvs-core
closed
In performance, no data is showing when location is an office
👹Bug Priority: high
**Bug description:** In performance, no data is showing when location is an office. And for office Completeness bloc should be hidden **Steps to reproduce:** 1. Login as any user other than field agent 2. Go to Performance 3. Select location as office and observe **Actual result:** 1) all values are 0% 2) Completeness block is showing **Expected result:** 1) should show actual values. If there is only 1 office in a district then all the values should be same as district 2) should hide completeness block **Screenshot:** ![office.png](https://images.zenhubusercontent.com/61920d086b30792f101fb9a3/267b3057-0435-4b05-b9cf-8da9315c0575) **Tested on:** https://login.farajaland-qa.opencrvs.org/
1.0
In performance, no data is showing when location is an office - **Bug description:** In performance, no data is showing when location is an office. And for office Completeness bloc should be hidden **Steps to reproduce:** 1. Login as any user other than field agent 2. Go to Performance 3. Select location as office and observe **Actual result:** 1) all values are 0% 2) Completeness block is showing **Expected result:** 1) should show actual values. If there is only 1 office in a district then all the values should be same as district 2) should hide completeness block **Screenshot:** ![office.png](https://images.zenhubusercontent.com/61920d086b30792f101fb9a3/267b3057-0435-4b05-b9cf-8da9315c0575) **Tested on:** https://login.farajaland-qa.opencrvs.org/
non_test
in performance no data is showing when location is an office bug description in performance no data is showing when location is an office and for office completeness bloc should be hidden steps to reproduce login as any user other than field agent go to performance select location as office and observe actual result all values are completeness block is showing expected result should show actual values if there is only office in a district then all the values should be same as district should hide completeness block screenshot tested on
0
50,417
10,500,534,066
IssuesEvent
2019-09-26 10:43:35
atomist/atomist-internal-sdm
https://api.github.com/repos/atomist/atomist-internal-sdm
closed
Code Inspection: npm audit on atomist/t29e48p34/npm-project-deps/master
bug code-inspection
### handlebars:<4.3.0 - _(error)_ [Prototype Pollution](https://npmjs.com/advisories/1164) _Upgrade to version 4.3.0 or later._ - `handlebars:4.2.0`: - `@atomist/sdm-pack-aspect>express-handlebars>handlebars` [atomist:code-inspection:atomist/t29e48p34/npm-project-deps/master=@atomist/atomist-sdm]
1.0
Code Inspection: npm audit on atomist/t29e48p34/npm-project-deps/master - ### handlebars:<4.3.0 - _(error)_ [Prototype Pollution](https://npmjs.com/advisories/1164) _Upgrade to version 4.3.0 or later._ - `handlebars:4.2.0`: - `@atomist/sdm-pack-aspect>express-handlebars>handlebars` [atomist:code-inspection:atomist/t29e48p34/npm-project-deps/master=@atomist/atomist-sdm]
non_test
code inspection npm audit on atomist npm project deps master handlebars error upgrade to version or later handlebars atomist sdm pack aspect express handlebars handlebars
0
343,665
30,682,389,137
IssuesEvent
2023-07-26 10:00:22
iho-ohi/S-101_Portrayal-Catalogue
https://api.github.com/repos/iho-ohi/S-101_Portrayal-Catalogue
closed
New symbol for Berth features with categoryOfCargo = 7 (Dangerous or Hazardous) - ENCWG7-5.3_2022 [PSWG #112]
enhancement test PC 1.1.0
SPEC at: https://github.com/S-101-Portrayal-subWG/Working-Documents/issues/112#issuecomment-1381193928
1.0
New symbol for Berth features with categoryOfCargo = 7 (Dangerous or Hazardous) - ENCWG7-5.3_2022 [PSWG #112] - SPEC at: https://github.com/S-101-Portrayal-subWG/Working-Documents/issues/112#issuecomment-1381193928
test
new symbol for berth features with categoryofcargo dangerous or hazardous spec at
1
75,027
7,458,753,560
IssuesEvent
2018-03-30 12:08:42
SunwellTracker/issues
https://api.github.com/repos/SunwellTracker/issues
closed
[Warrior]Deep Wounds
Works locally | Requires testing
Decription: The damage from warriors talent "Deep Wounds" isnt affected by resilience or any other dmg reduce effect. And the Base dmg is to High. 3,4k gs Warriors can stack that talent to 1300dmg per tick. On retail with Shadowmourne warrior i got that wounds on 1400dmg per tick in pve. Quote: PATCH 3.3.3: The damage from this talent is now affected by resilience as intended. helping source: http://wowwiki.wikia.com/wiki/Deep_Wounds How it works: Deep wounds arent effected by resilience or for expample painsub, dispersion, shadowform, shieldwall and all other things. So in Dispersion (90% dmg reduce) it does every 1 sec 600 dmg+. How it should work: That talent should do less dmg + should be affected by resilience and other dmg reduce effects. Source (you should point out proofs of your report, please give us some source): http://wowwiki.wikia.com/wiki/Deep_Wounds There you can see the fact that it should be affected by resilience. Cant find a source about that it should do less dmg but there is something wrong when warriors do the most dmg with deepwounds in a whole bossfight and if the wounds do the same dmg as a 6.8k gs warrior with shadowmourne. Pls analyse that talent with dev team :)
1.0
[Warrior]Deep Wounds - Decription: The damage from warriors talent "Deep Wounds" isnt affected by resilience or any other dmg reduce effect. And the Base dmg is to High. 3,4k gs Warriors can stack that talent to 1300dmg per tick. On retail with Shadowmourne warrior i got that wounds on 1400dmg per tick in pve. Quote: PATCH 3.3.3: The damage from this talent is now affected by resilience as intended. helping source: http://wowwiki.wikia.com/wiki/Deep_Wounds How it works: Deep wounds arent effected by resilience or for expample painsub, dispersion, shadowform, shieldwall and all other things. So in Dispersion (90% dmg reduce) it does every 1 sec 600 dmg+. How it should work: That talent should do less dmg + should be affected by resilience and other dmg reduce effects. Source (you should point out proofs of your report, please give us some source): http://wowwiki.wikia.com/wiki/Deep_Wounds There you can see the fact that it should be affected by resilience. Cant find a source about that it should do less dmg but there is something wrong when warriors do the most dmg with deepwounds in a whole bossfight and if the wounds do the same dmg as a 6.8k gs warrior with shadowmourne. Pls analyse that talent with dev team :)
test
deep wounds decription the damage from warriors talent deep wounds isnt affected by resilience or any other dmg reduce effect and the base dmg is to high gs warriors can stack that talent to per tick on retail with shadowmourne warrior i got that wounds on per tick in pve quote patch the damage from this talent is now affected by resilience as intended helping source how it works deep wounds arent effected by resilience or for expample painsub dispersion shadowform shieldwall and all other things so in dispersion dmg reduce it does every sec dmg how it should work that talent should do less dmg should be affected by resilience and other dmg reduce effects source you should point out proofs of your report please give us some source there you can see the fact that it should be affected by resilience cant find a source about that it should do less dmg but there is something wrong when warriors do the most dmg with deepwounds in a whole bossfight and if the wounds do the same dmg as a gs warrior with shadowmourne pls analyse that talent with dev team
1
122,505
10,225,332,902
IssuesEvent
2019-08-16 14:55:06
phetsims/energy-forms-and-changes
https://api.github.com/repos/phetsims/energy-forms-and-changes
opened
CT cannot read property name of undefined
type:automated-testing
``` energy-forms-and-changes : accessibility-fuzz : built : run Query: fuzz&a11y&memoryLimit=1000 Uncaught TypeError: Cannot read property 'name' of undefined TypeError: Cannot read property 'name' of undefined at Object.getHeatTransferFactor (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:410992) at A.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:1431513) at e.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:1561001) at e.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:1560656) at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:3236413 at n.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:425381) at t.stepSimulation (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:3247843) at t.stepOneFrame (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:3247650) at t.runAnimationLoop (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:3247511) id: Bayes Chrome Approximately 8/16/2019, 3:36:51 AM energy-forms-and-changes : accessibility-fuzz : require.js : run Query: brand=phet&ea&fuzz&a11y&memoryLimit=1000 Uncaught Error: Assertion failed: invalid category1: undefined Error: Assertion failed: invalid category1: undefined at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/assert/js/assert.js:22:13) at Object.getHeatTransferFactor (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/common/model/HeatTransferConstants.js?:77:17) at BeakerHeater.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/BeakerHeater.js?:169:54) at SystemsModel.stepModel (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/SystemsModel.js?:188:12) at SystemsModel.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/SystemsModel.js?:169:14) at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:217:22 at Action.execute (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/Action.js?:244:20) at Sim.stepSimulation (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:933:33) at Sim.stepOneFrame (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:914:14) at Sim.runAnimationLoop (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:897:14) id: Bayes Chrome Approximately 8/16/2019, 3:36:51 AM energy-forms-and-changes : accessibility-fuzzBoard : built : run Query: fuzzBoard&a11y&memoryLimit=1000 Uncaught TypeError: Cannot read property 'name' of undefined TypeError: Cannot read property 'name' of undefined at Object.getHeatTransferFactor (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:410992) at A.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:1431513) at e.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:1561001) at e.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:1560656) at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:3236413 at n.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:425381) at t.stepSimulation (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:3247843) at t.stepOneFrame (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:3247650) at t.runAnimationLoop (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:3247511) id: Bayes Chrome Approximately 8/16/2019, 3:36:51 AM energy-forms-and-changes : fuzz : require.js : run Query: brand=phet&ea&fuzz&memoryLimit=1000 Uncaught Error: Assertion failed: value failed isValidValue: -3.1164 Error: Assertion failed: value failed isValidValue: -3.1164 at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/assert/js/assert.js:22:13) at Object.isValueValid (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/ValidatorDef.js?:267:41) at validate (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/validate.js?:28:20) at NumberProperty.assertNumberPropertyValidateValue (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/NumberProperty.js?:67:26) at NumberProperty.set (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/NumberProperty.js?:89:54) at EFACIntroModel.fallToSurface (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/intro/model/EFACIntroModel.js?:743:47) at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/intro/model/EFACIntroModel.js?:338:16 at Array.forEach (<anonymous>) at EFACIntroModel.stepModel (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/intro/model/EFACIntroModel.js?:332:30) at EFACIntroModel.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/intro/model/EFACIntroModel.js?:314:14) id: Bayes Chrome Approximately 8/16/2019, 3:36:51 AM energy-forms-and-changes : xss-fuzz : run Query: brand=phet&ea&fuzz&stringTest=xss&memoryLimit=1000 Uncaught Error: Assertion failed: invalid category1: undefined Error: Assertion failed: invalid category1: undefined at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/assert/js/assert.js:22:13) at Object.getHeatTransferFactor (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/common/model/HeatTransferConstants.js?:77:17) at BeakerHeater.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/BeakerHeater.js?:169:54) at SystemsModel.stepModel (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/SystemsModel.js?:188:12) at SystemsModel.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/SystemsModel.js?:169:14) at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:217:22 at Action.execute (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/Action.js?:244:20) at Sim.stepSimulation (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:933:33) at Sim.stepOneFrame (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:914:14) at Sim.runAnimationLoop (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:897:14) id: Bayes Chrome Approximately 8/16/2019, 3:36:51 AM ```
1.0
CT cannot read property name of undefined - ``` energy-forms-and-changes : accessibility-fuzz : built : run Query: fuzz&a11y&memoryLimit=1000 Uncaught TypeError: Cannot read property 'name' of undefined TypeError: Cannot read property 'name' of undefined at Object.getHeatTransferFactor (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:410992) at A.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:1431513) at e.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:1561001) at e.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:1560656) at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:3236413 at n.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:425381) at t.stepSimulation (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:3247843) at t.stepOneFrame (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:3247650) at t.runAnimationLoop (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzz&a11y&memoryLimit=1000:1802:3247511) id: Bayes Chrome Approximately 8/16/2019, 3:36:51 AM energy-forms-and-changes : accessibility-fuzz : require.js : run Query: brand=phet&ea&fuzz&a11y&memoryLimit=1000 Uncaught Error: Assertion failed: invalid category1: undefined Error: Assertion failed: invalid category1: undefined at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/assert/js/assert.js:22:13) at Object.getHeatTransferFactor (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/common/model/HeatTransferConstants.js?:77:17) at BeakerHeater.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/BeakerHeater.js?:169:54) at SystemsModel.stepModel (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/SystemsModel.js?:188:12) at SystemsModel.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/SystemsModel.js?:169:14) at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:217:22 at Action.execute (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/Action.js?:244:20) at Sim.stepSimulation (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:933:33) at Sim.stepOneFrame (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:914:14) at Sim.runAnimationLoop (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:897:14) id: Bayes Chrome Approximately 8/16/2019, 3:36:51 AM energy-forms-and-changes : accessibility-fuzzBoard : built : run Query: fuzzBoard&a11y&memoryLimit=1000 Uncaught TypeError: Cannot read property 'name' of undefined TypeError: Cannot read property 'name' of undefined at Object.getHeatTransferFactor (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:410992) at A.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:1431513) at e.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:1561001) at e.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:1560656) at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:3236413 at n.value (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:425381) at t.stepSimulation (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:3247843) at t.stepOneFrame (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:3247650) at t.runAnimationLoop (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/build/phet/energy-forms-and-changes_en_phet.html?postMessageOnLoad&postMessageOnError&postMessageOnBeforeUnload&fuzzBoard&a11y&memoryLimit=1000:1802:3247511) id: Bayes Chrome Approximately 8/16/2019, 3:36:51 AM energy-forms-and-changes : fuzz : require.js : run Query: brand=phet&ea&fuzz&memoryLimit=1000 Uncaught Error: Assertion failed: value failed isValidValue: -3.1164 Error: Assertion failed: value failed isValidValue: -3.1164 at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/assert/js/assert.js:22:13) at Object.isValueValid (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/ValidatorDef.js?:267:41) at validate (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/validate.js?:28:20) at NumberProperty.assertNumberPropertyValidateValue (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/NumberProperty.js?:67:26) at NumberProperty.set (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/NumberProperty.js?:89:54) at EFACIntroModel.fallToSurface (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/intro/model/EFACIntroModel.js?:743:47) at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/intro/model/EFACIntroModel.js?:338:16 at Array.forEach (<anonymous>) at EFACIntroModel.stepModel (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/intro/model/EFACIntroModel.js?:332:30) at EFACIntroModel.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/intro/model/EFACIntroModel.js?:314:14) id: Bayes Chrome Approximately 8/16/2019, 3:36:51 AM energy-forms-and-changes : xss-fuzz : run Query: brand=phet&ea&fuzz&stringTest=xss&memoryLimit=1000 Uncaught Error: Assertion failed: invalid category1: undefined Error: Assertion failed: invalid category1: undefined at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/assert/js/assert.js:22:13) at Object.getHeatTransferFactor (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/common/model/HeatTransferConstants.js?:77:17) at BeakerHeater.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/BeakerHeater.js?:169:54) at SystemsModel.stepModel (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/SystemsModel.js?:188:12) at SystemsModel.step (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/energy-forms-and-changes/js/systems/model/SystemsModel.js?:169:14) at https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:217:22 at Action.execute (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/axon/js/Action.js?:244:20) at Sim.stepSimulation (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:933:33) at Sim.stepOneFrame (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:914:14) at Sim.runAnimationLoop (https://bayes.colorado.edu/continuous-testing/snapshot-1565948211353/joist/js/Sim.js?:897:14) id: Bayes Chrome Approximately 8/16/2019, 3:36:51 AM ```
test
ct cannot read property name of undefined energy forms and changes accessibility fuzz built run query fuzz memorylimit uncaught typeerror cannot read property name of undefined typeerror cannot read property name of undefined at object getheattransferfactor at a value at e value at e value at at n value at t stepsimulation at t steponeframe at t runanimationloop id bayes chrome approximately am energy forms and changes accessibility fuzz require js run query brand phet ea fuzz memorylimit uncaught error assertion failed invalid undefined error assertion failed invalid undefined at window assertions assertfunction at object getheattransferfactor at beakerheater step at systemsmodel stepmodel at systemsmodel step at at action execute at sim stepsimulation at sim steponeframe at sim runanimationloop id bayes chrome approximately am energy forms and changes accessibility fuzzboard built run query fuzzboard memorylimit uncaught typeerror cannot read property name of undefined typeerror cannot read property name of undefined at object getheattransferfactor at a value at e value at e value at at n value at t stepsimulation at t steponeframe at t runanimationloop id bayes chrome approximately am energy forms and changes fuzz require js run query brand phet ea fuzz memorylimit uncaught error assertion failed value failed isvalidvalue error assertion failed value failed isvalidvalue at window assertions assertfunction at object isvaluevalid at validate at numberproperty assertnumberpropertyvalidatevalue at numberproperty set at efacintromodel falltosurface at at array foreach at efacintromodel stepmodel at efacintromodel step id bayes chrome approximately am energy forms and changes xss fuzz run query brand phet ea fuzz stringtest xss memorylimit uncaught error assertion failed invalid undefined error assertion failed invalid undefined at window assertions assertfunction at object getheattransferfactor at beakerheater step at systemsmodel stepmodel at systemsmodel step at at action execute at sim stepsimulation at sim steponeframe at sim runanimationloop id bayes chrome approximately am
1
194,831
14,689,666,811
IssuesEvent
2021-01-02 11:09:16
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
kalyabin/vcsview: git_test.go; 53 LoC
fresh medium test
Found a possible issue in [kalyabin/vcsview](https://www.github.com/kalyabin/vcsview) at [git_test.go](https://github.com/kalyabin/vcsview/blob/cfe5e010e0206d6b98caad942e0322b2150ef7cd/git_test.go#L414-L466) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable key used in defer or goroutine at line 437 [Click here to see the code in its original context.](https://github.com/kalyabin/vcsview/blob/cfe5e010e0206d6b98caad942e0322b2150ef7cd/git_test.go#L414-L466) <details> <summary>Click here to show the 53 line(s) of Go which triggered the analyzer.</summary> ```go for key, testCase := range cases { var ( gotError error gotCommit int ) result := make(chan Commit, 1) e := g.ReadHistory(testCase.repoPath, testCase.path, testCase.branch, testCase.offset, testCase.limit, result) wg := sync.WaitGroup{} wg.Add(1) go func() { defer wg.Done() loop: for { select { case <-e.ctx.Done(): close(result) break loop case commit := <- result: gotCommit++ if commit.Id() == "" { t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty identifier", key, testCase) } if len(commit.Parents()) == 0 { t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty parents", key, testCase) } if commit.Message() == "" { t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty message", key, testCase) } if commit.Author().String() == "" { t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty author", key, testCase) } if commit.Date().Unix() < 0 { t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty date", key, testCase) } } } }() gotError = e.Run() wg.Wait() if gotError != nil { t.Fatalf("[%d] Git.ReadHistory(%v) has error: %v, want no errors", key, testCase, gotError) } if gotCommit != testCase.limit { t.Fatalf("[%d] Git.ReadHistory(%v) got %v commits, want: %v", key, testCase, gotCommit, testCase.limit) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: cfe5e010e0206d6b98caad942e0322b2150ef7cd
1.0
kalyabin/vcsview: git_test.go; 53 LoC - Found a possible issue in [kalyabin/vcsview](https://www.github.com/kalyabin/vcsview) at [git_test.go](https://github.com/kalyabin/vcsview/blob/cfe5e010e0206d6b98caad942e0322b2150ef7cd/git_test.go#L414-L466) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable key used in defer or goroutine at line 437 [Click here to see the code in its original context.](https://github.com/kalyabin/vcsview/blob/cfe5e010e0206d6b98caad942e0322b2150ef7cd/git_test.go#L414-L466) <details> <summary>Click here to show the 53 line(s) of Go which triggered the analyzer.</summary> ```go for key, testCase := range cases { var ( gotError error gotCommit int ) result := make(chan Commit, 1) e := g.ReadHistory(testCase.repoPath, testCase.path, testCase.branch, testCase.offset, testCase.limit, result) wg := sync.WaitGroup{} wg.Add(1) go func() { defer wg.Done() loop: for { select { case <-e.ctx.Done(): close(result) break loop case commit := <- result: gotCommit++ if commit.Id() == "" { t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty identifier", key, testCase) } if len(commit.Parents()) == 0 { t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty parents", key, testCase) } if commit.Message() == "" { t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty message", key, testCase) } if commit.Author().String() == "" { t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty author", key, testCase) } if commit.Date().Unix() < 0 { t.Fatalf("[%d] Git.ReadHistory(%v) commit has empty date", key, testCase) } } } }() gotError = e.Run() wg.Wait() if gotError != nil { t.Fatalf("[%d] Git.ReadHistory(%v) has error: %v, want no errors", key, testCase, gotError) } if gotCommit != testCase.limit { t.Fatalf("[%d] Git.ReadHistory(%v) got %v commits, want: %v", key, testCase, gotCommit, testCase.limit) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: cfe5e010e0206d6b98caad942e0322b2150ef7cd
test
kalyabin vcsview git test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable key used in defer or goroutine at line click here to show the line s of go which triggered the analyzer go for key testcase range cases var goterror error gotcommit int result make chan commit e g readhistory testcase repopath testcase path testcase branch testcase offset testcase limit result wg sync waitgroup wg add go func defer wg done loop for select case e ctx done close result break loop case commit result gotcommit if commit id t fatalf git readhistory v commit has empty identifier key testcase if len commit parents t fatalf git readhistory v commit has empty parents key testcase if commit message t fatalf git readhistory v commit has empty message key testcase if commit author string t fatalf git readhistory v commit has empty author key testcase if commit date unix t fatalf git readhistory v commit has empty date key testcase goterror e run wg wait if goterror nil t fatalf git readhistory v has error v want no errors key testcase goterror if gotcommit testcase limit t fatalf git readhistory v got v commits want v key testcase gotcommit testcase limit leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
288,055
24,882,269,637
IssuesEvent
2022-10-28 03:01:04
MPMG-DCC-UFMG/F01
https://api.github.com/repos/MPMG-DCC-UFMG/F01
closed
Teste de generalizacao para a tag Orçamento - Legislação - Jequitinhonha
generalization test development template - Memory (66) tag - Orçamento subtag - Legislação
DoD: Realizar o teste de Generalização do validador da tag Orçamento - Legislação para o Município de Jequitinhonha.
1.0
Teste de generalizacao para a tag Orçamento - Legislação - Jequitinhonha - DoD: Realizar o teste de Generalização do validador da tag Orçamento - Legislação para o Município de Jequitinhonha.
test
teste de generalizacao para a tag orçamento legislação jequitinhonha dod realizar o teste de generalização do validador da tag orçamento legislação para o município de jequitinhonha
1
131,939
18,262,075,889
IssuesEvent
2021-10-04 00:55:31
ghc-dev/Steve-Chambers-MD
https://api.github.com/repos/ghc-dev/Steve-Chambers-MD
opened
CVE-2020-13956 (Medium) detected in httpclient-4.5.10.jar
security vulnerability
## CVE-2020-13956 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>httpclient-4.5.10.jar</b></p></summary> <p>Apache HttpComponents Client</p> <p>Path to dependency file: Steve-Chambers-MD/build.gradle</p> <p>Path to vulnerable library: hes/modules-2/files-2.1/org.apache.httpcomponents/httpclient/4.5.10/7ca2e4276f4ef95e4db725a8cd4a1d1e7585b9e5/httpclient-4.5.10.jar</p> <p> Dependency Hierarchy: - :x: **httpclient-4.5.10.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Steve-Chambers-MD/commit/a94641a73803572678190ca83e1504bf0ea37d1f">a94641a73803572678190ca83e1504bf0ea37d1f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution. <p>Publish Date: 2020-12-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956>CVE-2020-13956</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.httpcomponents","packageName":"httpclient","packageVersion":"4.5.10","packageFilePaths":["/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"org.apache.httpcomponents:httpclient:4.5.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-13956","vulnerabilityDetails":"Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-13956 (Medium) detected in httpclient-4.5.10.jar - ## CVE-2020-13956 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>httpclient-4.5.10.jar</b></p></summary> <p>Apache HttpComponents Client</p> <p>Path to dependency file: Steve-Chambers-MD/build.gradle</p> <p>Path to vulnerable library: hes/modules-2/files-2.1/org.apache.httpcomponents/httpclient/4.5.10/7ca2e4276f4ef95e4db725a8cd4a1d1e7585b9e5/httpclient-4.5.10.jar</p> <p> Dependency Hierarchy: - :x: **httpclient-4.5.10.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Steve-Chambers-MD/commit/a94641a73803572678190ca83e1504bf0ea37d1f">a94641a73803572678190ca83e1504bf0ea37d1f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution. <p>Publish Date: 2020-12-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956>CVE-2020-13956</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2020-13956</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.httpcomponents","packageName":"httpclient","packageVersion":"4.5.10","packageFilePaths":["/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"org.apache.httpcomponents:httpclient:4.5.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.httpcomponents:httpclient:4.5.13;org.apache.httpcomponents:httpclient-osgi:4.5.13;org.apache.httpcomponents.client5:httpclient5:5.0.3;org.apache.httpcomponents.client5:httpclient5-osgi:5.0.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-13956","vulnerabilityDetails":"Apache HttpClient versions prior to version 4.5.13 and 5.0.3 can misinterpret malformed authority component in request URIs passed to the library as java.net.URI object and pick the wrong target host for request execution.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13956","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_test
cve medium detected in httpclient jar cve medium severity vulnerability vulnerable library httpclient jar apache httpcomponents client path to dependency file steve chambers md build gradle path to vulnerable library hes modules files org apache httpcomponents httpclient httpclient jar dependency hierarchy x httpclient jar vulnerable library found in head commit a href found in base branch master vulnerability details apache httpclient versions prior to version and can misinterpret malformed authority component in request uris passed to the library as java net uri object and pick the wrong target host for request execution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache httpcomponents httpclient org apache httpcomponents httpclient osgi org apache httpcomponents org apache httpcomponents osgi isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org apache httpcomponents httpclient isminimumfixversionavailable true minimumfixversion org apache httpcomponents httpclient org apache httpcomponents httpclient osgi org apache httpcomponents org apache httpcomponents osgi basebranches vulnerabilityidentifier cve vulnerabilitydetails apache httpclient versions prior to version and can misinterpret malformed authority component in request uris passed to the library as java net uri object and pick the wrong target host for request execution vulnerabilityurl
0
311,451
26,792,744,537
IssuesEvent
2023-02-01 09:42:47
akademia-envelo-3/MELO-front
https://api.github.com/repos/akademia-envelo-3/MELO-front
closed
[MKT-05] Uszczegółowienie makiety admina odnośnie kategorii
frontend UI sp1 test-ok
### Dodatkowe informacje Należy stworzyć makietę szczegółową dla widoku kategorii dla admina. Makieta ta składać się będzie z widoku wszystkich kategorii, widoku edycji kategorii oraz widoku dodania nowej kategorii. Makieta ma przedstawiać zarówno wersję widoku dla desktop oraz mobile. ### Makiety [Makieta 01](https://www.figma.com/file/ybxNkntmTWg7yLBzHhBx4R/MELO?node-id=333%3A8921&t=wFubEuu6vVmKgmdq-0) [Makieta 02](https://www.figma.com/file/ybxNkntmTWg7yLBzHhBx4R/MELO?node-id=333%3A8316&t=wFubEuu6vVmKgmdq-0) [Makieta 03](https://www.figma.com/file/ybxNkntmTWg7yLBzHhBx4R/MELO?node-id=394%3A8841&t=wFubEuu6vVmKgmdq-0) ### Kryteria akceptacji 1. Została utworzona makieta szczegółowa dla widoku wszystkich kategorii 2. Została utworzona makieta szczegółowa dla dodania nowej kategorii 3. Została utworzona makieta szczegółowa dla widoku edycji kategorii
1.0
[MKT-05] Uszczegółowienie makiety admina odnośnie kategorii - ### Dodatkowe informacje Należy stworzyć makietę szczegółową dla widoku kategorii dla admina. Makieta ta składać się będzie z widoku wszystkich kategorii, widoku edycji kategorii oraz widoku dodania nowej kategorii. Makieta ma przedstawiać zarówno wersję widoku dla desktop oraz mobile. ### Makiety [Makieta 01](https://www.figma.com/file/ybxNkntmTWg7yLBzHhBx4R/MELO?node-id=333%3A8921&t=wFubEuu6vVmKgmdq-0) [Makieta 02](https://www.figma.com/file/ybxNkntmTWg7yLBzHhBx4R/MELO?node-id=333%3A8316&t=wFubEuu6vVmKgmdq-0) [Makieta 03](https://www.figma.com/file/ybxNkntmTWg7yLBzHhBx4R/MELO?node-id=394%3A8841&t=wFubEuu6vVmKgmdq-0) ### Kryteria akceptacji 1. Została utworzona makieta szczegółowa dla widoku wszystkich kategorii 2. Została utworzona makieta szczegółowa dla dodania nowej kategorii 3. Została utworzona makieta szczegółowa dla widoku edycji kategorii
test
uszczegółowienie makiety admina odnośnie kategorii dodatkowe informacje należy stworzyć makietę szczegółową dla widoku kategorii dla admina makieta ta składać się będzie z widoku wszystkich kategorii widoku edycji kategorii oraz widoku dodania nowej kategorii makieta ma przedstawiać zarówno wersję widoku dla desktop oraz mobile makiety kryteria akceptacji została utworzona makieta szczegółowa dla widoku wszystkich kategorii została utworzona makieta szczegółowa dla dodania nowej kategorii została utworzona makieta szczegółowa dla widoku edycji kategorii
1
129,308
10,569,527,896
IssuesEvent
2019-10-06 20:07:59
mgba-emu/mgba
https://api.github.com/repos/mgba-emu/mgba
closed
[GBA] Volume of audio channels changes frequently
category:needs retest platform:GBA severity:regression
Compared to mGBA 0.6.3, the development build's volume for GBA games seems to frequently change in an odd way, causing some of the audio channels to sound louder than usual in comparison for brief moments. This can be noticed in games like the Sonic Advance games. Also, the following is probably unrelated to this issue, but the fps counter in the development build indicates that the fps is hovering around 65.6 fps even though I have the target fps set to the native amount.
1.0
[GBA] Volume of audio channels changes frequently - Compared to mGBA 0.6.3, the development build's volume for GBA games seems to frequently change in an odd way, causing some of the audio channels to sound louder than usual in comparison for brief moments. This can be noticed in games like the Sonic Advance games. Also, the following is probably unrelated to this issue, but the fps counter in the development build indicates that the fps is hovering around 65.6 fps even though I have the target fps set to the native amount.
test
volume of audio channels changes frequently compared to mgba the development build s volume for gba games seems to frequently change in an odd way causing some of the audio channels to sound louder than usual in comparison for brief moments this can be noticed in games like the sonic advance games also the following is probably unrelated to this issue but the fps counter in the development build indicates that the fps is hovering around fps even though i have the target fps set to the native amount
1
5,903
2,798,041,191
IssuesEvent
2015-05-12 16:34:37
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
Test failure in CI build 2339
PTAL test-failure
The following test appears to have failed: [#2339](https://circleci.com/gh/cockroachdb/cockroach/2339): ``` level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="770" msg="node 100000001: group 5 raft ready" level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="775" msg="HardState updated: {Term:6 Vote:4294967297 Commit:13 XXX_unrecognized:[]}" level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="778" msg="New Entry[0]: 6/13 EntryNormal 00000000000000006b4ee14837a7cb41: raft_id:5 origin_node_id:4294967297 cmd:<internal_resolve_intent:<header:<timestamp:<wall_time:0 logical:933 > cmd_id:<wall_time:0 random:0 > key:\"\\0" level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="781" msg="Committed Entry[0]: 6/13 EntryNormal 00000000000000006b4ee14837a7cb41: raft_id:5 origin_node_id:4294967297 cmd:<internal_resolve_intent:<header:<timestamp:<wall_time:0 logical:933 > cmd_id:<wall_time:0 random:0 > key:\"\\0" ================== WARNING: DATA RACE Read by goroutine 79: github.com/cockroachdb/cockroach/storage.(*rangeStats).GetSize() /go/src/github.com/cockroachdb/cockroach/storage/stats.go:68 +0x96 github.com/cockroachdb/cockroach/storage.(*splitQueue).process() /go/src/github.com/cockroachdb/cockroach/storage/split_queue.go:103 +0x81d github.com/cockroachdb/cockroach/storage.(*baseQueue).processOne() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:261 +0x5ce github.com/cockroachdb/cockroach/storage.func·011() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:214 +0x1b7 github.com/cockroachdb/cockroach/util.func·001() /go/src/github.com/cockroachdb/cockroach/util/stopper.go:73 +0x7f Previous write by goroutine 87: github.com/cockroachdb/cockroach/proto.(*MVCCStats).Add() /go/src/github.com/cockroachdb/cockroach/proto/data.go:630 +0xa6 github.com/cockroachdb/cockroach/storage.(*rangeStats).MergeMVCCStats() /go/src/github.com/cockroachdb/cockroach/storage/stats.go:82 +0x25c github.com/cockroachdb/cockroach/storage.(*Range).applyRaftCommand() /go/src/github.com/cockroachdb/cockroach/storage/range.go:845 +0xca3 github.com/cockroachdb/cockroach/storage.(*Range).processRaftCommand() /go/src/github.com/cockroachdb/cockroach/storage/range.go:767 +0x3b5 github.com/cockroachdb/cockroach/storage.func·036() /go/src/github.com/cockroachdb/cockroach/storage/store.go:1339 +0xce8 -- /go/src/github.com/cockroachdb/cockroach/kv/split_test.go:137 +0x7d testing.tRunner() /usr/src/go/src/testing/testing.go:447 +0x133 ================== ================== WARNING: DATA RACE Read by goroutine 79: github.com/cockroachdb/cockroach/storage.(*rangeStats).GetSize() /go/src/github.com/cockroachdb/cockroach/storage/stats.go:68 +0xae github.com/cockroachdb/cockroach/storage.(*splitQueue).process() /go/src/github.com/cockroachdb/cockroach/storage/split_queue.go:103 +0x81d github.com/cockroachdb/cockroach/storage.(*baseQueue).processOne() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:261 +0x5ce github.com/cockroachdb/cockroach/storage.func·011() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:214 +0x1b7 github.com/cockroachdb/cockroach/util.func·001() /go/src/github.com/cockroachdb/cockroach/util/stopper.go:73 +0x7f Previous write by goroutine 87: github.com/cockroachdb/cockroach/proto.(*MVCCStats).Add() /go/src/github.com/cockroachdb/cockroach/proto/data.go:631 +0x107 github.com/cockroachdb/cockroach/storage.(*rangeStats).MergeMVCCStats() /go/src/github.com/cockroachdb/cockroach/storage/stats.go:82 +0x25c github.com/cockroachdb/cockroach/storage.(*Range).applyRaftCommand() /go/src/github.com/cockroachdb/cockroach/storage/range.go:845 +0xca3 github.com/cockroachdb/cockroach/storage.(*Range).processRaftCommand() /go/src/github.com/cockroachdb/cockroach/storage/range.go:767 +0x3b5 github.com/cockroachdb/cockroach/storage.func·036() /go/src/github.com/cockroachdb/cockroach/storage/store.go:1339 +0xce8 level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="770" msg="node 100000001: group 5 raft ready" level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="775" msg="HardState updated: {Term:6 Vote:4294967297 Commit:13 XXX_unrecognized:[]}" level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="778" msg="New Entry[0]: 6/13 EntryNormal 00000000000000006b4ee14837a7cb41: raft_id:5 origin_node_id:4294967297 cmd:<internal_resolve_intent:<header:<timestamp:<wall_time:0 logical:933 > cmd_id:<wall_time:0 random:0 > key:\"\\0" level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="781" msg="Committed Entry[0]: 6/13 EntryNormal 00000000000000006b4ee14837a7cb41: raft_id:5 origin_node_id:4294967297 cmd:<internal_resolve_intent:<header:<timestamp:<wall_time:0 logical:933 > cmd_id:<wall_time:0 random:0 > key:\"\\0" ================== WARNING: DATA RACE Read by goroutine 79: github.com/cockroachdb/cockroach/storage.(*rangeStats).GetSize() /go/src/github.com/cockroachdb/cockroach/storage/stats.go:68 +0x96 github.com/cockroachdb/cockroach/storage.(*splitQueue).process() /go/src/github.com/cockroachdb/cockroach/storage/split_queue.go:103 +0x81d github.com/cockroachdb/cockroach/storage.(*baseQueue).processOne() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:261 +0x5ce github.com/cockroachdb/cockroach/storage.func·011() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:214 +0x1b7 github.com/cockroachdb/cockroach/util.func·001() /go/src/github.com/cockroachdb/cockroach/util/stopper.go:73 +0x7f Previous write by goroutine 87: github.com/cockroachdb/cockroach/proto.(*MVCCStats).Add() /go/src/github.com/cockroachdb/cockroach/proto/data.go:630 +0xa6 github.com/cockroachdb/cockroach/storage.(*rangeStats).MergeMVCCStats() /go/src/github.com/cockroachdb/cockroach/storage/stats.go:82 +0x25c github.com/cockroachdb/cockroach/storage.(*Range).applyRaftCommand() /go/src/github.com/cockroachdb/cockroach/storage/range.go:845 +0xca3 github.com/cockroachdb/cockroach/storage.(*Range).processRaftCommand() /go/src/github.com/cockroachdb/cockroach/storage/range.go:767 +0x3b5 github.com/cockroachdb/cockroach/storage.func·036() /go/src/github.com/cockroachdb/cockroach/storage/store.go:1339 +0xce8 -- /go/src/github.com/cockroachdb/cockroach/kv/split_test.go:137 +0x7d testing.tRunner() /usr/src/go/src/testing/testing.go:447 +0x133 ================== ================== WARNING: DATA RACE Read by goroutine 79: github.com/cockroachdb/cockroach/storage.(*rangeStats).GetSize() /go/src/github.com/cockroachdb/cockroach/storage/stats.go:68 +0xae github.com/cockroachdb/cockroach/storage.(*splitQueue).process() /go/src/github.com/cockroachdb/cockroach/storage/split_queue.go:103 +0x81d github.com/cockroachdb/cockroach/storage.(*baseQueue).processOne() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:261 +0x5ce github.com/cockroachdb/cockroach/storage.func·011() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:214 +0x1b7 github.com/cockroachdb/cockroach/util.func·001() /go/src/github.com/cockroachdb/cockroach/util/stopper.go:73 +0x7f Previous write by goroutine 87: github.com/cockroachdb/cockroach/proto.(*MVCCStats).Add() /go/src/github.com/cockroachdb/cockroach/proto/data.go:631 +0x107 github.com/cockroachdb/cockroach/storage.(*rangeStats).MergeMVCCStats() /go/src/github.com/cockroachdb/cockroach/storage/stats.go:82 +0x25c github.com/cockroachdb/cockroach/storage.(*Range).applyRaftCommand() /go/src/github.com/cockroachdb/cockroach/storage/range.go:845 +0xca3 github.com/cockroachdb/cockroach/storage.(*Range).processRaftCommand() /go/src/github.com/cockroachdb/cockroach/storage/range.go:767 +0x3b5 github.com/cockroachdb/cockroach/storage.func·036() /go/src/github.com/cockroachdb/cockroach/storage/store.go:1339 +0xce8 ``` Please assign, take a look and update the issue accordingly.
1.0
Test failure in CI build 2339 - The following test appears to have failed: [#2339](https://circleci.com/gh/cockroachdb/cockroach/2339): ``` level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="770" msg="node 100000001: group 5 raft ready" level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="775" msg="HardState updated: {Term:6 Vote:4294967297 Commit:13 XXX_unrecognized:[]}" level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="778" msg="New Entry[0]: 6/13 EntryNormal 00000000000000006b4ee14837a7cb41: raft_id:5 origin_node_id:4294967297 cmd:<internal_resolve_intent:<header:<timestamp:<wall_time:0 logical:933 > cmd_id:<wall_time:0 random:0 > key:\"\\0" level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="781" msg="Committed Entry[0]: 6/13 EntryNormal 00000000000000006b4ee14837a7cb41: raft_id:5 origin_node_id:4294967297 cmd:<internal_resolve_intent:<header:<timestamp:<wall_time:0 logical:933 > cmd_id:<wall_time:0 random:0 > key:\"\\0" ================== WARNING: DATA RACE Read by goroutine 79: github.com/cockroachdb/cockroach/storage.(*rangeStats).GetSize() /go/src/github.com/cockroachdb/cockroach/storage/stats.go:68 +0x96 github.com/cockroachdb/cockroach/storage.(*splitQueue).process() /go/src/github.com/cockroachdb/cockroach/storage/split_queue.go:103 +0x81d github.com/cockroachdb/cockroach/storage.(*baseQueue).processOne() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:261 +0x5ce github.com/cockroachdb/cockroach/storage.func·011() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:214 +0x1b7 github.com/cockroachdb/cockroach/util.func·001() /go/src/github.com/cockroachdb/cockroach/util/stopper.go:73 +0x7f Previous write by goroutine 87: github.com/cockroachdb/cockroach/proto.(*MVCCStats).Add() /go/src/github.com/cockroachdb/cockroach/proto/data.go:630 +0xa6 github.com/cockroachdb/cockroach/storage.(*rangeStats).MergeMVCCStats() /go/src/github.com/cockroachdb/cockroach/storage/stats.go:82 +0x25c github.com/cockroachdb/cockroach/storage.(*Range).applyRaftCommand() /go/src/github.com/cockroachdb/cockroach/storage/range.go:845 +0xca3 github.com/cockroachdb/cockroach/storage.(*Range).processRaftCommand() /go/src/github.com/cockroachdb/cockroach/storage/range.go:767 +0x3b5 github.com/cockroachdb/cockroach/storage.func·036() /go/src/github.com/cockroachdb/cockroach/storage/store.go:1339 +0xce8 -- /go/src/github.com/cockroachdb/cockroach/kv/split_test.go:137 +0x7d testing.tRunner() /usr/src/go/src/testing/testing.go:447 +0x133 ================== ================== WARNING: DATA RACE Read by goroutine 79: github.com/cockroachdb/cockroach/storage.(*rangeStats).GetSize() /go/src/github.com/cockroachdb/cockroach/storage/stats.go:68 +0xae github.com/cockroachdb/cockroach/storage.(*splitQueue).process() /go/src/github.com/cockroachdb/cockroach/storage/split_queue.go:103 +0x81d github.com/cockroachdb/cockroach/storage.(*baseQueue).processOne() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:261 +0x5ce github.com/cockroachdb/cockroach/storage.func·011() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:214 +0x1b7 github.com/cockroachdb/cockroach/util.func·001() /go/src/github.com/cockroachdb/cockroach/util/stopper.go:73 +0x7f Previous write by goroutine 87: github.com/cockroachdb/cockroach/proto.(*MVCCStats).Add() /go/src/github.com/cockroachdb/cockroach/proto/data.go:631 +0x107 github.com/cockroachdb/cockroach/storage.(*rangeStats).MergeMVCCStats() /go/src/github.com/cockroachdb/cockroach/storage/stats.go:82 +0x25c github.com/cockroachdb/cockroach/storage.(*Range).applyRaftCommand() /go/src/github.com/cockroachdb/cockroach/storage/range.go:845 +0xca3 github.com/cockroachdb/cockroach/storage.(*Range).processRaftCommand() /go/src/github.com/cockroachdb/cockroach/storage/range.go:767 +0x3b5 github.com/cockroachdb/cockroach/storage.func·036() /go/src/github.com/cockroachdb/cockroach/storage/store.go:1339 +0xce8 level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="770" msg="node 100000001: group 5 raft ready" level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="775" msg="HardState updated: {Term:6 Vote:4294967297 Commit:13 XXX_unrecognized:[]}" level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="778" msg="New Entry[0]: 6/13 EntryNormal 00000000000000006b4ee14837a7cb41: raft_id:5 origin_node_id:4294967297 cmd:<internal_resolve_intent:<header:<timestamp:<wall_time:0 logical:933 > cmd_id:<wall_time:0 random:0 > key:\"\\0" level="INFO" time="2015-05-12T15:32:36Z" file="multiraft.go" line="781" msg="Committed Entry[0]: 6/13 EntryNormal 00000000000000006b4ee14837a7cb41: raft_id:5 origin_node_id:4294967297 cmd:<internal_resolve_intent:<header:<timestamp:<wall_time:0 logical:933 > cmd_id:<wall_time:0 random:0 > key:\"\\0" ================== WARNING: DATA RACE Read by goroutine 79: github.com/cockroachdb/cockroach/storage.(*rangeStats).GetSize() /go/src/github.com/cockroachdb/cockroach/storage/stats.go:68 +0x96 github.com/cockroachdb/cockroach/storage.(*splitQueue).process() /go/src/github.com/cockroachdb/cockroach/storage/split_queue.go:103 +0x81d github.com/cockroachdb/cockroach/storage.(*baseQueue).processOne() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:261 +0x5ce github.com/cockroachdb/cockroach/storage.func·011() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:214 +0x1b7 github.com/cockroachdb/cockroach/util.func·001() /go/src/github.com/cockroachdb/cockroach/util/stopper.go:73 +0x7f Previous write by goroutine 87: github.com/cockroachdb/cockroach/proto.(*MVCCStats).Add() /go/src/github.com/cockroachdb/cockroach/proto/data.go:630 +0xa6 github.com/cockroachdb/cockroach/storage.(*rangeStats).MergeMVCCStats() /go/src/github.com/cockroachdb/cockroach/storage/stats.go:82 +0x25c github.com/cockroachdb/cockroach/storage.(*Range).applyRaftCommand() /go/src/github.com/cockroachdb/cockroach/storage/range.go:845 +0xca3 github.com/cockroachdb/cockroach/storage.(*Range).processRaftCommand() /go/src/github.com/cockroachdb/cockroach/storage/range.go:767 +0x3b5 github.com/cockroachdb/cockroach/storage.func·036() /go/src/github.com/cockroachdb/cockroach/storage/store.go:1339 +0xce8 -- /go/src/github.com/cockroachdb/cockroach/kv/split_test.go:137 +0x7d testing.tRunner() /usr/src/go/src/testing/testing.go:447 +0x133 ================== ================== WARNING: DATA RACE Read by goroutine 79: github.com/cockroachdb/cockroach/storage.(*rangeStats).GetSize() /go/src/github.com/cockroachdb/cockroach/storage/stats.go:68 +0xae github.com/cockroachdb/cockroach/storage.(*splitQueue).process() /go/src/github.com/cockroachdb/cockroach/storage/split_queue.go:103 +0x81d github.com/cockroachdb/cockroach/storage.(*baseQueue).processOne() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:261 +0x5ce github.com/cockroachdb/cockroach/storage.func·011() /go/src/github.com/cockroachdb/cockroach/storage/queue.go:214 +0x1b7 github.com/cockroachdb/cockroach/util.func·001() /go/src/github.com/cockroachdb/cockroach/util/stopper.go:73 +0x7f Previous write by goroutine 87: github.com/cockroachdb/cockroach/proto.(*MVCCStats).Add() /go/src/github.com/cockroachdb/cockroach/proto/data.go:631 +0x107 github.com/cockroachdb/cockroach/storage.(*rangeStats).MergeMVCCStats() /go/src/github.com/cockroachdb/cockroach/storage/stats.go:82 +0x25c github.com/cockroachdb/cockroach/storage.(*Range).applyRaftCommand() /go/src/github.com/cockroachdb/cockroach/storage/range.go:845 +0xca3 github.com/cockroachdb/cockroach/storage.(*Range).processRaftCommand() /go/src/github.com/cockroachdb/cockroach/storage/range.go:767 +0x3b5 github.com/cockroachdb/cockroach/storage.func·036() /go/src/github.com/cockroachdb/cockroach/storage/store.go:1339 +0xce8 ``` Please assign, take a look and update the issue accordingly.
test
test failure in ci build the following test appears to have failed level info time file multiraft go line msg node group raft ready level info time file multiraft go line msg hardstate updated term vote commit xxx unrecognized level info time file multiraft go line msg new entry entrynormal raft id origin node id cmd cmd id key level info time file multiraft go line msg committed entry entrynormal raft id origin node id cmd cmd id key warning data race read by goroutine github com cockroachdb cockroach storage rangestats getsize go src github com cockroachdb cockroach storage stats go github com cockroachdb cockroach storage splitqueue process go src github com cockroachdb cockroach storage split queue go github com cockroachdb cockroach storage basequeue processone go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach storage func· go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util func· go src github com cockroachdb cockroach util stopper go previous write by goroutine github com cockroachdb cockroach proto mvccstats add go src github com cockroachdb cockroach proto data go github com cockroachdb cockroach storage rangestats mergemvccstats go src github com cockroachdb cockroach storage stats go github com cockroachdb cockroach storage range applyraftcommand go src github com cockroachdb cockroach storage range go github com cockroachdb cockroach storage range processraftcommand go src github com cockroachdb cockroach storage range go github com cockroachdb cockroach storage func· go src github com cockroachdb cockroach storage store go go src github com cockroachdb cockroach kv split test go testing trunner usr src go src testing testing go warning data race read by goroutine github com cockroachdb cockroach storage rangestats getsize go src github com cockroachdb cockroach storage stats go github com cockroachdb cockroach storage splitqueue process go src github com cockroachdb cockroach storage split queue go github com cockroachdb cockroach storage basequeue processone go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach storage func· go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util func· go src github com cockroachdb cockroach util stopper go previous write by goroutine github com cockroachdb cockroach proto mvccstats add go src github com cockroachdb cockroach proto data go github com cockroachdb cockroach storage rangestats mergemvccstats go src github com cockroachdb cockroach storage stats go github com cockroachdb cockroach storage range applyraftcommand go src github com cockroachdb cockroach storage range go github com cockroachdb cockroach storage range processraftcommand go src github com cockroachdb cockroach storage range go github com cockroachdb cockroach storage func· go src github com cockroachdb cockroach storage store go level info time file multiraft go line msg node group raft ready level info time file multiraft go line msg hardstate updated term vote commit xxx unrecognized level info time file multiraft go line msg new entry entrynormal raft id origin node id cmd cmd id key level info time file multiraft go line msg committed entry entrynormal raft id origin node id cmd cmd id key warning data race read by goroutine github com cockroachdb cockroach storage rangestats getsize go src github com cockroachdb cockroach storage stats go github com cockroachdb cockroach storage splitqueue process go src github com cockroachdb cockroach storage split queue go github com cockroachdb cockroach storage basequeue processone go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach storage func· go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util func· go src github com cockroachdb cockroach util stopper go previous write by goroutine github com cockroachdb cockroach proto mvccstats add go src github com cockroachdb cockroach proto data go github com cockroachdb cockroach storage rangestats mergemvccstats go src github com cockroachdb cockroach storage stats go github com cockroachdb cockroach storage range applyraftcommand go src github com cockroachdb cockroach storage range go github com cockroachdb cockroach storage range processraftcommand go src github com cockroachdb cockroach storage range go github com cockroachdb cockroach storage func· go src github com cockroachdb cockroach storage store go go src github com cockroachdb cockroach kv split test go testing trunner usr src go src testing testing go warning data race read by goroutine github com cockroachdb cockroach storage rangestats getsize go src github com cockroachdb cockroach storage stats go github com cockroachdb cockroach storage splitqueue process go src github com cockroachdb cockroach storage split queue go github com cockroachdb cockroach storage basequeue processone go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach storage func· go src github com cockroachdb cockroach storage queue go github com cockroachdb cockroach util func· go src github com cockroachdb cockroach util stopper go previous write by goroutine github com cockroachdb cockroach proto mvccstats add go src github com cockroachdb cockroach proto data go github com cockroachdb cockroach storage rangestats mergemvccstats go src github com cockroachdb cockroach storage stats go github com cockroachdb cockroach storage range applyraftcommand go src github com cockroachdb cockroach storage range go github com cockroachdb cockroach storage range processraftcommand go src github com cockroachdb cockroach storage range go github com cockroachdb cockroach storage func· go src github com cockroachdb cockroach storage store go please assign take a look and update the issue accordingly
1
169,626
13,155,092,586
IssuesEvent
2020-08-10 08:12:59
ekaterinailin/AltaiPony
https://api.github.com/repos/ekaterinailin/AltaiPony
closed
Test the preservation of units throughout the pipeline
testing
<!-- Fill in the information below before opening an issue. --> #### What needs to be created or improved? <!-- Provide a clear and concise description of the issue. --> Write tests that explicitly check units of outputs of functions if the input units are known. **Question**: Would this be a good use case for decorators? (@gully) #### Can you provide an example? <!-- Provide a link or minimal code snippet that demonstrates the issue. --> ```python def test_units(): lc = get_k2sc_lc() flares = lc.find_flares(*args, **kwargs) test_find_flares(lc, flares) #OR @testunits def find_flares(*args, **kwargs): [...] return flares ``` #### What is the goal / expected behaviour? <!-- Describe the behavior you expected and how it differs from the behavior observed in the example. --> Time is given in days, but ED is typically measured in seconds, frequencies can be _per hour_ or _per year_, amplitued can be relative or e-/s.
1.0
Test the preservation of units throughout the pipeline - <!-- Fill in the information below before opening an issue. --> #### What needs to be created or improved? <!-- Provide a clear and concise description of the issue. --> Write tests that explicitly check units of outputs of functions if the input units are known. **Question**: Would this be a good use case for decorators? (@gully) #### Can you provide an example? <!-- Provide a link or minimal code snippet that demonstrates the issue. --> ```python def test_units(): lc = get_k2sc_lc() flares = lc.find_flares(*args, **kwargs) test_find_flares(lc, flares) #OR @testunits def find_flares(*args, **kwargs): [...] return flares ``` #### What is the goal / expected behaviour? <!-- Describe the behavior you expected and how it differs from the behavior observed in the example. --> Time is given in days, but ED is typically measured in seconds, frequencies can be _per hour_ or _per year_, amplitued can be relative or e-/s.
test
test the preservation of units throughout the pipeline what needs to be created or improved write tests that explicitly check units of outputs of functions if the input units are known question would this be a good use case for decorators gully can you provide an example python def test units lc get lc flares lc find flares args kwargs test find flares lc flares or testunits def find flares args kwargs return flares what is the goal expected behaviour time is given in days but ed is typically measured in seconds frequencies can be per hour or per year amplitued can be relative or e s
1
46,417
11,840,195,642
IssuesEvent
2020-03-23 18:25:42
apache/incubator-mxnet
https://api.github.com/repos/apache/incubator-mxnet
closed
USE_TVM_OP=1 build broken with DMLC_LOG_FATAL_THROW=0
Bug Build TVM OP
## Description Setting `-DDMLC_LOG_FATAL_THROW=0` is useful for debugging with `gdb` or `lldb`. Unfortunately we can't compile MXNet with `USE_TVM_OP=1` in that case. ### Error Message ``` [658/658] Linking CXX shared library libmxnet.so FAILED: libmxnet.so : && /usr/bin/c++ -fPIC -Wall -Wno-sign-compare -O3 -std=c++11 -fopenmp -shared -Wl,-soname,libmxnet.so -o libmxnet.so CMakeFiles/mxnet.dir/dummy.c.o -L/home/ubuntu/src/mxnet-master/build/3rdparty/tvm -Wl,-rpath,/usr/local/lib:/home/ubuntu/src/mxnet-master/build/3rdparty/tvm: -Wl,--whole-archive libmxnet.a -Wl,--no-whole-archive libmxnet.a 3rdparty/dmlc-core/libdmlc.a -ldl 3rdparty/mkldnn/src/libdnnl.a -lpthread -Wl,--start-group /opt/intel/mkl/lib/intel64/libmkl_intel_lp64.a /opt/intel/mkl/lib/intel64/libmkl_intel_thread.a /opt/intel/mkl/lib/intel64/libmkl_core.a /usr/local/lib/libiomp5.so -Wl,--end-group /usr/lib/x86_64-linux-gnu/librt.so /usr/lib/x86_64-linux-gnu/libopencv_highgui.so.3.2.0 /usr/lib/x86_64-linux-gnu/libopencv_imgcodecs.so.3.2.0 /usr/lib/x86_64-linux-gnu/libopencv_imgproc.so.3.2.0 /usr/lib/x86_64-linux-gnu/libopencv_core.so.3.2.0 -llapack -ltvm_runtime /usr/lib/gcc/x86_64-linux-gnu/7/libgomp.so -lpthread -lrt && cd /home/ubuntu/src/mxnet-master/build && /home/ubuntu/.pyenv/versions/3.8.2-debug/lib/python3.8/site-packages/cmake/data/bin/cmake -E copy /home/ubuntu/src/mxnet-master/build/3rdparty/mkldnn/include/dnnl_config.h /home/ubuntu/src/mxnet-master/include/mkldnn/ && /home/ubuntu/.pyenv/versions/3.8.2-debug/lib/python3.8/site-packages/cmake/data/bin/cmake -E copy /home/ubuntu/src/mxnet-master/build/3rdparty/mkldnn/include/dnnl_version.h /home/ubuntu/src/mxnet-master/include/mkldnn/ && cd /home/ubuntu/src/mxnet-master/build && /home/ubuntu/.pyenv/versions/3.8.2-debug/lib/python3.8/site-packages/cmake/data/bin/cmake -E env PYTHONPATH="/home/ubuntu/src/mxnet-master/3rdparty/tvm/python:/home/ubuntu/src/mxnet-master/3rdparty/tvm/topi/python:/home/ubuntu/src/mxnet-master/contrib" LD_LIBRARY_PATH=/home/ubuntu/src/mxnet-master/build:/home/ubuntu/src/mxnet-master/build/3rdparty/tvm:/usr/local/cuda/lib64:/usr/local/lib:/usr/lib:/usr/local/cuda/extras/CUPTI/lib64:/usr/local/mpi/lib:/home/ubuntu/local/lib:/home/ubuntu/local/lib:/usr/local/cuda/lib64:/usr/local/lib:/usr/lib:/usr/local/cuda/extras/CUPTI/lib64:/usr/local/mpi/lib:/home/ubuntu/local/lib /home/ubuntu/.pyenv/versions/3.8.2-debug/bin/python3.8 /home/ubuntu/src/mxnet-master/contrib/tvmop/compile.py -o/home/ubuntu/src/mxnet-master/build --config /home/ubuntu/src/mxnet-master/build/tvmop.conf -L /home/ubuntu/src/mxnet-master/build/3rdparty/tvm [19:56:21] ../3rdparty/tvm/src/node/reflection.cc:113: AttributeError: Variable object has no attributed shape Child aborted ninja: build stopped: subcommand failed. ``` ## To Reproduce ### Steps to reproduce 1. Apply patch below 2. Compile with `cmake -GNinja -DUSE_CUDA=0 -DUSE_TVM_OP=ON ..; ninja` ``` diff diff --git a/CMakeLists.txt b/CMakeLists.txt index 365da2d90..b4c9d267b 100644 --- a/CMakeLists.txt +++ b/CMakeLists.txt @@ -151,7 +151,7 @@ endif() add_definitions(-DDMLC_MODERN_THREAD_LOCAL=0) # disable stack trace in exception by default. add_definitions(-DDMLC_LOG_STACK_TRACE_SIZE=0) -add_definitions(-DDMLC_LOG_FATAL_THROW=1) +add_definitions(-DDMLC_LOG_FATAL_THROW=0) if(MSVC) add_definitions(-DWIN32_LEAN_AND_MEAN) ```
1.0
USE_TVM_OP=1 build broken with DMLC_LOG_FATAL_THROW=0 - ## Description Setting `-DDMLC_LOG_FATAL_THROW=0` is useful for debugging with `gdb` or `lldb`. Unfortunately we can't compile MXNet with `USE_TVM_OP=1` in that case. ### Error Message ``` [658/658] Linking CXX shared library libmxnet.so FAILED: libmxnet.so : && /usr/bin/c++ -fPIC -Wall -Wno-sign-compare -O3 -std=c++11 -fopenmp -shared -Wl,-soname,libmxnet.so -o libmxnet.so CMakeFiles/mxnet.dir/dummy.c.o -L/home/ubuntu/src/mxnet-master/build/3rdparty/tvm -Wl,-rpath,/usr/local/lib:/home/ubuntu/src/mxnet-master/build/3rdparty/tvm: -Wl,--whole-archive libmxnet.a -Wl,--no-whole-archive libmxnet.a 3rdparty/dmlc-core/libdmlc.a -ldl 3rdparty/mkldnn/src/libdnnl.a -lpthread -Wl,--start-group /opt/intel/mkl/lib/intel64/libmkl_intel_lp64.a /opt/intel/mkl/lib/intel64/libmkl_intel_thread.a /opt/intel/mkl/lib/intel64/libmkl_core.a /usr/local/lib/libiomp5.so -Wl,--end-group /usr/lib/x86_64-linux-gnu/librt.so /usr/lib/x86_64-linux-gnu/libopencv_highgui.so.3.2.0 /usr/lib/x86_64-linux-gnu/libopencv_imgcodecs.so.3.2.0 /usr/lib/x86_64-linux-gnu/libopencv_imgproc.so.3.2.0 /usr/lib/x86_64-linux-gnu/libopencv_core.so.3.2.0 -llapack -ltvm_runtime /usr/lib/gcc/x86_64-linux-gnu/7/libgomp.so -lpthread -lrt && cd /home/ubuntu/src/mxnet-master/build && /home/ubuntu/.pyenv/versions/3.8.2-debug/lib/python3.8/site-packages/cmake/data/bin/cmake -E copy /home/ubuntu/src/mxnet-master/build/3rdparty/mkldnn/include/dnnl_config.h /home/ubuntu/src/mxnet-master/include/mkldnn/ && /home/ubuntu/.pyenv/versions/3.8.2-debug/lib/python3.8/site-packages/cmake/data/bin/cmake -E copy /home/ubuntu/src/mxnet-master/build/3rdparty/mkldnn/include/dnnl_version.h /home/ubuntu/src/mxnet-master/include/mkldnn/ && cd /home/ubuntu/src/mxnet-master/build && /home/ubuntu/.pyenv/versions/3.8.2-debug/lib/python3.8/site-packages/cmake/data/bin/cmake -E env PYTHONPATH="/home/ubuntu/src/mxnet-master/3rdparty/tvm/python:/home/ubuntu/src/mxnet-master/3rdparty/tvm/topi/python:/home/ubuntu/src/mxnet-master/contrib" LD_LIBRARY_PATH=/home/ubuntu/src/mxnet-master/build:/home/ubuntu/src/mxnet-master/build/3rdparty/tvm:/usr/local/cuda/lib64:/usr/local/lib:/usr/lib:/usr/local/cuda/extras/CUPTI/lib64:/usr/local/mpi/lib:/home/ubuntu/local/lib:/home/ubuntu/local/lib:/usr/local/cuda/lib64:/usr/local/lib:/usr/lib:/usr/local/cuda/extras/CUPTI/lib64:/usr/local/mpi/lib:/home/ubuntu/local/lib /home/ubuntu/.pyenv/versions/3.8.2-debug/bin/python3.8 /home/ubuntu/src/mxnet-master/contrib/tvmop/compile.py -o/home/ubuntu/src/mxnet-master/build --config /home/ubuntu/src/mxnet-master/build/tvmop.conf -L /home/ubuntu/src/mxnet-master/build/3rdparty/tvm [19:56:21] ../3rdparty/tvm/src/node/reflection.cc:113: AttributeError: Variable object has no attributed shape Child aborted ninja: build stopped: subcommand failed. ``` ## To Reproduce ### Steps to reproduce 1. Apply patch below 2. Compile with `cmake -GNinja -DUSE_CUDA=0 -DUSE_TVM_OP=ON ..; ninja` ``` diff diff --git a/CMakeLists.txt b/CMakeLists.txt index 365da2d90..b4c9d267b 100644 --- a/CMakeLists.txt +++ b/CMakeLists.txt @@ -151,7 +151,7 @@ endif() add_definitions(-DDMLC_MODERN_THREAD_LOCAL=0) # disable stack trace in exception by default. add_definitions(-DDMLC_LOG_STACK_TRACE_SIZE=0) -add_definitions(-DDMLC_LOG_FATAL_THROW=1) +add_definitions(-DDMLC_LOG_FATAL_THROW=0) if(MSVC) add_definitions(-DWIN32_LEAN_AND_MEAN) ```
non_test
use tvm op build broken with dmlc log fatal throw description setting ddmlc log fatal throw is useful for debugging with gdb or lldb unfortunately we can t compile mxnet with use tvm op in that case error message linking cxx shared library libmxnet so failed libmxnet so usr bin c fpic wall wno sign compare std c fopenmp shared wl soname libmxnet so o libmxnet so cmakefiles mxnet dir dummy c o l home ubuntu src mxnet master build tvm wl rpath usr local lib home ubuntu src mxnet master build tvm wl whole archive libmxnet a wl no whole archive libmxnet a dmlc core libdmlc a ldl mkldnn src libdnnl a lpthread wl start group opt intel mkl lib libmkl intel a opt intel mkl lib libmkl intel thread a opt intel mkl lib libmkl core a usr local lib so wl end group usr lib linux gnu librt so usr lib linux gnu libopencv highgui so usr lib linux gnu libopencv imgcodecs so usr lib linux gnu libopencv imgproc so usr lib linux gnu libopencv core so llapack ltvm runtime usr lib gcc linux gnu libgomp so lpthread lrt cd home ubuntu src mxnet master build home ubuntu pyenv versions debug lib site packages cmake data bin cmake e copy home ubuntu src mxnet master build mkldnn include dnnl config h home ubuntu src mxnet master include mkldnn home ubuntu pyenv versions debug lib site packages cmake data bin cmake e copy home ubuntu src mxnet master build mkldnn include dnnl version h home ubuntu src mxnet master include mkldnn cd home ubuntu src mxnet master build home ubuntu pyenv versions debug lib site packages cmake data bin cmake e env pythonpath home ubuntu src mxnet master tvm python home ubuntu src mxnet master tvm topi python home ubuntu src mxnet master contrib ld library path home ubuntu src mxnet master build home ubuntu src mxnet master build tvm usr local cuda usr local lib usr lib usr local cuda extras cupti usr local mpi lib home ubuntu local lib home ubuntu local lib usr local cuda usr local lib usr lib usr local cuda extras cupti usr local mpi lib home ubuntu local lib home ubuntu pyenv versions debug bin home ubuntu src mxnet master contrib tvmop compile py o home ubuntu src mxnet master build config home ubuntu src mxnet master build tvmop conf l home ubuntu src mxnet master build tvm tvm src node reflection cc attributeerror variable object has no attributed shape child aborted ninja build stopped subcommand failed to reproduce steps to reproduce apply patch below compile with cmake gninja duse cuda duse tvm op on ninja diff diff git a cmakelists txt b cmakelists txt index a cmakelists txt b cmakelists txt endif add definitions ddmlc modern thread local disable stack trace in exception by default add definitions ddmlc log stack trace size add definitions ddmlc log fatal throw add definitions ddmlc log fatal throw if msvc add definitions lean and mean
0
363,805
10,755,416,028
IssuesEvent
2019-10-31 09:05:36
acidanthera/bugtracker
https://api.github.com/repos/acidanthera/bugtracker
closed
VoodooPS2, Thinkpad T440s, 3 physical button not working
enhancement priority:normal project:ps2
hi, hava an t440s with the trackpad from a t450( extra 3physical buttons). all gestures are working fine except the three button. in consol.log if I press on elf the three button it produce standard 17:20:16.553291 +0200 kernel synaptics_parse_hw_state: finger 0 pressure 0 width 8. thx [debug_2105.zip](https://github.com/acidanthera/bugtracker/files/3108267/debug_2105.zip)
1.0
VoodooPS2, Thinkpad T440s, 3 physical button not working - hi, hava an t440s with the trackpad from a t450( extra 3physical buttons). all gestures are working fine except the three button. in consol.log if I press on elf the three button it produce standard 17:20:16.553291 +0200 kernel synaptics_parse_hw_state: finger 0 pressure 0 width 8. thx [debug_2105.zip](https://github.com/acidanthera/bugtracker/files/3108267/debug_2105.zip)
non_test
thinkpad physical button not working hi hava an with the trackpad from a extra buttons all gestures are working fine except the three button in consol log if i press on elf the three button it produce standard kernel synaptics parse hw state finger pressure width thx
0
433,197
12,503,507,124
IssuesEvent
2020-06-02 07:22:24
metal3-io/metal3-dev-env
https://api.github.com/repos/metal3-io/metal3-dev-env
closed
Add cloud init log out put
kind/feature priority/important-longterm
In the CI, some jobs are failing and the output does not reveal much information. This [PR](281) is a case in point. It was failing due to missing binaries and the outputs did not show that. We were required to login to the provisioned machines and get the logs. It would be good the content of ```/var/log/cloud-init-output.log``` is also shown at the end as it would make pin pointing errors easy.
1.0
Add cloud init log out put - In the CI, some jobs are failing and the output does not reveal much information. This [PR](281) is a case in point. It was failing due to missing binaries and the outputs did not show that. We were required to login to the provisioned machines and get the logs. It would be good the content of ```/var/log/cloud-init-output.log``` is also shown at the end as it would make pin pointing errors easy.
non_test
add cloud init log out put in the ci some jobs are failing and the output does not reveal much information this is a case in point it was failing due to missing binaries and the outputs did not show that we were required to login to the provisioned machines and get the logs it would be good the content of var log cloud init output log is also shown at the end as it would make pin pointing errors easy
0
119,536
10,057,783,106
IssuesEvent
2019-07-22 12:31:18
ICIJ/datashare
https://api.github.com/repos/ICIJ/datashare
closed
Put 'Tags' above 'Details'
bug front need testing
In order to avoid horizontal scroll bars and a blank space behing 'Tags', we can put 'Tags' above 'Details' as it should not take too much room. <img width="835" alt="Screenshot 2019-07-15 at 13 56 32" src="https://user-images.githubusercontent.com/17233829/61214545-74a41c80-a708-11e9-8256-65bd7bd463cd.png">
1.0
Put 'Tags' above 'Details' - In order to avoid horizontal scroll bars and a blank space behing 'Tags', we can put 'Tags' above 'Details' as it should not take too much room. <img width="835" alt="Screenshot 2019-07-15 at 13 56 32" src="https://user-images.githubusercontent.com/17233829/61214545-74a41c80-a708-11e9-8256-65bd7bd463cd.png">
test
put tags above details in order to avoid horizontal scroll bars and a blank space behing tags we can put tags above details as it should not take too much room img width alt screenshot at src
1
46,469
2,957,910,188
IssuesEvent
2015-07-08 18:41:49
CenterForOpenScience/osf.io
https://api.github.com/repos/CenterForOpenScience/osf.io
closed
ID requested for linkedin profile does not lead to public profile
2 - ready bug: production intern priority - high
From the community repo: https://github.com/CenterForOpenScience/community/issues/269 ## Current behavior When adding a Linkedin profile, the user is shown the beginning of a URL that they are expected to fill out with their "profile ID" (https://www.linkedin.com/profile/view?id=) ![screen shot 2015-07-01 at 9 42 16 am](https://cloud.githubusercontent.com/assets/4965832/8455840/95abd6ca-1fd5-11e5-8ac2-09b39642b386.png) My Linkedin user ID is davidtmellor, which does not successfully complete the expected URL, it results in a 404 error on Linkedin. The correct URL that I think would complete it is: ttps://www.linkedin.com/profile/view?id=77582777 but that URL only works if the user is signed into Linkedin, a visitor to my OSF profile who clicks on that link would be prompted to sign into Linkedin before seeing my Linkefin profile. The suggest change below allows a user to access my Linkedin profile without signing up for Linkedin. ![screen shot 2015-07-01 at 9 42 38 am](https://cloud.githubusercontent.com/assets/4965832/8455846/9ef85c94-1fd5-11e5-8ef0-0f8e2756500d.png) The public profile also shows a number (the ID) instead of their user ID, which would be slightly more welcoming: ![screen shot 2015-07-01 at 9 42 09 am](https://cloud.githubusercontent.com/assets/4965832/8455862/b31c0630-1fd5-11e5-85f1-2734ef3efa1e.png) ## Suggested behavior Allow the user to enter in their Linkedin ID to complete this URL: https://www.linkedin.com/in/USERID example: https://www.linkedin.com/in/davidtmellor. Show this on the public profile OR Allow user to enter in any URL they want for their Linkedin profile.
1.0
ID requested for linkedin profile does not lead to public profile - From the community repo: https://github.com/CenterForOpenScience/community/issues/269 ## Current behavior When adding a Linkedin profile, the user is shown the beginning of a URL that they are expected to fill out with their "profile ID" (https://www.linkedin.com/profile/view?id=) ![screen shot 2015-07-01 at 9 42 16 am](https://cloud.githubusercontent.com/assets/4965832/8455840/95abd6ca-1fd5-11e5-8ac2-09b39642b386.png) My Linkedin user ID is davidtmellor, which does not successfully complete the expected URL, it results in a 404 error on Linkedin. The correct URL that I think would complete it is: ttps://www.linkedin.com/profile/view?id=77582777 but that URL only works if the user is signed into Linkedin, a visitor to my OSF profile who clicks on that link would be prompted to sign into Linkedin before seeing my Linkefin profile. The suggest change below allows a user to access my Linkedin profile without signing up for Linkedin. ![screen shot 2015-07-01 at 9 42 38 am](https://cloud.githubusercontent.com/assets/4965832/8455846/9ef85c94-1fd5-11e5-8ef0-0f8e2756500d.png) The public profile also shows a number (the ID) instead of their user ID, which would be slightly more welcoming: ![screen shot 2015-07-01 at 9 42 09 am](https://cloud.githubusercontent.com/assets/4965832/8455862/b31c0630-1fd5-11e5-85f1-2734ef3efa1e.png) ## Suggested behavior Allow the user to enter in their Linkedin ID to complete this URL: https://www.linkedin.com/in/USERID example: https://www.linkedin.com/in/davidtmellor. Show this on the public profile OR Allow user to enter in any URL they want for their Linkedin profile.
non_test
id requested for linkedin profile does not lead to public profile from the community repo current behavior when adding a linkedin profile the user is shown the beginning of a url that they are expected to fill out with their profile id my linkedin user id is davidtmellor which does not successfully complete the expected url it results in a error on linkedin the correct url that i think would complete it is ttps but that url only works if the user is signed into linkedin a visitor to my osf profile who clicks on that link would be prompted to sign into linkedin before seeing my linkefin profile the suggest change below allows a user to access my linkedin profile without signing up for linkedin the public profile also shows a number the id instead of their user id which would be slightly more welcoming suggested behavior allow the user to enter in their linkedin id to complete this url example show this on the public profile or allow user to enter in any url they want for their linkedin profile
0
50,481
6,097,900,444
IssuesEvent
2017-06-20 05:33:36
c2corg/v6_ui
https://api.github.com/repos/c2corg/v6_ui
closed
Unsupported L# tags in v6.1.10
ready for testing
For some reason L# are not correctly supported in release v6.1.10 (v6.1.9 was OK but with a simpler support I think). It can be observed in the demo. For instance http://www.demov6.camptocamp.org/routes/368130/fr/vallee-d-ailefroide-poire-d-ailefroide-bonne-poire I have switched the redis cache off on the demo to make sure rendered documents are not retrieved from the cache. The weird thing seems that sometimes it works, sometimes it doesn't. Well it seems that it works until the dcument is edited, than it's broken. @olaurendeau Would you have an idea?
1.0
Unsupported L# tags in v6.1.10 - For some reason L# are not correctly supported in release v6.1.10 (v6.1.9 was OK but with a simpler support I think). It can be observed in the demo. For instance http://www.demov6.camptocamp.org/routes/368130/fr/vallee-d-ailefroide-poire-d-ailefroide-bonne-poire I have switched the redis cache off on the demo to make sure rendered documents are not retrieved from the cache. The weird thing seems that sometimes it works, sometimes it doesn't. Well it seems that it works until the dcument is edited, than it's broken. @olaurendeau Would you have an idea?
test
unsupported l tags in for some reason l are not correctly supported in release was ok but with a simpler support i think it can be observed in the demo for instance i have switched the redis cache off on the demo to make sure rendered documents are not retrieved from the cache the weird thing seems that sometimes it works sometimes it doesn t well it seems that it works until the dcument is edited than it s broken olaurendeau would you have an idea
1
118,100
17,576,352,460
IssuesEvent
2021-08-15 17:32:48
turkdevops/brackets
https://api.github.com/repos/turkdevops/brackets
reopened
CVE-2017-16026 (Medium) detected in request-2.42.0.tgz, request-2.9.153.tgz
security vulnerability
## CVE-2017-16026 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>request-2.42.0.tgz</b>, <b>request-2.9.153.tgz</b></p></summary> <p> <details><summary><b>request-2.42.0.tgz</b></p></summary> <p>Simplified HTTP request client.</p> <p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.42.0.tgz">https://registry.npmjs.org/request/-/request-2.42.0.tgz</a></p> <p>Path to dependency file: brackets/package.json</p> <p>Path to vulnerable library: brackets/node_modules/phantomjs/node_modules/request/package.json</p> <p> Dependency Hierarchy: - phantomjs-1.9.18.tgz (Root Library) - :x: **request-2.42.0.tgz** (Vulnerable Library) </details> <details><summary><b>request-2.9.153.tgz</b></p></summary> <p>Simplified HTTP request client.</p> <p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.9.153.tgz">https://registry.npmjs.org/request/-/request-2.9.153.tgz</a></p> <p>Path to dependency file: brackets/src/extensions/default/JavaScriptQuickEdit/unittest-files/jquery-ui/package.json</p> <p>Path to vulnerable library: /src/extensions/default/JavaScriptQuickEdit/unittest-files/jquery-ui/node_modules/request/package.json</p> <p> Dependency Hierarchy: - :x: **request-2.9.153.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/turkdevops/brackets/commit/38b389e26019954346ecebf98a10c2f5fa9a0488">38b389e26019954346ecebf98a10c2f5fa9a0488</a></p> <p>Found in base branch: <b>checkTravis</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Request is an http client. If a request is made using ```multipart```, and the body type is a ```number```, then the specified number of non-zero memory is passed in the body. This affects Request >=2.2.6 <2.47.0 || >2.51.0 <=2.67.0. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16026>CVE-2017-16026</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-16026">https://nvd.nist.gov/vuln/detail/CVE-2017-16026</a></p> <p>Release Date: 2018-06-04</p> <p>Fix Resolution: 2.47.1,2.67.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-16026 (Medium) detected in request-2.42.0.tgz, request-2.9.153.tgz - ## CVE-2017-16026 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>request-2.42.0.tgz</b>, <b>request-2.9.153.tgz</b></p></summary> <p> <details><summary><b>request-2.42.0.tgz</b></p></summary> <p>Simplified HTTP request client.</p> <p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.42.0.tgz">https://registry.npmjs.org/request/-/request-2.42.0.tgz</a></p> <p>Path to dependency file: brackets/package.json</p> <p>Path to vulnerable library: brackets/node_modules/phantomjs/node_modules/request/package.json</p> <p> Dependency Hierarchy: - phantomjs-1.9.18.tgz (Root Library) - :x: **request-2.42.0.tgz** (Vulnerable Library) </details> <details><summary><b>request-2.9.153.tgz</b></p></summary> <p>Simplified HTTP request client.</p> <p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.9.153.tgz">https://registry.npmjs.org/request/-/request-2.9.153.tgz</a></p> <p>Path to dependency file: brackets/src/extensions/default/JavaScriptQuickEdit/unittest-files/jquery-ui/package.json</p> <p>Path to vulnerable library: /src/extensions/default/JavaScriptQuickEdit/unittest-files/jquery-ui/node_modules/request/package.json</p> <p> Dependency Hierarchy: - :x: **request-2.9.153.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/turkdevops/brackets/commit/38b389e26019954346ecebf98a10c2f5fa9a0488">38b389e26019954346ecebf98a10c2f5fa9a0488</a></p> <p>Found in base branch: <b>checkTravis</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Request is an http client. If a request is made using ```multipart```, and the body type is a ```number```, then the specified number of non-zero memory is passed in the body. This affects Request >=2.2.6 <2.47.0 || >2.51.0 <=2.67.0. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16026>CVE-2017-16026</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-16026">https://nvd.nist.gov/vuln/detail/CVE-2017-16026</a></p> <p>Release Date: 2018-06-04</p> <p>Fix Resolution: 2.47.1,2.67.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in request tgz request tgz cve medium severity vulnerability vulnerable libraries request tgz request tgz request tgz simplified http request client library home page a href path to dependency file brackets package json path to vulnerable library brackets node modules phantomjs node modules request package json dependency hierarchy phantomjs tgz root library x request tgz vulnerable library request tgz simplified http request client library home page a href path to dependency file brackets src extensions default javascriptquickedit unittest files jquery ui package json path to vulnerable library src extensions default javascriptquickedit unittest files jquery ui node modules request package json dependency hierarchy x request tgz vulnerable library found in head commit a href found in base branch checktravis vulnerability details request is an http client if a request is made using multipart and the body type is a number then the specified number of non zero memory is passed in the body this affects request publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
287,288
24,822,078,303
IssuesEvent
2022-10-25 17:18:05
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
Shader visually differs from playing scene and in-editor
topic:rendering needs testing topic:3d
### Godot version 4.0beta3 ### System information PopOS 22.04, 5.19.0-76051900-generic kernel, Ryzen 6800h, Radeon 6700m, amdgpu, Vulkan ### Issue description In editor: ![image](https://user-images.githubusercontent.com/15805689/197404425-85f6f7ed-1b9a-4082-a1f6-ff5feaffbb34.png) ![image](https://user-images.githubusercontent.com/15805689/197405689-da787f06-0134-4194-9937-f6727c728db3.png) When playing scene: ![image](https://user-images.githubusercontent.com/15805689/197404410-6dbcf103-902d-400d-8328-90e52071f6c6.png) ### Steps to reproduce Modifying shaders update in editor but don't show the same in game. It's not clear which is visually correct. Far and near of camera are set to the same as in editor. ### Minimal reproduction project Reproduction Repo: https://github.com/Mikeysax/example-project-for-godot-bug-67807 It appears like the noise being used for the foam is never applied: ![image](https://user-images.githubusercontent.com/15805689/197404139-eb3c2f36-bc0a-4fd2-a856-3e7a92e90978.png) Shader Code: https://gist.github.com/Mikeysax/a97f14e8f1d00a53c3ba83d964eeb350
1.0
Shader visually differs from playing scene and in-editor - ### Godot version 4.0beta3 ### System information PopOS 22.04, 5.19.0-76051900-generic kernel, Ryzen 6800h, Radeon 6700m, amdgpu, Vulkan ### Issue description In editor: ![image](https://user-images.githubusercontent.com/15805689/197404425-85f6f7ed-1b9a-4082-a1f6-ff5feaffbb34.png) ![image](https://user-images.githubusercontent.com/15805689/197405689-da787f06-0134-4194-9937-f6727c728db3.png) When playing scene: ![image](https://user-images.githubusercontent.com/15805689/197404410-6dbcf103-902d-400d-8328-90e52071f6c6.png) ### Steps to reproduce Modifying shaders update in editor but don't show the same in game. It's not clear which is visually correct. Far and near of camera are set to the same as in editor. ### Minimal reproduction project Reproduction Repo: https://github.com/Mikeysax/example-project-for-godot-bug-67807 It appears like the noise being used for the foam is never applied: ![image](https://user-images.githubusercontent.com/15805689/197404139-eb3c2f36-bc0a-4fd2-a856-3e7a92e90978.png) Shader Code: https://gist.github.com/Mikeysax/a97f14e8f1d00a53c3ba83d964eeb350
test
shader visually differs from playing scene and in editor godot version system information popos generic kernel ryzen radeon amdgpu vulkan issue description in editor when playing scene steps to reproduce modifying shaders update in editor but don t show the same in game it s not clear which is visually correct far and near of camera are set to the same as in editor minimal reproduction project reproduction repo it appears like the noise being used for the foam is never applied shader code
1
133,197
12,534,458,856
IssuesEvent
2020-06-04 19:28:29
spring-projects/spring-boot
https://api.github.com/repos/spring-projects/spring-boot
closed
Use highlightjs for syntax highlighting in HTML documentation
type: documentation
It looks like it got lost when the build was converted to Gradle.
1.0
Use highlightjs for syntax highlighting in HTML documentation - It looks like it got lost when the build was converted to Gradle.
non_test
use highlightjs for syntax highlighting in html documentation it looks like it got lost when the build was converted to gradle
0
312,816
23,444,530,534
IssuesEvent
2022-08-15 18:12:13
edgexfoundry/edgex-docs
https://api.github.com/repos/edgexfoundry/edgex-docs
closed
Document Device System Events
documentation enhancement
See implementation PRs for reference: https://github.com/edgexfoundry/edgex-go/pull/4101 https://github.com/edgexfoundry/edgex-go/pull/4121 And the ADR here: https://docs.edgexfoundry.org/2.3/design/adr/0024-system-events/
1.0
Document Device System Events - See implementation PRs for reference: https://github.com/edgexfoundry/edgex-go/pull/4101 https://github.com/edgexfoundry/edgex-go/pull/4121 And the ADR here: https://docs.edgexfoundry.org/2.3/design/adr/0024-system-events/
non_test
document device system events see implementation prs for reference and the adr here
0
320,076
27,420,002,996
IssuesEvent
2023-03-01 16:06:32
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
opened
Fix math.test_tensorflow_squared_difference
TensorFlow Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4301928602/jobs/7499787609" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|None |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4301928602/jobs/7499798636" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|None <details> <summary>Not found</summary> Not found </details> <details> <summary>Not found</summary> Not found </details>
1.0
Fix math.test_tensorflow_squared_difference - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4301928602/jobs/7499787609" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|None |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4301928602/jobs/7499798636" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|None <details> <summary>Not found</summary> Not found </details> <details> <summary>Not found</summary> Not found </details>
test
fix math test tensorflow squared difference tensorflow img src torch none numpy img src jax none not found not found not found not found
1
20,289
2,622,438,446
IssuesEvent
2015-03-04 02:44:47
calblueprint/PHC
https://api.github.com/repos/calblueprint/PHC
opened
standardize all text and views
medium priority UI/UX
standardize text size, font, format, style. standardize button sizes and other view sizes.
1.0
standardize all text and views - standardize text size, font, format, style. standardize button sizes and other view sizes.
non_test
standardize all text and views standardize text size font format style standardize button sizes and other view sizes
0
68,139
8,221,632,785
IssuesEvent
2018-09-06 03:05:37
rbastic/dyndao
https://api.github.com/repos/rbastic/dyndao
closed
Consider adding column Length (and Precision, for floats) as a parameter to the Is*Type() functions
design
There are situations where these things are important (due to ambiguity in type affinity.)
1.0
Consider adding column Length (and Precision, for floats) as a parameter to the Is*Type() functions - There are situations where these things are important (due to ambiguity in type affinity.)
non_test
consider adding column length and precision for floats as a parameter to the is type functions there are situations where these things are important due to ambiguity in type affinity
0
276,291
8,596,906,768
IssuesEvent
2018-11-15 17:07:28
InfiniteFlightAirportEditing/Airports
https://api.github.com/repos/InfiniteFlightAirportEditing/Airports
opened
OAHR-Herat-HERAT-AFGHANISTAN
Being Redone Low Priority
# Airport Name Herat # Country? Afghanistan # Improvements that need to be made? from scratch # Are you working on this airport? Yes # Airport Priority? (A380, 10000ft+ Runway) 9880ft Rwy
1.0
OAHR-Herat-HERAT-AFGHANISTAN - # Airport Name Herat # Country? Afghanistan # Improvements that need to be made? from scratch # Are you working on this airport? Yes # Airport Priority? (A380, 10000ft+ Runway) 9880ft Rwy
non_test
oahr herat herat afghanistan airport name herat country afghanistan improvements that need to be made from scratch are you working on this airport yes airport priority runway rwy
0
101,802
11,258,338,685
IssuesEvent
2020-01-13 04:08:57
KisaraBlue/ManaClash
https://api.github.com/repos/KisaraBlue/ManaClash
opened
Swagger UI documentation
documentation
Documentation can be generated in about an hour of work. I will do it on January 13th, after the deadline passes (I have an exam at 10:45 AM and I need to try to get a few hours of sleep before that, so I'll add documentation same day at night).
1.0
Swagger UI documentation - Documentation can be generated in about an hour of work. I will do it on January 13th, after the deadline passes (I have an exam at 10:45 AM and I need to try to get a few hours of sleep before that, so I'll add documentation same day at night).
non_test
swagger ui documentation documentation can be generated in about an hour of work i will do it on january after the deadline passes i have an exam at am and i need to try to get a few hours of sleep before that so i ll add documentation same day at night
0
226,473
17,352,981,655
IssuesEvent
2021-07-29 11:05:48
FreshRSS/FreshRSS
https://api.github.com/repos/FreshRSS/FreshRSS
closed
[Feature] Support for Raspberry Pi
Documentation :books:
**Is your feature request related to a problem? Please describe.** Hi, there. I ran FreshRSS on my cloud VPS for a couple of months and everything worked fine. After some time I decided to buy Raspberry Pi and move all my cloud self-hosted services there. I had some issues even with PostgreSQL but found [a solution](https://blog.samcater.com/fix-workaround-rpi4-docker-libseccomp2-docker-20/) that fixed that issue. Regardless of it FreshRSS app still can't start (I am running it in a container, `docker-compose.yml` is taken from [here](https://github.com/FreshRSS/FreshRSS/blob/edge/Docker/docker-compose.yml) without additional proxies or other stuff)... ``` CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 33ab4f55a320 freshrss/freshrss:latest "./Docker/entrypoint…" 25 seconds ago Restarting (1) 6 seconds ago freshrss-app f5dafa34d425 postgres:12-alpine "docker-entrypoint.s…" 27 seconds ago Up 25 seconds 0.0.0.0:5432->5432/tcp freshrss-db ``` Here are logs of FreshRSS app: ``` Attaching to freshrss-app freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error ``` **Describe the solution you'd like** It would be nice to have FreshRSS app able to run on Raspberry Pi since it is common platform for serving different services at home. **Additional context** Since I'm pretty sure it is not problem with FreshRSS app itself, but some underlying stuff with linux and so on (which I'm not so familiar with) I considered this GitHub issue to be a feature request and not bug. In the end it would be useful to have the solution written in the documentation. So it is like cry for a help plus documentation feature request😅
1.0
[Feature] Support for Raspberry Pi - **Is your feature request related to a problem? Please describe.** Hi, there. I ran FreshRSS on my cloud VPS for a couple of months and everything worked fine. After some time I decided to buy Raspberry Pi and move all my cloud self-hosted services there. I had some issues even with PostgreSQL but found [a solution](https://blog.samcater.com/fix-workaround-rpi4-docker-libseccomp2-docker-20/) that fixed that issue. Regardless of it FreshRSS app still can't start (I am running it in a container, `docker-compose.yml` is taken from [here](https://github.com/FreshRSS/FreshRSS/blob/edge/Docker/docker-compose.yml) without additional proxies or other stuff)... ``` CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 33ab4f55a320 freshrss/freshrss:latest "./Docker/entrypoint…" 25 seconds ago Restarting (1) 6 seconds ago freshrss-app f5dafa34d425 postgres:12-alpine "docker-entrypoint.s…" 27 seconds ago Up 25 seconds 0.0.0.0:5432->5432/tcp freshrss-db ``` Here are logs of FreshRSS app: ``` Attaching to freshrss-app freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error freshrss-app | standard_init_linux.go:219: exec user process caused: exec format error ``` **Describe the solution you'd like** It would be nice to have FreshRSS app able to run on Raspberry Pi since it is common platform for serving different services at home. **Additional context** Since I'm pretty sure it is not problem with FreshRSS app itself, but some underlying stuff with linux and so on (which I'm not so familiar with) I considered this GitHub issue to be a feature request and not bug. In the end it would be useful to have the solution written in the documentation. So it is like cry for a help plus documentation feature request😅
non_test
support for raspberry pi is your feature request related to a problem please describe hi there i ran freshrss on my cloud vps for a couple of months and everything worked fine after some time i decided to buy raspberry pi and move all my cloud self hosted services there i had some issues even with postgresql but found that fixed that issue regardless of it freshrss app still can t start i am running it in a container docker compose yml is taken from without additional proxies or other stuff container id image command created status ports names freshrss freshrss latest docker entrypoint… seconds ago restarting seconds ago freshrss app postgres alpine docker entrypoint s… seconds ago up seconds tcp freshrss db here are logs of freshrss app attaching to freshrss app freshrss app standard init linux go exec user process caused exec format error freshrss app standard init linux go exec user process caused exec format error freshrss app standard init linux go exec user process caused exec format error freshrss app standard init linux go exec user process caused exec format error freshrss app standard init linux go exec user process caused exec format error freshrss app standard init linux go exec user process caused exec format error freshrss app standard init linux go exec user process caused exec format error describe the solution you d like it would be nice to have freshrss app able to run on raspberry pi since it is common platform for serving different services at home additional context since i m pretty sure it is not problem with freshrss app itself but some underlying stuff with linux and so on which i m not so familiar with i considered this github issue to be a feature request and not bug in the end it would be useful to have the solution written in the documentation so it is like cry for a help plus documentation feature request😅
0
3,532
2,678,027,248
IssuesEvent
2015-03-26 07:12:09
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
test failure #914
test-failure
The following test appears to have failed: [#914](https://circleci.com/gh/cockroachdb/cockroach/914): ``` I0326 05:26:29.194389 81 multiraft.go:449] node 1: group 1 got message 8->1 MsgHeartbeatResp Term:6 Log:0/0 I0326 05:26:29.194408 81 multiraft.go:449] node 1: group 1 got message 4->1 MsgHeartbeatResp Term:6 Log:0/0 I0326 05:26:29.194422 81 multiraft.go:449] node 1: group 1 got message 5->1 MsgAppResp Term:6 Log:0/11 I0326 05:26:29.194440 81 multiraft.go:449] node 1: group 1 got message 6->1 MsgAppResp Term:6 Log:0/11 E0326 05:26:29.687754 81 heartbeat_test.go:51] timeout when reading from intercept channel panic: test timed out after 30s goroutine 384 [running]: testing.func·008() /usr/src/go/src/testing/testing.go:681 +0x12f created by time.goFunc /usr/src/go/src/time/sleep.go:129 +0x4b goroutine 1 [chan receive]: testing.RunTests(0x9a5d20, 0xb0d340, 0x7, 0x7, 0xc207ffc301) /usr/src/go/src/testing/testing.go:556 +0xad6 -- goroutine 345 [select]: github.com/cockroachdb/cockroach/multiraft.func·009() /go/src/github.com/cockroachdb/cockroach/multiraft/events_test.go:51 +0x3fc created by github.com/cockroachdb/cockroach/multiraft.(*eventDemux).start /go/src/github.com/cockroachdb/cockroach/multiraft/events_test.go:72 +0x8c FAIL github.com/cockroachdb/cockroach/multiraft 30.023s === RUN TestMemoryStorage --- PASS: TestMemoryStorage (0.00s) PASS ok github.com/cockroachdb/cockroach/multiraft/storagetest 0.004s === RUN TestClientCmdIDIsEmpty --- PASS: TestClientCmdIDIsEmpty (0.00s) === RUN TestResponseHeaderSetGoError --- PASS: TestResponseHeaderSetGoError (0.00s) === RUN TestResponseHeaderNilError --- PASS: TestResponseHeaderNilError (0.00s) I0326 05:26:29.194389 81 multiraft.go:449] node 1: group 1 got message 8->1 MsgHeartbeatResp Term:6 Log:0/0 I0326 05:26:29.194408 81 multiraft.go:449] node 1: group 1 got message 4->1 MsgHeartbeatResp Term:6 Log:0/0 I0326 05:26:29.194422 81 multiraft.go:449] node 1: group 1 got message 5->1 MsgAppResp Term:6 Log:0/11 I0326 05:26:29.194440 81 multiraft.go:449] node 1: group 1 got message 6->1 MsgAppResp Term:6 Log:0/11 E0326 05:26:29.687754 81 heartbeat_test.go:51] timeout when reading from intercept channel panic: test timed out after 30s goroutine 384 [running]: testing.func·008() /usr/src/go/src/testing/testing.go:681 +0x12f created by time.goFunc /usr/src/go/src/time/sleep.go:129 +0x4b goroutine 1 [chan receive]: testing.RunTests(0x9a5d20, 0xb0d340, 0x7, 0x7, 0xc207ffc301) /usr/src/go/src/testing/testing.go:556 +0xad6 -- goroutine 345 [select]: github.com/cockroachdb/cockroach/multiraft.func·009() /go/src/github.com/cockroachdb/cockroach/multiraft/events_test.go:51 +0x3fc created by github.com/cockroachdb/cockroach/multiraft.(*eventDemux).start /go/src/github.com/cockroachdb/cockroach/multiraft/events_test.go:72 +0x8c FAIL github.com/cockroachdb/cockroach/multiraft 30.023s === RUN TestMemoryStorage --- PASS: TestMemoryStorage (0.00s) PASS ok github.com/cockroachdb/cockroach/multiraft/storagetest 0.004s === RUN TestClientCmdIDIsEmpty --- PASS: TestClientCmdIDIsEmpty (0.00s) === RUN TestResponseHeaderSetGoError --- PASS: TestResponseHeaderSetGoError (0.00s) === RUN TestResponseHeaderNilError --- PASS: TestResponseHeaderNilError (0.00s) ``` Please assign, take a look and update the issue accordingly.
1.0
test failure #914 - The following test appears to have failed: [#914](https://circleci.com/gh/cockroachdb/cockroach/914): ``` I0326 05:26:29.194389 81 multiraft.go:449] node 1: group 1 got message 8->1 MsgHeartbeatResp Term:6 Log:0/0 I0326 05:26:29.194408 81 multiraft.go:449] node 1: group 1 got message 4->1 MsgHeartbeatResp Term:6 Log:0/0 I0326 05:26:29.194422 81 multiraft.go:449] node 1: group 1 got message 5->1 MsgAppResp Term:6 Log:0/11 I0326 05:26:29.194440 81 multiraft.go:449] node 1: group 1 got message 6->1 MsgAppResp Term:6 Log:0/11 E0326 05:26:29.687754 81 heartbeat_test.go:51] timeout when reading from intercept channel panic: test timed out after 30s goroutine 384 [running]: testing.func·008() /usr/src/go/src/testing/testing.go:681 +0x12f created by time.goFunc /usr/src/go/src/time/sleep.go:129 +0x4b goroutine 1 [chan receive]: testing.RunTests(0x9a5d20, 0xb0d340, 0x7, 0x7, 0xc207ffc301) /usr/src/go/src/testing/testing.go:556 +0xad6 -- goroutine 345 [select]: github.com/cockroachdb/cockroach/multiraft.func·009() /go/src/github.com/cockroachdb/cockroach/multiraft/events_test.go:51 +0x3fc created by github.com/cockroachdb/cockroach/multiraft.(*eventDemux).start /go/src/github.com/cockroachdb/cockroach/multiraft/events_test.go:72 +0x8c FAIL github.com/cockroachdb/cockroach/multiraft 30.023s === RUN TestMemoryStorage --- PASS: TestMemoryStorage (0.00s) PASS ok github.com/cockroachdb/cockroach/multiraft/storagetest 0.004s === RUN TestClientCmdIDIsEmpty --- PASS: TestClientCmdIDIsEmpty (0.00s) === RUN TestResponseHeaderSetGoError --- PASS: TestResponseHeaderSetGoError (0.00s) === RUN TestResponseHeaderNilError --- PASS: TestResponseHeaderNilError (0.00s) I0326 05:26:29.194389 81 multiraft.go:449] node 1: group 1 got message 8->1 MsgHeartbeatResp Term:6 Log:0/0 I0326 05:26:29.194408 81 multiraft.go:449] node 1: group 1 got message 4->1 MsgHeartbeatResp Term:6 Log:0/0 I0326 05:26:29.194422 81 multiraft.go:449] node 1: group 1 got message 5->1 MsgAppResp Term:6 Log:0/11 I0326 05:26:29.194440 81 multiraft.go:449] node 1: group 1 got message 6->1 MsgAppResp Term:6 Log:0/11 E0326 05:26:29.687754 81 heartbeat_test.go:51] timeout when reading from intercept channel panic: test timed out after 30s goroutine 384 [running]: testing.func·008() /usr/src/go/src/testing/testing.go:681 +0x12f created by time.goFunc /usr/src/go/src/time/sleep.go:129 +0x4b goroutine 1 [chan receive]: testing.RunTests(0x9a5d20, 0xb0d340, 0x7, 0x7, 0xc207ffc301) /usr/src/go/src/testing/testing.go:556 +0xad6 -- goroutine 345 [select]: github.com/cockroachdb/cockroach/multiraft.func·009() /go/src/github.com/cockroachdb/cockroach/multiraft/events_test.go:51 +0x3fc created by github.com/cockroachdb/cockroach/multiraft.(*eventDemux).start /go/src/github.com/cockroachdb/cockroach/multiraft/events_test.go:72 +0x8c FAIL github.com/cockroachdb/cockroach/multiraft 30.023s === RUN TestMemoryStorage --- PASS: TestMemoryStorage (0.00s) PASS ok github.com/cockroachdb/cockroach/multiraft/storagetest 0.004s === RUN TestClientCmdIDIsEmpty --- PASS: TestClientCmdIDIsEmpty (0.00s) === RUN TestResponseHeaderSetGoError --- PASS: TestResponseHeaderSetGoError (0.00s) === RUN TestResponseHeaderNilError --- PASS: TestResponseHeaderNilError (0.00s) ``` Please assign, take a look and update the issue accordingly.
test
test failure the following test appears to have failed multiraft go node group got message msgheartbeatresp term log multiraft go node group got message msgheartbeatresp term log multiraft go node group got message msgappresp term log multiraft go node group got message msgappresp term log heartbeat test go timeout when reading from intercept channel panic test timed out after goroutine testing func· usr src go src testing testing go created by time gofunc usr src go src time sleep go goroutine testing runtests usr src go src testing testing go goroutine github com cockroachdb cockroach multiraft func· go src github com cockroachdb cockroach multiraft events test go created by github com cockroachdb cockroach multiraft eventdemux start go src github com cockroachdb cockroach multiraft events test go fail github com cockroachdb cockroach multiraft run testmemorystorage pass testmemorystorage pass ok github com cockroachdb cockroach multiraft storagetest run testclientcmdidisempty pass testclientcmdidisempty run testresponseheadersetgoerror pass testresponseheadersetgoerror run testresponseheadernilerror pass testresponseheadernilerror multiraft go node group got message msgheartbeatresp term log multiraft go node group got message msgheartbeatresp term log multiraft go node group got message msgappresp term log multiraft go node group got message msgappresp term log heartbeat test go timeout when reading from intercept channel panic test timed out after goroutine testing func· usr src go src testing testing go created by time gofunc usr src go src time sleep go goroutine testing runtests usr src go src testing testing go goroutine github com cockroachdb cockroach multiraft func· go src github com cockroachdb cockroach multiraft events test go created by github com cockroachdb cockroach multiraft eventdemux start go src github com cockroachdb cockroach multiraft events test go fail github com cockroachdb cockroach multiraft run testmemorystorage pass testmemorystorage pass ok github com cockroachdb cockroach multiraft storagetest run testclientcmdidisempty pass testclientcmdidisempty run testresponseheadersetgoerror pass testresponseheadersetgoerror run testresponseheadernilerror pass testresponseheadernilerror please assign take a look and update the issue accordingly
1
126,949
17,146,277,094
IssuesEvent
2021-07-13 14:54:29
wordpress-mobile/gutenberg-mobile
https://api.github.com/repos/wordpress-mobile/gutenberg-mobile
closed
Add Layout Grid - 1st iteration: Read only version
Blocks [Status] Needs Design
What we are aiming on this iteration is rendering the grid correctly on mobile in different content size options but not yet allowing the users to change settings. There will be a [separate effort](https://github.com/wordpress-mobile/gutenberg-mobile/issues/2516) to decide for which sites we should enable this block on mobile. @enejb After making an initial investigation please drop a comment explaining how you'll approach developing this, what your sub-tasks will look like. Thanks! @iamthomasbishop We need some directions about UX, we are planning to start working on this ~Aug 10-13. So the sooner we can clarify somethings the more efficient our work will be. Thanks!
1.0
Add Layout Grid - 1st iteration: Read only version - What we are aiming on this iteration is rendering the grid correctly on mobile in different content size options but not yet allowing the users to change settings. There will be a [separate effort](https://github.com/wordpress-mobile/gutenberg-mobile/issues/2516) to decide for which sites we should enable this block on mobile. @enejb After making an initial investigation please drop a comment explaining how you'll approach developing this, what your sub-tasks will look like. Thanks! @iamthomasbishop We need some directions about UX, we are planning to start working on this ~Aug 10-13. So the sooner we can clarify somethings the more efficient our work will be. Thanks!
non_test
add layout grid iteration read only version what we are aiming on this iteration is rendering the grid correctly on mobile in different content size options but not yet allowing the users to change settings there will be a to decide for which sites we should enable this block on mobile enejb after making an initial investigation please drop a comment explaining how you ll approach developing this what your sub tasks will look like thanks iamthomasbishop we need some directions about ux we are planning to start working on this aug so the sooner we can clarify somethings the more efficient our work will be thanks
0
26,719
6,796,334,460
IssuesEvent
2017-11-01 18:39:06
Dijji/RepairTasks
https://api.github.com/repos/Dijji/RepairTasks
closed
wind 10 to win 7 still errors
CodePlexMigrationInitiated Impact: Unassigned
I couldn't make the unzip and remove function work. Still orig 5 errors. After scan, comment was not formatted. I am a klutx, so maybe its me. thankks #### This work item was migrated from CodePlex CodePlex work item ID: '3' Vote count: '1'
1.0
wind 10 to win 7 still errors - I couldn't make the unzip and remove function work. Still orig 5 errors. After scan, comment was not formatted. I am a klutx, so maybe its me. thankks #### This work item was migrated from CodePlex CodePlex work item ID: '3' Vote count: '1'
non_test
wind to win still errors i couldn t make the unzip and remove function work still orig errors after scan comment was not formatted i am a klutx so maybe its me thankks this work item was migrated from codeplex codeplex work item id vote count
0
510,759
14,815,778,583
IssuesEvent
2021-01-14 07:57:28
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
lms.autozone.com - site is not usable
browser-firefox engine-gecko os-linux priority-normal type-flash
<!-- @browser: Firefox 68.0 --> <!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/65511 --> **URL**: https://lms.autozone.com/Kview/CustomCodeBehind/base/courseware/scorm/scorm12courseframe.aspx **Browser / Version**: Firefox 68.0 **Operating System**: Linux **Tested Another Browser**: Yes Other **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/1/dc4c0dd8-565a-496f-a6fb-728343d627c9.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200409090751</li><li>channel: default</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2021/1/db35845e-e7e9-464d-bf24-f51f2cda37a1) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
lms.autozone.com - site is not usable - <!-- @browser: Firefox 68.0 --> <!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/65511 --> **URL**: https://lms.autozone.com/Kview/CustomCodeBehind/base/courseware/scorm/scorm12courseframe.aspx **Browser / Version**: Firefox 68.0 **Operating System**: Linux **Tested Another Browser**: Yes Other **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/1/dc4c0dd8-565a-496f-a6fb-728343d627c9.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200409090751</li><li>channel: default</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2021/1/db35845e-e7e9-464d-bf24-f51f2cda37a1) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
lms autozone com site is not usable url browser version firefox operating system linux tested another browser yes other problem type site is not usable description page not loading correctly steps to reproduce view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel default hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
18,572
3,072,603,030
IssuesEvent
2015-08-19 17:45:34
CocoaPods/CocoaPods
https://api.github.com/repos/CocoaPods/CocoaPods
closed
Missing Required Modules
s2:confirmed t2:defect
I amusing the 0.36.0 beta 2 version of the plugin and working to create an app using an Extension. I want my app framework to be the only part that uses certain cocoa pods. However when I try to import my AppKit into my main project (using swift) the complier complains that it is missing the required modules, which happen to be the cocoapods that I am not including in my main app. ![screen shot 2015-02-01 at 4 48 32 pm](https://cloud.githubusercontent.com/assets/1010384/5994697/57765326-aa32-11e4-9b69-e8e50f1d84a3.png) ![screen shot 2015-02-01 at 4 50 00 pm](https://cloud.githubusercontent.com/assets/1010384/5994700/9af657ae-aa32-11e4-8aa5-9e71d2969ad8.png)
1.0
Missing Required Modules - I amusing the 0.36.0 beta 2 version of the plugin and working to create an app using an Extension. I want my app framework to be the only part that uses certain cocoa pods. However when I try to import my AppKit into my main project (using swift) the complier complains that it is missing the required modules, which happen to be the cocoapods that I am not including in my main app. ![screen shot 2015-02-01 at 4 48 32 pm](https://cloud.githubusercontent.com/assets/1010384/5994697/57765326-aa32-11e4-9b69-e8e50f1d84a3.png) ![screen shot 2015-02-01 at 4 50 00 pm](https://cloud.githubusercontent.com/assets/1010384/5994700/9af657ae-aa32-11e4-8aa5-9e71d2969ad8.png)
non_test
missing required modules i amusing the beta version of the plugin and working to create an app using an extension i want my app framework to be the only part that uses certain cocoa pods however when i try to import my appkit into my main project using swift the complier complains that it is missing the required modules which happen to be the cocoapods that i am not including in my main app
0
76,604
7,541,418,562
IssuesEvent
2018-04-17 09:43:04
ISISScientificComputing/autoreduce
https://api.github.com/repos/ISISScientificComputing/autoreduce
closed
Static analysis
Maintenance Testing
This should be added after we have integrated Travis to the Repo. I have already tested that this works, but there are a notable amount of checks to fix - however they are mostly white space at a glance. Pylint or Flake8 can be used for these they both use the same standard (PEP8) so should produce roughly the same results. As a result of this issue, the warnings will also have to be fixed!
1.0
Static analysis - This should be added after we have integrated Travis to the Repo. I have already tested that this works, but there are a notable amount of checks to fix - however they are mostly white space at a glance. Pylint or Flake8 can be used for these they both use the same standard (PEP8) so should produce roughly the same results. As a result of this issue, the warnings will also have to be fixed!
test
static analysis this should be added after we have integrated travis to the repo i have already tested that this works but there are a notable amount of checks to fix however they are mostly white space at a glance pylint or can be used for these they both use the same standard so should produce roughly the same results as a result of this issue the warnings will also have to be fixed
1
318,445
27,304,568,627
IssuesEvent
2023-02-24 06:53:01
lightningnetwork/lnd
https://api.github.com/repos/lightningnetwork/lnd
closed
itest-flake: co-op channel fail during update chan policy test teardown
test flake itests btcd
At the very end of the test, for `btcd` backend (so far): ``` test_harness.go:88: Failed: (update channel policy): exited with error: *errors.errorString unable to close channel: error while waiting for broadcast tx: wanted aedef3c01bb9e4c046acb0dddd6aaf65476a9a5c2cd7500ab41b06909ccc7e49, found 0 txs in mempool: [] /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:242 (0xf32df8) closeChannelAndAssertType: t.Fatalf("unable to close channel: %v", err) /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:217 (0xf3ae7f) testUpdateChannelPolicy: return closeChannelAndAssertType(ctx, t, net, node, fundingChanPoint, false, force) /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:2058 (0xf3ae31) testUpdateChannelPolicy: closeChannelAndAssert(ctxt, t, net, net.Alice, chanPoint, false) /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/test_harness.go:112 (0xeeb1ee) (*harnessTest).RunTestCase: testCase.test(h.lndHarness, h) /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:11702 (0xf9ceaf) TestLightningNetworkDaemon.func4: ht.RunTestCase(testCase) /home/travis/.gimme/versions/go1.16.3.linux.amd64/src/testing/testing.go:1193 (0x51c72f) tRunner: fn(t) /home/travis/.gimme/versions/go1.16.3.linux.amd64/src/runtime/asm_amd64.s:1371 (0x471c01) goexit: BYTE $0x90 // NOP === CONT TestLightningNetworkDaemon lnd_test.go:11710: Failure time: 2021-07-03 13:30:47.057 --- FAIL: TestLightningNetworkDaemon (806.43s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc (596.71s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy (232.50s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/local_force_close_immediate_expiry (27.72s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/receiver_chain_claim (30.49s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/local_force_close_on-chain_htlc_timeout (26.10s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/remote_force_close_on-chain_htlc_timeout (25.80s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/local_chain_claim (45.21s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/remote_chain_claim (34.75s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/local_htlc_aggregation (37.98s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors (343.77s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/local_force_close_immediate_expiry (44.10s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/receiver_chain_claim (38.92s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/local_force_close_on-chain_htlc_timeout (50.20s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/remote_force_close_on-chain_htlc_timeout (33.90s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/local_chain_claim (68.85s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/remote_chain_claim (57.60s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/local_htlc_aggregation (45.52s) --- PASS: TestLightningNetworkDaemon/02-of-82/btcd/sweep_coins (6.09s) --- PASS: TestLightningNetworkDaemon/03-of-82/btcd/recovery_info (14.40s) --- PASS: TestLightningNetworkDaemon/04-of-82/btcd/onchain_fund_recovery (29.89s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow (66.18s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=legacy,dave_commit=legacy (7.02s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=legacy,dave_commit=tweakless (7.02s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=legacy,dave_commit=anchors (6.99s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=tweakless,dave_commit=legacy (6.99s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=tweakless,dave_commit=tweakless (6.98s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=tweakless,dave_commit=anchors (6.99s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=anchors,dave_commit=legacy (7.04s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=anchors,dave_commit=tweakless (6.99s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=anchors,dave_commit=anchors (6.97s) --- PASS: TestLightningNetworkDaemon/06-of-82/btcd/unconfirmed_channel_funding (7.32s) --- FAIL: TestLightningNetworkDaemon/07-of-82/btcd/update_channel_policy (74.45s) ```
2.0
itest-flake: co-op channel fail during update chan policy test teardown - At the very end of the test, for `btcd` backend (so far): ``` test_harness.go:88: Failed: (update channel policy): exited with error: *errors.errorString unable to close channel: error while waiting for broadcast tx: wanted aedef3c01bb9e4c046acb0dddd6aaf65476a9a5c2cd7500ab41b06909ccc7e49, found 0 txs in mempool: [] /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:242 (0xf32df8) closeChannelAndAssertType: t.Fatalf("unable to close channel: %v", err) /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:217 (0xf3ae7f) testUpdateChannelPolicy: return closeChannelAndAssertType(ctx, t, net, node, fundingChanPoint, false, force) /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:2058 (0xf3ae31) testUpdateChannelPolicy: closeChannelAndAssert(ctxt, t, net, net.Alice, chanPoint, false) /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/test_harness.go:112 (0xeeb1ee) (*harnessTest).RunTestCase: testCase.test(h.lndHarness, h) /home/travis/gopath/src/github.com/lightningnetwork/lnd/lntest/itest/lnd_test.go:11702 (0xf9ceaf) TestLightningNetworkDaemon.func4: ht.RunTestCase(testCase) /home/travis/.gimme/versions/go1.16.3.linux.amd64/src/testing/testing.go:1193 (0x51c72f) tRunner: fn(t) /home/travis/.gimme/versions/go1.16.3.linux.amd64/src/runtime/asm_amd64.s:1371 (0x471c01) goexit: BYTE $0x90 // NOP === CONT TestLightningNetworkDaemon lnd_test.go:11710: Failure time: 2021-07-03 13:30:47.057 --- FAIL: TestLightningNetworkDaemon (806.43s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc (596.71s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy (232.50s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/local_force_close_immediate_expiry (27.72s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/receiver_chain_claim (30.49s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/local_force_close_on-chain_htlc_timeout (26.10s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/remote_force_close_on-chain_htlc_timeout (25.80s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/local_chain_claim (45.21s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/remote_chain_claim (34.75s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=legacy/local_htlc_aggregation (37.98s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors (343.77s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/local_force_close_immediate_expiry (44.10s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/receiver_chain_claim (38.92s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/local_force_close_on-chain_htlc_timeout (50.20s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/remote_force_close_on-chain_htlc_timeout (33.90s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/local_chain_claim (68.85s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/remote_chain_claim (57.60s) --- PASS: TestLightningNetworkDaemon/01-of-82/btcd/test_multi-hop_htlc/committype=anchors/local_htlc_aggregation (45.52s) --- PASS: TestLightningNetworkDaemon/02-of-82/btcd/sweep_coins (6.09s) --- PASS: TestLightningNetworkDaemon/03-of-82/btcd/recovery_info (14.40s) --- PASS: TestLightningNetworkDaemon/04-of-82/btcd/onchain_fund_recovery (29.89s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow (66.18s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=legacy,dave_commit=legacy (7.02s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=legacy,dave_commit=tweakless (7.02s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=legacy,dave_commit=anchors (6.99s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=tweakless,dave_commit=legacy (6.99s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=tweakless,dave_commit=tweakless (6.98s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=tweakless,dave_commit=anchors (6.99s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=anchors,dave_commit=legacy (7.04s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=anchors,dave_commit=tweakless (6.99s) --- PASS: TestLightningNetworkDaemon/05-of-82/btcd/basic_funding_flow/carol_commit=anchors,dave_commit=anchors (6.97s) --- PASS: TestLightningNetworkDaemon/06-of-82/btcd/unconfirmed_channel_funding (7.32s) --- FAIL: TestLightningNetworkDaemon/07-of-82/btcd/update_channel_policy (74.45s) ```
test
itest flake co op channel fail during update chan policy test teardown at the very end of the test for btcd backend so far test harness go failed update channel policy exited with error errors errorstring unable to close channel error while waiting for broadcast tx wanted found txs in mempool home travis gopath src github com lightningnetwork lnd lntest itest lnd test go closechannelandasserttype t fatalf unable to close channel v err home travis gopath src github com lightningnetwork lnd lntest itest lnd test go testupdatechannelpolicy return closechannelandasserttype ctx t net node fundingchanpoint false force home travis gopath src github com lightningnetwork lnd lntest itest lnd test go testupdatechannelpolicy closechannelandassert ctxt t net net alice chanpoint false home travis gopath src github com lightningnetwork lnd lntest itest test harness go harnesstest runtestcase testcase test h lndharness h home travis gopath src github com lightningnetwork lnd lntest itest lnd test go testlightningnetworkdaemon ht runtestcase testcase home travis gimme versions linux src testing testing go trunner fn t home travis gimme versions linux src runtime asm s goexit byte nop cont testlightningnetworkdaemon lnd test go failure time fail testlightningnetworkdaemon pass testlightningnetworkdaemon of btcd test multi hop htlc pass testlightningnetworkdaemon of btcd test multi hop htlc committype legacy pass testlightningnetworkdaemon of btcd test multi hop htlc committype legacy local force close immediate expiry pass testlightningnetworkdaemon of btcd test multi hop htlc committype legacy receiver chain claim pass testlightningnetworkdaemon of btcd test multi hop htlc committype legacy local force close on chain htlc timeout pass testlightningnetworkdaemon of btcd test multi hop htlc committype legacy remote force close on chain htlc timeout pass testlightningnetworkdaemon of btcd test multi hop htlc committype legacy local chain claim pass testlightningnetworkdaemon of btcd test multi hop htlc committype legacy remote chain claim pass testlightningnetworkdaemon of btcd test multi hop htlc committype legacy local htlc aggregation pass testlightningnetworkdaemon of btcd test multi hop htlc committype anchors pass testlightningnetworkdaemon of btcd test multi hop htlc committype anchors local force close immediate expiry pass testlightningnetworkdaemon of btcd test multi hop htlc committype anchors receiver chain claim pass testlightningnetworkdaemon of btcd test multi hop htlc committype anchors local force close on chain htlc timeout pass testlightningnetworkdaemon of btcd test multi hop htlc committype anchors remote force close on chain htlc timeout pass testlightningnetworkdaemon of btcd test multi hop htlc committype anchors local chain claim pass testlightningnetworkdaemon of btcd test multi hop htlc committype anchors remote chain claim pass testlightningnetworkdaemon of btcd test multi hop htlc committype anchors local htlc aggregation pass testlightningnetworkdaemon of btcd sweep coins pass testlightningnetworkdaemon of btcd recovery info pass testlightningnetworkdaemon of btcd onchain fund recovery pass testlightningnetworkdaemon of btcd basic funding flow pass testlightningnetworkdaemon of btcd basic funding flow carol commit legacy dave commit legacy pass testlightningnetworkdaemon of btcd basic funding flow carol commit legacy dave commit tweakless pass testlightningnetworkdaemon of btcd basic funding flow carol commit legacy dave commit anchors pass testlightningnetworkdaemon of btcd basic funding flow carol commit tweakless dave commit legacy pass testlightningnetworkdaemon of btcd basic funding flow carol commit tweakless dave commit tweakless pass testlightningnetworkdaemon of btcd basic funding flow carol commit tweakless dave commit anchors pass testlightningnetworkdaemon of btcd basic funding flow carol commit anchors dave commit legacy pass testlightningnetworkdaemon of btcd basic funding flow carol commit anchors dave commit tweakless pass testlightningnetworkdaemon of btcd basic funding flow carol commit anchors dave commit anchors pass testlightningnetworkdaemon of btcd unconfirmed channel funding fail testlightningnetworkdaemon of btcd update channel policy
1
349,699
31,822,190,424
IssuesEvent
2023-09-14 03:54:37
longhorn/longhorn
https://api.github.com/repos/longhorn/longhorn
opened
[TEST][FEATURE] SPDK volumes support volume backup/restore
kind/test
adding/updating auto e2e test cases for https://github.com/longhorn/longhorn/issues/6138 if they can be automated cc @longhorn/qa
1.0
[TEST][FEATURE] SPDK volumes support volume backup/restore - adding/updating auto e2e test cases for https://github.com/longhorn/longhorn/issues/6138 if they can be automated cc @longhorn/qa
test
spdk volumes support volume backup restore adding updating auto test cases for if they can be automated cc longhorn qa
1
1,898
2,578,422,325
IssuesEvent
2015-02-12 23:08:06
Elgg/Elgg
https://api.github.com/repos/Elgg/Elgg
closed
Running unit tests using Developer Tools
tests
It looks like we are shipping the distribution without dev dependencies included, so unit tests won't run unless dependencies are installed using composer. Perhaps worth noting this somewhere.
1.0
Running unit tests using Developer Tools - It looks like we are shipping the distribution without dev dependencies included, so unit tests won't run unless dependencies are installed using composer. Perhaps worth noting this somewhere.
test
running unit tests using developer tools it looks like we are shipping the distribution without dev dependencies included so unit tests won t run unless dependencies are installed using composer perhaps worth noting this somewhere
1
230,854
18,719,852,325
IssuesEvent
2021-11-03 10:30:33
WordPress/gutenberg
https://api.github.com/repos/WordPress/gutenberg
closed
Cover Block min height unit "vh" missing / only "vw" selectable
[Type] Bug Needs Testing [Block] Cover
### Description I have noticed that you cannot select "vh" as unit for the min-height block setting of cover blocks, only "vh" (besides px etc.) ![image](https://user-images.githubusercontent.com/25924499/139244184-6c1ae634-79e8-4453-9dc7-ebda0485448c.png) ![image](https://user-images.githubusercontent.com/25924499/139244256-1a90e621-581f-431f-862f-30e035eb558d.png) ### Step-by-step reproduction instructions - Editor view - Insert Cover Block - Select unit list for min height ### Screenshots, screen recording, code snippet _No response_ ### Environment info WordPress 5.8.1 Gutenberg Version 11.8.0 (Plugin) ### Please confirm that you have searched existing issues in the repo. Yes ### Please confirm that you have tested with all plugins deactivated except Gutenberg. Yes
1.0
Cover Block min height unit "vh" missing / only "vw" selectable - ### Description I have noticed that you cannot select "vh" as unit for the min-height block setting of cover blocks, only "vh" (besides px etc.) ![image](https://user-images.githubusercontent.com/25924499/139244184-6c1ae634-79e8-4453-9dc7-ebda0485448c.png) ![image](https://user-images.githubusercontent.com/25924499/139244256-1a90e621-581f-431f-862f-30e035eb558d.png) ### Step-by-step reproduction instructions - Editor view - Insert Cover Block - Select unit list for min height ### Screenshots, screen recording, code snippet _No response_ ### Environment info WordPress 5.8.1 Gutenberg Version 11.8.0 (Plugin) ### Please confirm that you have searched existing issues in the repo. Yes ### Please confirm that you have tested with all plugins deactivated except Gutenberg. Yes
test
cover block min height unit vh missing only vw selectable description i have noticed that you cannot select vh as unit for the min height block setting of cover blocks only vh besides px etc step by step reproduction instructions editor view insert cover block select unit list for min height screenshots screen recording code snippet no response environment info wordpress gutenberg version plugin please confirm that you have searched existing issues in the repo yes please confirm that you have tested with all plugins deactivated except gutenberg yes
1
265,225
20,074,203,099
IssuesEvent
2022-02-04 10:47:41
HWR-Berlin-SWE-II-Gruppe-2-Team-3-2022/Dokumentation
https://api.github.com/repos/HWR-Berlin-SWE-II-Gruppe-2-Team-3-2022/Dokumentation
closed
[Documentation] Textueller Projekt-Teaser
Documentation
## Detailed Description Projekt-Teaser in Form einer textuellen Kurzbeschreibung des Projektthemas (200-250 Wörter), Abgabe eines Dokuments im PDF-Format. (5 / 100 Punkte) ### Tasks - [x] Wortanzahl erreichen - [x] "Corporate Design" etablieren - [x] Logo anpassen
1.0
[Documentation] Textueller Projekt-Teaser - ## Detailed Description Projekt-Teaser in Form einer textuellen Kurzbeschreibung des Projektthemas (200-250 Wörter), Abgabe eines Dokuments im PDF-Format. (5 / 100 Punkte) ### Tasks - [x] Wortanzahl erreichen - [x] "Corporate Design" etablieren - [x] Logo anpassen
non_test
textueller projekt teaser detailed description projekt teaser in form einer textuellen kurzbeschreibung des projektthemas wörter abgabe eines dokuments im pdf format punkte tasks wortanzahl erreichen corporate design etablieren logo anpassen
0
163,204
12,707,279,524
IssuesEvent
2020-06-23 08:39:40
wazuh/wazuh-kibana-app
https://api.github.com/repos/wazuh/wazuh-kibana-app
closed
Testing WUI 3.13.0 - 7.7.0/7.7.1
testing
All the tests must be done on every common browser and using the Open Distro plugin. - [x] Chrome - [x] Firefox - [x] Safari
1.0
Testing WUI 3.13.0 - 7.7.0/7.7.1 - All the tests must be done on every common browser and using the Open Distro plugin. - [x] Chrome - [x] Firefox - [x] Safari
test
testing wui all the tests must be done on every common browser and using the open distro plugin chrome firefox safari
1
230,662
18,681,839,874
IssuesEvent
2021-11-01 07:09:12
RasaHQ/rasa
https://api.github.com/repos/RasaHQ/rasa
opened
Scheduled Model Regression Test Failed
type:bug :bug: tool:model-regression-tests
*This PR is automatically created by the Scheduled Model Regression Test workflow. Checkout the Github Action Run [here](https://github.com/RasaHQ/rasa/actions/runs/1405766870).* <br> --- <br> **Description of Problem:** <br> Scheduled Model Regression Test failed. <br> **Configuration**: `Rules + TED` <br> **Dataset**: `retail-demo`
1.0
Scheduled Model Regression Test Failed - *This PR is automatically created by the Scheduled Model Regression Test workflow. Checkout the Github Action Run [here](https://github.com/RasaHQ/rasa/actions/runs/1405766870).* <br> --- <br> **Description of Problem:** <br> Scheduled Model Regression Test failed. <br> **Configuration**: `Rules + TED` <br> **Dataset**: `retail-demo`
test
scheduled model regression test failed this pr is automatically created by the scheduled model regression test workflow checkout the github action run description of problem scheduled model regression test failed configuration rules ted dataset retail demo
1
642,257
20,871,923,513
IssuesEvent
2022-03-22 12:46:49
datavisyn/tdp_core
https://api.github.com/repos/datavisyn/tdp_core
closed
Add checksum to score states
type: feature priority: low
In case a score is restored using the score parameters stored in the provenance state, it would be useful to store a checksum that tells the user if the re-calculated scores are the same as the original ones. When the data in the database changes, this will not the case.
1.0
Add checksum to score states - In case a score is restored using the score parameters stored in the provenance state, it would be useful to store a checksum that tells the user if the re-calculated scores are the same as the original ones. When the data in the database changes, this will not the case.
non_test
add checksum to score states in case a score is restored using the score parameters stored in the provenance state it would be useful to store a checksum that tells the user if the re calculated scores are the same as the original ones when the data in the database changes this will not the case
0
19,063
3,749,579,549
IssuesEvent
2016-03-11 00:40:14
Microsoft/RTVS
https://api.github.com/repos/Microsoft/RTVS
closed
Need better output for unit-tests in case of RHost start failure
area:REPL-Backend type:enhancement type:test issue
Currently it isn't visible in test output and require debugging
1.0
Need better output for unit-tests in case of RHost start failure - Currently it isn't visible in test output and require debugging
test
need better output for unit tests in case of rhost start failure currently it isn t visible in test output and require debugging
1
14,074
3,374,045,686
IssuesEvent
2015-11-24 10:59:57
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
opened
ISISPowderDiffractionPolarisTest: add a test case with ExistingV=yes
Component: Diffraction Group: Powder Quality: System Tests
The test added in #14534 checks the longer workflow (with ExistingV=no) with simple calibration files. We need to check with Ron what would be good calibration files from the last or a recent cycle to use for a test with `ExistingV=no` (that is, assuming a first calibration has been done at the beginning of the cycle and its results are ready to be re-used).
1.0
ISISPowderDiffractionPolarisTest: add a test case with ExistingV=yes - The test added in #14534 checks the longer workflow (with ExistingV=no) with simple calibration files. We need to check with Ron what would be good calibration files from the last or a recent cycle to use for a test with `ExistingV=no` (that is, assuming a first calibration has been done at the beginning of the cycle and its results are ready to be re-used).
test
isispowderdiffractionpolaristest add a test case with existingv yes the test added in checks the longer workflow with existingv no with simple calibration files we need to check with ron what would be good calibration files from the last or a recent cycle to use for a test with existingv no that is assuming a first calibration has been done at the beginning of the cycle and its results are ready to be re used
1
176,350
13,637,642,812
IssuesEvent
2020-09-25 08:10:11
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
opened
tests: arch: arm: arm_thread_swap: fails with bus fault
area: Tests bug
**Describe the bug** tests/arch/arm/arm_thread_swap/ fails with bus fault on nucleo_l552ze_q **To Reproduce** Steps to reproduce the behavior: 1. sanitycheck -N --device-testing --hardware-map ../map.yaml -c -p nucleo_l552ze_q -T tests/arch/arm/arm_thread_swap/ 2. See error **Expected behavior** test passed **Logs and console output** *** Booting Zephyr OS build v2.4.0-rc2-80-g1c3659519f69 *** ``` Running test suite arm_thread_swap =================================================================== START - test_arm_thread_swap E: ***** BUS FAULT ***** E: Precise data bus error E: BFAR Address: 0x0 E: r0/a1: 0x20000334 r1/a2: 0x00000000 r2/a3: 0x00000040 E: r3/a4: 0x20000334 r12/ip: 0x00000000 r14/lr: 0x08001705 E: xpsr: 0x81000200 E: s[ 0]: 0x00000000 s[ 1]: 0x00000000 s[ 2]: 0x00000000 s[ 3]: 0x00000000 E: s[ 4]: 0x00000000 s[ 5]: 0x00000000 s[ 6]: 0x00000000 s[ 7]: 0x00000000 E: s[ 8]: 0x00000000 s[ 9]: 0x00000000 s[10]: 0x00000000 s[11]: 0x00000000 E: s[12]: 0x00000000 s[13]: 0x00000000 s[14]: 0x00000000 s[15]: 0x00000000 E: fpscr: 0x00000000 E: Faulting instruction address (r15/pc): 0x0800c104 E: >>> ZEPHYR FATAL ERROR 0: CPU exception on CPU 0 E: Current thread: 0x20000280 (unknown) E: Halting system ``` **Environment (please complete the following information):** - OS: Linux - Toolchain Zephyr SDK - Commit SHA 1c3659519f69e5f121b9f90b23b89a6cbbd26f8e
1.0
tests: arch: arm: arm_thread_swap: fails with bus fault - **Describe the bug** tests/arch/arm/arm_thread_swap/ fails with bus fault on nucleo_l552ze_q **To Reproduce** Steps to reproduce the behavior: 1. sanitycheck -N --device-testing --hardware-map ../map.yaml -c -p nucleo_l552ze_q -T tests/arch/arm/arm_thread_swap/ 2. See error **Expected behavior** test passed **Logs and console output** *** Booting Zephyr OS build v2.4.0-rc2-80-g1c3659519f69 *** ``` Running test suite arm_thread_swap =================================================================== START - test_arm_thread_swap E: ***** BUS FAULT ***** E: Precise data bus error E: BFAR Address: 0x0 E: r0/a1: 0x20000334 r1/a2: 0x00000000 r2/a3: 0x00000040 E: r3/a4: 0x20000334 r12/ip: 0x00000000 r14/lr: 0x08001705 E: xpsr: 0x81000200 E: s[ 0]: 0x00000000 s[ 1]: 0x00000000 s[ 2]: 0x00000000 s[ 3]: 0x00000000 E: s[ 4]: 0x00000000 s[ 5]: 0x00000000 s[ 6]: 0x00000000 s[ 7]: 0x00000000 E: s[ 8]: 0x00000000 s[ 9]: 0x00000000 s[10]: 0x00000000 s[11]: 0x00000000 E: s[12]: 0x00000000 s[13]: 0x00000000 s[14]: 0x00000000 s[15]: 0x00000000 E: fpscr: 0x00000000 E: Faulting instruction address (r15/pc): 0x0800c104 E: >>> ZEPHYR FATAL ERROR 0: CPU exception on CPU 0 E: Current thread: 0x20000280 (unknown) E: Halting system ``` **Environment (please complete the following information):** - OS: Linux - Toolchain Zephyr SDK - Commit SHA 1c3659519f69e5f121b9f90b23b89a6cbbd26f8e
test
tests arch arm arm thread swap fails with bus fault describe the bug tests arch arm arm thread swap fails with bus fault on nucleo q to reproduce steps to reproduce the behavior sanitycheck n device testing hardware map map yaml c p nucleo q t tests arch arm arm thread swap see error expected behavior test passed logs and console output booting zephyr os build running test suite arm thread swap start test arm thread swap e bus fault e precise data bus error e bfar address e e ip lr e xpsr e s s s s e s s s s e s s s s e s s s s e fpscr e faulting instruction address pc e zephyr fatal error cpu exception on cpu e current thread unknown e halting system environment please complete the following information os linux toolchain zephyr sdk commit sha
1
762,138
26,709,895,996
IssuesEvent
2023-01-27 22:13:49
UTDNebula/planner
https://api.github.com/repos/UTDNebula/planner
closed
[Feature Request] Taking into Account Course Difficulty When Generating Degree Plans
Type: Feature Request Category: User Experience Status: On Hold Product Priority: Low Size: 8
# Overview In cooperation with Trends, Planner could create degree plans that take into account/visualize course difficulty. A star rating could be shown under each course to show relative difficulty in order to assist the user further when deciding which classes to take at certain times. Further down the line, it would be nice if when Planner generates a degree plan, it adds courses in a way that the average difficulty of the semester is relatively close to each other. This would allow users who want to add a minor/extra courses later on to see the best opportunity to take them. # Proposal First would be working with trends to track a relative difficulty for a course (using grade distribution, professor grades, etc.) Then assigning a value to each course would be necessary. Following that, displaying the difficulty would be a start and then taking an average difficulty for a semester and displaying to the user would also be helpful. Later implementations can begin messing around with the generation of degree plans in a way that the remaining courses for someone is organized such that the difficulty is not outrageously different from semester to semester. ## Supplementary Information Feel free to respond to this for more context.
1.0
[Feature Request] Taking into Account Course Difficulty When Generating Degree Plans - # Overview In cooperation with Trends, Planner could create degree plans that take into account/visualize course difficulty. A star rating could be shown under each course to show relative difficulty in order to assist the user further when deciding which classes to take at certain times. Further down the line, it would be nice if when Planner generates a degree plan, it adds courses in a way that the average difficulty of the semester is relatively close to each other. This would allow users who want to add a minor/extra courses later on to see the best opportunity to take them. # Proposal First would be working with trends to track a relative difficulty for a course (using grade distribution, professor grades, etc.) Then assigning a value to each course would be necessary. Following that, displaying the difficulty would be a start and then taking an average difficulty for a semester and displaying to the user would also be helpful. Later implementations can begin messing around with the generation of degree plans in a way that the remaining courses for someone is organized such that the difficulty is not outrageously different from semester to semester. ## Supplementary Information Feel free to respond to this for more context.
non_test
taking into account course difficulty when generating degree plans overview in cooperation with trends planner could create degree plans that take into account visualize course difficulty a star rating could be shown under each course to show relative difficulty in order to assist the user further when deciding which classes to take at certain times further down the line it would be nice if when planner generates a degree plan it adds courses in a way that the average difficulty of the semester is relatively close to each other this would allow users who want to add a minor extra courses later on to see the best opportunity to take them proposal first would be working with trends to track a relative difficulty for a course using grade distribution professor grades etc then assigning a value to each course would be necessary following that displaying the difficulty would be a start and then taking an average difficulty for a semester and displaying to the user would also be helpful later implementations can begin messing around with the generation of degree plans in a way that the remaining courses for someone is organized such that the difficulty is not outrageously different from semester to semester supplementary information feel free to respond to this for more context
0
336,470
30,195,477,313
IssuesEvent
2023-07-04 20:22:35
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: copy/bank/rows=100000,nodes=5,txn=true failed
C-test-failure O-robot A-testing O-roachtest T-kv branch-release-23.1
roachtest.copy/bank/rows=100000,nodes=5,txn=true [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/10756610?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/10756610?buildTab=artifacts#/copy/bank/rows=100000,nodes=5,txn=true) on release-23.1 @ [e12e85479312972b551677203849d29aeb38ad5f](https://github.com/cockroachdb/cockroach/commits/e12e85479312972b551677203849d29aeb38ad5f): ``` (cluster.go:2247).Run: output in run_063348.848455572_n1_workload-fixtures-lo: ./workload fixtures load bank --rows=100000 --payload-bytes=100 --seed 1 {pgurl:1} returned: COMMAND_PROBLEM: exit status 1 (monitor.go:137).Wait: monitor failure: monitor task failed: t.Fatal() was called test artifacts and logs in: /artifacts/copy/bank/rows=100000_nodes=5_txn=true/run_1 ``` <p>Parameters: <code>ROACHTEST_arch=amd64</code> , <code>ROACHTEST_cloud=aws</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*copy/bank/rows=100000,nodes=5,txn=true.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-29307
3.0
roachtest: copy/bank/rows=100000,nodes=5,txn=true failed - roachtest.copy/bank/rows=100000,nodes=5,txn=true [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/10756610?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/10756610?buildTab=artifacts#/copy/bank/rows=100000,nodes=5,txn=true) on release-23.1 @ [e12e85479312972b551677203849d29aeb38ad5f](https://github.com/cockroachdb/cockroach/commits/e12e85479312972b551677203849d29aeb38ad5f): ``` (cluster.go:2247).Run: output in run_063348.848455572_n1_workload-fixtures-lo: ./workload fixtures load bank --rows=100000 --payload-bytes=100 --seed 1 {pgurl:1} returned: COMMAND_PROBLEM: exit status 1 (monitor.go:137).Wait: monitor failure: monitor task failed: t.Fatal() was called test artifacts and logs in: /artifacts/copy/bank/rows=100000_nodes=5_txn=true/run_1 ``` <p>Parameters: <code>ROACHTEST_arch=amd64</code> , <code>ROACHTEST_cloud=aws</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*copy/bank/rows=100000,nodes=5,txn=true.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-29307
test
roachtest copy bank rows nodes txn true failed roachtest copy bank rows nodes txn true with on release cluster go run output in run workload fixtures lo workload fixtures load bank rows payload bytes seed pgurl returned command problem exit status monitor go wait monitor failure monitor task failed t fatal was called test artifacts and logs in artifacts copy bank rows nodes txn true run parameters roachtest arch roachtest cloud aws roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb kv triage jira issue crdb
1
236,986
7,754,586,259
IssuesEvent
2018-05-31 07:19:10
molgenis/molgenis
https://api.github.com/repos/molgenis/molgenis
closed
Internal server error when idAttribute of refEntity is visible == false and no other labelAttribute specified
2.1.0 4.1 5.2 7.0.0-RC bug mod:data-importer mod:dataexplorer priority-first
### How to Reproduce Create EMX model with one entity referencing to another, where idAttribute of the reference entity is set on visible == false (and no other attribute is specified as labelAttribute) ### Expected behavior The importer will give a warning or error ### Observed behavior The EMX is correctly uploaded, but you get an internal server error 500 when selecting the entity -Marieke has an example if this is not clear enough ;-)
1.0
Internal server error when idAttribute of refEntity is visible == false and no other labelAttribute specified - ### How to Reproduce Create EMX model with one entity referencing to another, where idAttribute of the reference entity is set on visible == false (and no other attribute is specified as labelAttribute) ### Expected behavior The importer will give a warning or error ### Observed behavior The EMX is correctly uploaded, but you get an internal server error 500 when selecting the entity -Marieke has an example if this is not clear enough ;-)
non_test
internal server error when idattribute of refentity is visible false and no other labelattribute specified how to reproduce create emx model with one entity referencing to another where idattribute of the reference entity is set on visible false and no other attribute is specified as labelattribute expected behavior the importer will give a warning or error observed behavior the emx is correctly uploaded but you get an internal server error when selecting the entity marieke has an example if this is not clear enough
0
11,714
14,546,541,572
IssuesEvent
2020-12-15 21:24:17
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
No Support for Generic Environment Resources
Pri2 devops-cicd-process/tech devops/prod product-feedback
I see that Azure DevOps yaml supports Kubernetes and Virtual Machines for deployment tracability in the form of "resources". Is there a reason there doesnt just exists a "generic" resource that is created if the resource is referenced through yml but doesnt exists yet? The same as environments are managed. That way jobs and commits will be recorded by resource based on the name of the resources. Example: ` environment: environment_name.resource_name ` And the environment **environment_name** is created and the "generic" resource is created **resource_name** --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 77d95db6-9983-7346-d0eb-4b7443e4e252 * Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087 * Content: [Environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops) * Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
No Support for Generic Environment Resources - I see that Azure DevOps yaml supports Kubernetes and Virtual Machines for deployment tracability in the form of "resources". Is there a reason there doesnt just exists a "generic" resource that is created if the resource is referenced through yml but doesnt exists yet? The same as environments are managed. That way jobs and commits will be recorded by resource based on the name of the resources. Example: ` environment: environment_name.resource_name ` And the environment **environment_name** is created and the "generic" resource is created **resource_name** --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 77d95db6-9983-7346-d0eb-4b7443e4e252 * Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087 * Content: [Environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops) * Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
non_test
no support for generic environment resources i see that azure devops yaml supports kubernetes and virtual machines for deployment tracability in the form of resources is there a reason there doesnt just exists a generic resource that is created if the resource is referenced through yml but doesnt exists yet the same as environments are managed that way jobs and commits will be recorded by resource based on the name of the resources example environment environment name resource name and the environment environment name is created and the generic resource is created resource name document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
0
249,000
21,093,293,186
IssuesEvent
2022-04-04 07:57:44
blumilksoftware/meetup
https://api.github.com/repos/blumilksoftware/meetup
opened
Initialize Behat
environment feature tests
With Behat tests initially provided in #45, it would be nice to have a way to run these scenarios. Behat should be added with newest version. Contexts should be properly configured and ready to be extended. The aim of this task is to setup Behat and is not to have all tests green.
1.0
Initialize Behat - With Behat tests initially provided in #45, it would be nice to have a way to run these scenarios. Behat should be added with newest version. Contexts should be properly configured and ready to be extended. The aim of this task is to setup Behat and is not to have all tests green.
test
initialize behat with behat tests initially provided in it would be nice to have a way to run these scenarios behat should be added with newest version contexts should be properly configured and ready to be extended the aim of this task is to setup behat and is not to have all tests green
1