Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
450,491
31,926,738,244
IssuesEvent
2023-09-19 02:46:44
apmwebdev/spelling-bee
https://api.github.com/repos/apmwebdev/spelling-bee
opened
Add basic technical documentation
documentation
In addition to the readme and in-app help, there should be some basic technical documentation about the structure and features of the app. This is especially important for hint profiles/panels, as there are different hint panel types and unique settings for each type.
1.0
Add basic technical documentation - In addition to the readme and in-app help, there should be some basic technical documentation about the structure and features of the app. This is especially important for hint profiles/panels, as there are different hint panel types and unique settings for each type.
non_defect
add basic technical documentation in addition to the readme and in app help there should be some basic technical documentation about the structure and features of the app this is especially important for hint profiles panels as there are different hint panel types and unique settings for each type
0
591,671
17,858,409,074
IssuesEvent
2021-09-05 13:42:00
francheska-vicente/cssweng
https://api.github.com/repos/francheska-vicente/cssweng
closed
Reservation that was booked should be removed from the reservation list
bug priority: medium issue: back-end severity: high
### Summary: - After booking a reservation, it should be removed from the reservation list. ### Steps to Reproduce: 1. Create a Booking from a Reservation 2. Go to the Main Reservation page ### Visual Proof: ![image](https://user-images.githubusercontent.com/75743382/131985958-f114a503-6003-4a07-b6e6-ce7703310738.png) ### Expected Results: - Reservation is removed from the list of reservations ### Actual Results: - Reservation is still there | Additional Information | | | ----------- | ----------- | | Platform | V8 engine (Google) | | Operating System | Windows 10 |
1.0
Reservation that was booked should be removed from the reservation list - ### Summary: - After booking a reservation, it should be removed from the reservation list. ### Steps to Reproduce: 1. Create a Booking from a Reservation 2. Go to the Main Reservation page ### Visual Proof: ![image](https://user-images.githubusercontent.com/75743382/131985958-f114a503-6003-4a07-b6e6-ce7703310738.png) ### Expected Results: - Reservation is removed from the list of reservations ### Actual Results: - Reservation is still there | Additional Information | | | ----------- | ----------- | | Platform | V8 engine (Google) | | Operating System | Windows 10 |
non_defect
reservation that was booked should be removed from the reservation list summary after booking a reservation it should be removed from the reservation list steps to reproduce create a booking from a reservation go to the main reservation page visual proof expected results reservation is removed from the list of reservations actual results reservation is still there additional information platform engine google operating system windows
0
148,740
13,246,101,177
IssuesEvent
2020-08-19 15:14:45
nearform/mira
https://api.github.com/repos/nearform/mira
closed
Generate API Docs
documentation
**Is your feature request related to a problem? Please describe.** It can be hard to see the scope of the Mira API and the elements involved without API docs. **Describe the solution you'd like** We should publish the TSDoc comments to a browsable format. We will use https://typedoc.org/ to generate this documentation, e.g. `typedoc --out docs --name 'src'`. We can't use the current `docs` root as per this example though, due to the existing content. Therefore, ideally we need to find a way to include the API docs within that tree and thereby have the api docs included in the docset on netlify. **Describe alternatives you've considered** See #36. If it is not possible to include the API docs in the main doc set then we should take a quick look at alternative tools. If all else fails then we should just make it easy for the developer to generate the api docs locally. **Additional context** This follows on from issue #36 , to add appropriate comments to the code
1.0
Generate API Docs - **Is your feature request related to a problem? Please describe.** It can be hard to see the scope of the Mira API and the elements involved without API docs. **Describe the solution you'd like** We should publish the TSDoc comments to a browsable format. We will use https://typedoc.org/ to generate this documentation, e.g. `typedoc --out docs --name 'src'`. We can't use the current `docs` root as per this example though, due to the existing content. Therefore, ideally we need to find a way to include the API docs within that tree and thereby have the api docs included in the docset on netlify. **Describe alternatives you've considered** See #36. If it is not possible to include the API docs in the main doc set then we should take a quick look at alternative tools. If all else fails then we should just make it easy for the developer to generate the api docs locally. **Additional context** This follows on from issue #36 , to add appropriate comments to the code
non_defect
generate api docs is your feature request related to a problem please describe it can be hard to see the scope of the mira api and the elements involved without api docs describe the solution you d like we should publish the tsdoc comments to a browsable format we will use to generate this documentation e g typedoc out docs name src we can t use the current docs root as per this example though due to the existing content therefore ideally we need to find a way to include the api docs within that tree and thereby have the api docs included in the docset on netlify describe alternatives you ve considered see if it is not possible to include the api docs in the main doc set then we should take a quick look at alternative tools if all else fails then we should just make it easy for the developer to generate the api docs locally additional context this follows on from issue to add appropriate comments to the code
0
60,897
17,023,552,088
IssuesEvent
2021-07-03 02:36:34
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
"yes HSE" in search results
Component: nominatim Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 7.27pm, Friday, 12th February 2010]** While search for "HSE, Dublin Road, Kilkenny, County Kilkenny, Leinster, Ireland" you can get a strangely formated result.. That has a "yes" in the beginning. Here are some results I got by searching for "h" ``` Yes HSE, Dublin Road, Kilkenny, County Kilkenny, Leinster, Ireland Yes Hoe, Pursers Lane, Peaslake, Surrey, South East England, United Kingdom Yes East House, Newbiggin, County Durham, North East England, United Kingdom Yes house, Prizren, Komuna e Prizrenit, Serbia ```
1.0
"yes HSE" in search results - **[Submitted to the original trac issue database at 7.27pm, Friday, 12th February 2010]** While search for "HSE, Dublin Road, Kilkenny, County Kilkenny, Leinster, Ireland" you can get a strangely formated result.. That has a "yes" in the beginning. Here are some results I got by searching for "h" ``` Yes HSE, Dublin Road, Kilkenny, County Kilkenny, Leinster, Ireland Yes Hoe, Pursers Lane, Peaslake, Surrey, South East England, United Kingdom Yes East House, Newbiggin, County Durham, North East England, United Kingdom Yes house, Prizren, Komuna e Prizrenit, Serbia ```
defect
yes hse in search results while search for hse dublin road kilkenny county kilkenny leinster ireland you can get a strangely formated result that has a yes in the beginning here are some results i got by searching for h yes hse dublin road kilkenny county kilkenny leinster ireland yes hoe pursers lane peaslake surrey south east england united kingdom yes east house newbiggin county durham north east england united kingdom yes house prizren komuna e prizrenit serbia
1
6,186
2,610,222,904
IssuesEvent
2015-02-26 19:10:44
chrsmith/somefinders
https://api.github.com/repos/chrsmith/somefinders
opened
rld.dll файл для mega armpes 2013.txt
auto-migrated Priority-Medium Type-Defect
``` '''Геодар Савельев''' Привет всем не подскажите где можно найти .rld.dll файл для mega armpes 2013.txt. как то выкладывали уже '''Гораций Молчанов''' Качай тут http://bit.ly/16QsD3X '''Вальтер Кулаков''' Просит ввести номер мобилы!Не опасно ли это? '''Баграт Крылов''' Неа все ок у меня ничего не списало '''Вольмир Калашников''' Неа все ок у меня ничего не списало Информация о файле: rld.dll файл для mega armpes 2013.txt Загружен: В этом месяце Скачан раз: 262 Рейтинг: 850 Средняя скорость скачивания: 1178 Похожих файлов: 34 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 18 Dec 2013 at 5:16
1.0
rld.dll файл для mega armpes 2013.txt - ``` '''Геодар Савельев''' Привет всем не подскажите где можно найти .rld.dll файл для mega armpes 2013.txt. как то выкладывали уже '''Гораций Молчанов''' Качай тут http://bit.ly/16QsD3X '''Вальтер Кулаков''' Просит ввести номер мобилы!Не опасно ли это? '''Баграт Крылов''' Неа все ок у меня ничего не списало '''Вольмир Калашников''' Неа все ок у меня ничего не списало Информация о файле: rld.dll файл для mega armpes 2013.txt Загружен: В этом месяце Скачан раз: 262 Рейтинг: 850 Средняя скорость скачивания: 1178 Похожих файлов: 34 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 18 Dec 2013 at 5:16
defect
rld dll файл для mega armpes txt геодар савельев привет всем не подскажите где можно найти rld dll файл для mega armpes txt как то выкладывали уже гораций молчанов качай тут вальтер кулаков просит ввести номер мобилы не опасно ли это баграт крылов неа все ок у меня ничего не списало вольмир калашников неа все ок у меня ничего не списало информация о файле rld dll файл для mega armpes txt загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at
1
421,048
28,307,802,100
IssuesEvent
2023-04-10 12:48:20
bounswe/bounswe2023group6
https://api.github.com/repos/bounswe/bounswe2023group6
opened
Forming an individual report for milestone 1
type: documentation priority: high status: inprogress area: milestone
### Problem As a part of the milestone report 1, each team member is required to create an Individual Contribution Report. ### Solution I will create my individual contribution report . The contribution report will include: - Member: Info about myself (name, group) - Responsibilities: Overall description of the responsibilities that are assigned to me - Main contributions: Overall description of my contributions to the project until - Milestone. Provide the links to your work and related issues in your project repository, - Management related significant issues: My issues that contribute to the management of our software project (Please only lists the most-significant ones, we DO track all your issues separately.) - Additional information (optional): Further relevant information related to my contributions ### Documentation _No response_ ### Additional notes _No response_ ### Reviewers _No response_ ### Deadline 10.04.2023 - Monday - 23.59
1.0
Forming an individual report for milestone 1 - ### Problem As a part of the milestone report 1, each team member is required to create an Individual Contribution Report. ### Solution I will create my individual contribution report . The contribution report will include: - Member: Info about myself (name, group) - Responsibilities: Overall description of the responsibilities that are assigned to me - Main contributions: Overall description of my contributions to the project until - Milestone. Provide the links to your work and related issues in your project repository, - Management related significant issues: My issues that contribute to the management of our software project (Please only lists the most-significant ones, we DO track all your issues separately.) - Additional information (optional): Further relevant information related to my contributions ### Documentation _No response_ ### Additional notes _No response_ ### Reviewers _No response_ ### Deadline 10.04.2023 - Monday - 23.59
non_defect
forming an individual report for milestone problem as a part of the milestone report each team member is required to create an individual contribution report solution i will create my individual contribution report the contribution report will include member info about myself name group responsibilities overall description of the responsibilities that are assigned to me main contributions overall description of my contributions to the project until milestone provide the links to your work and related issues in your project repository management related significant issues my issues that contribute to the management of our software project please only lists the most significant ones we do track all your issues separately additional information optional further relevant information related to my contributions documentation no response additional notes no response reviewers no response deadline monday
0
65,475
19,534,969,200
IssuesEvent
2021-12-31 03:32:47
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
opened
[🐛 Bug]:
I-defect needs-triaging
### What happened? # Getting `Error: ECONNRESET read` while trying to connect to a Selenium Standalone running on a Docker container on Windows 10 host. I am running a Selenium Standalone Server as a Docker container on my Windows 10 host. On my host, I have a Node.js script that uses selenium-webdriver package to connect to this Selenium Server container, but the script constantly returns this error: ``` ECONNRESET read ECONNRESET at ClientRequest.<anonymous> (C:\path-to-my-script\node_modules\selenium-webdriver\http\index.js:297:15) at ClientRequest.emit (events.js:400:28) at Socket.socketErrorListener (_http_client.js:475:9) at Socket.emit (events.js:400:28) at emitErrorNT (internal/streams/destroy.js:106:8) at emitErrorCloseNT (internal/streams/destroy.js:74:3) at processTicksAndRejections (internal/process/task_queues.js:82:21) ``` *Please note that I've added a new line of `console.log(e)` after index.js:279 that makes the line numbers different than the original library. For your information, the output of that line shows:* ``` Error: read ECONNRESET at TCP.onStreamRead (internal/stream_base_commons.js:209:20) { errno: -4077, code: 'ECONNRESET', syscall: 'read' } ``` ## A little Background about my experiments on this issue: ### A brief quote about the meaning behind this error: > "ECONNRESET" means the other side of the TCP conversation abruptly closed its end of the connection. ([citation link](https://www.quora.com/What-does-%E2%80%9CError-read-ECONNRESET%E2%80%9D-mean)) ### Using a different Webdriver (PHP Webdriver): When I use PHP Webdriver from (https://github.com/php-webdriver/php-webdriver), everything works perfectly fine. A new instance of a Firefox browser runs and navigates through the URL I want fully-functional, but when I use this library on Node.js, I only get this error, and the Selenium Server does react to anything (The library cannot make this request, therefore throws this ECONNRESET exception). ### Running this Selenium Docker container on a Linux CentOS 7 host: When I run this Node.js script to connect to the Selenium Standalone Docker container on a CentOS 7 virtual-machine host, everything works perfectly fine. Therefore, it seems to be a problem with this library communicating with Windows [WSL2](https://docs.microsoft.com/en-us/windows/wsl/about). ## My Node.js Code: ``` const {Builder, By} = require('selenium-webdriver'); (function run(){ try { let driver = await new Builder() .forBrowser('firefox') .usingServer(`http://localhost:4444/wd/hub`) .build(); console.log("Done"); } catch(e) { console.log(e); } await driver.get('https://google.com'); })(); ``` ## My Selenium Standalone bash script: `java -jar selenium-server-4.0.0.jar standalone --override-max-sessions true --max-sessions 20 --session-timeout 1800` ## My Dockerfile: *Link: https://github.com/iliaamiri/selenium-on-Linux-setup-dependencies/blob/main/Dockerfile* ## My scripts to run this Docker Container ### Script for Linux host: `docker run -it -p 4444:4444 -e DISPLAY=$DISPLAY -v /tmp/.X11-unix:/tmp/.X11-unix redbolder/selenium_standalone_firefox_centos` ### Script for Windows host: `docker run -it -p 4445:4444 -e DISPLAY=192.168.1.2:0.0 -v c:/tmp/X11-unix:/tmp/.X11-unix redbolder/selenium_standalone_firefox_centos` ## Every information about the technologies I have mentioned above: **Selenium Standalone Version:** 4.0.0 **Java Version:** JRE 8u311 Linux x64 **Linux Virtual Host:** CentOS 7 (Core) **Docker on Linux Virtual Host:** Docker Engine - Community - 20.10.10 **Docker on Windows Host:** Docker Engine - Community - 20.10.11 **Windows Host:** Windows 10 Pro **Docker Container Operation System:** CentOS 8 **Node.js Version:** 14.17.6 ### How can we reproduce the issue? ```shell Please refer to this Github Repository: https://github.com/iliaamiri/selenium-webdriver-javascript-ECONNRESET-read-issue ``` ### Relevant log output ```shell Error: ECONNRESET read ECONNRESET at ClientRequest.<anonymous> (C:\xampp\htdocs\Github-Repositories\a-web-bot-for-booking-the-closest-possible-appointment-on-ICBC-Driver-Licensing\node_modules\selenium-webdriver\http\index.js:297:15) at ClientRequest.emit (events.js:400:28) at Socket.socketErrorListener (_http_client.js:475:9) at Socket.emit (events.js:400:28) at emitErrorNT (internal/streams/destroy.js:106:8) at emitErrorCloseNT (internal/streams/destroy.js:74:3) at processTicksAndRejections (internal/process/task_queues.js:82:21) ``` ### Operating System CentOS 8 Docker Container on Windows 10 Host ### Selenium version Java 4.0.0 ### What are the browser(s) and version(s) where you see this issue? Firefox 91.4.0esr ### What are the browser driver(s) and version(s) where you see this issue? geckodriver 0.30.0 ### Are you using Selenium Grid? _No response_
1.0
[🐛 Bug]: - ### What happened? # Getting `Error: ECONNRESET read` while trying to connect to a Selenium Standalone running on a Docker container on Windows 10 host. I am running a Selenium Standalone Server as a Docker container on my Windows 10 host. On my host, I have a Node.js script that uses selenium-webdriver package to connect to this Selenium Server container, but the script constantly returns this error: ``` ECONNRESET read ECONNRESET at ClientRequest.<anonymous> (C:\path-to-my-script\node_modules\selenium-webdriver\http\index.js:297:15) at ClientRequest.emit (events.js:400:28) at Socket.socketErrorListener (_http_client.js:475:9) at Socket.emit (events.js:400:28) at emitErrorNT (internal/streams/destroy.js:106:8) at emitErrorCloseNT (internal/streams/destroy.js:74:3) at processTicksAndRejections (internal/process/task_queues.js:82:21) ``` *Please note that I've added a new line of `console.log(e)` after index.js:279 that makes the line numbers different than the original library. For your information, the output of that line shows:* ``` Error: read ECONNRESET at TCP.onStreamRead (internal/stream_base_commons.js:209:20) { errno: -4077, code: 'ECONNRESET', syscall: 'read' } ``` ## A little Background about my experiments on this issue: ### A brief quote about the meaning behind this error: > "ECONNRESET" means the other side of the TCP conversation abruptly closed its end of the connection. ([citation link](https://www.quora.com/What-does-%E2%80%9CError-read-ECONNRESET%E2%80%9D-mean)) ### Using a different Webdriver (PHP Webdriver): When I use PHP Webdriver from (https://github.com/php-webdriver/php-webdriver), everything works perfectly fine. A new instance of a Firefox browser runs and navigates through the URL I want fully-functional, but when I use this library on Node.js, I only get this error, and the Selenium Server does react to anything (The library cannot make this request, therefore throws this ECONNRESET exception). ### Running this Selenium Docker container on a Linux CentOS 7 host: When I run this Node.js script to connect to the Selenium Standalone Docker container on a CentOS 7 virtual-machine host, everything works perfectly fine. Therefore, it seems to be a problem with this library communicating with Windows [WSL2](https://docs.microsoft.com/en-us/windows/wsl/about). ## My Node.js Code: ``` const {Builder, By} = require('selenium-webdriver'); (function run(){ try { let driver = await new Builder() .forBrowser('firefox') .usingServer(`http://localhost:4444/wd/hub`) .build(); console.log("Done"); } catch(e) { console.log(e); } await driver.get('https://google.com'); })(); ``` ## My Selenium Standalone bash script: `java -jar selenium-server-4.0.0.jar standalone --override-max-sessions true --max-sessions 20 --session-timeout 1800` ## My Dockerfile: *Link: https://github.com/iliaamiri/selenium-on-Linux-setup-dependencies/blob/main/Dockerfile* ## My scripts to run this Docker Container ### Script for Linux host: `docker run -it -p 4444:4444 -e DISPLAY=$DISPLAY -v /tmp/.X11-unix:/tmp/.X11-unix redbolder/selenium_standalone_firefox_centos` ### Script for Windows host: `docker run -it -p 4445:4444 -e DISPLAY=192.168.1.2:0.0 -v c:/tmp/X11-unix:/tmp/.X11-unix redbolder/selenium_standalone_firefox_centos` ## Every information about the technologies I have mentioned above: **Selenium Standalone Version:** 4.0.0 **Java Version:** JRE 8u311 Linux x64 **Linux Virtual Host:** CentOS 7 (Core) **Docker on Linux Virtual Host:** Docker Engine - Community - 20.10.10 **Docker on Windows Host:** Docker Engine - Community - 20.10.11 **Windows Host:** Windows 10 Pro **Docker Container Operation System:** CentOS 8 **Node.js Version:** 14.17.6 ### How can we reproduce the issue? ```shell Please refer to this Github Repository: https://github.com/iliaamiri/selenium-webdriver-javascript-ECONNRESET-read-issue ``` ### Relevant log output ```shell Error: ECONNRESET read ECONNRESET at ClientRequest.<anonymous> (C:\xampp\htdocs\Github-Repositories\a-web-bot-for-booking-the-closest-possible-appointment-on-ICBC-Driver-Licensing\node_modules\selenium-webdriver\http\index.js:297:15) at ClientRequest.emit (events.js:400:28) at Socket.socketErrorListener (_http_client.js:475:9) at Socket.emit (events.js:400:28) at emitErrorNT (internal/streams/destroy.js:106:8) at emitErrorCloseNT (internal/streams/destroy.js:74:3) at processTicksAndRejections (internal/process/task_queues.js:82:21) ``` ### Operating System CentOS 8 Docker Container on Windows 10 Host ### Selenium version Java 4.0.0 ### What are the browser(s) and version(s) where you see this issue? Firefox 91.4.0esr ### What are the browser driver(s) and version(s) where you see this issue? geckodriver 0.30.0 ### Are you using Selenium Grid? _No response_
defect
what happened getting error econnreset read while trying to connect to a selenium standalone running on a docker container on windows host i am running a selenium standalone server as a docker container on my windows host on my host i have a node js script that uses selenium webdriver package to connect to this selenium server container but the script constantly returns this error econnreset read econnreset at clientrequest c path to my script node modules selenium webdriver http index js at clientrequest emit events js at socket socketerrorlistener http client js at socket emit events js at emiterrornt internal streams destroy js at emiterrorclosent internal streams destroy js at processticksandrejections internal process task queues js please note that i ve added a new line of console log e after index js that makes the line numbers different than the original library for your information the output of that line shows error read econnreset at tcp onstreamread internal stream base commons js errno code econnreset syscall read a little background about my experiments on this issue a brief quote about the meaning behind this error econnreset means the other side of the tcp conversation abruptly closed its end of the connection using a different webdriver php webdriver when i use php webdriver from everything works perfectly fine a new instance of a firefox browser runs and navigates through the url i want fully functional but when i use this library on node js i only get this error and the selenium server does react to anything the library cannot make this request therefore throws this econnreset exception running this selenium docker container on a linux centos host when i run this node js script to connect to the selenium standalone docker container on a centos virtual machine host everything works perfectly fine therefore it seems to be a problem with this library communicating with windows my node js code const builder by require selenium webdriver function run try let driver await new builder forbrowser firefox usingserver build console log done catch e console log e await driver get my selenium standalone bash script java jar selenium server jar standalone override max sessions true max sessions session timeout my dockerfile link my scripts to run this docker container script for linux host docker run it p e display display v tmp unix tmp unix redbolder selenium standalone firefox centos script for windows host docker run it p e display v c tmp unix tmp unix redbolder selenium standalone firefox centos every information about the technologies i have mentioned above selenium standalone version java version jre linux linux virtual host centos core docker on linux virtual host docker engine community docker on windows host docker engine community windows host windows pro docker container operation system centos node js version how can we reproduce the issue shell please refer to this github repository relevant log output shell error econnreset read econnreset at clientrequest c xampp htdocs github repositories a web bot for booking the closest possible appointment on icbc driver licensing node modules selenium webdriver http index js at clientrequest emit events js at socket socketerrorlistener http client js at socket emit events js at emiterrornt internal streams destroy js at emiterrorclosent internal streams destroy js at processticksandrejections internal process task queues js operating system centos docker container on windows host selenium version java what are the browser s and version s where you see this issue firefox what are the browser driver s and version s where you see this issue geckodriver are you using selenium grid no response
1
44,983
12,506,007,750
IssuesEvent
2020-06-02 11:48:55
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
QuorumListener is triggered during member startup but before it has joined a cluster
Type: Defect
<!-- Thanks for reporting your issue. Please share with us the following information, to help us resolve your issue quickly and efficiently. --> **Describe the bug** Tested on 3.12.2 which is customer's version. In a cluster with Quorum defined, without touching any maps, just calling `Hazelcast.newHazelcastInstance();` causes `QuorumListener` to be triggered, before the member joins anyone and forms the cluster initially. **Expected behavior** We shouldn’t fire any quorum listener events after a member starts until quorum is met initially. **To Reproduce** Steps to reproduce the behavior: 1. Define a quorum of 3 with a listener attached ``` <quorum enabled="true" name="myquorum"> <quorum-size>3</quorum-size> <quorum-listeners> <quorum-listener>MyQuoLis</quorum-listener> </quorum-listeners> </quorum> ``` 2. Start one member 3. Observe how member logs a quorum event before join: ``` INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Hazelcast 3.12.2 (20190802 - e34b163) starting at [127.0.0.1]:5701 Jun 01, 2020 4:22:40 PM com.hazelcast.system INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Copyright (c) 2008-2019, Hazelcast, Inc. All Rights Reserved. Jun 01, 2020 4:22:40 PM com.hazelcast.spi.impl.operationservice.impl.BackpressureRegulator INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Backpressure is disabled Jun 01, 2020 4:22:40 PM com.hazelcast.instance.Node INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Creating TcpIpJoiner Jun 01, 2020 4:22:40 PM com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Starting 8 partition threads and 5 generic threads (1 dedicated for priority tasks) Jun 01, 2020 4:22:40 PM com.hazelcast.internal.diagnostics.Diagnostics INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Diagnostics disabled. To enable add -Dhazelcast.diagnostics.enabled=true to the JVM arguments. Jun 01, 2020 4:22:40 PM com.hazelcast.core.LifecycleService INFO: [127.0.0.1]:5701 [sertug] [3.12.2] [127.0.0.1]:5701 is STARTING Listener triggered: QuorumEvent{threshold=3, currentMembers=com.hazelcast.internal.cluster.impl.MemberSelectingCollection@1d30e322, presence=false} Jun 01, 2020 4:22:40 PM com.hazelcast.nio.tcp.TcpIpConnector INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Connecting to /127.0.0.1:5703, timeout: 10000, bind-any: true Jun 01, 2020 4:22:40 PM com.hazelcast.nio.tcp.TcpIpConnector INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Connecting to /127.0.0.1:5702, timeout: 10000, bind-any: true Jun 01, 2020 4:22:40 PM com.hazelcast.nio.tcp.TcpIpConnector INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Could not connect to: /127.0.0.1:5702. Reason: SocketException[Connection refused to address /127.0.0.1:5702] Jun 01, 2020 4:22:40 PM com.hazelcast.nio.tcp.TcpIpConnector INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Could not connect to: /127.0.0.1:5703. Reason: SocketException[Connection refused to address /127.0.0.1:5703] Jun 01, 2020 4:22:40 PM com.hazelcast.cluster.impl.TcpIpJoiner INFO: [127.0.0.1]:5701 [sertug] [3.12.2] [127.0.0.1]:5703 is added to the blacklist. Jun 01, 2020 4:22:40 PM com.hazelcast.cluster.impl.TcpIpJoiner INFO: [127.0.0.1]:5701 [sertug] [3.12.2] [127.0.0.1]:5702 is added to the blacklist. Jun 01, 2020 4:22:41 PM com.hazelcast.internal.cluster.ClusterService INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Members {size:1, ver:1} [ Member [127.0.0.1]:5701 - 3a89efeb-9da0-4a08-8b54-f6f7632387d5 this ] ```
1.0
QuorumListener is triggered during member startup but before it has joined a cluster - <!-- Thanks for reporting your issue. Please share with us the following information, to help us resolve your issue quickly and efficiently. --> **Describe the bug** Tested on 3.12.2 which is customer's version. In a cluster with Quorum defined, without touching any maps, just calling `Hazelcast.newHazelcastInstance();` causes `QuorumListener` to be triggered, before the member joins anyone and forms the cluster initially. **Expected behavior** We shouldn’t fire any quorum listener events after a member starts until quorum is met initially. **To Reproduce** Steps to reproduce the behavior: 1. Define a quorum of 3 with a listener attached ``` <quorum enabled="true" name="myquorum"> <quorum-size>3</quorum-size> <quorum-listeners> <quorum-listener>MyQuoLis</quorum-listener> </quorum-listeners> </quorum> ``` 2. Start one member 3. Observe how member logs a quorum event before join: ``` INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Hazelcast 3.12.2 (20190802 - e34b163) starting at [127.0.0.1]:5701 Jun 01, 2020 4:22:40 PM com.hazelcast.system INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Copyright (c) 2008-2019, Hazelcast, Inc. All Rights Reserved. Jun 01, 2020 4:22:40 PM com.hazelcast.spi.impl.operationservice.impl.BackpressureRegulator INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Backpressure is disabled Jun 01, 2020 4:22:40 PM com.hazelcast.instance.Node INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Creating TcpIpJoiner Jun 01, 2020 4:22:40 PM com.hazelcast.spi.impl.operationexecutor.impl.OperationExecutorImpl INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Starting 8 partition threads and 5 generic threads (1 dedicated for priority tasks) Jun 01, 2020 4:22:40 PM com.hazelcast.internal.diagnostics.Diagnostics INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Diagnostics disabled. To enable add -Dhazelcast.diagnostics.enabled=true to the JVM arguments. Jun 01, 2020 4:22:40 PM com.hazelcast.core.LifecycleService INFO: [127.0.0.1]:5701 [sertug] [3.12.2] [127.0.0.1]:5701 is STARTING Listener triggered: QuorumEvent{threshold=3, currentMembers=com.hazelcast.internal.cluster.impl.MemberSelectingCollection@1d30e322, presence=false} Jun 01, 2020 4:22:40 PM com.hazelcast.nio.tcp.TcpIpConnector INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Connecting to /127.0.0.1:5703, timeout: 10000, bind-any: true Jun 01, 2020 4:22:40 PM com.hazelcast.nio.tcp.TcpIpConnector INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Connecting to /127.0.0.1:5702, timeout: 10000, bind-any: true Jun 01, 2020 4:22:40 PM com.hazelcast.nio.tcp.TcpIpConnector INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Could not connect to: /127.0.0.1:5702. Reason: SocketException[Connection refused to address /127.0.0.1:5702] Jun 01, 2020 4:22:40 PM com.hazelcast.nio.tcp.TcpIpConnector INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Could not connect to: /127.0.0.1:5703. Reason: SocketException[Connection refused to address /127.0.0.1:5703] Jun 01, 2020 4:22:40 PM com.hazelcast.cluster.impl.TcpIpJoiner INFO: [127.0.0.1]:5701 [sertug] [3.12.2] [127.0.0.1]:5703 is added to the blacklist. Jun 01, 2020 4:22:40 PM com.hazelcast.cluster.impl.TcpIpJoiner INFO: [127.0.0.1]:5701 [sertug] [3.12.2] [127.0.0.1]:5702 is added to the blacklist. Jun 01, 2020 4:22:41 PM com.hazelcast.internal.cluster.ClusterService INFO: [127.0.0.1]:5701 [sertug] [3.12.2] Members {size:1, ver:1} [ Member [127.0.0.1]:5701 - 3a89efeb-9da0-4a08-8b54-f6f7632387d5 this ] ```
defect
quorumlistener is triggered during member startup but before it has joined a cluster thanks for reporting your issue please share with us the following information to help us resolve your issue quickly and efficiently describe the bug tested on which is customer s version in a cluster with quorum defined without touching any maps just calling hazelcast newhazelcastinstance causes quorumlistener to be triggered before the member joins anyone and forms the cluster initially expected behavior we shouldn’t fire any quorum listener events after a member starts until quorum is met initially to reproduce steps to reproduce the behavior define a quorum of with a listener attached myquolis start one member observe how member logs a quorum event before join info hazelcast starting at jun pm com hazelcast system info copyright c hazelcast inc all rights reserved jun pm com hazelcast spi impl operationservice impl backpressureregulator info backpressure is disabled jun pm com hazelcast instance node info creating tcpipjoiner jun pm com hazelcast spi impl operationexecutor impl operationexecutorimpl info starting partition threads and generic threads dedicated for priority tasks jun pm com hazelcast internal diagnostics diagnostics info diagnostics disabled to enable add dhazelcast diagnostics enabled true to the jvm arguments jun pm com hazelcast core lifecycleservice info is starting listener triggered quorumevent threshold currentmembers com hazelcast internal cluster impl memberselectingcollection presence false jun pm com hazelcast nio tcp tcpipconnector info connecting to timeout bind any true jun pm com hazelcast nio tcp tcpipconnector info connecting to timeout bind any true jun pm com hazelcast nio tcp tcpipconnector info could not connect to reason socketexception jun pm com hazelcast nio tcp tcpipconnector info could not connect to reason socketexception jun pm com hazelcast cluster impl tcpipjoiner info is added to the blacklist jun pm com hazelcast cluster impl tcpipjoiner info is added to the blacklist jun pm com hazelcast internal cluster clusterservice info members size ver member this
1
13,530
3,737,384,952
IssuesEvent
2016-03-08 19:05:59
playframework/playframework
https://api.github.com/repos/playframework/playframework
closed
[doc] [2.5.x] WS pages still use iteratees rather than Akka Streams
documentation
### Play Version (2.5.x / etc) 2.5.x ### API (Scala / Java / Neither / Both) Both ### Expected Behavior Play WS should discuss processing large responses in the context of Akka Streams ### Actual Behavior Still talks about iteratees: * https://www.playframework.com/documentation/2.5.x/ScalaWS#Processing-large-responses * https://www.playframework.com/documentation/2.5.x/JavaWS#Processing-large-responses
1.0
[doc] [2.5.x] WS pages still use iteratees rather than Akka Streams - ### Play Version (2.5.x / etc) 2.5.x ### API (Scala / Java / Neither / Both) Both ### Expected Behavior Play WS should discuss processing large responses in the context of Akka Streams ### Actual Behavior Still talks about iteratees: * https://www.playframework.com/documentation/2.5.x/ScalaWS#Processing-large-responses * https://www.playframework.com/documentation/2.5.x/JavaWS#Processing-large-responses
non_defect
ws pages still use iteratees rather than akka streams play version x etc x api scala java neither both both expected behavior play ws should discuss processing large responses in the context of akka streams actual behavior still talks about iteratees
0
499,681
14,475,992,363
IssuesEvent
2020-12-10 03:00:05
cjs8487/SS-Randomizer-Tracker
https://api.github.com/repos/cjs8487/SS-Randomizer-Tracker
closed
Logic can have duplicate entries
Medium Priority bug
It is possible for logic parsing to result in the same logical requirement appearing multiple times in the list, especially if macros are not fully simplified. If this cannot (or will not) be fixed in the rando macro location, the parsed list will need to be pruned for duplicates
1.0
Logic can have duplicate entries - It is possible for logic parsing to result in the same logical requirement appearing multiple times in the list, especially if macros are not fully simplified. If this cannot (or will not) be fixed in the rando macro location, the parsed list will need to be pruned for duplicates
non_defect
logic can have duplicate entries it is possible for logic parsing to result in the same logical requirement appearing multiple times in the list especially if macros are not fully simplified if this cannot or will not be fixed in the rando macro location the parsed list will need to be pruned for duplicates
0
5,716
2,610,213,985
IssuesEvent
2015-02-26 19:08:18
chrsmith/somefinders
https://api.github.com/repos/chrsmith/somefinders
opened
tiger ev-220 инструкция.doc
auto-migrated Priority-Medium Type-Defect
``` '''Анри Лебедев''' День добрый никак не могу найти .tiger ev-220 инструкция.doc. как то выкладывали уже '''Арвид Кондратьев''' Вот хороший сайт где можно скачать http://bit.ly/17Kd1mn '''Георгий Новиков''' Спасибо вроде то но просит телефон вводить '''Владлен Никитин''' Не это не влияет на баланс '''Альфред Родионов''' Неа все ок у меня ничего не списало Информация о файле: tiger ev-220 инструкция.doc Загружен: В этом месяце Скачан раз: 1319 Рейтинг: 909 Средняя скорость скачивания: 1180 Похожих файлов: 33 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 12:22
1.0
tiger ev-220 инструкция.doc - ``` '''Анри Лебедев''' День добрый никак не могу найти .tiger ev-220 инструкция.doc. как то выкладывали уже '''Арвид Кондратьев''' Вот хороший сайт где можно скачать http://bit.ly/17Kd1mn '''Георгий Новиков''' Спасибо вроде то но просит телефон вводить '''Владлен Никитин''' Не это не влияет на баланс '''Альфред Родионов''' Неа все ок у меня ничего не списало Информация о файле: tiger ev-220 инструкция.doc Загружен: В этом месяце Скачан раз: 1319 Рейтинг: 909 Средняя скорость скачивания: 1180 Похожих файлов: 33 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 12:22
defect
tiger ev инструкция doc анри лебедев день добрый никак не могу найти tiger ev инструкция doc как то выкладывали уже арвид кондратьев вот хороший сайт где можно скачать георгий новиков спасибо вроде то но просит телефон вводить владлен никитин не это не влияет на баланс альфред родионов неа все ок у меня ничего не списало информация о файле tiger ev инструкция doc загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at
1
71,319
23,537,825,319
IssuesEvent
2022-08-20 00:20:04
jccastillo0007/eFacturaT
https://api.github.com/repos/jccastillo0007/eFacturaT
closed
Conector 4.0 - Documento de pago, no está calculando correctamente los importeDR
resolved defect
Finkok dice que tenemos una cierta variación, y por eso lo rechaza. (Fhoy igual lo rechazó). Aquí te dejo el XML que se generó desde el conector. En un excel pongo únicamente las baseDR y los importesDR que se reportan al XML (en amarillo). En ese mismo excel, pongo en verde el cálculo de tomar la base y multiplicarlo por la tasa del iva, en este caso 0.08. Y también una columna indicando al diferencia. Pues es muy leve, pero el PAC (ambos) lo rechazan. Hay que revisar el cálculo del importeDR, cómo sale? [fixSabadoSergio.xml.txt](https://github.com/jccastillo0007/eFacturaT/files/9135963/fixSabadoSergio.xml.txt) [fixSabadoSergio.xlsx.txt](https://github.com/jccastillo0007/eFacturaT/files/9135964/fixSabadoSergio.xlsx.txt)
1.0
Conector 4.0 - Documento de pago, no está calculando correctamente los importeDR - Finkok dice que tenemos una cierta variación, y por eso lo rechaza. (Fhoy igual lo rechazó). Aquí te dejo el XML que se generó desde el conector. En un excel pongo únicamente las baseDR y los importesDR que se reportan al XML (en amarillo). En ese mismo excel, pongo en verde el cálculo de tomar la base y multiplicarlo por la tasa del iva, en este caso 0.08. Y también una columna indicando al diferencia. Pues es muy leve, pero el PAC (ambos) lo rechazan. Hay que revisar el cálculo del importeDR, cómo sale? [fixSabadoSergio.xml.txt](https://github.com/jccastillo0007/eFacturaT/files/9135963/fixSabadoSergio.xml.txt) [fixSabadoSergio.xlsx.txt](https://github.com/jccastillo0007/eFacturaT/files/9135964/fixSabadoSergio.xlsx.txt)
defect
conector documento de pago no está calculando correctamente los importedr finkok dice que tenemos una cierta variación y por eso lo rechaza fhoy igual lo rechazó aquí te dejo el xml que se generó desde el conector en un excel pongo únicamente las basedr y los importesdr que se reportan al xml en amarillo en ese mismo excel pongo en verde el cálculo de tomar la base y multiplicarlo por la tasa del iva en este caso y también una columna indicando al diferencia pues es muy leve pero el pac ambos lo rechazan hay que revisar el cálculo del importedr cómo sale
1
143,890
22,200,223,523
IssuesEvent
2022-06-07 10:29:33
scholokov/qax-portal
https://api.github.com/repos/scholokov/qax-portal
opened
New design for Portal (test)
design
To create example of new design in purple theme for Portal project
1.0
New design for Portal (test) - To create example of new design in purple theme for Portal project
non_defect
new design for portal test to create example of new design in purple theme for portal project
0
63,761
17,930,994,344
IssuesEvent
2021-09-10 09:13:09
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
searching encrypted messages on the desktop app seems to not work anymore
T-Defect X-Needs-Info A-Search A-Indexing
i cant seem to search anymore it provides 0 results in encrypted chats, i searched for an existing issue but couldnt find one so if one already exists my bad couldnt find it. it also says Use the Desktop app to search encrypted messages, even though im already using the desktop app
1.0
searching encrypted messages on the desktop app seems to not work anymore - i cant seem to search anymore it provides 0 results in encrypted chats, i searched for an existing issue but couldnt find one so if one already exists my bad couldnt find it. it also says Use the Desktop app to search encrypted messages, even though im already using the desktop app
defect
searching encrypted messages on the desktop app seems to not work anymore i cant seem to search anymore it provides results in encrypted chats i searched for an existing issue but couldnt find one so if one already exists my bad couldnt find it it also says use the desktop app to search encrypted messages even though im already using the desktop app
1
259,447
27,621,898,822
IssuesEvent
2023-03-10 01:20:04
nidhi7598/linux-3.0.35
https://api.github.com/repos/nidhi7598/linux-3.0.35
closed
CVE-2011-4917 (Medium) detected in linuxlinux-3.0.40, linuxlinux-3.0.40 - autoclosed
Mend: dependency security vulnerability
## CVE-2011-4917 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxlinux-3.0.40</b>, <b>linuxlinux-3.0.40</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel through 3.1 there is an information disclosure issue via /proc/stat. <p>Publish Date: 2022-04-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2011-4917>CVE-2011-4917</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.suse.com/show_bug.cgi?id=738749">https://bugzilla.suse.com/show_bug.cgi?id=738749</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: v3.3-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2011-4917 (Medium) detected in linuxlinux-3.0.40, linuxlinux-3.0.40 - autoclosed - ## CVE-2011-4917 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxlinux-3.0.40</b>, <b>linuxlinux-3.0.40</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel through 3.1 there is an information disclosure issue via /proc/stat. <p>Publish Date: 2022-04-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2011-4917>CVE-2011-4917</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.suse.com/show_bug.cgi?id=738749">https://bugzilla.suse.com/show_bug.cgi?id=738749</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: v3.3-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in linuxlinux linuxlinux autoclosed cve medium severity vulnerability vulnerable libraries linuxlinux linuxlinux vulnerability details in the linux kernel through there is an information disclosure issue via proc stat publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
179,647
6,627,487,861
IssuesEvent
2017-09-23 03:31:34
strongloop/loopback-connector-cloudant
https://api.github.com/repos/strongloop/loopback-connector-cloudant
closed
specify unique identifier `uuid` in a property
feature needs-doc needs-priority stale team-apex
### Feature - [x] Feature request ### Background of feature After triaging/helping solve some user's problem, I feel people use cloudant connector **are not aware of couchdb is not designed to change same document frequently and multiple times**, couchdb stores status change by creating different documents and including the same unique id to tell that they are attached to the same item, not updating the same document. Take the following scenario as an example: - A customer places an order, when he/she selects all items and clicks button `place order`, a **purchase** document is created as: ```js { "_id": "023f7a21dbe8a4177a2816e4ad1ea27e", "model": "purchase", "order_id": "320afa89017426b994162ab004ce3383", "basket": [ { "product_id": "A56" }, { "product_id": "B32" } ] } ``` *Please note the purchase document has a property called `order_id` to refer to a specific order, and the `model` property represents its loopback model name* - And then clicks button `checkout` to make payment, two **payment** documents(since there are two items in the order), with same `order_id` as the purchase document, are created as: ```js { "_id": "bf70c30ea5d8c3cd088fef98ad678e9e", "model": "payment", "account_id": "985522332", "order_id": "320afa89017426b994162ab004ce3383", "value": 6.46, "method": "credit card", "payment_reference": "AB9977G244FF2F667" } { "_id": "12c0ea6cd3d2c6e3b1d34442aea6a2d9", "type": "payment", "account_id": "985522332", "order_id": "320afa89017426b994162ab004ce3383", "value": 20.00, "method": "voucher", "payment_reference": "Q88775662377224" } ``` So in this case, we use `purchase` and `payment` as loopback models to represent the events of an order, and **an unique identifier(uuid) `order_id`** to make sure related events refer to a certain order. User can see the status of an order by creating a view of everything you know about an `order_id` as a ledger containing necessary informations. How to define those informations are beyond the topic of this issue, and well documented in Cloudant#transaction's doc(check the following link): ### Additional information (Node.js version, LoopBack version, etc) The example above is from Cloudant#transaction doc, for details, please refer to https://docs.cloudant.com/transactions.html ### Proposal of implementation loopback model supports defining a property as `uuid`, see https://loopback.io/doc/en/lb3/Model-definition-JSON-file.html#general-property-properties While I am not sure does it treat `uuid` the same way as what cloudant does with endpoint `GET _uuids`, will investigate more on that. - If there is difference between them, we need a function in `lib/cloudant.js` to generate an uuid. - If they are same, then this story would be a doc story.
1.0
specify unique identifier `uuid` in a property - ### Feature - [x] Feature request ### Background of feature After triaging/helping solve some user's problem, I feel people use cloudant connector **are not aware of couchdb is not designed to change same document frequently and multiple times**, couchdb stores status change by creating different documents and including the same unique id to tell that they are attached to the same item, not updating the same document. Take the following scenario as an example: - A customer places an order, when he/she selects all items and clicks button `place order`, a **purchase** document is created as: ```js { "_id": "023f7a21dbe8a4177a2816e4ad1ea27e", "model": "purchase", "order_id": "320afa89017426b994162ab004ce3383", "basket": [ { "product_id": "A56" }, { "product_id": "B32" } ] } ``` *Please note the purchase document has a property called `order_id` to refer to a specific order, and the `model` property represents its loopback model name* - And then clicks button `checkout` to make payment, two **payment** documents(since there are two items in the order), with same `order_id` as the purchase document, are created as: ```js { "_id": "bf70c30ea5d8c3cd088fef98ad678e9e", "model": "payment", "account_id": "985522332", "order_id": "320afa89017426b994162ab004ce3383", "value": 6.46, "method": "credit card", "payment_reference": "AB9977G244FF2F667" } { "_id": "12c0ea6cd3d2c6e3b1d34442aea6a2d9", "type": "payment", "account_id": "985522332", "order_id": "320afa89017426b994162ab004ce3383", "value": 20.00, "method": "voucher", "payment_reference": "Q88775662377224" } ``` So in this case, we use `purchase` and `payment` as loopback models to represent the events of an order, and **an unique identifier(uuid) `order_id`** to make sure related events refer to a certain order. User can see the status of an order by creating a view of everything you know about an `order_id` as a ledger containing necessary informations. How to define those informations are beyond the topic of this issue, and well documented in Cloudant#transaction's doc(check the following link): ### Additional information (Node.js version, LoopBack version, etc) The example above is from Cloudant#transaction doc, for details, please refer to https://docs.cloudant.com/transactions.html ### Proposal of implementation loopback model supports defining a property as `uuid`, see https://loopback.io/doc/en/lb3/Model-definition-JSON-file.html#general-property-properties While I am not sure does it treat `uuid` the same way as what cloudant does with endpoint `GET _uuids`, will investigate more on that. - If there is difference between them, we need a function in `lib/cloudant.js` to generate an uuid. - If they are same, then this story would be a doc story.
non_defect
specify unique identifier uuid in a property feature feature request background of feature after triaging helping solve some user s problem i feel people use cloudant connector are not aware of couchdb is not designed to change same document frequently and multiple times couchdb stores status change by creating different documents and including the same unique id to tell that they are attached to the same item not updating the same document take the following scenario as an example a customer places an order when he she selects all items and clicks button place order a purchase document is created as js id model purchase order id basket product id product id please note the purchase document has a property called order id to refer to a specific order and the model property represents its loopback model name and then clicks button checkout to make payment two payment documents since there are two items in the order with same order id as the purchase document are created as js id model payment account id order id value method credit card payment reference id type payment account id order id value method voucher payment reference so in this case we use purchase and payment as loopback models to represent the events of an order and an unique identifier uuid order id to make sure related events refer to a certain order user can see the status of an order by creating a view of everything you know about an order id as a ledger containing necessary informations how to define those informations are beyond the topic of this issue and well documented in cloudant transaction s doc check the following link additional information node js version loopback version etc the example above is from cloudant transaction doc for details please refer to proposal of implementation loopback model supports defining a property as uuid see while i am not sure does it treat uuid the same way as what cloudant does with endpoint get uuids will investigate more on that if there is difference between them we need a function in lib cloudant js to generate an uuid if they are same then this story would be a doc story
0
41,759
10,594,109,138
IssuesEvent
2019-10-09 16:04:24
CocoaPods/CocoaPods
https://api.github.com/repos/CocoaPods/CocoaPods
closed
pod repo update crash on 1.8.x
d1:easy s2:confirmed s7:workaround available t2:defect
<!-- ℹ Please fill out this template when filing an issue. All lines beginning with an ℹ symbol instruct you with what info we expect. Before you start, are you using the latest CocoaPods release? A lot changes with Xcode releases that are not backwards compatible. Not an issue about the CocoaPods command line app? Please file an issue in the appropriate repo - https://github.com/CocoaPods Issues are for feature requests, and bugs; questions should go to Stack Overflow Using CocoaPods <= 0.39: https://blog.cocoapods.org/Sharding/ Using Xcode 10.1: Requires CocoaPods 1.6.0 or above. Issue with Nanaimo not loading: Please run `[sudo] gem uninstall nanaimo` and remove all but the latest version. Issues with `pod search`? Try deleting your cache `rm -rf ~/Library/Caches/CocoaPods`first. --> * [x] I've read and understood the [*CONTRIBUTING* guidelines and have done my best effort to follow](https://github.com/CocoaPods/CocoaPods/blob/master/CONTRIBUTING.md). # Report I know CocoaPods 1.8 use the CDN by default, but since `pod setup` does nothing, if you have a Podfile with `source 'https://github.com/CocoaPods/Specs.git'`, when you run `pod repo update` it will fail. `pod setup` should continue working or there should be a new command to get the git based source downloaded. ## What did you do? I have an autogenerated Podfile with `source 'https://github.com/CocoaPods/Specs.git'`, when I run `pod repo update` it fails. ## What did you expect to happen? To get the git sources as on previous versions ## What happened instead? Got this error: No such file or directory ~/.cocoapods/repos/Spec_Lock ## CocoaPods Environment ### Stack ``` CocoaPods : 1.8.0 Ruby : ruby 2.3.7p456 (2018-03-28 revision 63024) [x86_64-darwin17] RubyGems : 2.5.2.3 Host : Mac OS X 10.14.6 (18G95) Xcode : 11.0 (11A420a) Git : git version 2.21.0 (Apple Git-122) Ruby lib dir : /Users/jcesarmobile/.rbenv/versions/2.3.7/lib Repositories: ``` ### Plugins ``` cocoapods-deintegrate : 1.0.4 cocoapods-plugins : 1.0.0 cocoapods-search : 1.0.0 cocoapods-stats : 1.1.0 cocoapods-trunk : 1.4.0 cocoapods-try : 1.1.0 ```
1.0
pod repo update crash on 1.8.x - <!-- ℹ Please fill out this template when filing an issue. All lines beginning with an ℹ symbol instruct you with what info we expect. Before you start, are you using the latest CocoaPods release? A lot changes with Xcode releases that are not backwards compatible. Not an issue about the CocoaPods command line app? Please file an issue in the appropriate repo - https://github.com/CocoaPods Issues are for feature requests, and bugs; questions should go to Stack Overflow Using CocoaPods <= 0.39: https://blog.cocoapods.org/Sharding/ Using Xcode 10.1: Requires CocoaPods 1.6.0 or above. Issue with Nanaimo not loading: Please run `[sudo] gem uninstall nanaimo` and remove all but the latest version. Issues with `pod search`? Try deleting your cache `rm -rf ~/Library/Caches/CocoaPods`first. --> * [x] I've read and understood the [*CONTRIBUTING* guidelines and have done my best effort to follow](https://github.com/CocoaPods/CocoaPods/blob/master/CONTRIBUTING.md). # Report I know CocoaPods 1.8 use the CDN by default, but since `pod setup` does nothing, if you have a Podfile with `source 'https://github.com/CocoaPods/Specs.git'`, when you run `pod repo update` it will fail. `pod setup` should continue working or there should be a new command to get the git based source downloaded. ## What did you do? I have an autogenerated Podfile with `source 'https://github.com/CocoaPods/Specs.git'`, when I run `pod repo update` it fails. ## What did you expect to happen? To get the git sources as on previous versions ## What happened instead? Got this error: No such file or directory ~/.cocoapods/repos/Spec_Lock ## CocoaPods Environment ### Stack ``` CocoaPods : 1.8.0 Ruby : ruby 2.3.7p456 (2018-03-28 revision 63024) [x86_64-darwin17] RubyGems : 2.5.2.3 Host : Mac OS X 10.14.6 (18G95) Xcode : 11.0 (11A420a) Git : git version 2.21.0 (Apple Git-122) Ruby lib dir : /Users/jcesarmobile/.rbenv/versions/2.3.7/lib Repositories: ``` ### Plugins ``` cocoapods-deintegrate : 1.0.4 cocoapods-plugins : 1.0.0 cocoapods-search : 1.0.0 cocoapods-stats : 1.1.0 cocoapods-trunk : 1.4.0 cocoapods-try : 1.1.0 ```
defect
pod repo update crash on x ℹ please fill out this template when filing an issue all lines beginning with an ℹ symbol instruct you with what info we expect before you start are you using the latest cocoapods release a lot changes with xcode releases that are not backwards compatible not an issue about the cocoapods command line app please file an issue in the appropriate repo issues are for feature requests and bugs questions should go to stack overflow using cocoapods using xcode requires cocoapods or above issue with nanaimo not loading please run gem uninstall nanaimo and remove all but the latest version issues with pod search try deleting your cache rm rf library caches cocoapods first i ve read and understood the report i know cocoapods use the cdn by default but since pod setup does nothing if you have a podfile with source when you run pod repo update it will fail pod setup should continue working or there should be a new command to get the git based source downloaded what did you do i have an autogenerated podfile with source when i run pod repo update it fails what did you expect to happen to get the git sources as on previous versions what happened instead got this error no such file or directory cocoapods repos spec lock cocoapods environment stack cocoapods ruby ruby revision rubygems host mac os x xcode git git version apple git ruby lib dir users jcesarmobile rbenv versions lib repositories plugins cocoapods deintegrate cocoapods plugins cocoapods search cocoapods stats cocoapods trunk cocoapods try
1
78,708
27,725,639,175
IssuesEvent
2023-03-15 01:50:12
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Possible data leak
T-Defect Security
### Steps to reproduce 1. Open terminal and go to `/home/[user]/.config/Element/Local Storage/leveldb` 2. Type `strings` command on log files. 3. You will see traces of talks. ### Outcome #### What did you expect? I expect Element clear any possible log with info that could be extracted by someone and break the privacy/security/anonimous ### Operating system Parrot OS 5.2 ### Application version 1.11.24 ### How did you install the app? _No response_ ### Homeserver _No response_ ### Will you send logs? No
1.0
Possible data leak - ### Steps to reproduce 1. Open terminal and go to `/home/[user]/.config/Element/Local Storage/leveldb` 2. Type `strings` command on log files. 3. You will see traces of talks. ### Outcome #### What did you expect? I expect Element clear any possible log with info that could be extracted by someone and break the privacy/security/anonimous ### Operating system Parrot OS 5.2 ### Application version 1.11.24 ### How did you install the app? _No response_ ### Homeserver _No response_ ### Will you send logs? No
defect
possible data leak steps to reproduce open terminal and go to home config element local storage leveldb type strings command on log files you will see traces of talks outcome what did you expect i expect element clear any possible log with info that could be extracted by someone and break the privacy security anonimous operating system parrot os application version how did you install the app no response homeserver no response will you send logs no
1
494,575
14,260,732,824
IssuesEvent
2020-11-20 10:15:12
mir-dataset-loaders/mirdata
https://api.github.com/repos/mir-dataset-loaders/mirdata
closed
Metadata RWC collection broken?
bug priority
I think maybe something changed in the way we're uncompressing when we download that broke the metadata (if I download it manually it works). This is the error I get when I try to load`rwc` files: ``` Traceback (most recent call last): File "/Users/mf3734/Documents/repos/mirdata/scripts/update_index.py", line 199, in <module> main() File "/Users/mf3734/Documents/repos/mirdata/scripts/update_index.py", line 196, in main test_track_load([d for d in DATASETS if 'rwc_classical' not in d]) File "/Users/mf3734/Documents/repos/mirdata/scripts/update_index.py", line 179, in test_track_load dataset.load_tracks() File "/Users/mf3734/Documents/repos/mirdata/mirdata/core.py", line 126, in load_tracks return {track_id: self.track(track_id) for track_id in self.track_ids} File "/Users/mf3734/Documents/repos/mirdata/mirdata/core.py", line 126, in <dictcomp> return {track_id: self.track(track_id) for track_id in self.track_ids} File "/Users/mf3734/Documents/repos/mirdata/mirdata/core.py", line 65, in <lambda> self.track = lambda track_id: self._track(track_id) File "/Users/mf3734/Documents/repos/mirdata/mirdata/core.py", line 115, in _track return self._track_object(track_id, self.data_home) File "/Users/mf3734/Documents/repos/mirdata/mirdata/datasets/rwc_jazz.py", line 173, in __init__ metadata = DATA.metadata(data_home) File "/Users/mf3734/Documents/repos/mirdata/mirdata/utils.py", line 250, in metadata self._metadata = self.metadata_load_fn(data_home) File "/Users/mf3734/Documents/repos/mirdata/mirdata/datasets/rwc_jazz.py", line 104, in _load_metadata dialect = csv.Sniffer().sniff(fhandle.read(1024)) File "/Users/mf3734/anaconda3/envs/mirdata/lib/python3.7/codecs.py", line 322, in decode (result, consumed) = self._buffer_decode(data, self.errors, final) UnicodeDecodeError: 'utf-8' codec can't decode byte 0xdb in position 12: invalid continuation byte Process finished with exit code 1 ``` I'm wondering why do the tests pass, am I getting confused? Labeling it as a question for now
1.0
Metadata RWC collection broken? - I think maybe something changed in the way we're uncompressing when we download that broke the metadata (if I download it manually it works). This is the error I get when I try to load`rwc` files: ``` Traceback (most recent call last): File "/Users/mf3734/Documents/repos/mirdata/scripts/update_index.py", line 199, in <module> main() File "/Users/mf3734/Documents/repos/mirdata/scripts/update_index.py", line 196, in main test_track_load([d for d in DATASETS if 'rwc_classical' not in d]) File "/Users/mf3734/Documents/repos/mirdata/scripts/update_index.py", line 179, in test_track_load dataset.load_tracks() File "/Users/mf3734/Documents/repos/mirdata/mirdata/core.py", line 126, in load_tracks return {track_id: self.track(track_id) for track_id in self.track_ids} File "/Users/mf3734/Documents/repos/mirdata/mirdata/core.py", line 126, in <dictcomp> return {track_id: self.track(track_id) for track_id in self.track_ids} File "/Users/mf3734/Documents/repos/mirdata/mirdata/core.py", line 65, in <lambda> self.track = lambda track_id: self._track(track_id) File "/Users/mf3734/Documents/repos/mirdata/mirdata/core.py", line 115, in _track return self._track_object(track_id, self.data_home) File "/Users/mf3734/Documents/repos/mirdata/mirdata/datasets/rwc_jazz.py", line 173, in __init__ metadata = DATA.metadata(data_home) File "/Users/mf3734/Documents/repos/mirdata/mirdata/utils.py", line 250, in metadata self._metadata = self.metadata_load_fn(data_home) File "/Users/mf3734/Documents/repos/mirdata/mirdata/datasets/rwc_jazz.py", line 104, in _load_metadata dialect = csv.Sniffer().sniff(fhandle.read(1024)) File "/Users/mf3734/anaconda3/envs/mirdata/lib/python3.7/codecs.py", line 322, in decode (result, consumed) = self._buffer_decode(data, self.errors, final) UnicodeDecodeError: 'utf-8' codec can't decode byte 0xdb in position 12: invalid continuation byte Process finished with exit code 1 ``` I'm wondering why do the tests pass, am I getting confused? Labeling it as a question for now
non_defect
metadata rwc collection broken i think maybe something changed in the way we re uncompressing when we download that broke the metadata if i download it manually it works this is the error i get when i try to load rwc files traceback most recent call last file users documents repos mirdata scripts update index py line in main file users documents repos mirdata scripts update index py line in main test track load file users documents repos mirdata scripts update index py line in test track load dataset load tracks file users documents repos mirdata mirdata core py line in load tracks return track id self track track id for track id in self track ids file users documents repos mirdata mirdata core py line in return track id self track track id for track id in self track ids file users documents repos mirdata mirdata core py line in self track lambda track id self track track id file users documents repos mirdata mirdata core py line in track return self track object track id self data home file users documents repos mirdata mirdata datasets rwc jazz py line in init metadata data metadata data home file users documents repos mirdata mirdata utils py line in metadata self metadata self metadata load fn data home file users documents repos mirdata mirdata datasets rwc jazz py line in load metadata dialect csv sniffer sniff fhandle read file users envs mirdata lib codecs py line in decode result consumed self buffer decode data self errors final unicodedecodeerror utf codec can t decode byte in position invalid continuation byte process finished with exit code i m wondering why do the tests pass am i getting confused labeling it as a question for now
0
16,461
21,387,394,744
IssuesEvent
2022-04-21 01:08:00
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Cannot use Sub-Query referencing a GUI question and a date based filter
Type:Bug Priority:P2 Querying/Processor Querying/Parameters & Variables Querying/Nested Queries .Backend .Reproduced
**Describe the bug** When creating a SQL question, which uses a [Saved Question](https://www.metabase.com/docs/latest/users-guide/referencing-saved-questions-in-queries.html) that is GUI based and a Date filter or Field Filter on a date column, then query fails. It works with Text and Number filters. Workaround is to convert the GUI question to SQL, then it will work. **To Reproduce** 1. Question > Sample > Products - save as "Q1" 2. SQL query > Sample - replace `114` with the ID of Q1 from step 1 - set `filter` as Field Filter of Orders.CreatedAt or a Date filter ``` with exclude_products as {{#114}} select count(*) from orders where true [[and {{filter}}]] -- Field Filter, but will also fail with Date filter: [[and created_at={{datevariable}}]] ``` 3. Run the query - it works ![image](https://user-images.githubusercontent.com/1447303/160118915-681495c3-2a7a-45a6-a216-1419e7f37431.png) 4. Input anything in the filter widget and the query will fail with `Error building query parameter map: Error determining value for parameter \"#114\": The sub-query from referenced question #114 failed with the following error: \"Output of unwrap-field-clause does not match schema: \\n\\n\\t [(not (present? :field)) (not (present? \\\"id-or-name\\\")) (not (present? \\\"options\\\"))] \\n\\n\"` <details><summary>Full stacktrace</summary> ``` 2022-03-25 13:04:43,661 ERROR middleware.catch-exceptions :: Error processing query: null {:database_id 4, :started_at #t "2022-03-25T13:04:42.528037+01:00[Europe/Copenhagen]", :via [{:status :failed, :class clojure.lang.ExceptionInfo, :error "The sub-query from referenced question #114 failed with the following error: \"Output of unwrap-field-clause does not match schema: \\n\\n\\t [(not (present? :field)) (not (present? \\\"id-or-name\\\")) (not (present? \\\"options\\\"))] \\n\\n\"", :stacktrace ["--> driver.common.parameters.values$fn__78245$parse_tag__78238__78250$fn__78254.invoke(values.clj:134)" "driver.common.parameters.values$fn__78245$parse_tag__78238__78250.invoke(values.clj:121)" "driver.common.parameters.values$fn__78500$value_for_tag__78505$fn__78506.invoke(values.clj:307)" "driver.common.parameters.values$fn__78500$value_for_tag__78505.invoke(values.clj:302)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529$fn__78533$iter__78535__78539$fn__78540.invoke(values.clj:326)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529$fn__78533.invoke(values.clj:325)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529.invoke(values.clj:316)" "driver.sql$fn__78608$substitute_native_parameters__78601__78613$fn__78617.invoke(sql.clj:46)" "driver.sql$fn__78608$substitute_native_parameters__78601__78613.invoke(sql.clj:42)" "query_processor.middleware.parameters.native$expand_inner.invokeStatic(native.clj:39)" "query_processor.middleware.parameters.native$expand_inner.invoke(native.clj:30)" "query_processor.middleware.parameters$expand_one.invokeStatic(parameters.clj:49)" "query_processor.middleware.parameters$expand_one.invoke(parameters.clj:40)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "mbql.util.match.impl$replace_in_collection$iter__27400__27404$fn__27405.invoke(impl.cljc:44)" "mbql.util.match.impl$replace_in_collection.invokeStatic(impl.cljc:43)" "mbql.util.match.impl$replace_in_collection.invoke(impl.cljc:38)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:55)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_parameters.invokeStatic(parameters.clj:76)" "query_processor.middleware.parameters$expand_parameters.invoke(parameters.clj:72)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604$fn__48605.invoke(parameters.clj:81)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604.invoke(parameters.clj:78)" "query_processor.middleware.parameters$substitute_parameters$fn__48624.invoke(parameters.clj:109)" "query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__45984.invoke(resolve_referenced.clj:79)" "query_processor.middleware.expand_macros$expand_macros$fn__52276.invoke(expand_macros.clj:184)" "query_processor.middleware.add_timezone_info$add_timezone_info$fn__48404.invoke(add_timezone_info.clj:15)" "query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__51227.invoke(splice_params_in_response.clj:32)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643$fn__50648.invoke(resolve_database_and_driver.clj:35)" "driver$do_with_driver.invokeStatic(driver.clj:60)" "driver$do_with_driver.invoke(driver.clj:56)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643.invoke(resolve_database_and_driver.clj:34)" "query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__46530.invoke(fetch_source_query.clj:286)" "query_processor.middleware.store$initialize_store$fn__46721$fn__46722.invoke(store.clj:11)" "query_processor.store$do_with_store.invokeStatic(store.clj:44)" "query_processor.store$do_with_store.invoke(store.clj:38)" "query_processor.middleware.store$initialize_store$fn__46721.invoke(store.clj:10)" "query_processor.middleware.validate$validate_query$fn__50978.invoke(validate.clj:10)" "query_processor.middleware.normalize_query$normalize$fn__50985.invoke(normalize_query.clj:22)" "query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__48350.invoke(add_rows_truncated.clj:35)" "query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49641.invoke(results_metadata.clj:82)" "query_processor.middleware.constraints$add_default_userland_constraints$fn__48368.invoke(constraints.clj:42)" "query_processor.middleware.process_userland_query$process_userland_query$fn__50919.invoke(process_userland_query.clj:146)" "query_processor.middleware.catch_exceptions$catch_exceptions$fn__51307.invoke(catch_exceptions.clj:169)" "query_processor.reducible$async_qp$qp_STAR___43282$thunk__43283.invoke(reducible.clj:103)" "query_processor.reducible$async_qp$qp_STAR___43282.invoke(reducible.clj:109)" "query_processor.reducible$sync_qp$qp_STAR___43291$fn__43294.invoke(reducible.clj:135)" "query_processor.reducible$sync_qp$qp_STAR___43291.invoke(reducible.clj:134)" "query_processor$process_userland_query.invokeStatic(query_processor.clj:247)" "query_processor$process_userland_query.doInvoke(query_processor.clj:243)" "query_processor$fn__52324$process_query_and_save_execution_BANG___52333$fn__52336.invoke(query_processor.clj:258)" "query_processor$fn__52324$process_query_and_save_execution_BANG___52333.invoke(query_processor.clj:251)" "query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377$fn__52380.invoke(query_processor.clj:270)" "query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377.invoke(query_processor.clj:263)" "api.dataset$run_query_async$fn__65303.invoke(dataset.clj:69)" "query_processor.streaming$streaming_response_STAR_$fn__38418$fn__38419.invoke(streaming.clj:162)" "query_processor.streaming$streaming_response_STAR_$fn__38418.invoke(streaming.clj:161)" "async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)" "async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)" "async.streaming_response$do_f_async$task__26888.invoke(streaming_response.clj:84)"], :error_type :invalid-parameter, :ex-data {:card-query-error? true, :card-id 114, :tag {:id "8c468524-e0f4-aaf7-3757-f84141a5e27d", :name "#114", :display-name "#114", :type :card, :card-id 114}, :type :invalid-parameter}} {:status :failed, :class clojure.lang.ExceptionInfo, :error "Error determining value for parameter \"#114\": The sub-query from referenced question #114 failed with the following error: \"Output of unwrap-field-clause does not match schema: \\n\\n\\t [(not (present? :field)) (not (present? \\\"id-or-name\\\")) (not (present? \\\"options\\\"))] \\n\\n\"", :stacktrace ["--> driver.common.parameters.values$fn__78500$value_for_tag__78505$fn__78506.invoke(values.clj:309)" "driver.common.parameters.values$fn__78500$value_for_tag__78505.invoke(values.clj:302)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529$fn__78533$iter__78535__78539$fn__78540.invoke(values.clj:326)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529$fn__78533.invoke(values.clj:325)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529.invoke(values.clj:316)" "driver.sql$fn__78608$substitute_native_parameters__78601__78613$fn__78617.invoke(sql.clj:46)" "driver.sql$fn__78608$substitute_native_parameters__78601__78613.invoke(sql.clj:42)" "query_processor.middleware.parameters.native$expand_inner.invokeStatic(native.clj:39)" "query_processor.middleware.parameters.native$expand_inner.invoke(native.clj:30)" "query_processor.middleware.parameters$expand_one.invokeStatic(parameters.clj:49)" "query_processor.middleware.parameters$expand_one.invoke(parameters.clj:40)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "mbql.util.match.impl$replace_in_collection$iter__27400__27404$fn__27405.invoke(impl.cljc:44)" "mbql.util.match.impl$replace_in_collection.invokeStatic(impl.cljc:43)" "mbql.util.match.impl$replace_in_collection.invoke(impl.cljc:38)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:55)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_parameters.invokeStatic(parameters.clj:76)" "query_processor.middleware.parameters$expand_parameters.invoke(parameters.clj:72)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604$fn__48605.invoke(parameters.clj:81)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604.invoke(parameters.clj:78)" "query_processor.middleware.parameters$substitute_parameters$fn__48624.invoke(parameters.clj:109)" "query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__45984.invoke(resolve_referenced.clj:79)" "query_processor.middleware.expand_macros$expand_macros$fn__52276.invoke(expand_macros.clj:184)" "query_processor.middleware.add_timezone_info$add_timezone_info$fn__48404.invoke(add_timezone_info.clj:15)" "query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__51227.invoke(splice_params_in_response.clj:32)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643$fn__50648.invoke(resolve_database_and_driver.clj:35)" "driver$do_with_driver.invokeStatic(driver.clj:60)" "driver$do_with_driver.invoke(driver.clj:56)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643.invoke(resolve_database_and_driver.clj:34)" "query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__46530.invoke(fetch_source_query.clj:286)" "query_processor.middleware.store$initialize_store$fn__46721$fn__46722.invoke(store.clj:11)" "query_processor.store$do_with_store.invokeStatic(store.clj:44)" "query_processor.store$do_with_store.invoke(store.clj:38)" "query_processor.middleware.store$initialize_store$fn__46721.invoke(store.clj:10)" "query_processor.middleware.validate$validate_query$fn__50978.invoke(validate.clj:10)" "query_processor.middleware.normalize_query$normalize$fn__50985.invoke(normalize_query.clj:22)" "query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__48350.invoke(add_rows_truncated.clj:35)" "query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49641.invoke(results_metadata.clj:82)" "query_processor.middleware.constraints$add_default_userland_constraints$fn__48368.invoke(constraints.clj:42)" "query_processor.middleware.process_userland_query$process_userland_query$fn__50919.invoke(process_userland_query.clj:146)" "query_processor.middleware.catch_exceptions$catch_exceptions$fn__51307.invoke(catch_exceptions.clj:169)" "query_processor.reducible$async_qp$qp_STAR___43282$thunk__43283.invoke(reducible.clj:103)" "query_processor.reducible$async_qp$qp_STAR___43282.invoke(reducible.clj:109)" "query_processor.reducible$sync_qp$qp_STAR___43291$fn__43294.invoke(reducible.clj:135)" "query_processor.reducible$sync_qp$qp_STAR___43291.invoke(reducible.clj:134)" "query_processor$process_userland_query.invokeStatic(query_processor.clj:247)" "query_processor$process_userland_query.doInvoke(query_processor.clj:243)" "query_processor$fn__52324$process_query_and_save_execution_BANG___52333$fn__52336.invoke(query_processor.clj:258)" "query_processor$fn__52324$process_query_and_save_execution_BANG___52333.invoke(query_processor.clj:251)" "query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377$fn__52380.invoke(query_processor.clj:270)" "query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377.invoke(query_processor.clj:263)" "api.dataset$run_query_async$fn__65303.invoke(dataset.clj:69)" "query_processor.streaming$streaming_response_STAR_$fn__38418$fn__38419.invoke(streaming.clj:162)" "query_processor.streaming$streaming_response_STAR_$fn__38418.invoke(streaming.clj:161)" "async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)" "async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)" "async.streaming_response$do_f_async$task__26888.invoke(streaming_response.clj:84)"], :error_type :invalid-parameter, :ex-data {:tag {:id "8c468524-e0f4-aaf7-3757-f84141a5e27d", :name "#114", :display-name "#114", :type :card, :card-id 114}, :type :invalid-parameter}} {:status :failed, :class clojure.lang.ExceptionInfo, :error "Error building query parameter map: Error determining value for parameter \"#114\": The sub-query from referenced question #114 failed with the following error: \"Output of unwrap-field-clause does not match schema: \\n\\n\\t [(not (present? :field)) (not (present? \\\"id-or-name\\\")) (not (present? \\\"options\\\"))] \\n\\n\"", :stacktrace ["--> driver.common.parameters.values$fn__78524$query__GT_params_map__78529$fn__78533.invoke(values.clj:334)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529.invoke(values.clj:316)" "driver.sql$fn__78608$substitute_native_parameters__78601__78613$fn__78617.invoke(sql.clj:46)" "driver.sql$fn__78608$substitute_native_parameters__78601__78613.invoke(sql.clj:42)" "query_processor.middleware.parameters.native$expand_inner.invokeStatic(native.clj:39)" "query_processor.middleware.parameters.native$expand_inner.invoke(native.clj:30)" "query_processor.middleware.parameters$expand_one.invokeStatic(parameters.clj:49)" "query_processor.middleware.parameters$expand_one.invoke(parameters.clj:40)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "mbql.util.match.impl$replace_in_collection$iter__27400__27404$fn__27405.invoke(impl.cljc:44)" "mbql.util.match.impl$replace_in_collection.invokeStatic(impl.cljc:43)" "mbql.util.match.impl$replace_in_collection.invoke(impl.cljc:38)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:55)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_parameters.invokeStatic(parameters.clj:76)" "query_processor.middleware.parameters$expand_parameters.invoke(parameters.clj:72)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604$fn__48605.invoke(parameters.clj:81)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604.invoke(parameters.clj:78)" "query_processor.middleware.parameters$substitute_parameters$fn__48624.invoke(parameters.clj:109)" "query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__45984.invoke(resolve_referenced.clj:79)" "query_processor.middleware.expand_macros$expand_macros$fn__52276.invoke(expand_macros.clj:184)" "query_processor.middleware.add_timezone_info$add_timezone_info$fn__48404.invoke(add_timezone_info.clj:15)" "query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__51227.invoke(splice_params_in_response.clj:32)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643$fn__50648.invoke(resolve_database_and_driver.clj:35)" "driver$do_with_driver.invokeStatic(driver.clj:60)" "driver$do_with_driver.invoke(driver.clj:56)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643.invoke(resolve_database_and_driver.clj:34)" "query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__46530.invoke(fetch_source_query.clj:286)" "query_processor.middleware.store$initialize_store$fn__46721$fn__46722.invoke(store.clj:11)" "query_processor.store$do_with_store.invokeStatic(store.clj:44)" "query_processor.store$do_with_store.invoke(store.clj:38)" "query_processor.middleware.store$initialize_store$fn__46721.invoke(store.clj:10)" "query_processor.middleware.validate$validate_query$fn__50978.invoke(validate.clj:10)" "query_processor.middleware.normalize_query$normalize$fn__50985.invoke(normalize_query.clj:22)" "query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__48350.invoke(add_rows_truncated.clj:35)" "query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49641.invoke(results_metadata.clj:82)" "query_processor.middleware.constraints$add_default_userland_constraints$fn__48368.invoke(constraints.clj:42)" "query_processor.middleware.process_userland_query$process_userland_query$fn__50919.invoke(process_userland_query.clj:146)" "query_processor.middleware.catch_exceptions$catch_exceptions$fn__51307.invoke(catch_exceptions.clj:169)" "query_processor.reducible$async_qp$qp_STAR___43282$thunk__43283.invoke(reducible.clj:103)" "query_processor.reducible$async_qp$qp_STAR___43282.invoke(reducible.clj:109)" "query_processor.reducible$sync_qp$qp_STAR___43291$fn__43294.invoke(reducible.clj:135)" "query_processor.reducible$sync_qp$qp_STAR___43291.invoke(reducible.clj:134)" "query_processor$process_userland_query.invokeStatic(query_processor.clj:247)" "query_processor$process_userland_query.doInvoke(query_processor.clj:243)" "query_processor$fn__52324$process_query_and_save_execution_BANG___52333$fn__52336.invoke(query_processor.clj:258)" "query_processor$fn__52324$process_query_and_save_execution_BANG___52333.invoke(query_processor.clj:251)" "query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377$fn__52380.invoke(query_processor.clj:270)" "query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377.invoke(query_processor.clj:263)" "api.dataset$run_query_async$fn__65303.invoke(dataset.clj:69)" "query_processor.streaming$streaming_response_STAR_$fn__38418$fn__38419.invoke(streaming.clj:162)" "query_processor.streaming$streaming_response_STAR_$fn__38418.invoke(streaming.clj:161)" "async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)" "async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)" "async.streaming_response$do_f_async$task__26888.invoke(streaming_response.clj:84)"], :error_type :invalid-parameter, :ex-data {:type :invalid-parameter, :tags {"#114" {:id "8c468524-e0f4-aaf7-3757-f84141a5e27d", :name "#114", :display-name "#114", :type :card, :card-id 114}, "filter" {:id "79400356-2c24-782c-fa46-e9d29b9dacfb", :name "filter", :display-name "Filter", :type :dimension, :dimension [:field 86 nil], :widget-type :date/all-options, :default nil}}, :params [{:type :date/all-options, :value "past30years", :target [:dimension [:template-tag "filter"]], :id "79400356-2c24-782c-fa46-e9d29b9dacfb"}]}}], :error_type :invalid-query, :json_query {:database 4, :native {:template-tags {:#114 {:id "8c468524-e0f4-aaf7-3757-f84141a5e27d", :name "#114", :display-name "#114", :type "card", :card-id 114}, :filter {:id "79400356-2c24-782c-fa46-e9d29b9dacfb", :name "filter", :display-name "Filter", :type "dimension", :dimension ["field" 86 nil], :widget-type "date/all-options", :default nil}}, :query "with exclude_products as {{#114}}\nselect count(*) from orders where {{filter}}"}, :type "native", :parameters [{:type "date/all-options", :value "past30years", :target ["dimension" ["template-tag" "filter"]], :id "79400356-2c24-782c-fa46-e9d29b9dacfb"}], :middleware {:js-int-to-string? true, :add-default-userland-constraints? true}}, :status :failed, :class clojure.lang.ExceptionInfo, :stacktrace ["--> models.params$fn__38495$unwrap_field_clause__38500.invoke(params.clj:21)" "query_processor.middleware.parameters.mbql$fn__48480$build_filter_clause__48485$fn__48495.invoke(mbql.clj:57)" "query_processor.middleware.parameters.mbql$fn__48480$build_filter_clause__48485.invoke(mbql.clj:44)" "query_processor.middleware.parameters.mbql$expand.invokeStatic(mbql.clj:86)" "query_processor.middleware.parameters.mbql$expand.invoke(mbql.clj:72)" "query_processor.middleware.parameters$expand_mbql_params.invokeStatic(parameters.clj:36)" "query_processor.middleware.parameters$expand_mbql_params.invoke(parameters.clj:32)" "query_processor.middleware.parameters$expand_one.invokeStatic(parameters.clj:48)" "query_processor.middleware.parameters$expand_one.invoke(parameters.clj:40)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "mbql.util.match.impl$replace_in_collection$iter__27400__27404$fn__27405.invoke(impl.cljc:44)" "mbql.util.match.impl$replace_in_collection.invokeStatic(impl.cljc:43)" "mbql.util.match.impl$replace_in_collection.invoke(impl.cljc:38)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:55)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_parameters.invokeStatic(parameters.clj:76)" "query_processor.middleware.parameters$expand_parameters.invoke(parameters.clj:72)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604$fn__48605.invoke(parameters.clj:81)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604.invoke(parameters.clj:78)" "query_processor.middleware.parameters$substitute_parameters$fn__48624.invoke(parameters.clj:109)" "query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__45984.invoke(resolve_referenced.clj:79)" "query_processor.middleware.expand_macros$expand_macros$fn__52276.invoke(expand_macros.clj:184)" "query_processor.middleware.add_timezone_info$add_timezone_info$fn__48404.invoke(add_timezone_info.clj:15)" "query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__51227.invoke(splice_params_in_response.clj:32)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643$fn__50648.invoke(resolve_database_and_driver.clj:35)" "driver$do_with_driver.invokeStatic(driver.clj:60)" "driver$do_with_driver.invoke(driver.clj:56)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643.invoke(resolve_database_and_driver.clj:34)" "query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__46530.invoke(fetch_source_query.clj:286)" "query_processor.middleware.store$initialize_store$fn__46721$fn__46722.invoke(store.clj:11)" "query_processor.store$do_with_store.invokeStatic(store.clj:42)" "query_processor.store$do_with_store.invoke(store.clj:38)" "query_processor.middleware.store$initialize_store$fn__46721.invoke(store.clj:10)" "query_processor.middleware.validate$validate_query$fn__50978.invoke(validate.clj:10)" "query_processor.middleware.normalize_query$normalize$fn__50985.invoke(normalize_query.clj:22)" "query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__48350.invoke(add_rows_truncated.clj:35)" "query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49641.invoke(results_metadata.clj:82)" "query_processor.reducible$async_qp$qp_STAR___43282$thunk__43283.invoke(reducible.clj:103)" "query_processor.reducible$async_qp$qp_STAR___43282.invoke(reducible.clj:109)" "query_processor.reducible$sync_qp$qp_STAR___43291$fn__43294.invoke(reducible.clj:135)" "query_processor.reducible$sync_qp$qp_STAR___43291.invoke(reducible.clj:134)" "query_processor$preprocess_query.invokeStatic(query_processor.clj:172)" "query_processor$preprocess_query.invoke(query_processor.clj:164)" "query_processor$query__GT_native.invokeStatic(query_processor.clj:203)" "query_processor$query__GT_native.invoke(query_processor.clj:199)" "driver.common.parameters.values$fn__78245$parse_tag__78238__78250$fn__78254.invoke(values.clj:132)" "driver.common.parameters.values$fn__78245$parse_tag__78238__78250.invoke(values.clj:121)" "driver.common.parameters.values$fn__78500$value_for_tag__78505$fn__78506.invoke(values.clj:307)" "driver.common.parameters.values$fn__78500$value_for_tag__78505.invoke(values.clj:302)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529$fn__78533$iter__78535__78539$fn__78540.invoke(values.clj:326)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529$fn__78533.invoke(values.clj:325)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529.invoke(values.clj:316)" "driver.sql$fn__78608$substitute_native_parameters__78601__78613$fn__78617.invoke(sql.clj:46)" "driver.sql$fn__78608$substitute_native_parameters__78601__78613.invoke(sql.clj:42)" "query_processor.middleware.parameters.native$expand_inner.invokeStatic(native.clj:39)" "query_processor.middleware.parameters.native$expand_inner.invoke(native.clj:30)" "query_processor.middleware.parameters$expand_one.invokeStatic(parameters.clj:49)" "query_processor.middleware.parameters$expand_one.invoke(parameters.clj:40)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "mbql.util.match.impl$replace_in_collection$iter__27400__27404$fn__27405.invoke(impl.cljc:44)" "mbql.util.match.impl$replace_in_collection.invokeStatic(impl.cljc:43)" "mbql.util.match.impl$replace_in_collection.invoke(impl.cljc:38)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:55)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_parameters.invokeStatic(parameters.clj:76)" "query_processor.middleware.parameters$expand_parameters.invoke(parameters.clj:72)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604$fn__48605.invoke(parameters.clj:81)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604.invoke(parameters.clj:78)" "query_processor.middleware.parameters$substitute_parameters$fn__48624.invoke(parameters.clj:109)" "query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__45984.invoke(resolve_referenced.clj:79)" "query_processor.middleware.expand_macros$expand_macros$fn__52276.invoke(expand_macros.clj:184)" "query_processor.middleware.add_timezone_info$add_timezone_info$fn__48404.invoke(add_timezone_info.clj:15)" "query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__51227.invoke(splice_params_in_response.clj:32)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643$fn__50648.invoke(resolve_database_and_driver.clj:35)" "driver$do_with_driver.invokeStatic(driver.clj:60)" "driver$do_with_driver.invoke(driver.clj:56)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643.invoke(resolve_database_and_driver.clj:34)" "query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__46530.invoke(fetch_source_query.clj:286)" "query_processor.middleware.store$initialize_store$fn__46721$fn__46722.invoke(store.clj:11)" "query_processor.store$do_with_store.invokeStatic(store.clj:44)" "query_processor.store$do_with_store.invoke(store.clj:38)" "query_processor.middleware.store$initialize_store$fn__46721.invoke(store.clj:10)" "query_processor.middleware.validate$validate_query$fn__50978.invoke(validate.clj:10)" "query_processor.middleware.normalize_query$normalize$fn__50985.invoke(normalize_query.clj:22)" "query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__48350.invoke(add_rows_truncated.clj:35)" "query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49641.invoke(results_metadata.clj:82)" "query_processor.middleware.constraints$add_default_userland_constraints$fn__48368.invoke(constraints.clj:42)" "query_processor.middleware.process_userland_query$process_userland_query$fn__50919.invoke(process_userland_query.clj:146)" "query_processor.middleware.catch_exceptions$catch_exceptions$fn__51307.invoke(catch_exceptions.clj:169)" "query_processor.reducible$async_qp$qp_STAR___43282$thunk__43283.invoke(reducible.clj:103)" "query_processor.reducible$async_qp$qp_STAR___43282.invoke(reducible.clj:109)" "query_processor.reducible$sync_qp$qp_STAR___43291$fn__43294.invoke(reducible.clj:135)" "query_processor.reducible$sync_qp$qp_STAR___43291.invoke(reducible.clj:134)" "query_processor$process_userland_query.invokeStatic(query_processor.clj:247)" "query_processor$process_userland_query.doInvoke(query_processor.clj:243)" "query_processor$fn__52324$process_query_and_save_execution_BANG___52333$fn__52336.invoke(query_processor.clj:258)" "query_processor$fn__52324$process_query_and_save_execution_BANG___52333.invoke(query_processor.clj:251)" "query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377$fn__52380.invoke(query_processor.clj:270)" "query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377.invoke(query_processor.clj:263)" "api.dataset$run_query_async$fn__65303.invoke(dataset.clj:69)" "query_processor.streaming$streaming_response_STAR_$fn__38418$fn__38419.invoke(streaming.clj:162)" "query_processor.streaming$streaming_response_STAR_$fn__38418.invoke(streaming.clj:161)" "async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)" "async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)" "async.streaming_response$do_f_async$task__26888.invoke(streaming_response.clj:84)"], :card_id nil, :context :ad-hoc, :error "Output of unwrap-field-clause does not match schema: \n\n\t [(not (present? :field)) (not (present? \"id-or-name\")) (not (present? \"options\"))] \n\n", :row_count 0, :running_time 0, :ex-data {:type :schema.core/error, :value nil, :error [(not (present? :field)) (not (present? "id-or-name")) (not (present? "options"))]}, :data {:rows [], :cols []}} 2022-03-25 13:04:43,668 DEBUG middleware.log :: POST /api/dataset 202 [ASYNC: completed] 1.1 s (8 DB calls) App DB connections: 0/10 Jetty threads: 3/50 (3 idle, 0 queued) (56 total active threads) Queries in flight: 1 (0 queued) ``` </details> **Information about your Metabase Installation:** Tested 0.38.5 thru 0.42.3
1.0
Cannot use Sub-Query referencing a GUI question and a date based filter - **Describe the bug** When creating a SQL question, which uses a [Saved Question](https://www.metabase.com/docs/latest/users-guide/referencing-saved-questions-in-queries.html) that is GUI based and a Date filter or Field Filter on a date column, then query fails. It works with Text and Number filters. Workaround is to convert the GUI question to SQL, then it will work. **To Reproduce** 1. Question > Sample > Products - save as "Q1" 2. SQL query > Sample - replace `114` with the ID of Q1 from step 1 - set `filter` as Field Filter of Orders.CreatedAt or a Date filter ``` with exclude_products as {{#114}} select count(*) from orders where true [[and {{filter}}]] -- Field Filter, but will also fail with Date filter: [[and created_at={{datevariable}}]] ``` 3. Run the query - it works ![image](https://user-images.githubusercontent.com/1447303/160118915-681495c3-2a7a-45a6-a216-1419e7f37431.png) 4. Input anything in the filter widget and the query will fail with `Error building query parameter map: Error determining value for parameter \"#114\": The sub-query from referenced question #114 failed with the following error: \"Output of unwrap-field-clause does not match schema: \\n\\n\\t [(not (present? :field)) (not (present? \\\"id-or-name\\\")) (not (present? \\\"options\\\"))] \\n\\n\"` <details><summary>Full stacktrace</summary> ``` 2022-03-25 13:04:43,661 ERROR middleware.catch-exceptions :: Error processing query: null {:database_id 4, :started_at #t "2022-03-25T13:04:42.528037+01:00[Europe/Copenhagen]", :via [{:status :failed, :class clojure.lang.ExceptionInfo, :error "The sub-query from referenced question #114 failed with the following error: \"Output of unwrap-field-clause does not match schema: \\n\\n\\t [(not (present? :field)) (not (present? \\\"id-or-name\\\")) (not (present? \\\"options\\\"))] \\n\\n\"", :stacktrace ["--> driver.common.parameters.values$fn__78245$parse_tag__78238__78250$fn__78254.invoke(values.clj:134)" "driver.common.parameters.values$fn__78245$parse_tag__78238__78250.invoke(values.clj:121)" "driver.common.parameters.values$fn__78500$value_for_tag__78505$fn__78506.invoke(values.clj:307)" "driver.common.parameters.values$fn__78500$value_for_tag__78505.invoke(values.clj:302)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529$fn__78533$iter__78535__78539$fn__78540.invoke(values.clj:326)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529$fn__78533.invoke(values.clj:325)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529.invoke(values.clj:316)" "driver.sql$fn__78608$substitute_native_parameters__78601__78613$fn__78617.invoke(sql.clj:46)" "driver.sql$fn__78608$substitute_native_parameters__78601__78613.invoke(sql.clj:42)" "query_processor.middleware.parameters.native$expand_inner.invokeStatic(native.clj:39)" "query_processor.middleware.parameters.native$expand_inner.invoke(native.clj:30)" "query_processor.middleware.parameters$expand_one.invokeStatic(parameters.clj:49)" "query_processor.middleware.parameters$expand_one.invoke(parameters.clj:40)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "mbql.util.match.impl$replace_in_collection$iter__27400__27404$fn__27405.invoke(impl.cljc:44)" "mbql.util.match.impl$replace_in_collection.invokeStatic(impl.cljc:43)" "mbql.util.match.impl$replace_in_collection.invoke(impl.cljc:38)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:55)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_parameters.invokeStatic(parameters.clj:76)" "query_processor.middleware.parameters$expand_parameters.invoke(parameters.clj:72)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604$fn__48605.invoke(parameters.clj:81)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604.invoke(parameters.clj:78)" "query_processor.middleware.parameters$substitute_parameters$fn__48624.invoke(parameters.clj:109)" "query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__45984.invoke(resolve_referenced.clj:79)" "query_processor.middleware.expand_macros$expand_macros$fn__52276.invoke(expand_macros.clj:184)" "query_processor.middleware.add_timezone_info$add_timezone_info$fn__48404.invoke(add_timezone_info.clj:15)" "query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__51227.invoke(splice_params_in_response.clj:32)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643$fn__50648.invoke(resolve_database_and_driver.clj:35)" "driver$do_with_driver.invokeStatic(driver.clj:60)" "driver$do_with_driver.invoke(driver.clj:56)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643.invoke(resolve_database_and_driver.clj:34)" "query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__46530.invoke(fetch_source_query.clj:286)" "query_processor.middleware.store$initialize_store$fn__46721$fn__46722.invoke(store.clj:11)" "query_processor.store$do_with_store.invokeStatic(store.clj:44)" "query_processor.store$do_with_store.invoke(store.clj:38)" "query_processor.middleware.store$initialize_store$fn__46721.invoke(store.clj:10)" "query_processor.middleware.validate$validate_query$fn__50978.invoke(validate.clj:10)" "query_processor.middleware.normalize_query$normalize$fn__50985.invoke(normalize_query.clj:22)" "query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__48350.invoke(add_rows_truncated.clj:35)" "query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49641.invoke(results_metadata.clj:82)" "query_processor.middleware.constraints$add_default_userland_constraints$fn__48368.invoke(constraints.clj:42)" "query_processor.middleware.process_userland_query$process_userland_query$fn__50919.invoke(process_userland_query.clj:146)" "query_processor.middleware.catch_exceptions$catch_exceptions$fn__51307.invoke(catch_exceptions.clj:169)" "query_processor.reducible$async_qp$qp_STAR___43282$thunk__43283.invoke(reducible.clj:103)" "query_processor.reducible$async_qp$qp_STAR___43282.invoke(reducible.clj:109)" "query_processor.reducible$sync_qp$qp_STAR___43291$fn__43294.invoke(reducible.clj:135)" "query_processor.reducible$sync_qp$qp_STAR___43291.invoke(reducible.clj:134)" "query_processor$process_userland_query.invokeStatic(query_processor.clj:247)" "query_processor$process_userland_query.doInvoke(query_processor.clj:243)" "query_processor$fn__52324$process_query_and_save_execution_BANG___52333$fn__52336.invoke(query_processor.clj:258)" "query_processor$fn__52324$process_query_and_save_execution_BANG___52333.invoke(query_processor.clj:251)" "query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377$fn__52380.invoke(query_processor.clj:270)" "query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377.invoke(query_processor.clj:263)" "api.dataset$run_query_async$fn__65303.invoke(dataset.clj:69)" "query_processor.streaming$streaming_response_STAR_$fn__38418$fn__38419.invoke(streaming.clj:162)" "query_processor.streaming$streaming_response_STAR_$fn__38418.invoke(streaming.clj:161)" "async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)" "async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)" "async.streaming_response$do_f_async$task__26888.invoke(streaming_response.clj:84)"], :error_type :invalid-parameter, :ex-data {:card-query-error? true, :card-id 114, :tag {:id "8c468524-e0f4-aaf7-3757-f84141a5e27d", :name "#114", :display-name "#114", :type :card, :card-id 114}, :type :invalid-parameter}} {:status :failed, :class clojure.lang.ExceptionInfo, :error "Error determining value for parameter \"#114\": The sub-query from referenced question #114 failed with the following error: \"Output of unwrap-field-clause does not match schema: \\n\\n\\t [(not (present? :field)) (not (present? \\\"id-or-name\\\")) (not (present? \\\"options\\\"))] \\n\\n\"", :stacktrace ["--> driver.common.parameters.values$fn__78500$value_for_tag__78505$fn__78506.invoke(values.clj:309)" "driver.common.parameters.values$fn__78500$value_for_tag__78505.invoke(values.clj:302)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529$fn__78533$iter__78535__78539$fn__78540.invoke(values.clj:326)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529$fn__78533.invoke(values.clj:325)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529.invoke(values.clj:316)" "driver.sql$fn__78608$substitute_native_parameters__78601__78613$fn__78617.invoke(sql.clj:46)" "driver.sql$fn__78608$substitute_native_parameters__78601__78613.invoke(sql.clj:42)" "query_processor.middleware.parameters.native$expand_inner.invokeStatic(native.clj:39)" "query_processor.middleware.parameters.native$expand_inner.invoke(native.clj:30)" "query_processor.middleware.parameters$expand_one.invokeStatic(parameters.clj:49)" "query_processor.middleware.parameters$expand_one.invoke(parameters.clj:40)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "mbql.util.match.impl$replace_in_collection$iter__27400__27404$fn__27405.invoke(impl.cljc:44)" "mbql.util.match.impl$replace_in_collection.invokeStatic(impl.cljc:43)" "mbql.util.match.impl$replace_in_collection.invoke(impl.cljc:38)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:55)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_parameters.invokeStatic(parameters.clj:76)" "query_processor.middleware.parameters$expand_parameters.invoke(parameters.clj:72)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604$fn__48605.invoke(parameters.clj:81)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604.invoke(parameters.clj:78)" "query_processor.middleware.parameters$substitute_parameters$fn__48624.invoke(parameters.clj:109)" "query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__45984.invoke(resolve_referenced.clj:79)" "query_processor.middleware.expand_macros$expand_macros$fn__52276.invoke(expand_macros.clj:184)" "query_processor.middleware.add_timezone_info$add_timezone_info$fn__48404.invoke(add_timezone_info.clj:15)" "query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__51227.invoke(splice_params_in_response.clj:32)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643$fn__50648.invoke(resolve_database_and_driver.clj:35)" "driver$do_with_driver.invokeStatic(driver.clj:60)" "driver$do_with_driver.invoke(driver.clj:56)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643.invoke(resolve_database_and_driver.clj:34)" "query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__46530.invoke(fetch_source_query.clj:286)" "query_processor.middleware.store$initialize_store$fn__46721$fn__46722.invoke(store.clj:11)" "query_processor.store$do_with_store.invokeStatic(store.clj:44)" "query_processor.store$do_with_store.invoke(store.clj:38)" "query_processor.middleware.store$initialize_store$fn__46721.invoke(store.clj:10)" "query_processor.middleware.validate$validate_query$fn__50978.invoke(validate.clj:10)" "query_processor.middleware.normalize_query$normalize$fn__50985.invoke(normalize_query.clj:22)" "query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__48350.invoke(add_rows_truncated.clj:35)" "query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49641.invoke(results_metadata.clj:82)" "query_processor.middleware.constraints$add_default_userland_constraints$fn__48368.invoke(constraints.clj:42)" "query_processor.middleware.process_userland_query$process_userland_query$fn__50919.invoke(process_userland_query.clj:146)" "query_processor.middleware.catch_exceptions$catch_exceptions$fn__51307.invoke(catch_exceptions.clj:169)" "query_processor.reducible$async_qp$qp_STAR___43282$thunk__43283.invoke(reducible.clj:103)" "query_processor.reducible$async_qp$qp_STAR___43282.invoke(reducible.clj:109)" "query_processor.reducible$sync_qp$qp_STAR___43291$fn__43294.invoke(reducible.clj:135)" "query_processor.reducible$sync_qp$qp_STAR___43291.invoke(reducible.clj:134)" "query_processor$process_userland_query.invokeStatic(query_processor.clj:247)" "query_processor$process_userland_query.doInvoke(query_processor.clj:243)" "query_processor$fn__52324$process_query_and_save_execution_BANG___52333$fn__52336.invoke(query_processor.clj:258)" "query_processor$fn__52324$process_query_and_save_execution_BANG___52333.invoke(query_processor.clj:251)" "query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377$fn__52380.invoke(query_processor.clj:270)" "query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377.invoke(query_processor.clj:263)" "api.dataset$run_query_async$fn__65303.invoke(dataset.clj:69)" "query_processor.streaming$streaming_response_STAR_$fn__38418$fn__38419.invoke(streaming.clj:162)" "query_processor.streaming$streaming_response_STAR_$fn__38418.invoke(streaming.clj:161)" "async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)" "async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)" "async.streaming_response$do_f_async$task__26888.invoke(streaming_response.clj:84)"], :error_type :invalid-parameter, :ex-data {:tag {:id "8c468524-e0f4-aaf7-3757-f84141a5e27d", :name "#114", :display-name "#114", :type :card, :card-id 114}, :type :invalid-parameter}} {:status :failed, :class clojure.lang.ExceptionInfo, :error "Error building query parameter map: Error determining value for parameter \"#114\": The sub-query from referenced question #114 failed with the following error: \"Output of unwrap-field-clause does not match schema: \\n\\n\\t [(not (present? :field)) (not (present? \\\"id-or-name\\\")) (not (present? \\\"options\\\"))] \\n\\n\"", :stacktrace ["--> driver.common.parameters.values$fn__78524$query__GT_params_map__78529$fn__78533.invoke(values.clj:334)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529.invoke(values.clj:316)" "driver.sql$fn__78608$substitute_native_parameters__78601__78613$fn__78617.invoke(sql.clj:46)" "driver.sql$fn__78608$substitute_native_parameters__78601__78613.invoke(sql.clj:42)" "query_processor.middleware.parameters.native$expand_inner.invokeStatic(native.clj:39)" "query_processor.middleware.parameters.native$expand_inner.invoke(native.clj:30)" "query_processor.middleware.parameters$expand_one.invokeStatic(parameters.clj:49)" "query_processor.middleware.parameters$expand_one.invoke(parameters.clj:40)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "mbql.util.match.impl$replace_in_collection$iter__27400__27404$fn__27405.invoke(impl.cljc:44)" "mbql.util.match.impl$replace_in_collection.invokeStatic(impl.cljc:43)" "mbql.util.match.impl$replace_in_collection.invoke(impl.cljc:38)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:55)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_parameters.invokeStatic(parameters.clj:76)" "query_processor.middleware.parameters$expand_parameters.invoke(parameters.clj:72)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604$fn__48605.invoke(parameters.clj:81)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604.invoke(parameters.clj:78)" "query_processor.middleware.parameters$substitute_parameters$fn__48624.invoke(parameters.clj:109)" "query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__45984.invoke(resolve_referenced.clj:79)" "query_processor.middleware.expand_macros$expand_macros$fn__52276.invoke(expand_macros.clj:184)" "query_processor.middleware.add_timezone_info$add_timezone_info$fn__48404.invoke(add_timezone_info.clj:15)" "query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__51227.invoke(splice_params_in_response.clj:32)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643$fn__50648.invoke(resolve_database_and_driver.clj:35)" "driver$do_with_driver.invokeStatic(driver.clj:60)" "driver$do_with_driver.invoke(driver.clj:56)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643.invoke(resolve_database_and_driver.clj:34)" "query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__46530.invoke(fetch_source_query.clj:286)" "query_processor.middleware.store$initialize_store$fn__46721$fn__46722.invoke(store.clj:11)" "query_processor.store$do_with_store.invokeStatic(store.clj:44)" "query_processor.store$do_with_store.invoke(store.clj:38)" "query_processor.middleware.store$initialize_store$fn__46721.invoke(store.clj:10)" "query_processor.middleware.validate$validate_query$fn__50978.invoke(validate.clj:10)" "query_processor.middleware.normalize_query$normalize$fn__50985.invoke(normalize_query.clj:22)" "query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__48350.invoke(add_rows_truncated.clj:35)" "query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49641.invoke(results_metadata.clj:82)" "query_processor.middleware.constraints$add_default_userland_constraints$fn__48368.invoke(constraints.clj:42)" "query_processor.middleware.process_userland_query$process_userland_query$fn__50919.invoke(process_userland_query.clj:146)" "query_processor.middleware.catch_exceptions$catch_exceptions$fn__51307.invoke(catch_exceptions.clj:169)" "query_processor.reducible$async_qp$qp_STAR___43282$thunk__43283.invoke(reducible.clj:103)" "query_processor.reducible$async_qp$qp_STAR___43282.invoke(reducible.clj:109)" "query_processor.reducible$sync_qp$qp_STAR___43291$fn__43294.invoke(reducible.clj:135)" "query_processor.reducible$sync_qp$qp_STAR___43291.invoke(reducible.clj:134)" "query_processor$process_userland_query.invokeStatic(query_processor.clj:247)" "query_processor$process_userland_query.doInvoke(query_processor.clj:243)" "query_processor$fn__52324$process_query_and_save_execution_BANG___52333$fn__52336.invoke(query_processor.clj:258)" "query_processor$fn__52324$process_query_and_save_execution_BANG___52333.invoke(query_processor.clj:251)" "query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377$fn__52380.invoke(query_processor.clj:270)" "query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377.invoke(query_processor.clj:263)" "api.dataset$run_query_async$fn__65303.invoke(dataset.clj:69)" "query_processor.streaming$streaming_response_STAR_$fn__38418$fn__38419.invoke(streaming.clj:162)" "query_processor.streaming$streaming_response_STAR_$fn__38418.invoke(streaming.clj:161)" "async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)" "async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)" "async.streaming_response$do_f_async$task__26888.invoke(streaming_response.clj:84)"], :error_type :invalid-parameter, :ex-data {:type :invalid-parameter, :tags {"#114" {:id "8c468524-e0f4-aaf7-3757-f84141a5e27d", :name "#114", :display-name "#114", :type :card, :card-id 114}, "filter" {:id "79400356-2c24-782c-fa46-e9d29b9dacfb", :name "filter", :display-name "Filter", :type :dimension, :dimension [:field 86 nil], :widget-type :date/all-options, :default nil}}, :params [{:type :date/all-options, :value "past30years", :target [:dimension [:template-tag "filter"]], :id "79400356-2c24-782c-fa46-e9d29b9dacfb"}]}}], :error_type :invalid-query, :json_query {:database 4, :native {:template-tags {:#114 {:id "8c468524-e0f4-aaf7-3757-f84141a5e27d", :name "#114", :display-name "#114", :type "card", :card-id 114}, :filter {:id "79400356-2c24-782c-fa46-e9d29b9dacfb", :name "filter", :display-name "Filter", :type "dimension", :dimension ["field" 86 nil], :widget-type "date/all-options", :default nil}}, :query "with exclude_products as {{#114}}\nselect count(*) from orders where {{filter}}"}, :type "native", :parameters [{:type "date/all-options", :value "past30years", :target ["dimension" ["template-tag" "filter"]], :id "79400356-2c24-782c-fa46-e9d29b9dacfb"}], :middleware {:js-int-to-string? true, :add-default-userland-constraints? true}}, :status :failed, :class clojure.lang.ExceptionInfo, :stacktrace ["--> models.params$fn__38495$unwrap_field_clause__38500.invoke(params.clj:21)" "query_processor.middleware.parameters.mbql$fn__48480$build_filter_clause__48485$fn__48495.invoke(mbql.clj:57)" "query_processor.middleware.parameters.mbql$fn__48480$build_filter_clause__48485.invoke(mbql.clj:44)" "query_processor.middleware.parameters.mbql$expand.invokeStatic(mbql.clj:86)" "query_processor.middleware.parameters.mbql$expand.invoke(mbql.clj:72)" "query_processor.middleware.parameters$expand_mbql_params.invokeStatic(parameters.clj:36)" "query_processor.middleware.parameters$expand_mbql_params.invoke(parameters.clj:32)" "query_processor.middleware.parameters$expand_one.invokeStatic(parameters.clj:48)" "query_processor.middleware.parameters$expand_one.invoke(parameters.clj:40)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "mbql.util.match.impl$replace_in_collection$iter__27400__27404$fn__27405.invoke(impl.cljc:44)" "mbql.util.match.impl$replace_in_collection.invokeStatic(impl.cljc:43)" "mbql.util.match.impl$replace_in_collection.invoke(impl.cljc:38)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:55)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_parameters.invokeStatic(parameters.clj:76)" "query_processor.middleware.parameters$expand_parameters.invoke(parameters.clj:72)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604$fn__48605.invoke(parameters.clj:81)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604.invoke(parameters.clj:78)" "query_processor.middleware.parameters$substitute_parameters$fn__48624.invoke(parameters.clj:109)" "query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__45984.invoke(resolve_referenced.clj:79)" "query_processor.middleware.expand_macros$expand_macros$fn__52276.invoke(expand_macros.clj:184)" "query_processor.middleware.add_timezone_info$add_timezone_info$fn__48404.invoke(add_timezone_info.clj:15)" "query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__51227.invoke(splice_params_in_response.clj:32)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643$fn__50648.invoke(resolve_database_and_driver.clj:35)" "driver$do_with_driver.invokeStatic(driver.clj:60)" "driver$do_with_driver.invoke(driver.clj:56)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643.invoke(resolve_database_and_driver.clj:34)" "query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__46530.invoke(fetch_source_query.clj:286)" "query_processor.middleware.store$initialize_store$fn__46721$fn__46722.invoke(store.clj:11)" "query_processor.store$do_with_store.invokeStatic(store.clj:42)" "query_processor.store$do_with_store.invoke(store.clj:38)" "query_processor.middleware.store$initialize_store$fn__46721.invoke(store.clj:10)" "query_processor.middleware.validate$validate_query$fn__50978.invoke(validate.clj:10)" "query_processor.middleware.normalize_query$normalize$fn__50985.invoke(normalize_query.clj:22)" "query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__48350.invoke(add_rows_truncated.clj:35)" "query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49641.invoke(results_metadata.clj:82)" "query_processor.reducible$async_qp$qp_STAR___43282$thunk__43283.invoke(reducible.clj:103)" "query_processor.reducible$async_qp$qp_STAR___43282.invoke(reducible.clj:109)" "query_processor.reducible$sync_qp$qp_STAR___43291$fn__43294.invoke(reducible.clj:135)" "query_processor.reducible$sync_qp$qp_STAR___43291.invoke(reducible.clj:134)" "query_processor$preprocess_query.invokeStatic(query_processor.clj:172)" "query_processor$preprocess_query.invoke(query_processor.clj:164)" "query_processor$query__GT_native.invokeStatic(query_processor.clj:203)" "query_processor$query__GT_native.invoke(query_processor.clj:199)" "driver.common.parameters.values$fn__78245$parse_tag__78238__78250$fn__78254.invoke(values.clj:132)" "driver.common.parameters.values$fn__78245$parse_tag__78238__78250.invoke(values.clj:121)" "driver.common.parameters.values$fn__78500$value_for_tag__78505$fn__78506.invoke(values.clj:307)" "driver.common.parameters.values$fn__78500$value_for_tag__78505.invoke(values.clj:302)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529$fn__78533$iter__78535__78539$fn__78540.invoke(values.clj:326)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529$fn__78533.invoke(values.clj:325)" "driver.common.parameters.values$fn__78524$query__GT_params_map__78529.invoke(values.clj:316)" "driver.sql$fn__78608$substitute_native_parameters__78601__78613$fn__78617.invoke(sql.clj:46)" "driver.sql$fn__78608$substitute_native_parameters__78601__78613.invoke(sql.clj:42)" "query_processor.middleware.parameters.native$expand_inner.invokeStatic(native.clj:39)" "query_processor.middleware.parameters.native$expand_inner.invoke(native.clj:30)" "query_processor.middleware.parameters$expand_one.invokeStatic(parameters.clj:49)" "query_processor.middleware.parameters$expand_one.invoke(parameters.clj:40)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "mbql.util.match.impl$replace_in_collection$iter__27400__27404$fn__27405.invoke(impl.cljc:44)" "mbql.util.match.impl$replace_in_collection.invokeStatic(impl.cljc:43)" "mbql.util.match.impl$replace_in_collection.invoke(impl.cljc:38)" "query_processor.middleware.parameters$expand_all$replace_48583__48584.invoke(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:58)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:55)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:52)" "query_processor.middleware.parameters$expand_parameters.invokeStatic(parameters.clj:76)" "query_processor.middleware.parameters$expand_parameters.invoke(parameters.clj:72)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604$fn__48605.invoke(parameters.clj:81)" "query_processor.middleware.parameters$fn__48599$substitute_parameters_STAR___48604.invoke(parameters.clj:78)" "query_processor.middleware.parameters$substitute_parameters$fn__48624.invoke(parameters.clj:109)" "query_processor.middleware.resolve_referenced$resolve_referenced_card_resources$fn__45984.invoke(resolve_referenced.clj:79)" "query_processor.middleware.expand_macros$expand_macros$fn__52276.invoke(expand_macros.clj:184)" "query_processor.middleware.add_timezone_info$add_timezone_info$fn__48404.invoke(add_timezone_info.clj:15)" "query_processor.middleware.splice_params_in_response$splice_params_in_response$fn__51227.invoke(splice_params_in_response.clj:32)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643$fn__50648.invoke(resolve_database_and_driver.clj:35)" "driver$do_with_driver.invokeStatic(driver.clj:60)" "driver$do_with_driver.invoke(driver.clj:56)" "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__50643.invoke(resolve_database_and_driver.clj:34)" "query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__46530.invoke(fetch_source_query.clj:286)" "query_processor.middleware.store$initialize_store$fn__46721$fn__46722.invoke(store.clj:11)" "query_processor.store$do_with_store.invokeStatic(store.clj:44)" "query_processor.store$do_with_store.invoke(store.clj:38)" "query_processor.middleware.store$initialize_store$fn__46721.invoke(store.clj:10)" "query_processor.middleware.validate$validate_query$fn__50978.invoke(validate.clj:10)" "query_processor.middleware.normalize_query$normalize$fn__50985.invoke(normalize_query.clj:22)" "query_processor.middleware.add_rows_truncated$add_rows_truncated$fn__48350.invoke(add_rows_truncated.clj:35)" "query_processor.middleware.results_metadata$record_and_return_metadata_BANG_$fn__49641.invoke(results_metadata.clj:82)" "query_processor.middleware.constraints$add_default_userland_constraints$fn__48368.invoke(constraints.clj:42)" "query_processor.middleware.process_userland_query$process_userland_query$fn__50919.invoke(process_userland_query.clj:146)" "query_processor.middleware.catch_exceptions$catch_exceptions$fn__51307.invoke(catch_exceptions.clj:169)" "query_processor.reducible$async_qp$qp_STAR___43282$thunk__43283.invoke(reducible.clj:103)" "query_processor.reducible$async_qp$qp_STAR___43282.invoke(reducible.clj:109)" "query_processor.reducible$sync_qp$qp_STAR___43291$fn__43294.invoke(reducible.clj:135)" "query_processor.reducible$sync_qp$qp_STAR___43291.invoke(reducible.clj:134)" "query_processor$process_userland_query.invokeStatic(query_processor.clj:247)" "query_processor$process_userland_query.doInvoke(query_processor.clj:243)" "query_processor$fn__52324$process_query_and_save_execution_BANG___52333$fn__52336.invoke(query_processor.clj:258)" "query_processor$fn__52324$process_query_and_save_execution_BANG___52333.invoke(query_processor.clj:251)" "query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377$fn__52380.invoke(query_processor.clj:270)" "query_processor$fn__52368$process_query_and_save_with_max_results_constraints_BANG___52377.invoke(query_processor.clj:263)" "api.dataset$run_query_async$fn__65303.invoke(dataset.clj:69)" "query_processor.streaming$streaming_response_STAR_$fn__38418$fn__38419.invoke(streaming.clj:162)" "query_processor.streaming$streaming_response_STAR_$fn__38418.invoke(streaming.clj:161)" "async.streaming_response$do_f_STAR_.invokeStatic(streaming_response.clj:65)" "async.streaming_response$do_f_STAR_.invoke(streaming_response.clj:63)" "async.streaming_response$do_f_async$task__26888.invoke(streaming_response.clj:84)"], :card_id nil, :context :ad-hoc, :error "Output of unwrap-field-clause does not match schema: \n\n\t [(not (present? :field)) (not (present? \"id-or-name\")) (not (present? \"options\"))] \n\n", :row_count 0, :running_time 0, :ex-data {:type :schema.core/error, :value nil, :error [(not (present? :field)) (not (present? "id-or-name")) (not (present? "options"))]}, :data {:rows [], :cols []}} 2022-03-25 13:04:43,668 DEBUG middleware.log :: POST /api/dataset 202 [ASYNC: completed] 1.1 s (8 DB calls) App DB connections: 0/10 Jetty threads: 3/50 (3 idle, 0 queued) (56 total active threads) Queries in flight: 1 (0 queued) ``` </details> **Information about your Metabase Installation:** Tested 0.38.5 thru 0.42.3
non_defect
cannot use sub query referencing a gui question and a date based filter describe the bug when creating a sql question which uses a that is gui based and a date filter or field filter on a date column then query fails it works with text and number filters workaround is to convert the gui question to sql then it will work to reproduce question sample products save as sql query sample replace with the id of from step set filter as field filter of orders createdat or a date filter with exclude products as select count from orders where true field filter but will also fail with date filter run the query it works input anything in the filter widget and the query will fail with error building query parameter map error determining value for parameter the sub query from referenced question failed with the following error output of unwrap field clause does not match schema n n t n n full stacktrace error middleware catch exceptions error processing query null database id started at t via status failed class clojure lang exceptioninfo error the sub query from referenced question failed with the following error output of unwrap field clause does not match schema n n t n n stacktrace driver common parameters values fn parse tag fn invoke values clj driver common parameters values fn parse tag invoke values clj driver common parameters values fn value for tag fn invoke values clj driver common parameters values fn value for tag invoke values clj driver common parameters values fn query gt params map fn iter fn invoke values clj driver common parameters values fn query gt params map fn invoke values clj driver common parameters values fn query gt params map invoke values clj driver sql fn substitute native parameters fn invoke sql clj driver sql fn substitute native parameters invoke sql clj query processor middleware parameters native expand inner invokestatic native clj query processor middleware parameters native expand inner invoke native clj query processor middleware parameters expand one invokestatic parameters clj query processor middleware parameters expand one invoke parameters clj query processor middleware parameters expand all replace invoke parameters clj mbql util match impl replace in collection iter fn invoke impl cljc mbql util match impl replace in collection invokestatic impl cljc mbql util match impl replace in collection invoke impl cljc query processor middleware parameters expand all replace invoke parameters clj query processor middleware parameters expand all invokestatic parameters clj query processor middleware parameters expand all invoke parameters clj query processor middleware parameters expand all invokestatic parameters clj query processor middleware parameters expand all invoke parameters clj query processor middleware parameters expand parameters invokestatic parameters clj query processor middleware parameters expand parameters invoke parameters clj query processor middleware parameters fn substitute parameters star fn invoke parameters clj query processor middleware parameters fn substitute parameters star invoke parameters clj query processor middleware parameters substitute parameters fn invoke parameters clj query processor middleware resolve referenced resolve referenced card resources fn invoke resolve referenced clj query processor middleware expand macros expand macros fn invoke expand macros clj query processor middleware add timezone info add timezone info fn invoke add timezone info clj query processor middleware splice params in response splice params in response fn invoke splice params in response clj query processor middleware resolve database and driver resolve database and driver fn fn invoke resolve database and driver clj driver do with driver invokestatic driver clj driver do with driver invoke driver clj query processor middleware resolve database and driver resolve database and driver fn invoke resolve database and driver clj query processor middleware fetch source query resolve card id source tables fn invoke fetch source query clj query processor middleware store initialize store fn fn invoke store clj query processor store do with store invokestatic store clj query processor store do with store invoke store clj query processor middleware store initialize store fn invoke store clj query processor middleware validate validate query fn invoke validate clj query processor middleware normalize query normalize fn invoke normalize query clj query processor middleware add rows truncated add rows truncated fn invoke add rows truncated clj query processor middleware results metadata record and return metadata bang fn invoke results metadata clj query processor middleware constraints add default userland constraints fn invoke constraints clj query processor middleware process userland query process userland query fn invoke process userland query clj query processor middleware catch exceptions catch exceptions fn invoke catch exceptions clj query processor reducible async qp qp star thunk invoke reducible clj query processor reducible async qp qp star invoke reducible clj query processor reducible sync qp qp star fn invoke reducible clj query processor reducible sync qp qp star invoke reducible clj query processor process userland query invokestatic query processor clj query processor process userland query doinvoke query processor clj query processor fn process query and save execution bang fn invoke query processor clj query processor fn process query and save execution bang invoke query processor clj query processor fn process query and save with max results constraints bang fn invoke query processor clj query processor fn process query and save with max results constraints bang invoke query processor clj api dataset run query async fn invoke dataset clj query processor streaming streaming response star fn fn invoke streaming clj query processor streaming streaming response star fn invoke streaming clj async streaming response do f star invokestatic streaming response clj async streaming response do f star invoke streaming response clj async streaming response do f async task invoke streaming response clj error type invalid parameter ex data card query error true card id tag id name display name type card card id type invalid parameter status failed class clojure lang exceptioninfo error error determining value for parameter the sub query from referenced question failed with the following error output of unwrap field clause does not match schema n n t n n stacktrace driver common parameters values fn value for tag fn invoke values clj driver common parameters values fn value for tag invoke values clj driver common parameters values fn query gt params map fn iter fn invoke values clj driver common parameters values fn query gt params map fn invoke values clj driver common parameters values fn query gt params map invoke values clj driver sql fn substitute native parameters fn invoke sql clj driver sql fn substitute native parameters invoke sql clj query processor middleware parameters native expand inner invokestatic native clj query processor middleware parameters native expand inner invoke native clj query processor middleware parameters expand one invokestatic parameters clj query processor middleware parameters expand one invoke parameters clj query processor middleware parameters expand all replace invoke parameters clj mbql util match impl replace in collection iter fn invoke impl cljc mbql util match impl replace in collection invokestatic impl cljc mbql util match impl replace in collection invoke impl cljc query processor middleware parameters expand all replace invoke parameters clj query processor middleware parameters expand all invokestatic parameters clj query processor middleware parameters expand all invoke parameters clj query processor middleware parameters expand all invokestatic parameters clj query processor middleware parameters expand all invoke parameters clj query processor middleware parameters expand parameters invokestatic parameters clj query processor middleware parameters expand parameters invoke parameters clj query processor middleware parameters fn substitute parameters star fn invoke parameters clj query processor middleware parameters fn substitute parameters star invoke parameters clj query processor middleware parameters substitute parameters fn invoke parameters clj query processor middleware resolve referenced resolve referenced card resources fn invoke resolve referenced clj query processor middleware expand macros expand macros fn invoke expand macros clj query processor middleware add timezone info add timezone info fn invoke add timezone info clj query processor middleware splice params in response splice params in response fn invoke splice params in response clj query processor middleware resolve database and driver resolve database and driver fn fn invoke resolve database and driver clj driver do with driver invokestatic driver clj driver do with driver invoke driver clj query processor middleware resolve database and driver resolve database and driver fn invoke resolve database and driver clj query processor middleware fetch source query resolve card id source tables fn invoke fetch source query clj query processor middleware store initialize store fn fn invoke store clj query processor store do with store invokestatic store clj query processor store do with store invoke store clj query processor middleware store initialize store fn invoke store clj query processor middleware validate validate query fn invoke validate clj query processor middleware normalize query normalize fn invoke normalize query clj query processor middleware add rows truncated add rows truncated fn invoke add rows truncated clj query processor middleware results metadata record and return metadata bang fn invoke results metadata clj query processor middleware constraints add default userland constraints fn invoke constraints clj query processor middleware process userland query process userland query fn invoke process userland query clj query processor middleware catch exceptions catch exceptions fn invoke catch exceptions clj query processor reducible async qp qp star thunk invoke reducible clj query processor reducible async qp qp star invoke reducible clj query processor reducible sync qp qp star fn invoke reducible clj query processor reducible sync qp qp star invoke reducible clj query processor process userland query invokestatic query processor clj query processor process userland query doinvoke query processor clj query processor fn process query and save execution bang fn invoke query processor clj query processor fn process query and save execution bang invoke query processor clj query processor fn process query and save with max results constraints bang fn invoke query processor clj query processor fn process query and save with max results constraints bang invoke query processor clj api dataset run query async fn invoke dataset clj query processor streaming streaming response star fn fn invoke streaming clj query processor streaming streaming response star fn invoke streaming clj async streaming response do f star invokestatic streaming response clj async streaming response do f star invoke streaming response clj async streaming response do f async task invoke streaming response clj error type invalid parameter ex data tag id name display name type card card id type invalid parameter status failed class clojure lang exceptioninfo error error building query parameter map error determining value for parameter the sub query from referenced question failed with the following error output of unwrap field clause does not match schema n n t n n stacktrace driver common parameters values fn query gt params map fn invoke values clj driver common parameters values fn query gt params map invoke values clj driver sql fn substitute native parameters fn invoke sql clj driver sql fn substitute native parameters invoke sql clj query processor middleware parameters native expand inner invokestatic native clj query processor middleware parameters native expand inner invoke native clj query processor middleware parameters expand one invokestatic parameters clj query processor middleware parameters expand one invoke parameters clj query processor middleware parameters expand all replace invoke parameters clj mbql util match impl replace in collection iter fn invoke impl cljc mbql util match impl replace in collection invokestatic impl cljc mbql util match impl replace in collection invoke impl cljc query processor middleware parameters expand all replace invoke parameters clj query processor middleware parameters expand all invokestatic parameters clj query processor middleware parameters expand all invoke parameters clj query processor middleware parameters expand all invokestatic parameters clj query processor middleware parameters expand all invoke parameters clj query processor middleware parameters expand parameters invokestatic parameters clj query processor middleware parameters expand parameters invoke parameters clj query processor middleware parameters fn substitute parameters star fn invoke parameters clj query processor middleware parameters fn substitute parameters star invoke parameters clj query processor middleware parameters substitute parameters fn invoke parameters clj query processor middleware resolve referenced resolve referenced card resources fn invoke resolve referenced clj query processor middleware expand macros expand macros fn invoke expand macros clj query processor middleware add timezone info add timezone info fn invoke add timezone info clj query processor middleware splice params in response splice params in response fn invoke splice params in response clj query processor middleware resolve database and driver resolve database and driver fn fn invoke resolve database and driver clj driver do with driver invokestatic driver clj driver do with driver invoke driver clj query processor middleware resolve database and driver resolve database and driver fn invoke resolve database and driver clj query processor middleware fetch source query resolve card id source tables fn invoke fetch source query clj query processor middleware store initialize store fn fn invoke store clj query processor store do with store invokestatic store clj query processor store do with store invoke store clj query processor middleware store initialize store fn invoke store clj query processor middleware validate validate query fn invoke validate clj query processor middleware normalize query normalize fn invoke normalize query clj query processor middleware add rows truncated add rows truncated fn invoke add rows truncated clj query processor middleware results metadata record and return metadata bang fn invoke results metadata clj query processor middleware constraints add default userland constraints fn invoke constraints clj query processor middleware process userland query process userland query fn invoke process userland query clj query processor middleware catch exceptions catch exceptions fn invoke catch exceptions clj query processor reducible async qp qp star thunk invoke reducible clj query processor reducible async qp qp star invoke reducible clj query processor reducible sync qp qp star fn invoke reducible clj query processor reducible sync qp qp star invoke reducible clj query processor process userland query invokestatic query processor clj query processor process userland query doinvoke query processor clj query processor fn process query and save execution bang fn invoke query processor clj query processor fn process query and save execution bang invoke query processor clj query processor fn process query and save with max results constraints bang fn invoke query processor clj query processor fn process query and save with max results constraints bang invoke query processor clj api dataset run query async fn invoke dataset clj query processor streaming streaming response star fn fn invoke streaming clj query processor streaming streaming response star fn invoke streaming clj async streaming response do f star invokestatic streaming response clj async streaming response do f star invoke streaming response clj async streaming response do f async task invoke streaming response clj error type invalid parameter ex data type invalid parameter tags id name display name type card card id filter id name filter display name filter type dimension dimension widget type date all options default nil params type date all options value target id error type invalid query json query database native template tags id name display name type card card id filter id name filter display name filter type dimension dimension widget type date all options default nil query with exclude products as nselect count from orders where filter type native parameters type date all options value target id middleware js int to string true add default userland constraints true status failed class clojure lang exceptioninfo stacktrace models params fn unwrap field clause invoke params clj query processor middleware parameters mbql fn build filter clause fn invoke mbql clj query processor middleware parameters mbql fn build filter clause invoke mbql clj query processor middleware parameters mbql expand invokestatic mbql clj query processor middleware parameters mbql expand invoke mbql clj query processor middleware parameters expand mbql params invokestatic parameters clj query processor middleware parameters expand mbql params invoke parameters clj query processor middleware parameters expand one invokestatic parameters clj query processor middleware parameters expand one invoke parameters clj query processor middleware parameters expand all replace invoke parameters clj mbql util match impl replace in collection iter fn invoke impl cljc mbql util match impl replace in collection invokestatic impl cljc mbql util match impl replace in collection invoke impl cljc query processor middleware parameters expand all replace invoke parameters clj query processor middleware parameters expand all invokestatic parameters clj query processor middleware parameters expand all invoke parameters clj query processor middleware parameters expand all invokestatic parameters clj query processor middleware parameters expand all invoke parameters clj query processor middleware parameters expand parameters invokestatic parameters clj query processor middleware parameters expand parameters invoke parameters clj query processor middleware parameters fn substitute parameters star fn invoke parameters clj query processor middleware parameters fn substitute parameters star invoke parameters clj query processor middleware parameters substitute parameters fn invoke parameters clj query processor middleware resolve referenced resolve referenced card resources fn invoke resolve referenced clj query processor middleware expand macros expand macros fn invoke expand macros clj query processor middleware add timezone info add timezone info fn invoke add timezone info clj query processor middleware splice params in response splice params in response fn invoke splice params in response clj query processor middleware resolve database and driver resolve database and driver fn fn invoke resolve database and driver clj driver do with driver invokestatic driver clj driver do with driver invoke driver clj query processor middleware resolve database and driver resolve database and driver fn invoke resolve database and driver clj query processor middleware fetch source query resolve card id source tables fn invoke fetch source query clj query processor middleware store initialize store fn fn invoke store clj query processor store do with store invokestatic store clj query processor store do with store invoke store clj query processor middleware store initialize store fn invoke store clj query processor middleware validate validate query fn invoke validate clj query processor middleware normalize query normalize fn invoke normalize query clj query processor middleware add rows truncated add rows truncated fn invoke add rows truncated clj query processor middleware results metadata record and return metadata bang fn invoke results metadata clj query processor reducible async qp qp star thunk invoke reducible clj query processor reducible async qp qp star invoke reducible clj query processor reducible sync qp qp star fn invoke reducible clj query processor reducible sync qp qp star invoke reducible clj query processor preprocess query invokestatic query processor clj query processor preprocess query invoke query processor clj query processor query gt native invokestatic query processor clj query processor query gt native invoke query processor clj driver common parameters values fn parse tag fn invoke values clj driver common parameters values fn parse tag invoke values clj driver common parameters values fn value for tag fn invoke values clj driver common parameters values fn value for tag invoke values clj driver common parameters values fn query gt params map fn iter fn invoke values clj driver common parameters values fn query gt params map fn invoke values clj driver common parameters values fn query gt params map invoke values clj driver sql fn substitute native parameters fn invoke sql clj driver sql fn substitute native parameters invoke sql clj query processor middleware parameters native expand inner invokestatic native clj query processor middleware parameters native expand inner invoke native clj query processor middleware parameters expand one invokestatic parameters clj query processor middleware parameters expand one invoke parameters clj query processor middleware parameters expand all replace invoke parameters clj mbql util match impl replace in collection iter fn invoke impl cljc mbql util match impl replace in collection invokestatic impl cljc mbql util match impl replace in collection invoke impl cljc query processor middleware parameters expand all replace invoke parameters clj query processor middleware parameters expand all invokestatic parameters clj query processor middleware parameters expand all invoke parameters clj query processor middleware parameters expand all invokestatic parameters clj query processor middleware parameters expand all invoke parameters clj query processor middleware parameters expand parameters invokestatic parameters clj query processor middleware parameters expand parameters invoke parameters clj query processor middleware parameters fn substitute parameters star fn invoke parameters clj query processor middleware parameters fn substitute parameters star invoke parameters clj query processor middleware parameters substitute parameters fn invoke parameters clj query processor middleware resolve referenced resolve referenced card resources fn invoke resolve referenced clj query processor middleware expand macros expand macros fn invoke expand macros clj query processor middleware add timezone info add timezone info fn invoke add timezone info clj query processor middleware splice params in response splice params in response fn invoke splice params in response clj query processor middleware resolve database and driver resolve database and driver fn fn invoke resolve database and driver clj driver do with driver invokestatic driver clj driver do with driver invoke driver clj query processor middleware resolve database and driver resolve database and driver fn invoke resolve database and driver clj query processor middleware fetch source query resolve card id source tables fn invoke fetch source query clj query processor middleware store initialize store fn fn invoke store clj query processor store do with store invokestatic store clj query processor store do with store invoke store clj query processor middleware store initialize store fn invoke store clj query processor middleware validate validate query fn invoke validate clj query processor middleware normalize query normalize fn invoke normalize query clj query processor middleware add rows truncated add rows truncated fn invoke add rows truncated clj query processor middleware results metadata record and return metadata bang fn invoke results metadata clj query processor middleware constraints add default userland constraints fn invoke constraints clj query processor middleware process userland query process userland query fn invoke process userland query clj query processor middleware catch exceptions catch exceptions fn invoke catch exceptions clj query processor reducible async qp qp star thunk invoke reducible clj query processor reducible async qp qp star invoke reducible clj query processor reducible sync qp qp star fn invoke reducible clj query processor reducible sync qp qp star invoke reducible clj query processor process userland query invokestatic query processor clj query processor process userland query doinvoke query processor clj query processor fn process query and save execution bang fn invoke query processor clj query processor fn process query and save execution bang invoke query processor clj query processor fn process query and save with max results constraints bang fn invoke query processor clj query processor fn process query and save with max results constraints bang invoke query processor clj api dataset run query async fn invoke dataset clj query processor streaming streaming response star fn fn invoke streaming clj query processor streaming streaming response star fn invoke streaming clj async streaming response do f star invokestatic streaming response clj async streaming response do f star invoke streaming response clj async streaming response do f async task invoke streaming response clj card id nil context ad hoc error output of unwrap field clause does not match schema n n t n n row count running time ex data type schema core error value nil error data rows cols debug middleware log post api dataset s db calls app db connections jetty threads idle queued total active threads queries in flight queued information about your metabase installation tested thru
0
712,650
24,502,306,402
IssuesEvent
2022-10-10 13:43:18
AY2223S1-CS2103-F14-3/tp
https://api.github.com/repos/AY2223S1-CS2103-F14-3/tp
reopened
As a user with many application emails, I want to add details such as location, date, time and contact person/information for an interview
type.Story priority.High
...so that I need not sieve through my emails to remind myself of these details.
1.0
As a user with many application emails, I want to add details such as location, date, time and contact person/information for an interview - ...so that I need not sieve through my emails to remind myself of these details.
non_defect
as a user with many application emails i want to add details such as location date time and contact person information for an interview so that i need not sieve through my emails to remind myself of these details
0
54,729
6,845,665,116
IssuesEvent
2017-11-13 09:12:16
mozilla/OpenDesign
https://api.github.com/repos/mozilla/OpenDesign
closed
Mana Icon/Branding
Logo design needed
## Goal: An icon/new branding for [Mana](https://mana.mozilla.org/) ## Info: This request has come out of the current (work in progress) redesign of the SSO (Okta) Dashboard. See below. Currently all Mozilla sites are represented by the black highlight Zilla treatment. This is a placeholder until icons are available. Yulia (Brand) is the originator of this request and should be kept apprised of progress/requests for feedback and direction. ![landing](https://cloud.githubusercontent.com/assets/18686986/24779166/6aadde9a-1ae3-11e7-9396-e1891293ac83.png) ## Style Information: Single color icon that can be used with or without a wordmark is preferable. Please default to Yulia for further direction. Also, the site owner/team should be consulted for conceptual direction. ## Deadline: Initial Beta release for the SSO (Auth0) Dashboard is slated for April 21st 2017. If this can be ready by then for initial release implementation, fantastic! Though understandable if that timeline is too aggressive and can not be met. We will have regular releases/feature enhancements and can integrate icon(s) updates as they are available. ## Tag: Design Needed
1.0
Mana Icon/Branding - ## Goal: An icon/new branding for [Mana](https://mana.mozilla.org/) ## Info: This request has come out of the current (work in progress) redesign of the SSO (Okta) Dashboard. See below. Currently all Mozilla sites are represented by the black highlight Zilla treatment. This is a placeholder until icons are available. Yulia (Brand) is the originator of this request and should be kept apprised of progress/requests for feedback and direction. ![landing](https://cloud.githubusercontent.com/assets/18686986/24779166/6aadde9a-1ae3-11e7-9396-e1891293ac83.png) ## Style Information: Single color icon that can be used with or without a wordmark is preferable. Please default to Yulia for further direction. Also, the site owner/team should be consulted for conceptual direction. ## Deadline: Initial Beta release for the SSO (Auth0) Dashboard is slated for April 21st 2017. If this can be ready by then for initial release implementation, fantastic! Though understandable if that timeline is too aggressive and can not be met. We will have regular releases/feature enhancements and can integrate icon(s) updates as they are available. ## Tag: Design Needed
non_defect
mana icon branding goal an icon new branding for info this request has come out of the current work in progress redesign of the sso okta dashboard see below currently all mozilla sites are represented by the black highlight zilla treatment this is a placeholder until icons are available yulia brand is the originator of this request and should be kept apprised of progress requests for feedback and direction style information single color icon that can be used with or without a wordmark is preferable please default to yulia for further direction also the site owner team should be consulted for conceptual direction deadline initial beta release for the sso dashboard is slated for april if this can be ready by then for initial release implementation fantastic though understandable if that timeline is too aggressive and can not be met we will have regular releases feature enhancements and can integrate icon s updates as they are available tag design needed
0
783,074
27,517,535,558
IssuesEvent
2023-03-06 13:01:26
AY2223S2-CS2113T-T09-3/tp
https://api.github.com/repos/AY2223S2-CS2113T-T09-3/tp
opened
Find recipes that I can make with my current ingredients and tools
type.Story priority.High
... so that I can cook without having to go shopping each time
1.0
Find recipes that I can make with my current ingredients and tools - ... so that I can cook without having to go shopping each time
non_defect
find recipes that i can make with my current ingredients and tools so that i can cook without having to go shopping each time
0
15,240
19,178,875,190
IssuesEvent
2021-12-04 03:09:55
ooi-data/CE01ISSP-SP001-08-FLORTJ000-recovered_cspp-flort_sample
https://api.github.com/repos/ooi-data/CE01ISSP-SP001-08-FLORTJ000-recovered_cspp-flort_sample
opened
🛑 Processing failed: TypeError
process
## Overview `TypeError` found in `processing_task` task during run ended on 2021-12-04T03:09:54.779447. ## Details Flow name: `CE01ISSP-SP001-08-FLORTJ000-recovered_cspp-flort_sample` Task name: `processing_task` Error type: `TypeError` Error message: int() argument must be a string, a bytes-like object or a number, not 'NoneType' <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 157, in processing process_dataset( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 146, in process_dataset append_to_zarr(mod_ds, store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 342, in append_to_zarr mod_ds = _prepare_ds_to_append(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 142, in _prepare_ds_to_append new_arr.fill( TypeError: int() argument must be a string, a bytes-like object or a number, not 'NoneType' ``` </details>
1.0
🛑 Processing failed: TypeError - ## Overview `TypeError` found in `processing_task` task during run ended on 2021-12-04T03:09:54.779447. ## Details Flow name: `CE01ISSP-SP001-08-FLORTJ000-recovered_cspp-flort_sample` Task name: `processing_task` Error type: `TypeError` Error message: int() argument must be a string, a bytes-like object or a number, not 'NoneType' <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 157, in processing process_dataset( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 146, in process_dataset append_to_zarr(mod_ds, store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 342, in append_to_zarr mod_ds = _prepare_ds_to_append(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 142, in _prepare_ds_to_append new_arr.fill( TypeError: int() argument must be a string, a bytes-like object or a number, not 'NoneType' ``` </details>
non_defect
🛑 processing failed typeerror overview typeerror found in processing task task during run ended on details flow name recovered cspp flort sample task name processing task error type typeerror error message int argument must be a string a bytes like object or a number not nonetype traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing process dataset file srv conda envs notebook lib site packages ooi harvester processor init py line in process dataset append to zarr mod ds store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr mod ds prepare ds to append store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in prepare ds to append new arr fill typeerror int argument must be a string a bytes like object or a number not nonetype
0
22,013
3,588,076,367
IssuesEvent
2016-01-30 19:35:48
bigbluebutton/bigbluebutton
https://api.github.com/repos/bigbluebutton/bigbluebutton
closed
Deskshare fails after 6-7 times of start-stop (v0.8)
Accepted Defect DeskShare (Java) Normal Priority
Originally reported on Google Code with ID 1389 ``` Deskshare fails after 6-7 start-stop actions. From Red5 logs it seems like servlet is still running after deskshare window is closed - at first it gets CONSUMER_CONNECT_PUSH event (and all others) once, then twice, trice, etc. Logs attached ``` Reported by `upcFrost` on 2013-01-25 17:02:55 <hr> * *Attachment: [FirstRun.txt](https://storage.googleapis.com/google-code-attachments/bigbluebutton/issue-1389/comment-0/FirstRun.txt)* * *Attachment: [SecondRun.txt](https://storage.googleapis.com/google-code-attachments/bigbluebutton/issue-1389/comment-0/SecondRun.txt)* * *Attachment: [ThirdRun.txt](https://storage.googleapis.com/google-code-attachments/bigbluebutton/issue-1389/comment-0/ThirdRun.txt)*
1.0
Deskshare fails after 6-7 times of start-stop (v0.8) - Originally reported on Google Code with ID 1389 ``` Deskshare fails after 6-7 start-stop actions. From Red5 logs it seems like servlet is still running after deskshare window is closed - at first it gets CONSUMER_CONNECT_PUSH event (and all others) once, then twice, trice, etc. Logs attached ``` Reported by `upcFrost` on 2013-01-25 17:02:55 <hr> * *Attachment: [FirstRun.txt](https://storage.googleapis.com/google-code-attachments/bigbluebutton/issue-1389/comment-0/FirstRun.txt)* * *Attachment: [SecondRun.txt](https://storage.googleapis.com/google-code-attachments/bigbluebutton/issue-1389/comment-0/SecondRun.txt)* * *Attachment: [ThirdRun.txt](https://storage.googleapis.com/google-code-attachments/bigbluebutton/issue-1389/comment-0/ThirdRun.txt)*
defect
deskshare fails after times of start stop originally reported on google code with id deskshare fails after start stop actions from logs it seems like servlet is still running after deskshare window is closed at first it gets consumer connect push event and all others once then twice trice etc logs attached reported by upcfrost on attachment attachment attachment
1
155,623
19,802,913,223
IssuesEvent
2022-01-19 01:11:07
joshbnewton-prioritize-campaign/AdvanceAuto_DataFastLane
https://api.github.com/repos/joshbnewton-prioritize-campaign/AdvanceAuto_DataFastLane
opened
CVE-2022-23302 (High) detected in log4j-1.2.17.jar
security vulnerability
## CVE-2022-23302 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary> <p>Apache Log4j 1.2</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p> <p> Dependency Hierarchy: - spark-core_2.12-3.0.1.jar (Root Library) - :x: **log4j-1.2.17.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> JMSSink in all versions of Log4j 1.x is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration or if the configuration references an LDAP service the attacker has access to. The attacker can provide a TopicConnectionFactoryBindingName configuration causing JMSSink to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-4104. Note this issue only affects Log4j 1.x when specifically configured to use JMSSink, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions. <p>Publish Date: 2022-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23302>CVE-2022-23302</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.17","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.spark:spark-core_2.12:3.0.1;log4j:log4j:1.2.17","isMinimumFixVersionAvailable":false,"isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2022-23302","vulnerabilityDetails":"JMSSink in all versions of Log4j 1.x is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration or if the configuration references an LDAP service the attacker has access to. The attacker can provide a TopicConnectionFactoryBindingName configuration causing JMSSink to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-4104. Note this issue only affects Log4j 1.x when specifically configured to use JMSSink, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23302","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2022-23302 (High) detected in log4j-1.2.17.jar - ## CVE-2022-23302 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary> <p>Apache Log4j 1.2</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p> <p> Dependency Hierarchy: - spark-core_2.12-3.0.1.jar (Root Library) - :x: **log4j-1.2.17.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> JMSSink in all versions of Log4j 1.x is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration or if the configuration references an LDAP service the attacker has access to. The attacker can provide a TopicConnectionFactoryBindingName configuration causing JMSSink to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-4104. Note this issue only affects Log4j 1.x when specifically configured to use JMSSink, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions. <p>Publish Date: 2022-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23302>CVE-2022-23302</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.17","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.spark:spark-core_2.12:3.0.1;log4j:log4j:1.2.17","isMinimumFixVersionAvailable":false,"isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2022-23302","vulnerabilityDetails":"JMSSink in all versions of Log4j 1.x is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration or if the configuration references an LDAP service the attacker has access to. The attacker can provide a TopicConnectionFactoryBindingName configuration causing JMSSink to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-4104. Note this issue only affects Log4j 1.x when specifically configured to use JMSSink, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23302","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_defect
cve high detected in jar cve high severity vulnerability vulnerable library jar apache path to dependency file pom xml path to vulnerable library home wss scanner repository jar dependency hierarchy spark core jar root library x jar vulnerable library found in base branch master vulnerability details jmssink in all versions of x is vulnerable to deserialization of untrusted data when the attacker has write access to the configuration or if the configuration references an ldap service the attacker has access to the attacker can provide a topicconnectionfactorybindingname configuration causing jmssink to perform jndi requests that result in remote code execution in a similar fashion to cve note this issue only affects x when specifically configured to use jmssink which is not the default apache reached end of life in august users should upgrade to as it addresses numerous other issues from the previous versions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org apache spark spark core isminimumfixversionavailable false isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails jmssink in all versions of x is vulnerable to deserialization of untrusted data when the attacker has write access to the configuration or if the configuration references an ldap service the attacker has access to the attacker can provide a topicconnectionfactorybindingname configuration causing jmssink to perform jndi requests that result in remote code execution in a similar fashion to cve note this issue only affects x when specifically configured to use jmssink which is not the default apache reached end of life in august users should upgrade to as it addresses numerous other issues from the previous versions vulnerabilityurl
0
20,184
6,826,839,061
IssuesEvent
2017-11-08 15:20:16
moby/moby
https://api.github.com/repos/moby/moby
opened
HEALTHCHECK curl results in OOM killer
area/builder
This is not strictly a docker issue, but perhaps the DOCKERFILE documentation about HEALTHCHECK using curl should be amended. I was mystified by this for a couple of days. Turns out that at least on CENTOS, frequent calls to curl result in memory exhaustion due to increase in kernel's dentry cache, so eventually the container ends up being OOM killed. Solution is to define this env variable NSS_SDB_USE_CACHE=YES Issue is documented here https://tools.knetik.io/blog/2014-05-16-optimizing-aws-nss-softoken https://serverfault.com/questions/561350/unusually-high-dentry-cache-usage
1.0
HEALTHCHECK curl results in OOM killer - This is not strictly a docker issue, but perhaps the DOCKERFILE documentation about HEALTHCHECK using curl should be amended. I was mystified by this for a couple of days. Turns out that at least on CENTOS, frequent calls to curl result in memory exhaustion due to increase in kernel's dentry cache, so eventually the container ends up being OOM killed. Solution is to define this env variable NSS_SDB_USE_CACHE=YES Issue is documented here https://tools.knetik.io/blog/2014-05-16-optimizing-aws-nss-softoken https://serverfault.com/questions/561350/unusually-high-dentry-cache-usage
non_defect
healthcheck curl results in oom killer this is not strictly a docker issue but perhaps the dockerfile documentation about healthcheck using curl should be amended i was mystified by this for a couple of days turns out that at least on centos frequent calls to curl result in memory exhaustion due to increase in kernel s dentry cache so eventually the container ends up being oom killed solution is to define this env variable nss sdb use cache yes issue is documented here
0
34,639
7,458,264,101
IssuesEvent
2018-03-30 09:27:25
kerdokullamae/test_koik_issued
https://api.github.com/repos/kerdokullamae/test_koik_issued
opened
Märksõnade järjestuse muutmine ei tööta
C: AIS P: normal T: defect
**Reported by eilika on 15 Oct 2014 07:42 UTC** Märksõnade (ainese, koha, valdkonna) järjestuse muutmine ei tööta: - Avan ainese märksõna "Aadlikonvent" muutmiseks. - Valin dropdowni "Tõsta" väärtuseks "Enne Alevid" ja salvestan. - Märksõna kuvatakse nimekirjas endisel kohal.
1.0
Märksõnade järjestuse muutmine ei tööta - **Reported by eilika on 15 Oct 2014 07:42 UTC** Märksõnade (ainese, koha, valdkonna) järjestuse muutmine ei tööta: - Avan ainese märksõna "Aadlikonvent" muutmiseks. - Valin dropdowni "Tõsta" väärtuseks "Enne Alevid" ja salvestan. - Märksõna kuvatakse nimekirjas endisel kohal.
defect
märksõnade järjestuse muutmine ei tööta reported by eilika on oct utc märksõnade ainese koha valdkonna järjestuse muutmine ei tööta avan ainese märksõna aadlikonvent muutmiseks valin dropdowni tõsta väärtuseks enne alevid ja salvestan märksõna kuvatakse nimekirjas endisel kohal
1
44,502
12,217,257,666
IssuesEvent
2020-05-01 16:49:03
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
closed
[cms-ci] PR builds failing with `SQLSTATE[40001]: [error] Serialization failure: 1213 Deadlock found when trying to get lock; try restarting transaction: INSERT INTO {cache_config}`
Critical defect DevOps
Multiple PRs (#1433, #1438, and #1304) are now failing with: > SQLSTATE[40001]: [error] Serialization failure: 1213 Deadlock found when trying to get lock; try restarting transaction: INSERT INTO {cache_config} I attempted to resolve by setting: `transaction-isolation = READ-COMMITTED` in /etc/my.cnf per https://www.drupal.org/project/drupal/issues/2833539#comment-12500749, which we have applied to the upstream ENVs RDS param group a while ago, and never applied to CI. But they are still happening now. > mysql> SELECT @@global.tx_isolation, @@session.tx_isolation; > +-----------------------+------------------------+ > | @@global.tx_isolation | @@session.tx_isolation | > +-----------------------+------------------------+ > | READ-COMMITTED | READ-COMMITTED | > +-----------------------+------------------------+
1.0
[cms-ci] PR builds failing with `SQLSTATE[40001]: [error] Serialization failure: 1213 Deadlock found when trying to get lock; try restarting transaction: INSERT INTO {cache_config}` - Multiple PRs (#1433, #1438, and #1304) are now failing with: > SQLSTATE[40001]: [error] Serialization failure: 1213 Deadlock found when trying to get lock; try restarting transaction: INSERT INTO {cache_config} I attempted to resolve by setting: `transaction-isolation = READ-COMMITTED` in /etc/my.cnf per https://www.drupal.org/project/drupal/issues/2833539#comment-12500749, which we have applied to the upstream ENVs RDS param group a while ago, and never applied to CI. But they are still happening now. > mysql> SELECT @@global.tx_isolation, @@session.tx_isolation; > +-----------------------+------------------------+ > | @@global.tx_isolation | @@session.tx_isolation | > +-----------------------+------------------------+ > | READ-COMMITTED | READ-COMMITTED | > +-----------------------+------------------------+
defect
pr builds failing with sqlstate serialization failure deadlock found when trying to get lock try restarting transaction insert into cache config multiple prs and are now failing with sqlstate serialization failure deadlock found when trying to get lock try restarting transaction insert into cache config i attempted to resolve by setting transaction isolation read committed in etc my cnf per which we have applied to the upstream envs rds param group a while ago and never applied to ci but they are still happening now mysql select global tx isolation session tx isolation global tx isolation session tx isolation read committed read committed
1
294,244
9,014,766,763
IssuesEvent
2019-02-05 23:33:55
kubeflow/testing
https://api.github.com/repos/kubeflow/testing
closed
Cron job to garbage collect test resources
priority/p2
Our tests create a bunch of resources - Argo workflows - Namespaces - GCE VMs - GKE clusters Most of these resources (exception is Argo) should be GC'd by the teardown steps in our tests. But some failures/bugs prevent resources from being GC'd. It would be good to have a cron job to periodically garbage collect old resources like namespaces.
1.0
Cron job to garbage collect test resources - Our tests create a bunch of resources - Argo workflows - Namespaces - GCE VMs - GKE clusters Most of these resources (exception is Argo) should be GC'd by the teardown steps in our tests. But some failures/bugs prevent resources from being GC'd. It would be good to have a cron job to periodically garbage collect old resources like namespaces.
non_defect
cron job to garbage collect test resources our tests create a bunch of resources argo workflows namespaces gce vms gke clusters most of these resources exception is argo should be gc d by the teardown steps in our tests but some failures bugs prevent resources from being gc d it would be good to have a cron job to periodically garbage collect old resources like namespaces
0
109,914
4,415,356,792
IssuesEvent
2016-08-14 01:18:24
jahirfiquitiva/IconShowcase
https://api.github.com/repos/jahirfiquitiva/IconShowcase
closed
Fix request section lag
bug enhancement help wanted high priority
It only happens when getting to the end of the list. And it wasn't present until a couple updates, even when the section code hasn't changed significatively. Android Design library issue?
1.0
Fix request section lag - It only happens when getting to the end of the list. And it wasn't present until a couple updates, even when the section code hasn't changed significatively. Android Design library issue?
non_defect
fix request section lag it only happens when getting to the end of the list and it wasn t present until a couple updates even when the section code hasn t changed significatively android design library issue
0
316,499
23,634,988,382
IssuesEvent
2022-08-25 12:37:58
Ocelot-Social-Community/Ocelot-Social
https://api.github.com/repos/Ocelot-Social-Community/Ocelot-Social
closed
💥 [DevOps] Fix typo in PULL_REQUEST_TEMPLATE.md file
documentation good first issue devops
## 💥 DevOps Ticket Fix typo `Pullrequest` -> `Pull Request`.
1.0
💥 [DevOps] Fix typo in PULL_REQUEST_TEMPLATE.md file - ## 💥 DevOps Ticket Fix typo `Pullrequest` -> `Pull Request`.
non_defect
💥 fix typo in pull request template md file 💥 devops ticket fix typo pullrequest pull request
0
49,483
13,186,775,551
IssuesEvent
2020-08-13 01:16:48
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
[dataclasses] I3MCTree read/write machine dependence (Trac #1503)
Incomplete Migration Migrated from Trac combo core defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1503">https://code.icecube.wisc.edu/ticket/1503</a>, reported by olivas and owned by david.schultz</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-18T21:14:15", "description": "Writing a tree on a 32-bit Ubuntu machine and reading on a 64-bit Ubuntu machine fails with the following message :\n\nIn [5]: frame = f.pop_frame()\n\nIn [6]: print frame\n[ I3Frame (DAQ):\n 'I3EventHeader' [DAQ] ==> I3EventHeader (99)\n 'I3MCTree' [DAQ] ==> TreeBase::Tree<I3Particle, I3ParticleID, __gnu_cxx::hash<I3ParticleID> > (418)\n 'I3MCWeightDict' [DAQ] ==> I3Map<string, double> (1110)\n 'MCTimeIncEventID' [DAQ] ==> I3PODHolder<bool> (27)\n 'NuGPrimary' [DAQ] ==> I3Particle (150)\n]\n\n\nIn [7]: mctree = frame[\"I3MCTree\"]\nWARN (dataclasses): unknown RDMC code \"-14\" cannot be converted to a I3Particle::ParticleType. It will appear as \"unknown\". (I3Particle.cxx:864 in void I3Particle::load(Archive&, unsigned int) [with Archive = boost::archive::portable_binary_iarchive])\nWARN (dataclasses): unknown RDMC code \"65535\" cannot be converted to a I3Particle::ParticleType. It will appear as \"unknown\". (I3Particle.cxx:864 in void I3Particle::load(Archive&, unsigned int) [with Archive = boost::archive::portable_binary_iarchive])\nWARN (dataclasses): unknown RDMC code \"2001098\" cannot be converted to a I3Particle::ParticleType. It will appear as \"unknown\". (I3Particle.cxx:864 in void I3Particle::load(Archive&, unsigned int) [with Archive = boost::archive::portable_binary_iarchive])\npython: /usr/include/boost/dynamic_bitset/dynamic_bitset.hpp:616: boost::dynamic_bitset<Block, Allocator>::~dynamic_bitset() [with Block = long unsigned int; Allocator = std::allocator<long unsigned int>]: Assertion `m_check_invariants()' failed.\n[1] 7057 abort (core dumped) ipython\n\n\nOther frame objects deserialize just fine.\n\nBoth machines are 14.04 LTS built against system packages. My guess is this is a 32/64 bit issue.", "reporter": "olivas", "cc": "", "resolution": "fixed", "_ts": "1458335655846260", "component": "combo core", "summary": "[dataclasses] I3MCTree read/write machine dependence", "priority": "major", "keywords": "", "time": "2016-01-11T03:06:13", "milestone": "", "owner": "david.schultz", "type": "defect" } ``` </p> </details>
1.0
[dataclasses] I3MCTree read/write machine dependence (Trac #1503) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1503">https://code.icecube.wisc.edu/ticket/1503</a>, reported by olivas and owned by david.schultz</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-18T21:14:15", "description": "Writing a tree on a 32-bit Ubuntu machine and reading on a 64-bit Ubuntu machine fails with the following message :\n\nIn [5]: frame = f.pop_frame()\n\nIn [6]: print frame\n[ I3Frame (DAQ):\n 'I3EventHeader' [DAQ] ==> I3EventHeader (99)\n 'I3MCTree' [DAQ] ==> TreeBase::Tree<I3Particle, I3ParticleID, __gnu_cxx::hash<I3ParticleID> > (418)\n 'I3MCWeightDict' [DAQ] ==> I3Map<string, double> (1110)\n 'MCTimeIncEventID' [DAQ] ==> I3PODHolder<bool> (27)\n 'NuGPrimary' [DAQ] ==> I3Particle (150)\n]\n\n\nIn [7]: mctree = frame[\"I3MCTree\"]\nWARN (dataclasses): unknown RDMC code \"-14\" cannot be converted to a I3Particle::ParticleType. It will appear as \"unknown\". (I3Particle.cxx:864 in void I3Particle::load(Archive&, unsigned int) [with Archive = boost::archive::portable_binary_iarchive])\nWARN (dataclasses): unknown RDMC code \"65535\" cannot be converted to a I3Particle::ParticleType. It will appear as \"unknown\". (I3Particle.cxx:864 in void I3Particle::load(Archive&, unsigned int) [with Archive = boost::archive::portable_binary_iarchive])\nWARN (dataclasses): unknown RDMC code \"2001098\" cannot be converted to a I3Particle::ParticleType. It will appear as \"unknown\". (I3Particle.cxx:864 in void I3Particle::load(Archive&, unsigned int) [with Archive = boost::archive::portable_binary_iarchive])\npython: /usr/include/boost/dynamic_bitset/dynamic_bitset.hpp:616: boost::dynamic_bitset<Block, Allocator>::~dynamic_bitset() [with Block = long unsigned int; Allocator = std::allocator<long unsigned int>]: Assertion `m_check_invariants()' failed.\n[1] 7057 abort (core dumped) ipython\n\n\nOther frame objects deserialize just fine.\n\nBoth machines are 14.04 LTS built against system packages. My guess is this is a 32/64 bit issue.", "reporter": "olivas", "cc": "", "resolution": "fixed", "_ts": "1458335655846260", "component": "combo core", "summary": "[dataclasses] I3MCTree read/write machine dependence", "priority": "major", "keywords": "", "time": "2016-01-11T03:06:13", "milestone": "", "owner": "david.schultz", "type": "defect" } ``` </p> </details>
defect
read write machine dependence trac migrated from json status closed changetime description writing a tree on a bit ubuntu machine and reading on a bit ubuntu machine fails with the following message n nin frame f pop frame n nin print frame n n treebase tree n n mctimeinceventid n nugprimary n n n nin mctree frame nwarn dataclasses unknown rdmc code cannot be converted to a particletype it will appear as unknown cxx in void load archive unsigned int nwarn dataclasses unknown rdmc code cannot be converted to a particletype it will appear as unknown cxx in void load archive unsigned int nwarn dataclasses unknown rdmc code cannot be converted to a particletype it will appear as unknown cxx in void load archive unsigned int npython usr include boost dynamic bitset dynamic bitset hpp boost dynamic bitset dynamic bitset assertion m check invariants failed n abort core dumped ipython n n nother frame objects deserialize just fine n nboth machines are lts built against system packages my guess is this is a bit issue reporter olivas cc resolution fixed ts component combo core summary read write machine dependence priority major keywords time milestone owner david schultz type defect
1
40,898
10,591,146,410
IssuesEvent
2019-10-09 10:13:30
gradle/gradle
https://api.github.com/repos/gradle/gradle
opened
Allow CopySpec to specify a different normalization
@build-cache from:member
While copying files (using `Copy` or `Zip` etc.) we currently treat the sources specified via `CopySpec`s as `PathSensitivity.RELATIVE`. However, sometimes we want to copy things where we know they are going to be used as a classpath, like when including JARs in a WAR file. In those cases we could specify `@Classpath` normalization and avoid recreating the WAR file every time one of the JARs is rebuilt without a (runtime) significant change. --- cc: @gradle/build-cache
1.0
Allow CopySpec to specify a different normalization - While copying files (using `Copy` or `Zip` etc.) we currently treat the sources specified via `CopySpec`s as `PathSensitivity.RELATIVE`. However, sometimes we want to copy things where we know they are going to be used as a classpath, like when including JARs in a WAR file. In those cases we could specify `@Classpath` normalization and avoid recreating the WAR file every time one of the JARs is rebuilt without a (runtime) significant change. --- cc: @gradle/build-cache
non_defect
allow copyspec to specify a different normalization while copying files using copy or zip etc we currently treat the sources specified via copyspec s as pathsensitivity relative however sometimes we want to copy things where we know they are going to be used as a classpath like when including jars in a war file in those cases we could specify classpath normalization and avoid recreating the war file every time one of the jars is rebuilt without a runtime significant change cc gradle build cache
0
282,809
24,498,050,768
IssuesEvent
2022-10-10 10:22:48
apache/shardingsphere
https://api.github.com/repos/apache/shardingsphere
opened
[DistSQL] Add SQL parser test case for `ClearHintStatement`
in: test feature: DistSQL hacktoberfest
### Background `CLEAR HINT` is a syntax in [DistSQL RAL](https://shardingsphere.apache.org/document/current/en/user-manual/shardingsphere-proxy/distsql/syntax/ral/), when `ShardingSphereSQLParserEngine` receives this SQL text, it parses it as `ClearHintStatement`. Now we need to add a test case for this parsing process to assert that the parsing is correct. ### Aim Add SQL parser test case for `ClearHintStatement`. - Affected files: - SQLParserTestCases (**modify**) - test/parser/src/main/resources/case/ral/updatable.xml (**modify**) - test/parser/src/main/resources/sql/supported/ral/updatable.xml (**modify**) - test/parser/src/main/java/org/apache/shardingsphere/test/sql/parser/parameterized/jaxb/cases/domain/statement/distsql/ral/ClearHintStatementTestCase.java (**create**) - test/parser/src/main/java/org/apache/shardingsphere/test/sql/parser/parameterized/asserts/statement/distsql/ral/impl/updatable/ClearHintStatementAssert.java (**create**) - UpdatableRALStatementAssert (**modify**) - After all, execute `DistSQLParserParameterizedTest#assertDistSQL()` for validation ### Basic Qualifications - Java - Maven ### Example FYI - ShowAllVariablesStatement - ShowAllVariablesStatementTestCase - ShowAllVariablesStatementAssert - SQLParserTestCases - search for "show-all-variables"
1.0
[DistSQL] Add SQL parser test case for `ClearHintStatement` - ### Background `CLEAR HINT` is a syntax in [DistSQL RAL](https://shardingsphere.apache.org/document/current/en/user-manual/shardingsphere-proxy/distsql/syntax/ral/), when `ShardingSphereSQLParserEngine` receives this SQL text, it parses it as `ClearHintStatement`. Now we need to add a test case for this parsing process to assert that the parsing is correct. ### Aim Add SQL parser test case for `ClearHintStatement`. - Affected files: - SQLParserTestCases (**modify**) - test/parser/src/main/resources/case/ral/updatable.xml (**modify**) - test/parser/src/main/resources/sql/supported/ral/updatable.xml (**modify**) - test/parser/src/main/java/org/apache/shardingsphere/test/sql/parser/parameterized/jaxb/cases/domain/statement/distsql/ral/ClearHintStatementTestCase.java (**create**) - test/parser/src/main/java/org/apache/shardingsphere/test/sql/parser/parameterized/asserts/statement/distsql/ral/impl/updatable/ClearHintStatementAssert.java (**create**) - UpdatableRALStatementAssert (**modify**) - After all, execute `DistSQLParserParameterizedTest#assertDistSQL()` for validation ### Basic Qualifications - Java - Maven ### Example FYI - ShowAllVariablesStatement - ShowAllVariablesStatementTestCase - ShowAllVariablesStatementAssert - SQLParserTestCases - search for "show-all-variables"
non_defect
add sql parser test case for clearhintstatement background clear hint is a syntax in when shardingspheresqlparserengine receives this sql text it parses it as clearhintstatement now we need to add a test case for this parsing process to assert that the parsing is correct aim add sql parser test case for clearhintstatement affected files sqlparsertestcases modify test parser src main resources case ral updatable xml modify test parser src main resources sql supported ral updatable xml modify test parser src main java org apache shardingsphere test sql parser parameterized jaxb cases domain statement distsql ral clearhintstatementtestcase java create test parser src main java org apache shardingsphere test sql parser parameterized asserts statement distsql ral impl updatable clearhintstatementassert java create updatableralstatementassert modify after all execute distsqlparserparameterizedtest assertdistsql for validation basic qualifications java maven example fyi showallvariablesstatement showallvariablesstatementtestcase showallvariablesstatementassert sqlparsertestcases search for show all variables
0
289,504
21,785,493,445
IssuesEvent
2022-05-14 04:06:17
Max-Rodriguez/Instagram-Offline
https://api.github.com/repos/Max-Rodriguez/Instagram-Offline
opened
Write entire project documentation, updated to latest code on master branch
documentation
Posted as a note to self for later :)
1.0
Write entire project documentation, updated to latest code on master branch - Posted as a note to self for later :)
non_defect
write entire project documentation updated to latest code on master branch posted as a note to self for later
0
108,333
16,766,576,422
IssuesEvent
2021-06-14 09:33:36
SmartBear/swaggerhub-cli
https://api.github.com/repos/SmartBear/swaggerhub-cli
closed
CVE-2020-28469 (Medium) detected in glob-parent-5.1.1.tgz - autoclosed
security vulnerability
## CVE-2020-28469 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-5.1.1.tgz</b></p></summary> <p>Extract the non-magic parent path from a glob string.</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz</a></p> <p>Path to dependency file: swaggerhub-cli/package.json</p> <p>Path to vulnerable library: swaggerhub-cli/node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - eslint-7.22.0.tgz (Root Library) - :x: **glob-parent-5.1.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator. <p>Publish Date: 2021-06-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469>CVE-2020-28469</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469</a></p> <p>Release Date: 2021-06-03</p> <p>Fix Resolution: glob-parent - 5.1.2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"glob-parent","packageVersion":"5.1.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"eslint:7.22.0;glob-parent:5.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"glob-parent - 5.1.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28469","vulnerabilityDetails":"This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-28469 (Medium) detected in glob-parent-5.1.1.tgz - autoclosed - ## CVE-2020-28469 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-5.1.1.tgz</b></p></summary> <p>Extract the non-magic parent path from a glob string.</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz</a></p> <p>Path to dependency file: swaggerhub-cli/package.json</p> <p>Path to vulnerable library: swaggerhub-cli/node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - eslint-7.22.0.tgz (Root Library) - :x: **glob-parent-5.1.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator. <p>Publish Date: 2021-06-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469>CVE-2020-28469</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469</a></p> <p>Release Date: 2021-06-03</p> <p>Fix Resolution: glob-parent - 5.1.2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"glob-parent","packageVersion":"5.1.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"eslint:7.22.0;glob-parent:5.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"glob-parent - 5.1.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28469","vulnerabilityDetails":"This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_defect
cve medium detected in glob parent tgz autoclosed cve medium severity vulnerability vulnerable library glob parent tgz extract the non magic parent path from a glob string library home page a href path to dependency file swaggerhub cli package json path to vulnerable library swaggerhub cli node modules glob parent package json dependency hierarchy eslint tgz root library x glob parent tgz vulnerable library found in base branch master vulnerability details this affects the package glob parent before the enclosure regex used to check for strings ending in enclosure containing path separator publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree eslint glob parent isminimumfixversionavailable true minimumfixversion glob parent basebranches vulnerabilityidentifier cve vulnerabilitydetails this affects the package glob parent before the enclosure regex used to check for strings ending in enclosure containing path separator vulnerabilityurl
0
45,541
12,838,691,559
IssuesEvent
2020-07-07 17:55:43
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
SQL Server 2017 CU21 installation error
T: Defect
### Expected behavior and actual behavior: SQL Server 2017 CU 21 apply patch from sql server 2017 CU20 fails with error: The operation failed because an index or statistics with name 'IX_EventMessageContext_EventMessageId' already exists on table 'internal.event_message_context' ### Steps to reproduce the problem (if possible, create an MCVE: https://github.com/jOOQ/jOOQ-mcve): Apply patch SQLServer2017-KB4557397-x64.exe over sql server 2017 Developer CU20 ### Versions: Microsoft SQL Server 2017 (RTM-CU20) (KB4541283) - 14.0.3294.2 (X64) Mar 13 2020 14:53:45 Copyright (C) 2017 Microsoft Corporation Developer Edition (64-bit) on Windows Server 2019 Standard 10.0 <X64> (Build 17763: ) (Hypervisor) - jOOQ: - Java: - Database (include vendor): SQL Server 2017 CU20 - OS: Windows Server 2019 Standard 10.0 - JDBC Driver (include name if inofficial driver):
1.0
SQL Server 2017 CU21 installation error - ### Expected behavior and actual behavior: SQL Server 2017 CU 21 apply patch from sql server 2017 CU20 fails with error: The operation failed because an index or statistics with name 'IX_EventMessageContext_EventMessageId' already exists on table 'internal.event_message_context' ### Steps to reproduce the problem (if possible, create an MCVE: https://github.com/jOOQ/jOOQ-mcve): Apply patch SQLServer2017-KB4557397-x64.exe over sql server 2017 Developer CU20 ### Versions: Microsoft SQL Server 2017 (RTM-CU20) (KB4541283) - 14.0.3294.2 (X64) Mar 13 2020 14:53:45 Copyright (C) 2017 Microsoft Corporation Developer Edition (64-bit) on Windows Server 2019 Standard 10.0 <X64> (Build 17763: ) (Hypervisor) - jOOQ: - Java: - Database (include vendor): SQL Server 2017 CU20 - OS: Windows Server 2019 Standard 10.0 - JDBC Driver (include name if inofficial driver):
defect
sql server installation error expected behavior and actual behavior sql server cu apply patch from sql server fails with error the operation failed because an index or statistics with name ix eventmessagecontext eventmessageid already exists on table internal event message context steps to reproduce the problem if possible create an mcve apply patch exe over sql server developer versions microsoft sql server rtm mar copyright c microsoft corporation developer edition bit on windows server standard build hypervisor jooq java database include vendor sql server os windows server standard jdbc driver include name if inofficial driver
1
184,015
31,806,204,508
IssuesEvent
2023-09-13 14:05:56
vr4vet/vr4vet
https://api.github.com/repos/vr4vet/vr4vet
closed
Design and concept for the Controller help
ux-ui 2D design
We need define how the controllers will be introduced and explain to the user, including the buttons and functions they have. The design might include, showing a digital twin versions of the correct controller, animations or pointers, voice over introduction, and links to the functions of the buttons (asking the user to teleport when explaining the teleport button).
1.0
Design and concept for the Controller help - We need define how the controllers will be introduced and explain to the user, including the buttons and functions they have. The design might include, showing a digital twin versions of the correct controller, animations or pointers, voice over introduction, and links to the functions of the buttons (asking the user to teleport when explaining the teleport button).
non_defect
design and concept for the controller help we need define how the controllers will be introduced and explain to the user including the buttons and functions they have the design might include showing a digital twin versions of the correct controller animations or pointers voice over introduction and links to the functions of the buttons asking the user to teleport when explaining the teleport button
0
33,875
2,773,157,590
IssuesEvent
2015-05-03 11:20:00
raspibo/eventman
https://api.github.com/repos/raspibo/eventman
closed
imported persons are not associated to event
bug priority: low
In our showcase environment imported persons are not associated to events.
1.0
imported persons are not associated to event - In our showcase environment imported persons are not associated to events.
non_defect
imported persons are not associated to event in our showcase environment imported persons are not associated to events
0
16,826
2,948,318,332
IssuesEvent
2015-07-06 01:26:28
Winetricks/winetricks
https://api.github.com/repos/Winetricks/winetricks
closed
Set up WINEPREFIX variable for "Browse files" option
auto-migrated Priority-Medium Type-Defect
``` I can't run any .exe program through "Browse files" because WINEPREFIX is not set. Steps to reproduce: 1. Run winetricks 2. Select custom prefix 3. Choose "Browse files" 4. Open terminal in your FM (e.g. F4 in Dolphin) 5. type echo $WINEPREFIX ``` Original issue reported on code.google.com by `PZinin` on 2 Sep 2012 at 6:56
1.0
Set up WINEPREFIX variable for "Browse files" option - ``` I can't run any .exe program through "Browse files" because WINEPREFIX is not set. Steps to reproduce: 1. Run winetricks 2. Select custom prefix 3. Choose "Browse files" 4. Open terminal in your FM (e.g. F4 in Dolphin) 5. type echo $WINEPREFIX ``` Original issue reported on code.google.com by `PZinin` on 2 Sep 2012 at 6:56
defect
set up wineprefix variable for browse files option i can t run any exe program through browse files because wineprefix is not set steps to reproduce run winetricks select custom prefix choose browse files open terminal in your fm e g in dolphin type echo wineprefix original issue reported on code google com by pzinin on sep at
1
45,601
18,765,395,470
IssuesEvent
2021-11-05 22:48:27
gradido/gradido
https://api.github.com/repos/gradido/gradido
closed
🔧 [Refactor] Implement Login Call createTransactions on Apollo
service: login server refactor service: backend
## 🔧 Refactor ticket <!-- Describe your issue in detail. Include screenshots if needed. Give us as much information as possible. Use a clear and concise description of what the problem is.--> Implement Login Call createTransactions on Apollo
2.0
🔧 [Refactor] Implement Login Call createTransactions on Apollo - ## 🔧 Refactor ticket <!-- Describe your issue in detail. Include screenshots if needed. Give us as much information as possible. Use a clear and concise description of what the problem is.--> Implement Login Call createTransactions on Apollo
non_defect
🔧 implement login call createtransactions on apollo 🔧 refactor ticket implement login call createtransactions on apollo
0
38,442
8,819,418,404
IssuesEvent
2018-12-31 20:00:23
hazelcast/hazelcast-nodejs-client
https://api.github.com/repos/hazelcast/hazelcast-nodejs-client
closed
map.get with non-existent key increases near cache entry count
Type: Defect
This bug should be fixed and a test should be added.
1.0
map.get with non-existent key increases near cache entry count - This bug should be fixed and a test should be added.
defect
map get with non existent key increases near cache entry count this bug should be fixed and a test should be added
1
187,345
14,427,551,346
IssuesEvent
2020-12-06 04:45:40
kalexmills/github-vet-tests-dec2020
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
closed
pbolla0818/oci_terraform: oci/limits_quota_test.go; 14 LoC
fresh small test
Found a possible issue in [pbolla0818/oci_terraform](https://www.github.com/pbolla0818/oci_terraform) at [oci/limits_quota_test.go](https://github.com/pbolla0818/oci_terraform/blob/c233d54c5fe32f12c234d6dceefba0a9b4ab3022/oci/limits_quota_test.go#L258-L271) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > reference to quotaId is reassigned at line 262 [Click here to see the code in its original context.](https://github.com/pbolla0818/oci_terraform/blob/c233d54c5fe32f12c234d6dceefba0a9b4ab3022/oci/limits_quota_test.go#L258-L271) <details> <summary>Click here to show the 14 line(s) of Go which triggered the analyzer.</summary> ```go for _, quotaId := range quotaIds { if ok := SweeperDefaultResourceId[quotaId]; !ok { deleteQuotaRequest := oci_limits.DeleteQuotaRequest{} deleteQuotaRequest.QuotaId = &quotaId deleteQuotaRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "limits") _, error := quotasClient.DeleteQuota(context.Background(), deleteQuotaRequest) if error != nil { fmt.Printf("Error deleting Quota %s %s, It is possible that the resource is already deleted. Please verify manually \n", quotaId, error) continue } } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: c233d54c5fe32f12c234d6dceefba0a9b4ab3022
1.0
pbolla0818/oci_terraform: oci/limits_quota_test.go; 14 LoC - Found a possible issue in [pbolla0818/oci_terraform](https://www.github.com/pbolla0818/oci_terraform) at [oci/limits_quota_test.go](https://github.com/pbolla0818/oci_terraform/blob/c233d54c5fe32f12c234d6dceefba0a9b4ab3022/oci/limits_quota_test.go#L258-L271) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > reference to quotaId is reassigned at line 262 [Click here to see the code in its original context.](https://github.com/pbolla0818/oci_terraform/blob/c233d54c5fe32f12c234d6dceefba0a9b4ab3022/oci/limits_quota_test.go#L258-L271) <details> <summary>Click here to show the 14 line(s) of Go which triggered the analyzer.</summary> ```go for _, quotaId := range quotaIds { if ok := SweeperDefaultResourceId[quotaId]; !ok { deleteQuotaRequest := oci_limits.DeleteQuotaRequest{} deleteQuotaRequest.QuotaId = &quotaId deleteQuotaRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "limits") _, error := quotasClient.DeleteQuota(context.Background(), deleteQuotaRequest) if error != nil { fmt.Printf("Error deleting Quota %s %s, It is possible that the resource is already deleted. Please verify manually \n", quotaId, error) continue } } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: c233d54c5fe32f12c234d6dceefba0a9b4ab3022
non_defect
oci terraform oci limits quota test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message reference to quotaid is reassigned at line click here to show the line s of go which triggered the analyzer go for quotaid range quotaids if ok sweeperdefaultresourceid ok deletequotarequest oci limits deletequotarequest deletequotarequest quotaid quotaid deletequotarequest requestmetadata retrypolicy getretrypolicy true limits error quotasclient deletequota context background deletequotarequest if error nil fmt printf error deleting quota s s it is possible that the resource is already deleted please verify manually n quotaid error continue leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
504,764
14,620,955,125
IssuesEvent
2020-12-22 20:41:41
googleapis/elixir-google-api
https://api.github.com/repos/googleapis/elixir-google-api
closed
Synthesis failed for Translate
api: translate autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate Translate. :broken_heart: Here's the output from running `synth.py`: ``` poison Compiling 4 files (.ex) warning: Integer.to_char_list/2 is deprecated. Use Integer.to_charlist/2 instead lib/poison/encoder.ex:173 Generated poison app ==> ssl_verify_fun Compiling 7 files (.erl) Generated ssl_verify_fun app ===> Compiling certifi ===> Compiling hackney ==> oauth2 Compiling 13 files (.ex) Generated oauth2 app ==> mime Compiling 2 files (.ex) Generated mime app ==> tesla Compiling 26 files (.ex) Generated tesla app ==> google_gax Compiling 5 files (.ex) Generated google_gax app ==> google_api_discovery Compiling 21 files (.ex) Generated google_api_discovery app ==> google_apis Compiling 28 files (.ex) Generated google_apis app 15:02:53.354 [info] FETCHING: https://translation.googleapis.com/$discovery/GOOGLE_REST_SIMPLE_URI?version=v2 15:02:53.472 [info] FETCHING: https://translation.googleapis.com/$discovery/rest?version=v2 15:02:53.476 [info] FOUND: https://translation.googleapis.com/$discovery/rest?version=v2 Revision check: old=20170525, new=20170525, generating=true Creating leading directories Writing DetectLanguageRequest to clients/translate/lib/google_api/translate/v2/model/detect_language_request.ex. Writing DetectionsListResponse to clients/translate/lib/google_api/translate/v2/model/detections_list_response.ex. Writing DetectionsResource to clients/translate/lib/google_api/translate/v2/model/detections_resource.ex. Writing GetSupportedLanguagesRequest to clients/translate/lib/google_api/translate/v2/model/get_supported_languages_request.ex. Writing LanguagesListResponse to clients/translate/lib/google_api/translate/v2/model/languages_list_response.ex. Writing LanguagesResource to clients/translate/lib/google_api/translate/v2/model/languages_resource.ex. Writing TranslateTextRequest to clients/translate/lib/google_api/translate/v2/model/translate_text_request.ex. Writing TranslationsListResponse to clients/translate/lib/google_api/translate/v2/model/translations_list_response.ex. Writing TranslationsResource to clients/translate/lib/google_api/translate/v2/model/translations_resource.ex. Writing Detections to clients/translate/lib/google_api/translate/v2/api/detections.ex. Writing Languages to clients/translate/lib/google_api/translate/v2/api/languages.ex. Writing Translations to clients/translate/lib/google_api/translate/v2/api/translations.ex. Writing connection.ex. Writing metadata.ex. Writing mix.exs Writing README.md Writing LICENSE Writing .gitignore Writing config/config.exs Writing test/test_helper.exs 15:02:53.556 [info] FETCHING: https://translation.googleapis.com/$discovery/GOOGLE_REST_SIMPLE_URI?version=v3 15:02:53.562 [info] FETCHING: https://translation.googleapis.com/$discovery/rest?version=v3 15:02:53.574 [info] FOUND: https://translation.googleapis.com/$discovery/rest?version=v3 Revision check: old=20201204, new=20201211, generating=true Creating leading directories Writing BatchTranslateTextRequest to clients/translate/lib/google_api/translate/v3/model/batch_translate_text_request.ex. Writing CancelOperationRequest to clients/translate/lib/google_api/translate/v3/model/cancel_operation_request.ex. Writing DetectLanguageRequest to clients/translate/lib/google_api/translate/v3/model/detect_language_request.ex. Writing DetectLanguageResponse to clients/translate/lib/google_api/translate/v3/model/detect_language_response.ex. Writing DetectedLanguage to clients/translate/lib/google_api/translate/v3/model/detected_language.ex. Writing Empty to clients/translate/lib/google_api/translate/v3/model/empty.ex. Writing GcsDestination to clients/translate/lib/google_api/translate/v3/model/gcs_destination.ex. Writing GcsSource to clients/translate/lib/google_api/translate/v3/model/gcs_source.ex. Writing Glossary to clients/translate/lib/google_api/translate/v3/model/glossary.ex. Writing GlossaryInputConfig to clients/translate/lib/google_api/translate/v3/model/glossary_input_config.ex. Writing InputConfig to clients/translate/lib/google_api/translate/v3/model/input_config.ex. Writing LanguageCodePair to clients/translate/lib/google_api/translate/v3/model/language_code_pair.ex. Writing LanguageCodesSet to clients/translate/lib/google_api/translate/v3/model/language_codes_set.ex. Writing ListGlossariesResponse to clients/translate/lib/google_api/translate/v3/model/list_glossaries_response.ex. Writing ListLocationsResponse to clients/translate/lib/google_api/translate/v3/model/list_locations_response.ex. Writing ListOperationsResponse to clients/translate/lib/google_api/translate/v3/model/list_operations_response.ex. Writing Location to clients/translate/lib/google_api/translate/v3/model/location.ex. Writing Operation to clients/translate/lib/google_api/translate/v3/model/operation.ex. Writing OutputConfig to clients/translate/lib/google_api/translate/v3/model/output_config.ex. Writing Status to clients/translate/lib/google_api/translate/v3/model/status.ex. Writing SupportedLanguage to clients/translate/lib/google_api/translate/v3/model/supported_language.ex. Writing SupportedLanguages to clients/translate/lib/google_api/translate/v3/model/supported_languages.ex. Writing TranslateTextGlossaryConfig to clients/translate/lib/google_api/translate/v3/model/translate_text_glossary_config.ex. Writing TranslateTextRequest to clients/translate/lib/google_api/translate/v3/model/translate_text_request.ex. Writing TranslateTextResponse to clients/translate/lib/google_api/translate/v3/model/translate_text_response.ex. Writing Translation to clients/translate/lib/google_api/translate/v3/model/translation.ex. Writing WaitOperationRequest to clients/translate/lib/google_api/translate/v3/model/wait_operation_request.ex. Writing Projects to clients/translate/lib/google_api/translate/v3/api/projects.ex. Writing connection.ex. Writing metadata.ex. Writing mix.exs Writing README.md Writing LICENSE Writing .gitignore Writing config/config.exs Writing test/test_helper.exs 15:02:54.113 [info] Found only discovery_revision and/or formatting changes. Not significant enough for a PR. fixing file permissions Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 252, in __exit__ self.observer.stop() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/utils/__init__.py", line 81, in stop self.on_thread_stop() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/api.py", line 361, in on_thread_stop self.unschedule_all() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/api.py", line 357, in unschedule_all self._clear_emitters() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/api.py", line 231, in _clear_emitters emitter.stop() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/utils/__init__.py", line 81, in stop self.on_thread_stop() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/inotify.py", line 121, in on_thread_stop self._inotify.close() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/inotify_buffer.py", line 50, in close self.stop() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/utils/__init__.py", line 81, in stop self.on_thread_stop() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/inotify_buffer.py", line 46, in on_thread_stop self._inotify.close() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/inotify_c.py", line 277, in close os.close(self._inotify_fd) OSError: [Errno 9] Bad file descriptor 2020-12-18 07:02:57,189 autosynth [ERROR] > Synthesis failed 2020-12-18 07:02:57,189 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 354, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 189, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 291, in _inner_main ).synthesize(synth_log_path / "sponge_log.log") File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/translate/synth.metadata', 'synth.py', '--', 'Translate']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/3a455424-7540-46a9-bb77-8265c8f04c06/targets/github%2Fsynthtool;config=default/tests;query=elixir-google-api;failed=false).
1.0
Synthesis failed for Translate - Hello! Autosynth couldn't regenerate Translate. :broken_heart: Here's the output from running `synth.py`: ``` poison Compiling 4 files (.ex) warning: Integer.to_char_list/2 is deprecated. Use Integer.to_charlist/2 instead lib/poison/encoder.ex:173 Generated poison app ==> ssl_verify_fun Compiling 7 files (.erl) Generated ssl_verify_fun app ===> Compiling certifi ===> Compiling hackney ==> oauth2 Compiling 13 files (.ex) Generated oauth2 app ==> mime Compiling 2 files (.ex) Generated mime app ==> tesla Compiling 26 files (.ex) Generated tesla app ==> google_gax Compiling 5 files (.ex) Generated google_gax app ==> google_api_discovery Compiling 21 files (.ex) Generated google_api_discovery app ==> google_apis Compiling 28 files (.ex) Generated google_apis app 15:02:53.354 [info] FETCHING: https://translation.googleapis.com/$discovery/GOOGLE_REST_SIMPLE_URI?version=v2 15:02:53.472 [info] FETCHING: https://translation.googleapis.com/$discovery/rest?version=v2 15:02:53.476 [info] FOUND: https://translation.googleapis.com/$discovery/rest?version=v2 Revision check: old=20170525, new=20170525, generating=true Creating leading directories Writing DetectLanguageRequest to clients/translate/lib/google_api/translate/v2/model/detect_language_request.ex. Writing DetectionsListResponse to clients/translate/lib/google_api/translate/v2/model/detections_list_response.ex. Writing DetectionsResource to clients/translate/lib/google_api/translate/v2/model/detections_resource.ex. Writing GetSupportedLanguagesRequest to clients/translate/lib/google_api/translate/v2/model/get_supported_languages_request.ex. Writing LanguagesListResponse to clients/translate/lib/google_api/translate/v2/model/languages_list_response.ex. Writing LanguagesResource to clients/translate/lib/google_api/translate/v2/model/languages_resource.ex. Writing TranslateTextRequest to clients/translate/lib/google_api/translate/v2/model/translate_text_request.ex. Writing TranslationsListResponse to clients/translate/lib/google_api/translate/v2/model/translations_list_response.ex. Writing TranslationsResource to clients/translate/lib/google_api/translate/v2/model/translations_resource.ex. Writing Detections to clients/translate/lib/google_api/translate/v2/api/detections.ex. Writing Languages to clients/translate/lib/google_api/translate/v2/api/languages.ex. Writing Translations to clients/translate/lib/google_api/translate/v2/api/translations.ex. Writing connection.ex. Writing metadata.ex. Writing mix.exs Writing README.md Writing LICENSE Writing .gitignore Writing config/config.exs Writing test/test_helper.exs 15:02:53.556 [info] FETCHING: https://translation.googleapis.com/$discovery/GOOGLE_REST_SIMPLE_URI?version=v3 15:02:53.562 [info] FETCHING: https://translation.googleapis.com/$discovery/rest?version=v3 15:02:53.574 [info] FOUND: https://translation.googleapis.com/$discovery/rest?version=v3 Revision check: old=20201204, new=20201211, generating=true Creating leading directories Writing BatchTranslateTextRequest to clients/translate/lib/google_api/translate/v3/model/batch_translate_text_request.ex. Writing CancelOperationRequest to clients/translate/lib/google_api/translate/v3/model/cancel_operation_request.ex. Writing DetectLanguageRequest to clients/translate/lib/google_api/translate/v3/model/detect_language_request.ex. Writing DetectLanguageResponse to clients/translate/lib/google_api/translate/v3/model/detect_language_response.ex. Writing DetectedLanguage to clients/translate/lib/google_api/translate/v3/model/detected_language.ex. Writing Empty to clients/translate/lib/google_api/translate/v3/model/empty.ex. Writing GcsDestination to clients/translate/lib/google_api/translate/v3/model/gcs_destination.ex. Writing GcsSource to clients/translate/lib/google_api/translate/v3/model/gcs_source.ex. Writing Glossary to clients/translate/lib/google_api/translate/v3/model/glossary.ex. Writing GlossaryInputConfig to clients/translate/lib/google_api/translate/v3/model/glossary_input_config.ex. Writing InputConfig to clients/translate/lib/google_api/translate/v3/model/input_config.ex. Writing LanguageCodePair to clients/translate/lib/google_api/translate/v3/model/language_code_pair.ex. Writing LanguageCodesSet to clients/translate/lib/google_api/translate/v3/model/language_codes_set.ex. Writing ListGlossariesResponse to clients/translate/lib/google_api/translate/v3/model/list_glossaries_response.ex. Writing ListLocationsResponse to clients/translate/lib/google_api/translate/v3/model/list_locations_response.ex. Writing ListOperationsResponse to clients/translate/lib/google_api/translate/v3/model/list_operations_response.ex. Writing Location to clients/translate/lib/google_api/translate/v3/model/location.ex. Writing Operation to clients/translate/lib/google_api/translate/v3/model/operation.ex. Writing OutputConfig to clients/translate/lib/google_api/translate/v3/model/output_config.ex. Writing Status to clients/translate/lib/google_api/translate/v3/model/status.ex. Writing SupportedLanguage to clients/translate/lib/google_api/translate/v3/model/supported_language.ex. Writing SupportedLanguages to clients/translate/lib/google_api/translate/v3/model/supported_languages.ex. Writing TranslateTextGlossaryConfig to clients/translate/lib/google_api/translate/v3/model/translate_text_glossary_config.ex. Writing TranslateTextRequest to clients/translate/lib/google_api/translate/v3/model/translate_text_request.ex. Writing TranslateTextResponse to clients/translate/lib/google_api/translate/v3/model/translate_text_response.ex. Writing Translation to clients/translate/lib/google_api/translate/v3/model/translation.ex. Writing WaitOperationRequest to clients/translate/lib/google_api/translate/v3/model/wait_operation_request.ex. Writing Projects to clients/translate/lib/google_api/translate/v3/api/projects.ex. Writing connection.ex. Writing metadata.ex. Writing mix.exs Writing README.md Writing LICENSE Writing .gitignore Writing config/config.exs Writing test/test_helper.exs 15:02:54.113 [info] Found only discovery_revision and/or formatting changes. Not significant enough for a PR. fixing file permissions Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 252, in __exit__ self.observer.stop() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/utils/__init__.py", line 81, in stop self.on_thread_stop() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/api.py", line 361, in on_thread_stop self.unschedule_all() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/api.py", line 357, in unschedule_all self._clear_emitters() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/api.py", line 231, in _clear_emitters emitter.stop() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/utils/__init__.py", line 81, in stop self.on_thread_stop() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/inotify.py", line 121, in on_thread_stop self._inotify.close() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/inotify_buffer.py", line 50, in close self.stop() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/utils/__init__.py", line 81, in stop self.on_thread_stop() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/inotify_buffer.py", line 46, in on_thread_stop self._inotify.close() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/watchdog/observers/inotify_c.py", line 277, in close os.close(self._inotify_fd) OSError: [Errno 9] Bad file descriptor 2020-12-18 07:02:57,189 autosynth [ERROR] > Synthesis failed 2020-12-18 07:02:57,189 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 354, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 189, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 291, in _inner_main ).synthesize(synth_log_path / "sponge_log.log") File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/translate/synth.metadata', 'synth.py', '--', 'Translate']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/3a455424-7540-46a9-bb77-8265c8f04c06/targets/github%2Fsynthtool;config=default/tests;query=elixir-google-api;failed=false).
non_defect
synthesis failed for translate hello autosynth couldn t regenerate translate broken heart here s the output from running synth py poison compiling files ex warning integer to char list is deprecated use integer to charlist instead lib poison encoder ex generated poison app ssl verify fun compiling files erl generated ssl verify fun app compiling certifi compiling hackney compiling files ex generated app mime compiling files ex generated mime app tesla compiling files ex generated tesla app google gax compiling files ex generated google gax app google api discovery compiling files ex generated google api discovery app google apis compiling files ex generated google apis app fetching fetching found revision check old new generating true creating leading directories writing detectlanguagerequest to clients translate lib google api translate model detect language request ex writing detectionslistresponse to clients translate lib google api translate model detections list response ex writing detectionsresource to clients translate lib google api translate model detections resource ex writing getsupportedlanguagesrequest to clients translate lib google api translate model get supported languages request ex writing languageslistresponse to clients translate lib google api translate model languages list response ex writing languagesresource to clients translate lib google api translate model languages resource ex writing translatetextrequest to clients translate lib google api translate model translate text request ex writing translationslistresponse to clients translate lib google api translate model translations list response ex writing translationsresource to clients translate lib google api translate model translations resource ex writing detections to clients translate lib google api translate api detections ex writing languages to clients translate lib google api translate api languages ex writing translations to clients translate lib google api translate api translations ex writing connection ex writing metadata ex writing mix exs writing readme md writing license writing gitignore writing config config exs writing test test helper exs fetching fetching found revision check old new generating true creating leading directories writing batchtranslatetextrequest to clients translate lib google api translate model batch translate text request ex writing canceloperationrequest to clients translate lib google api translate model cancel operation request ex writing detectlanguagerequest to clients translate lib google api translate model detect language request ex writing detectlanguageresponse to clients translate lib google api translate model detect language response ex writing detectedlanguage to clients translate lib google api translate model detected language ex writing empty to clients translate lib google api translate model empty ex writing gcsdestination to clients translate lib google api translate model gcs destination ex writing gcssource to clients translate lib google api translate model gcs source ex writing glossary to clients translate lib google api translate model glossary ex writing glossaryinputconfig to clients translate lib google api translate model glossary input config ex writing inputconfig to clients translate lib google api translate model input config ex writing languagecodepair to clients translate lib google api translate model language code pair ex writing languagecodesset to clients translate lib google api translate model language codes set ex writing listglossariesresponse to clients translate lib google api translate model list glossaries response ex writing listlocationsresponse to clients translate lib google api translate model list locations response ex writing listoperationsresponse to clients translate lib google api translate model list operations response ex writing location to clients translate lib google api translate model location ex writing operation to clients translate lib google api translate model operation ex writing outputconfig to clients translate lib google api translate model output config ex writing status to clients translate lib google api translate model status ex writing supportedlanguage to clients translate lib google api translate model supported language ex writing supportedlanguages to clients translate lib google api translate model supported languages ex writing translatetextglossaryconfig to clients translate lib google api translate model translate text glossary config ex writing translatetextrequest to clients translate lib google api translate model translate text request ex writing translatetextresponse to clients translate lib google api translate model translate text response ex writing translation to clients translate lib google api translate model translation ex writing waitoperationrequest to clients translate lib google api translate model wait operation request ex writing projects to clients translate lib google api translate api projects ex writing connection ex writing metadata ex writing mix exs writing readme md writing license writing gitignore writing config config exs writing test test helper exs found only discovery revision and or formatting changes not significant enough for a pr fixing file permissions traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file tmpfs src github synthtool synthtool metadata py line in exit self observer stop file tmpfs src github synthtool env lib site packages watchdog utils init py line in stop self on thread stop file tmpfs src github synthtool env lib site packages watchdog observers api py line in on thread stop self unschedule all file tmpfs src github synthtool env lib site packages watchdog observers api py line in unschedule all self clear emitters file tmpfs src github synthtool env lib site packages watchdog observers api py line in clear emitters emitter stop file tmpfs src github synthtool env lib site packages watchdog utils init py line in stop self on thread stop file tmpfs src github synthtool env lib site packages watchdog observers inotify py line in on thread stop self inotify close file tmpfs src github synthtool env lib site packages watchdog observers inotify buffer py line in close self stop file tmpfs src github synthtool env lib site packages watchdog utils init py line in stop self on thread stop file tmpfs src github synthtool env lib site packages watchdog observers inotify buffer py line in on thread stop self inotify close file tmpfs src github synthtool env lib site packages watchdog observers inotify c py line in close os close self inotify fd oserror bad file descriptor autosynth synthesis failed autosynth running git clean fdx removing pycache traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize synth log path sponge log log file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
0
25,582
4,403,443,506
IssuesEvent
2016-08-11 07:56:27
r89m/tinyurl-generator
https://api.github.com/repos/r89m/tinyurl-generator
closed
The icon for toolbar does not show in Firefox Nightly (15.0a1 (2012-05-27))
auto-migrated Priority-High Type-Defect
``` What steps will reproduce the problem? 1.Right click tool bar. 2.Icon for TinyURL option does not show toolbar button. 3. What is the expected output? What do you see instead? Expect to see icon to drag to tool bar What version of the product are you using? On what operating system? Latest public on Mac OS X Snow Leopard (10.6.8) Please provide any additional information below. ``` Original issue reported on code.google.com by `ACCRi...@gmail.com` on 28 May 2012 at 5:15
1.0
The icon for toolbar does not show in Firefox Nightly (15.0a1 (2012-05-27)) - ``` What steps will reproduce the problem? 1.Right click tool bar. 2.Icon for TinyURL option does not show toolbar button. 3. What is the expected output? What do you see instead? Expect to see icon to drag to tool bar What version of the product are you using? On what operating system? Latest public on Mac OS X Snow Leopard (10.6.8) Please provide any additional information below. ``` Original issue reported on code.google.com by `ACCRi...@gmail.com` on 28 May 2012 at 5:15
defect
the icon for toolbar does not show in firefox nightly what steps will reproduce the problem right click tool bar icon for tinyurl option does not show toolbar button what is the expected output what do you see instead expect to see icon to drag to tool bar what version of the product are you using on what operating system latest public on mac os x snow leopard please provide any additional information below original issue reported on code google com by accri gmail com on may at
1
37,703
8,476,857,591
IssuesEvent
2018-10-24 23:49:20
jccastillo0007/eFacturaT
https://api.github.com/repos/jccastillo0007/eFacturaT
opened
Web Ventas-Registrar CFDI externo, rfc receptor con & no lo puede importar
bug defect
Tenemos un caso con un cliente del grupo etsa, donde el rfc del receptor, incluye la secuencia de escape para el ampersand: cfdi:Receptor Rfc="S&amp;T980427D19" te lo mandé por correo. No pueden subir el XML, ya que al validarlo con el sat, marca error.
1.0
Web Ventas-Registrar CFDI externo, rfc receptor con & no lo puede importar - Tenemos un caso con un cliente del grupo etsa, donde el rfc del receptor, incluye la secuencia de escape para el ampersand: cfdi:Receptor Rfc="S&amp;T980427D19" te lo mandé por correo. No pueden subir el XML, ya que al validarlo con el sat, marca error.
defect
web ventas registrar cfdi externo rfc receptor con no lo puede importar tenemos un caso con un cliente del grupo etsa donde el rfc del receptor incluye la secuencia de escape para el ampersand cfdi receptor rfc s amp te lo mandé por correo no pueden subir el xml ya que al validarlo con el sat marca error
1
14,541
2,822,792,242
IssuesEvent
2015-05-21 02:41:31
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
closed
Quote are stripped in subquery when used in a virtualField
Defect models
Cake 2.6.4. Quotes are stripped from a subquery when that subquery is used in a virtualField. ```php $tmp = $ds->buildStatement( array( 'fields'=>array('Customer.name'), 'table'=>$ds->fullTableName($this->Contract->Customer), 'alias'=>'Customer', 'limit'=>1, 'conditions'=>array( "Contract.customer_id = Customer.id" ) ), $this->Contract ); $this->Contract->virtualFields = array( 'v_test'=>$tmp ); $t = $this->Contract->field('v_test', array('id'=>5)); ``` The buildStatement alone generates: ```sql SELECT `Customer`.`name` FROM `jwb_inventory`.`customers` AS `Customer` WHERE `Contract`.`customer_id` = `Customer`.`id` LIMIT 1 ``` But when it goes through as virtualfield (as in the above code example), the quotes on the table alias get stripped, which looks like: ```sql SELECT (SELECT `Customer`.`name` FROM `jwb_inventory`.`customers` AS Customer WHERE `Contract`.`customer_id` = `Customer`.`id` LIMIT 1) AS `Contract__v_test` FROM `jwb_inventory`.`contracts` AS `Contract` WHERE `id` = 5 LIMIT 1 ``` Notice there are no longer any quotes around "Customer" in the subquery. This can obviously cause issues when the alias used matches a reserved keyword.
1.0
Quote are stripped in subquery when used in a virtualField - Cake 2.6.4. Quotes are stripped from a subquery when that subquery is used in a virtualField. ```php $tmp = $ds->buildStatement( array( 'fields'=>array('Customer.name'), 'table'=>$ds->fullTableName($this->Contract->Customer), 'alias'=>'Customer', 'limit'=>1, 'conditions'=>array( "Contract.customer_id = Customer.id" ) ), $this->Contract ); $this->Contract->virtualFields = array( 'v_test'=>$tmp ); $t = $this->Contract->field('v_test', array('id'=>5)); ``` The buildStatement alone generates: ```sql SELECT `Customer`.`name` FROM `jwb_inventory`.`customers` AS `Customer` WHERE `Contract`.`customer_id` = `Customer`.`id` LIMIT 1 ``` But when it goes through as virtualfield (as in the above code example), the quotes on the table alias get stripped, which looks like: ```sql SELECT (SELECT `Customer`.`name` FROM `jwb_inventory`.`customers` AS Customer WHERE `Contract`.`customer_id` = `Customer`.`id` LIMIT 1) AS `Contract__v_test` FROM `jwb_inventory`.`contracts` AS `Contract` WHERE `id` = 5 LIMIT 1 ``` Notice there are no longer any quotes around "Customer" in the subquery. This can obviously cause issues when the alias used matches a reserved keyword.
defect
quote are stripped in subquery when used in a virtualfield cake quotes are stripped from a subquery when that subquery is used in a virtualfield php tmp ds buildstatement array fields array customer name table ds fulltablename this contract customer alias customer limit conditions array contract customer id customer id this contract this contract virtualfields array v test tmp t this contract field v test array id the buildstatement alone generates sql select customer name from jwb inventory customers as customer where contract customer id customer id limit but when it goes through as virtualfield as in the above code example the quotes on the table alias get stripped which looks like sql select select customer name from jwb inventory customers as customer where contract customer id customer id limit as contract v test from jwb inventory contracts as contract where id limit notice there are no longer any quotes around customer in the subquery this can obviously cause issues when the alias used matches a reserved keyword
1
17,347
3,000,841,459
IssuesEvent
2015-07-24 06:39:18
krasa/eclipse-code-formatter-intellij-plugin
https://api.github.com/repos/krasa/eclipse-code-formatter-intellij-plugin
closed
Eclipse code formatting failure in Android studio
auto-migrated Priority-Medium Type-Defect
``` Error shown : failed to format with Eclipse code formatter. Configured settings file does not exist, path="" ``` Original issue reported on code.google.com by `fa...@qburst.com` on 15 Jul 2015 at 9:56
1.0
Eclipse code formatting failure in Android studio - ``` Error shown : failed to format with Eclipse code formatter. Configured settings file does not exist, path="" ``` Original issue reported on code.google.com by `fa...@qburst.com` on 15 Jul 2015 at 9:56
defect
eclipse code formatting failure in android studio error shown failed to format with eclipse code formatter configured settings file does not exist path original issue reported on code google com by fa qburst com on jul at
1
141,967
5,447,701,651
IssuesEvent
2017-03-07 14:16:07
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
closed
[craftercms] Invoking update on a module updates all
bug Priority: Medium
When invoking update like this: `./gradlew -Penv=authoring -Pmodule=studio update` All modules got updated instead of just studio.
1.0
[craftercms] Invoking update on a module updates all - When invoking update like this: `./gradlew -Penv=authoring -Pmodule=studio update` All modules got updated instead of just studio.
non_defect
invoking update on a module updates all when invoking update like this gradlew penv authoring pmodule studio update all modules got updated instead of just studio
0
78,898
27,811,803,762
IssuesEvent
2023-03-18 07:39:59
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
closed
Issues when replacing a disk
Type: Defect Status: Stale
Hi, every time I try to replace a disk with a few errors, after resilver the new disk gets errors too. I have tried it with 4 different brand-new disks for two years! In this long time I hoped, that with a new release the issue will be solved and gone. But this is not so. Now there are more errors and the computer gets very slow with 100% CPU. My system: `uname -a` `Linux albiggen 5.4.143-gentoo #1 SMP Tue Sep 7 12:56:32 CEST 2021 x86_64 Intel(R) Xeon(R) Gold 6138 CPU @ 2.00GHz GenuineIntel GNU/Linux` Main board Supermicro X11DPG-QT MNL-1998 Controller: `lspci -knn -s 17:0.0` `17:00.0 Serial Attached SCSI controller [0107]: Broadcom / LSI SAS3224 PCI-Express Fusion-MPT SAS-3 [1000:00c4] (rev 01) Subsystem: Broadcom / LSI SAS3224 PCI-Express Fusion-MPT SAS-3 [1000:3190] Kernel driver in use: mpt3sas ` `zpool version` `zfs-2.1.1-r5-gentoo zfs-kmod-2.1.0-r0-gentoo` `zpool status dpool` ``` pool: dpool state: DEGRADED status: One or more devices is currently being resilvered. The pool will continue to function, possibly in a degraded state. action: Wait for the resilver to complete. scan: resilver in progress since Thu Nov 25 13:27:01 2021 39.2T scanned at 1.22G/s, 35.6T issued at 1.11G/s, 42.0T total 465G resilvered, 84.77% done, 01:38:20 to go config: NAME STATE READ WRITE CKSUM dpool DEGRADED 0 0 0 raidz2-0 ONLINE 0 0 0 ata-WDC_WD30EFRX-68AX9N0_WD-WCC1T1492478 ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N0FVSV47 ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N0KC6SUN ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N1HY49V8 ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N1LL8S72 ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N2YNSZZZ ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N4RPDK59 ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N5NZVVHH ONLINE 0 0 0 raidz2-1 DEGRADED 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N5YNUCTL ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N7KUPX7F ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WMC4N0N4SL9K ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WMC4N0N59P4W ONLINE 0 0 0 ata-WDC_WD80EFAX-68KNBN0_VDH9D96D ONLINE 0 0 0 replacing-5 UNAVAIL 4 10 29 insufficient replicas ata-WDC_WD40EFAX-68JH4N0_WD-WX12D10KPEPH FAULTED 0 1.30K 0 too many errors (resilvering) ata-WDC_WD40EFAX-68JH4N0_WD-WX21DB9KC4E6 FAULTED 4 13 0 too many errors (resilvering) ata-WDC_WD80EFAX-68KNBN0_VDH9DVZD ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WMC4N1376540 ONLINE 0 0 0 logs nvme0n1p6 ONLINE 0 0 0 nvme1n1p3 ONLINE 0 0 0 cache nvme1n1p2 ONLINE 0 0 0 nvme0n1p5 ONLINE 0 0 0 errors: No known data errors ``` I checked both disks with smartctl - short and long. No errors. How can I find the cause - or is it maybe a bug? I hope for your advice. Best regards and many thanks in advance
1.0
Issues when replacing a disk - Hi, every time I try to replace a disk with a few errors, after resilver the new disk gets errors too. I have tried it with 4 different brand-new disks for two years! In this long time I hoped, that with a new release the issue will be solved and gone. But this is not so. Now there are more errors and the computer gets very slow with 100% CPU. My system: `uname -a` `Linux albiggen 5.4.143-gentoo #1 SMP Tue Sep 7 12:56:32 CEST 2021 x86_64 Intel(R) Xeon(R) Gold 6138 CPU @ 2.00GHz GenuineIntel GNU/Linux` Main board Supermicro X11DPG-QT MNL-1998 Controller: `lspci -knn -s 17:0.0` `17:00.0 Serial Attached SCSI controller [0107]: Broadcom / LSI SAS3224 PCI-Express Fusion-MPT SAS-3 [1000:00c4] (rev 01) Subsystem: Broadcom / LSI SAS3224 PCI-Express Fusion-MPT SAS-3 [1000:3190] Kernel driver in use: mpt3sas ` `zpool version` `zfs-2.1.1-r5-gentoo zfs-kmod-2.1.0-r0-gentoo` `zpool status dpool` ``` pool: dpool state: DEGRADED status: One or more devices is currently being resilvered. The pool will continue to function, possibly in a degraded state. action: Wait for the resilver to complete. scan: resilver in progress since Thu Nov 25 13:27:01 2021 39.2T scanned at 1.22G/s, 35.6T issued at 1.11G/s, 42.0T total 465G resilvered, 84.77% done, 01:38:20 to go config: NAME STATE READ WRITE CKSUM dpool DEGRADED 0 0 0 raidz2-0 ONLINE 0 0 0 ata-WDC_WD30EFRX-68AX9N0_WD-WCC1T1492478 ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N0FVSV47 ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N0KC6SUN ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N1HY49V8 ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N1LL8S72 ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N2YNSZZZ ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N4RPDK59 ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N5NZVVHH ONLINE 0 0 0 raidz2-1 DEGRADED 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N5YNUCTL ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WCC4N7KUPX7F ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WMC4N0N4SL9K ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WMC4N0N59P4W ONLINE 0 0 0 ata-WDC_WD80EFAX-68KNBN0_VDH9D96D ONLINE 0 0 0 replacing-5 UNAVAIL 4 10 29 insufficient replicas ata-WDC_WD40EFAX-68JH4N0_WD-WX12D10KPEPH FAULTED 0 1.30K 0 too many errors (resilvering) ata-WDC_WD40EFAX-68JH4N0_WD-WX21DB9KC4E6 FAULTED 4 13 0 too many errors (resilvering) ata-WDC_WD80EFAX-68KNBN0_VDH9DVZD ONLINE 0 0 0 ata-WDC_WD30EFRX-68EUZN0_WD-WMC4N1376540 ONLINE 0 0 0 logs nvme0n1p6 ONLINE 0 0 0 nvme1n1p3 ONLINE 0 0 0 cache nvme1n1p2 ONLINE 0 0 0 nvme0n1p5 ONLINE 0 0 0 errors: No known data errors ``` I checked both disks with smartctl - short and long. No errors. How can I find the cause - or is it maybe a bug? I hope for your advice. Best regards and many thanks in advance
defect
issues when replacing a disk hi every time i try to replace a disk with a few errors after resilver the new disk gets errors too i have tried it with different brand new disks for two years in this long time i hoped that with a new release the issue will be solved and gone but this is not so now there are more errors and the computer gets very slow with cpu my system uname a linux albiggen gentoo smp tue sep cest intel r xeon r gold cpu genuineintel gnu linux main board supermicro qt mnl controller lspci knn s serial attached scsi controller broadcom lsi pci express fusion mpt sas rev subsystem broadcom lsi pci express fusion mpt sas kernel driver in use zpool version zfs gentoo zfs kmod gentoo zpool status dpool pool dpool state degraded status one or more devices is currently being resilvered the pool will continue to function possibly in a degraded state action wait for the resilver to complete scan resilver in progress since thu nov scanned at s issued at s total resilvered done to go config name state read write cksum dpool degraded online ata wdc wd online ata wdc wd online ata wdc wd online ata wdc wd online ata wdc wd online ata wdc wd online ata wdc wd online ata wdc wd online degraded ata wdc wd online ata wdc wd online ata wdc wd online ata wdc wd online ata wdc online replacing unavail insufficient replicas ata wdc wd faulted too many errors resilvering ata wdc wd faulted too many errors resilvering ata wdc online ata wdc wd online logs online online cache online online errors no known data errors i checked both disks with smartctl short and long no errors how can i find the cause or is it maybe a bug i hope for your advice best regards and many thanks in advance
1
43,477
11,727,593,440
IssuesEvent
2020-03-10 16:10:36
CocoaPods/CocoaPods
https://api.github.com/repos/CocoaPods/CocoaPods
closed
Cocoapod overwrite change made in project after post_install
t2:defect
* [x] I've read and understood the [*CONTRIBUTING guidelines and have done my best effort to follow](https://github.com/CocoaPods/CocoaPods/blob/master/CONTRIBUTING.md). # Report 🌈 When I modify the project in post_install, some of my change are overwrited after pod finish. ### Podfile ``` platform :ios, '8.0' post_install do |installer| test_proj = installer.aggregate_targets.find do |target| target.user_project.root_object.name == "Test" end.user_project test_target = test_proj.native_targets.find { |target| target.name == "Test"} to_del = test_proj.native_targets.find_all { |target| target.name.start_with?("NewTest") } test_proj.targets.delete(to_del) target = test_proj.new_target(test_target.symbol_type, "NewTest", test_target.platform_name, test_target.deployment_target) target.build_configurations.clear test_target.build_configurations.each do |item| build = test_proj.new(Xcodeproj::Project::XCBuildConfiguration) build.name = item.name build.base_configuration_reference = item.base_configuration_reference build.build_settings.update(item.build_settings) target.build_configurations << build end # copy build_phases target.build_phases.clear test_target.build_phases.each do |src| dst = test_proj.new(src.class) target.build_phases << dst if (src.is_a?(Xcodeproj::Project::PBXShellScriptBuildPhase)) dst.name = "#{src.name}" dst.input_paths = src.input_paths dst.output_paths = src.output_paths dst.shell_path = src.shell_path dst.shell_script = src.shell_script dst.show_env_vars_in_log = src.show_env_vars_in_log else src.files.each do |f| build_file = test_proj.new(Xcodeproj::Project::Object::PBXBuildFile) build_file.file_ref = f.file_ref dst.files << build_file end end end test_proj.save() sleep(100) end target 'Test' do end ``` ### Bug What I do is just copying the target. When the project is opened before sleep end, everything is correct. After, configuration file for my new target is missing, and some build phase are deleted. ## CocoaPods Environment ### Stack ``` CocoaPods : 1.4.0 Ruby : ruby 2.3.3p222 (2016-11-21 revision 56859) [universal.x86_64-darwin17] RubyGems : 2.5.2 Host : Mac OS X 10.13 (17A365) Xcode : 9.2 (9C40b) Git : git version 2.14.3 (Apple Git-98) Ruby lib dir : /System/Library/Frameworks/Ruby.framework/Versions/2.3/usr/lib Repositories : master - https://github.com/CocoaPods/Specs.git @ 21d6bc804ed8858b89e89cbe290b52faf0bf534a ``` ### Installation Source ``` Executable Path: /usr/local/bin/pod ``` ### Plugins ``` cocoapods-deintegrate : 1.0.2 cocoapods-plugins : 1.0.0 cocoapods-search : 1.0.0 cocoapods-stats : 1.0.0 cocoapods-trunk : 1.3.0 cocoapods-try : 1.1.0 ``` ## Project that demonstrates the issue Just create a empty project in Xcode and copy my podfile. 🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈
1.0
Cocoapod overwrite change made in project after post_install - * [x] I've read and understood the [*CONTRIBUTING guidelines and have done my best effort to follow](https://github.com/CocoaPods/CocoaPods/blob/master/CONTRIBUTING.md). # Report 🌈 When I modify the project in post_install, some of my change are overwrited after pod finish. ### Podfile ``` platform :ios, '8.0' post_install do |installer| test_proj = installer.aggregate_targets.find do |target| target.user_project.root_object.name == "Test" end.user_project test_target = test_proj.native_targets.find { |target| target.name == "Test"} to_del = test_proj.native_targets.find_all { |target| target.name.start_with?("NewTest") } test_proj.targets.delete(to_del) target = test_proj.new_target(test_target.symbol_type, "NewTest", test_target.platform_name, test_target.deployment_target) target.build_configurations.clear test_target.build_configurations.each do |item| build = test_proj.new(Xcodeproj::Project::XCBuildConfiguration) build.name = item.name build.base_configuration_reference = item.base_configuration_reference build.build_settings.update(item.build_settings) target.build_configurations << build end # copy build_phases target.build_phases.clear test_target.build_phases.each do |src| dst = test_proj.new(src.class) target.build_phases << dst if (src.is_a?(Xcodeproj::Project::PBXShellScriptBuildPhase)) dst.name = "#{src.name}" dst.input_paths = src.input_paths dst.output_paths = src.output_paths dst.shell_path = src.shell_path dst.shell_script = src.shell_script dst.show_env_vars_in_log = src.show_env_vars_in_log else src.files.each do |f| build_file = test_proj.new(Xcodeproj::Project::Object::PBXBuildFile) build_file.file_ref = f.file_ref dst.files << build_file end end end test_proj.save() sleep(100) end target 'Test' do end ``` ### Bug What I do is just copying the target. When the project is opened before sleep end, everything is correct. After, configuration file for my new target is missing, and some build phase are deleted. ## CocoaPods Environment ### Stack ``` CocoaPods : 1.4.0 Ruby : ruby 2.3.3p222 (2016-11-21 revision 56859) [universal.x86_64-darwin17] RubyGems : 2.5.2 Host : Mac OS X 10.13 (17A365) Xcode : 9.2 (9C40b) Git : git version 2.14.3 (Apple Git-98) Ruby lib dir : /System/Library/Frameworks/Ruby.framework/Versions/2.3/usr/lib Repositories : master - https://github.com/CocoaPods/Specs.git @ 21d6bc804ed8858b89e89cbe290b52faf0bf534a ``` ### Installation Source ``` Executable Path: /usr/local/bin/pod ``` ### Plugins ``` cocoapods-deintegrate : 1.0.2 cocoapods-plugins : 1.0.0 cocoapods-search : 1.0.0 cocoapods-stats : 1.0.0 cocoapods-trunk : 1.3.0 cocoapods-try : 1.1.0 ``` ## Project that demonstrates the issue Just create a empty project in Xcode and copy my podfile. 🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈
defect
cocoapod overwrite change made in project after post install i ve read and understood the report 🌈 when i modify the project in post install some of my change are overwrited after pod finish podfile platform ios post install do installer test proj installer aggregate targets find do target target user project root object name test end user project test target test proj native targets find target target name test to del test proj native targets find all target target name start with newtest test proj targets delete to del target test proj new target test target symbol type newtest test target platform name test target deployment target target build configurations clear test target build configurations each do item build test proj new xcodeproj project xcbuildconfiguration build name item name build base configuration reference item base configuration reference build build settings update item build settings target build configurations build end copy build phases target build phases clear test target build phases each do src dst test proj new src class target build phases dst if src is a xcodeproj project pbxshellscriptbuildphase dst name src name dst input paths src input paths dst output paths src output paths dst shell path src shell path dst shell script src shell script dst show env vars in log src show env vars in log else src files each do f build file test proj new xcodeproj project object pbxbuildfile build file file ref f file ref dst files build file end end end test proj save sleep end target test do end bug what i do is just copying the target when the project is opened before sleep end everything is correct after configuration file for my new target is missing and some build phase are deleted cocoapods environment stack cocoapods ruby ruby revision rubygems host mac os x xcode git git version apple git ruby lib dir system library frameworks ruby framework versions usr lib repositories master installation source executable path usr local bin pod plugins cocoapods deintegrate cocoapods plugins cocoapods search cocoapods stats cocoapods trunk cocoapods try project that demonstrates the issue just create a empty project in xcode and copy my podfile 🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈🌈
1
4,284
2,610,090,867
IssuesEvent
2015-02-26 18:27:28
chrsmith/dsdsdaadf
https://api.github.com/repos/chrsmith/dsdsdaadf
opened
深圳痘坑怎么祛除
auto-migrated Priority-Medium Type-Defect
``` 深圳痘坑怎么祛除【深圳韩方科颜全国热线400-869-1818,24小时 QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘�� �——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方� ��颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健 康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业�� �疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘� ��。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:44
1.0
深圳痘坑怎么祛除 - ``` 深圳痘坑怎么祛除【深圳韩方科颜全国热线400-869-1818,24小时 QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘�� �——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方� ��颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健 康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业�� �疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘� ��。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:44
defect
深圳痘坑怎么祛除 深圳痘坑怎么祛除【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘�� �——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方� ��颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健 康祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业�� �疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘� ��。 original issue reported on code google com by szft com on may at
1
1,225
2,601,761,013
IssuesEvent
2015-02-24 00:35:06
chrsmith/bwapi
https://api.github.com/repos/chrsmith/bwapi
closed
Save replay conflict causes crash
auto-migrated Milestone-Release Priority-High Type-Defect Usability
``` When Starcraft goes to save LastReplay.rep for multiple instances, they conflict and one instance will crash. It appears that the "save failure" functionality for outside of the game was not tested by Blizzard. Detour DeleteFile, CreateFile, and GetFileAttributes. Replace the LastReplay.rep with our auto-menu save_replay name, or append the name with (2), (3), etc. if it exists. ``` ----- Original issue reported on code.google.com by `AHeinerm` on 13 Mar 2011 at 12:55
1.0
Save replay conflict causes crash - ``` When Starcraft goes to save LastReplay.rep for multiple instances, they conflict and one instance will crash. It appears that the "save failure" functionality for outside of the game was not tested by Blizzard. Detour DeleteFile, CreateFile, and GetFileAttributes. Replace the LastReplay.rep with our auto-menu save_replay name, or append the name with (2), (3), etc. if it exists. ``` ----- Original issue reported on code.google.com by `AHeinerm` on 13 Mar 2011 at 12:55
defect
save replay conflict causes crash when starcraft goes to save lastreplay rep for multiple instances they conflict and one instance will crash it appears that the save failure functionality for outside of the game was not tested by blizzard detour deletefile createfile and getfileattributes replace the lastreplay rep with our auto menu save replay name or append the name with etc if it exists original issue reported on code google com by aheinerm on mar at
1
59,220
17,016,570,372
IssuesEvent
2021-07-02 12:54:42
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
opened
osmosis rii-task failed
Component: osmosis Priority: critical Type: defect
**[Submitted to the original trac issue database at 9.41am, Monday, 1st September 2014]** Hi, on update with: ${OSMOSIS} --rri workingDirectory=. --wxc update.osc.gz I get this error on different files: #### Sep 01, 2014 11:36:20 AM sun.net.www.http.HttpClient logFinest AM FEINSTEN: KeepAlive stream used: http://planet.openstreetmap.org/replication/minute/001/021/814.osc.gz Sep 01, 2014 11:36:20 AM sun.net.www.protocol.http.HttpURLConnection getInputStream FEIN: sun.net.www.MessageHeader@5fe36eb910 pairs: {null: HTTP/1.1 200 OK}{Date: Mon, 01 Sep 2014 09:36:20 GMT}{Server: Apache/2.2.22 (Ubuntu)}{Last-Modified: Wed, 27 Aug 2014 10:30:09 GMT}{ETag: "5b8ce49-a5e0-50199e783c919"}{Accept-Ranges: bytes}{Content-Length: 42464}{Keep-Alive: timeout=5, max=2}{Connection: Keep-Alive}{Content-Type: application/x-gzip} Sep 01, 2014 11:36:20 AM org.openstreetmap.osmosis.replication.v0_6.BaseReplicationDownloader download FEINER: Processing replication sequence 1021815. Sep 01, 2014 11:36:20 AM sun.net.www.protocol.http.HttpURLConnection plainConnect AM FEINSTEN: ProxySelector Request for http://planet.openstreetmap.org/replication/minute/001/021/815.state.txt Sep 01, 2014 11:36:20 AM sun.net.www.http.HttpClient logFinest AM FEINSTEN: KeepAlive stream retrieved from the cache, sun.net.www.http.HttpClient(http://planet.openstreetmap.org/replication/minute/001/021/814.osc.gz) Sep 01, 2014 11:36:20 AM sun.net.www.protocol.http.HttpURLConnection plainConnect AM FEINSTEN: Proxy used: DIRECT Sep 01, 2014 11:36:20 AM sun.net.www.protocol.http.HttpURLConnection writeRequests FEIN: sun.net.www.MessageHeader@64e052805 pairs: {GET /replication/minute/001/021/815.state.txt HTTP/1.1: null}{User-Agent: Java/1.7.0_65}{Host: planet.openstreetmap.org}{Accept: text/html, image/gif, image/jpeg, *; q=.2, */*; q=.2}{Connection: keep-alive} Sep 01, 2014 11:36:20 AM sun.net.www.protocol.http.HttpURLConnection getInputStream FEIN: sun.net.www.MessageHeader@4348853d11 pairs: {null: HTTP/1.1 200 OK}{Date: Mon, 01 Sep 2014 09:36:20 GMT}{Server: Apache/2.2.22 (Ubuntu)}{Last-Modified: Wed, 27 Aug 2014 10:31:02 GMT}{ETag: "5b8ce4d-a5-50199eab1d02c"}{Accept-Ranges: bytes}{Content-Length: 165}{Vary: Accept-Encoding}{Keep-Alive: timeout=5, max=1}{Connection: Keep-Alive}{Content-Type: text/plain; charset=utf-8} Sep 01, 2014 11:36:20 AM sun.net.www.protocol.http.HttpURLConnection plainConnect AM FEINSTEN: ProxySelector Request for http://planet.openstreetmap.org/replication/minute/001/021/815.osc.gz Sep 01, 2014 11:38:27 AM org.openstreetmap.osmosis.core.pipeline.common.ActiveTaskManager waitForCompletion SCHWERWIEGEND: Thread for task 1-rri failed org.openstreetmap.osmosis.core.OsmosisRuntimeException: Unable to read the changeset file 001/021/815.osc.gz from the server. at org.openstreetmap.osmosis.replication.v0_6.BaseReplicationDownloader.downloadReplicationFile(BaseReplicationDownloader.java:133) at org.openstreetmap.osmosis.replication.v0_6.BaseReplicationDownloader.download(BaseReplicationDownloader.java:262) at org.openstreetmap.osmosis.replication.v0_6.BaseReplicationDownloader.runImpl(BaseReplicationDownloader.java:302) at org.openstreetmap.osmosis.replication.v0_6.BaseReplicationDownloader.run(BaseReplicationDownloader.java:381) at java.lang.Thread.run(Thread.java:745) Caused by: java.net.ConnectException: Connection timed out at java.net.PlainSocketImpl.socketConnect(Native Method) at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:339) at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:200) at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:182) at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392) at java.net.Socket.connect(Socket.java:579) at sun.net.NetworkClient.doConnect(NetworkClient.java:175) at sun.net.www.http.HttpClient.openServer(HttpClient.java:432) at sun.net.www.http.HttpClient.openServer(HttpClient.java:527) at sun.net.www.http.HttpClient.<init>(HttpClient.java:211) at sun.net.www.http.HttpClient.New(HttpClient.java:308) at sun.net.www.http.HttpClient.New(HttpClient.java:326) at sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:996) at sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:932) at sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:850) at sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1300) at org.openstreetmap.osmosis.replication.v0_6.BaseReplicationDownloader.downloadReplicationFile(BaseReplicationDownloader.java:107) ... 4 more ####
1.0
osmosis rii-task failed - **[Submitted to the original trac issue database at 9.41am, Monday, 1st September 2014]** Hi, on update with: ${OSMOSIS} --rri workingDirectory=. --wxc update.osc.gz I get this error on different files: #### Sep 01, 2014 11:36:20 AM sun.net.www.http.HttpClient logFinest AM FEINSTEN: KeepAlive stream used: http://planet.openstreetmap.org/replication/minute/001/021/814.osc.gz Sep 01, 2014 11:36:20 AM sun.net.www.protocol.http.HttpURLConnection getInputStream FEIN: sun.net.www.MessageHeader@5fe36eb910 pairs: {null: HTTP/1.1 200 OK}{Date: Mon, 01 Sep 2014 09:36:20 GMT}{Server: Apache/2.2.22 (Ubuntu)}{Last-Modified: Wed, 27 Aug 2014 10:30:09 GMT}{ETag: "5b8ce49-a5e0-50199e783c919"}{Accept-Ranges: bytes}{Content-Length: 42464}{Keep-Alive: timeout=5, max=2}{Connection: Keep-Alive}{Content-Type: application/x-gzip} Sep 01, 2014 11:36:20 AM org.openstreetmap.osmosis.replication.v0_6.BaseReplicationDownloader download FEINER: Processing replication sequence 1021815. Sep 01, 2014 11:36:20 AM sun.net.www.protocol.http.HttpURLConnection plainConnect AM FEINSTEN: ProxySelector Request for http://planet.openstreetmap.org/replication/minute/001/021/815.state.txt Sep 01, 2014 11:36:20 AM sun.net.www.http.HttpClient logFinest AM FEINSTEN: KeepAlive stream retrieved from the cache, sun.net.www.http.HttpClient(http://planet.openstreetmap.org/replication/minute/001/021/814.osc.gz) Sep 01, 2014 11:36:20 AM sun.net.www.protocol.http.HttpURLConnection plainConnect AM FEINSTEN: Proxy used: DIRECT Sep 01, 2014 11:36:20 AM sun.net.www.protocol.http.HttpURLConnection writeRequests FEIN: sun.net.www.MessageHeader@64e052805 pairs: {GET /replication/minute/001/021/815.state.txt HTTP/1.1: null}{User-Agent: Java/1.7.0_65}{Host: planet.openstreetmap.org}{Accept: text/html, image/gif, image/jpeg, *; q=.2, */*; q=.2}{Connection: keep-alive} Sep 01, 2014 11:36:20 AM sun.net.www.protocol.http.HttpURLConnection getInputStream FEIN: sun.net.www.MessageHeader@4348853d11 pairs: {null: HTTP/1.1 200 OK}{Date: Mon, 01 Sep 2014 09:36:20 GMT}{Server: Apache/2.2.22 (Ubuntu)}{Last-Modified: Wed, 27 Aug 2014 10:31:02 GMT}{ETag: "5b8ce4d-a5-50199eab1d02c"}{Accept-Ranges: bytes}{Content-Length: 165}{Vary: Accept-Encoding}{Keep-Alive: timeout=5, max=1}{Connection: Keep-Alive}{Content-Type: text/plain; charset=utf-8} Sep 01, 2014 11:36:20 AM sun.net.www.protocol.http.HttpURLConnection plainConnect AM FEINSTEN: ProxySelector Request for http://planet.openstreetmap.org/replication/minute/001/021/815.osc.gz Sep 01, 2014 11:38:27 AM org.openstreetmap.osmosis.core.pipeline.common.ActiveTaskManager waitForCompletion SCHWERWIEGEND: Thread for task 1-rri failed org.openstreetmap.osmosis.core.OsmosisRuntimeException: Unable to read the changeset file 001/021/815.osc.gz from the server. at org.openstreetmap.osmosis.replication.v0_6.BaseReplicationDownloader.downloadReplicationFile(BaseReplicationDownloader.java:133) at org.openstreetmap.osmosis.replication.v0_6.BaseReplicationDownloader.download(BaseReplicationDownloader.java:262) at org.openstreetmap.osmosis.replication.v0_6.BaseReplicationDownloader.runImpl(BaseReplicationDownloader.java:302) at org.openstreetmap.osmosis.replication.v0_6.BaseReplicationDownloader.run(BaseReplicationDownloader.java:381) at java.lang.Thread.run(Thread.java:745) Caused by: java.net.ConnectException: Connection timed out at java.net.PlainSocketImpl.socketConnect(Native Method) at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:339) at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:200) at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:182) at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392) at java.net.Socket.connect(Socket.java:579) at sun.net.NetworkClient.doConnect(NetworkClient.java:175) at sun.net.www.http.HttpClient.openServer(HttpClient.java:432) at sun.net.www.http.HttpClient.openServer(HttpClient.java:527) at sun.net.www.http.HttpClient.<init>(HttpClient.java:211) at sun.net.www.http.HttpClient.New(HttpClient.java:308) at sun.net.www.http.HttpClient.New(HttpClient.java:326) at sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:996) at sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:932) at sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:850) at sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1300) at org.openstreetmap.osmosis.replication.v0_6.BaseReplicationDownloader.downloadReplicationFile(BaseReplicationDownloader.java:107) ... 4 more ####
defect
osmosis rii task failed hi on update with osmosis rri workingdirectory wxc update osc gz i get this error on different files sep am sun net logfinest am feinsten keepalive stream used sep am sun net getinputstream fein sun net pairs null http ok date mon sep gmt server apache ubuntu last modified wed aug gmt etag accept ranges bytes content length keep alive timeout max connection keep alive content type application x gzip sep am org openstreetmap osmosis replication basereplicationdownloader download feiner processing replication sequence sep am sun net plainconnect am feinsten proxyselector request for sep am sun net logfinest am feinsten keepalive stream retrieved from the cache sun net sep am sun net plainconnect am feinsten proxy used direct sep am sun net writerequests fein sun net pairs get replication minute state txt http null user agent java host planet openstreetmap org accept text html image gif image jpeg q q connection keep alive sep am sun net getinputstream fein sun net pairs null http ok date mon sep gmt server apache ubuntu last modified wed aug gmt etag accept ranges bytes content length vary accept encoding keep alive timeout max connection keep alive content type text plain charset utf sep am sun net plainconnect am feinsten proxyselector request for sep am org openstreetmap osmosis core pipeline common activetaskmanager waitforcompletion schwerwiegend thread for task rri failed org openstreetmap osmosis core osmosisruntimeexception unable to read the changeset file osc gz from the server at org openstreetmap osmosis replication basereplicationdownloader downloadreplicationfile basereplicationdownloader java at org openstreetmap osmosis replication basereplicationdownloader download basereplicationdownloader java at org openstreetmap osmosis replication basereplicationdownloader runimpl basereplicationdownloader java at org openstreetmap osmosis replication basereplicationdownloader run basereplicationdownloader java at java lang thread run thread java caused by java net connectexception connection timed out at java net plainsocketimpl socketconnect native method at java net abstractplainsocketimpl doconnect abstractplainsocketimpl java at java net abstractplainsocketimpl connecttoaddress abstractplainsocketimpl java at java net abstractplainsocketimpl connect abstractplainsocketimpl java at java net sockssocketimpl connect sockssocketimpl java at java net socket connect socket java at sun net networkclient doconnect networkclient java at sun net at sun net at sun net at sun net at sun net at sun net at sun net at sun net at sun net at org openstreetmap osmosis replication basereplicationdownloader downloadreplicationfile basereplicationdownloader java more
1
27,801
5,105,612,851
IssuesEvent
2017-01-05 08:20:39
BOINC/boinc
https://api.github.com/repos/BOINC/boinc
closed
update_versions overwrites files in the download directory
C: Server - Setup P: Major T: Defect
When adding new app versions, update versions will happily overwrite existing files, which will cause download errors due to invalid signatures. update_versions should check if a file exists in the download directory before copying. If it exists and is identical no copy should occur. If it exists and is not identical this should be considered an error, and update_versions should not insert that app version. If it does not exist then the file can be copied to the download directory.
1.0
update_versions overwrites files in the download directory - When adding new app versions, update versions will happily overwrite existing files, which will cause download errors due to invalid signatures. update_versions should check if a file exists in the download directory before copying. If it exists and is identical no copy should occur. If it exists and is not identical this should be considered an error, and update_versions should not insert that app version. If it does not exist then the file can be copied to the download directory.
defect
update versions overwrites files in the download directory when adding new app versions update versions will happily overwrite existing files which will cause download errors due to invalid signatures update versions should check if a file exists in the download directory before copying if it exists and is identical no copy should occur if it exists and is not identical this should be considered an error and update versions should not insert that app version if it does not exist then the file can be copied to the download directory
1
31,909
6,658,461,368
IssuesEvent
2017-09-30 19:34:30
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
[PrimeFaces showcase] chat example is not working
defect
## 1) Environment - PrimeFaces version: 6.1 - Atmosphere version: 2.4.12 - Does it work on the newest released PrimeFaces version? Version? No - Application server + version: GlassFish 4.1.2 - Affected browsers: Tested with Chrome only ## 2) Expected behavior I am running [Chat example in showcase](https://www.primefaces.org/showcase/push/chat.xhtml). 1. Users can login 2. Users are notified when new users log in ## 3) Actual behavior 1. Nothing happens when I click "Login" button. This is because `init` annotated with `@PostConstruct` is not called and `ChatView.users` is `null`. ( I fixed the above problem by assigning the instance as the instance variable is declared. I am not sure if this is the right way to fix it. ) 2. Users are not notified when new users log in. In ChatResource.java ``` @OnClose public void onClose(RemoteEndpoint r, EventBus eventBus) { ChatUsers users= (ChatUsers) ctx.getAttribute("chatUsers"); users.remove(username); eventBus.publish(room + "/*", new Message(String.format("%s has left the room", username), true)); } ``` This is because `eventBus` is `null`. ## 4) Steps to reproduce I used exactly the same code as [the chat example in PrimeFaces showcase](https://www.primefaces.org/showcase/push/chat.xhtml), except that I enabled the input textbox and button in the `chat.xhtml`. ## 5) Sample XHTML See [chat.xhtml](https://www.primefaces.org/showcase/push/chat.xhtml), and enable the textbox and button. ## 6) Sample bean See [Chat example](https://www.primefaces.org/showcase/push/chat.xhtml) ## 7) Question 1. What are the right way to fix the above problems? 2. Is it possible to inject CDI bean (For example, a class annotated with `@javax.enterprise.context.ApplicationScoped`) to the class annotated with `@PushEndPoint`? It seems that we cannot inject either JSF beans and CDI beans. And `ServletContext.getAttribute` can only access JSF beans. An example is as follows. ChatResource.java ``` @PushEndpoint("/{room}/{user}") @Singleton public class ChatResource { @PathParam("room") private String room; @PathParam("user") private String username; @Inject private ChatUsers users; ... } ``` ChatUsers.java ``` ... import javax.enterprise.context.ApplicationScoped; @ApplicationScoped public class ChatUsers implements Serializable { private List<String> users; @PostConstruct public void init() { users = new ArrayList<String>(); } ... } ```
1.0
[PrimeFaces showcase] chat example is not working - ## 1) Environment - PrimeFaces version: 6.1 - Atmosphere version: 2.4.12 - Does it work on the newest released PrimeFaces version? Version? No - Application server + version: GlassFish 4.1.2 - Affected browsers: Tested with Chrome only ## 2) Expected behavior I am running [Chat example in showcase](https://www.primefaces.org/showcase/push/chat.xhtml). 1. Users can login 2. Users are notified when new users log in ## 3) Actual behavior 1. Nothing happens when I click "Login" button. This is because `init` annotated with `@PostConstruct` is not called and `ChatView.users` is `null`. ( I fixed the above problem by assigning the instance as the instance variable is declared. I am not sure if this is the right way to fix it. ) 2. Users are not notified when new users log in. In ChatResource.java ``` @OnClose public void onClose(RemoteEndpoint r, EventBus eventBus) { ChatUsers users= (ChatUsers) ctx.getAttribute("chatUsers"); users.remove(username); eventBus.publish(room + "/*", new Message(String.format("%s has left the room", username), true)); } ``` This is because `eventBus` is `null`. ## 4) Steps to reproduce I used exactly the same code as [the chat example in PrimeFaces showcase](https://www.primefaces.org/showcase/push/chat.xhtml), except that I enabled the input textbox and button in the `chat.xhtml`. ## 5) Sample XHTML See [chat.xhtml](https://www.primefaces.org/showcase/push/chat.xhtml), and enable the textbox and button. ## 6) Sample bean See [Chat example](https://www.primefaces.org/showcase/push/chat.xhtml) ## 7) Question 1. What are the right way to fix the above problems? 2. Is it possible to inject CDI bean (For example, a class annotated with `@javax.enterprise.context.ApplicationScoped`) to the class annotated with `@PushEndPoint`? It seems that we cannot inject either JSF beans and CDI beans. And `ServletContext.getAttribute` can only access JSF beans. An example is as follows. ChatResource.java ``` @PushEndpoint("/{room}/{user}") @Singleton public class ChatResource { @PathParam("room") private String room; @PathParam("user") private String username; @Inject private ChatUsers users; ... } ``` ChatUsers.java ``` ... import javax.enterprise.context.ApplicationScoped; @ApplicationScoped public class ChatUsers implements Serializable { private List<String> users; @PostConstruct public void init() { users = new ArrayList<String>(); } ... } ```
defect
chat example is not working environment primefaces version atmosphere version does it work on the newest released primefaces version version no application server version glassfish affected browsers tested with chrome only expected behavior i am running users can login users are notified when new users log in actual behavior nothing happens when i click login button this is because init annotated with postconstruct is not called and chatview users is null i fixed the above problem by assigning the instance as the instance variable is declared i am not sure if this is the right way to fix it users are not notified when new users log in in chatresource java onclose public void onclose remoteendpoint r eventbus eventbus chatusers users chatusers ctx getattribute chatusers users remove username eventbus publish room new message string format s has left the room username true this is because eventbus is null steps to reproduce i used exactly the same code as except that i enabled the input textbox and button in the chat xhtml sample xhtml see and enable the textbox and button sample bean see question what are the right way to fix the above problems is it possible to inject cdi bean for example a class annotated with javax enterprise context applicationscoped to the class annotated with pushendpoint it seems that we cannot inject either jsf beans and cdi beans and servletcontext getattribute can only access jsf beans an example is as follows chatresource java pushendpoint room user singleton public class chatresource pathparam room private string room pathparam user private string username inject private chatusers users chatusers java import javax enterprise context applicationscoped applicationscoped public class chatusers implements serializable private list users postconstruct public void init users new arraylist
1
6,195
2,610,222,993
IssuesEvent
2015-02-26 19:10:44
chrsmith/somefinders
https://api.github.com/repos/chrsmith/somefinders
opened
инструкция zanussi fe 1002
auto-migrated Priority-Medium Type-Defect
``` '''Авдей Захаров''' Привет всем не подскажите где можно найти .инструкция zanussi fe 1002. где то видел уже '''Геодар Григорьев''' Качай тут http://bit.ly/1aMqgDD '''Аксён Носов''' Просит ввести номер мобилы!Не опасно ли это? '''Аполлинарий Ларионов''' Неа все ок у меня ничего не списало '''Артур Некрасов''' Не это не влияет на баланс Информация о файле: инструкция zanussi fe 1002 Загружен: В этом месяце Скачан раз: 412 Рейтинг: 258 Средняя скорость скачивания: 326 Похожих файлов: 33 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 18 Dec 2013 at 5:45
1.0
инструкция zanussi fe 1002 - ``` '''Авдей Захаров''' Привет всем не подскажите где можно найти .инструкция zanussi fe 1002. где то видел уже '''Геодар Григорьев''' Качай тут http://bit.ly/1aMqgDD '''Аксён Носов''' Просит ввести номер мобилы!Не опасно ли это? '''Аполлинарий Ларионов''' Неа все ок у меня ничего не списало '''Артур Некрасов''' Не это не влияет на баланс Информация о файле: инструкция zanussi fe 1002 Загружен: В этом месяце Скачан раз: 412 Рейтинг: 258 Средняя скорость скачивания: 326 Похожих файлов: 33 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 18 Dec 2013 at 5:45
defect
инструкция zanussi fe авдей захаров привет всем не подскажите где можно найти инструкция zanussi fe где то видел уже геодар григорьев качай тут аксён носов просит ввести номер мобилы не опасно ли это аполлинарий ларионов неа все ок у меня ничего не списало артур некрасов не это не влияет на баланс информация о файле инструкция zanussi fe загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at
1
36,530
7,978,960,340
IssuesEvent
2018-07-17 20:04:35
aleofreddi/svgpan
https://api.github.com/repos/aleofreddi/svgpan
closed
1.2.2 broken in Firefox when using Viewport (fix included)
Priority-Medium Type-Defect
``` While trying to use an SVG image with a Viewport set and svgpan 1.2.2 I'm unable to move/zoom the SVG. Firefox outputs lots of "matrix is undefined" errors to the console from the setCTM line in getRoot(). This appears to be because Firefox is behaving differently (to say Chrome) on the call to t.getCTM(). There is some discussion about the issue in a Firefox bug report at https://bugzilla.mozilla.org/show_bug.cgi?id=543965 As a result of that bug report I've found I can fix the issue by changing line 115 in version 1.2.2. From: setCTM(r, t.getCTM()); To: setCTM(r, t.getCTM() || t.getScreenCTM()); ``` Original issue reported on code.google.com by `graham_a...@hotmail.com` on 30 Sep 2011 at 9:15
1.0
1.2.2 broken in Firefox when using Viewport (fix included) - ``` While trying to use an SVG image with a Viewport set and svgpan 1.2.2 I'm unable to move/zoom the SVG. Firefox outputs lots of "matrix is undefined" errors to the console from the setCTM line in getRoot(). This appears to be because Firefox is behaving differently (to say Chrome) on the call to t.getCTM(). There is some discussion about the issue in a Firefox bug report at https://bugzilla.mozilla.org/show_bug.cgi?id=543965 As a result of that bug report I've found I can fix the issue by changing line 115 in version 1.2.2. From: setCTM(r, t.getCTM()); To: setCTM(r, t.getCTM() || t.getScreenCTM()); ``` Original issue reported on code.google.com by `graham_a...@hotmail.com` on 30 Sep 2011 at 9:15
defect
broken in firefox when using viewport fix included while trying to use an svg image with a viewport set and svgpan i m unable to move zoom the svg firefox outputs lots of matrix is undefined errors to the console from the setctm line in getroot this appears to be because firefox is behaving differently to say chrome on the call to t getctm there is some discussion about the issue in a firefox bug report at as a result of that bug report i ve found i can fix the issue by changing line in version from setctm r t getctm to setctm r t getctm t getscreenctm original issue reported on code google com by graham a hotmail com on sep at
1
77,572
27,056,330,774
IssuesEvent
2023-02-13 16:23:58
zed-industries/community
https://api.github.com/repos/zed-industries/community
opened
Find highlights persist even after terminal is cleared
defect terminal
### Check for existing issues - [X] Completed ### Describe the bug / provide steps to reproduce it A user brought this up ### Environment Zed: v0.73.3 (preview) OS: macOS 13.1.0 Memory: 64 GiB Architecture: aarch64 ### If applicable, add mockups / screenshots to help explain present your vision of the feature <img width="326" alt="SCR-20230213-ftj" src="https://user-images.githubusercontent.com/19867440/218513837-830fc216-fd68-4557-8b46-06f415209c27.png"> ### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue. If you only need the most recent lines, you can run the `zed: open log` command palette action to see the last 1000. _No response_
1.0
Find highlights persist even after terminal is cleared - ### Check for existing issues - [X] Completed ### Describe the bug / provide steps to reproduce it A user brought this up ### Environment Zed: v0.73.3 (preview) OS: macOS 13.1.0 Memory: 64 GiB Architecture: aarch64 ### If applicable, add mockups / screenshots to help explain present your vision of the feature <img width="326" alt="SCR-20230213-ftj" src="https://user-images.githubusercontent.com/19867440/218513837-830fc216-fd68-4557-8b46-06f415209c27.png"> ### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue. If you only need the most recent lines, you can run the `zed: open log` command palette action to see the last 1000. _No response_
defect
find highlights persist even after terminal is cleared check for existing issues completed describe the bug provide steps to reproduce it a user brought this up environment zed preview os macos memory gib architecture if applicable add mockups screenshots to help explain present your vision of the feature img width alt scr ftj src if applicable attach your library logs zed zed log file to this issue if you only need the most recent lines you can run the zed open log command palette action to see the last no response
1
45,151
12,602,759,234
IssuesEvent
2020-06-11 12:19:49
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Cannot use Meta.getSchema() for DROP SCHEMA statement
C: Functionality E: All Editions P: Medium T: Defect
As reported on the jOOQ user group: ---- ```java Schema schema = // from a call to meta() and finding the right schema db.getDSLContext().dropSchema(schema.getName()).execute();// works db.getDSLContext().dropSchema(schema).execute(); // does not work, see exception below db.getDSLContext().execute("drop schema jsl_db"); // works ``` The name of my database is "test" and the schema is obviously "jsl_db" ``` Exception in thread "main" org.jooq.exception.DataAccessException: SQL [drop schema "test"."jsl_db"]; ERROR: syntax error at or near "." Position: 19 at org.jooq_3.10.4.POSTGRES.debug(Unknown Source) at org.jooq.impl.Tools.translate(Tools.java:2239) at org.jooq.impl.DefaultExecuteContext.sqlException(DefaultExecuteContext.java:704) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:361) at jsl.utilities.reporting.JSLDatabase.dropJSLDbSchema(JSLDatabase.java:345) at jsl.utilities.reporting.JSLDatabase.createJSLDatabase(JSLDatabase.java:305) at jsl.utilities.reporting.JSLDatabase.createPostgresJSLDatabase(JSLDatabase.java:279) at jsl.utilities.reporting.JSLDatabase.createPostgresLocalHostJSLDatabase(JSLDatabase.java:253) at ex.queueing.DriverLicenseBureauWithQ.runExperiment(DriverLicenseBureauWithQ.java:264) at ex.queueing.DriverLicenseBureauWithQ.main(DriverLicenseBureauWithQ.java:207) Caused by: org.postgresql.util.PSQLException: ERROR: syntax error at or near "." Position: 19 at org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2433) at org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2178) at org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:306) at org.postgresql.jdbc.PgStatement.executeInternal(PgStatement.java:441) at org.postgresql.jdbc.PgStatement.execute(PgStatement.java:365) at org.postgresql.jdbc.PgPreparedStatement.executeWithFlags(PgPreparedStatement.java:155) at org.postgresql.jdbc.PgPreparedStatement.execute(PgPreparedStatement.java:144) at com.zaxxer.hikari.pool.ProxyPreparedStatement.execute(ProxyPreparedStatement.java:44) at com.zaxxer.hikari.pool.HikariProxyPreparedStatement.execute(HikariProxyPreparedStatement.java) at org.jooq.tools.jdbc.DefaultPreparedStatement.execute(DefaultPreparedStatement.java:209) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:429) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:347) ... 6 more ``` ---- See also: https://groups.google.com/forum/#!topic/jooq-user/tnMjwO-p8sM
1.0
Cannot use Meta.getSchema() for DROP SCHEMA statement - As reported on the jOOQ user group: ---- ```java Schema schema = // from a call to meta() and finding the right schema db.getDSLContext().dropSchema(schema.getName()).execute();// works db.getDSLContext().dropSchema(schema).execute(); // does not work, see exception below db.getDSLContext().execute("drop schema jsl_db"); // works ``` The name of my database is "test" and the schema is obviously "jsl_db" ``` Exception in thread "main" org.jooq.exception.DataAccessException: SQL [drop schema "test"."jsl_db"]; ERROR: syntax error at or near "." Position: 19 at org.jooq_3.10.4.POSTGRES.debug(Unknown Source) at org.jooq.impl.Tools.translate(Tools.java:2239) at org.jooq.impl.DefaultExecuteContext.sqlException(DefaultExecuteContext.java:704) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:361) at jsl.utilities.reporting.JSLDatabase.dropJSLDbSchema(JSLDatabase.java:345) at jsl.utilities.reporting.JSLDatabase.createJSLDatabase(JSLDatabase.java:305) at jsl.utilities.reporting.JSLDatabase.createPostgresJSLDatabase(JSLDatabase.java:279) at jsl.utilities.reporting.JSLDatabase.createPostgresLocalHostJSLDatabase(JSLDatabase.java:253) at ex.queueing.DriverLicenseBureauWithQ.runExperiment(DriverLicenseBureauWithQ.java:264) at ex.queueing.DriverLicenseBureauWithQ.main(DriverLicenseBureauWithQ.java:207) Caused by: org.postgresql.util.PSQLException: ERROR: syntax error at or near "." Position: 19 at org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2433) at org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2178) at org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:306) at org.postgresql.jdbc.PgStatement.executeInternal(PgStatement.java:441) at org.postgresql.jdbc.PgStatement.execute(PgStatement.java:365) at org.postgresql.jdbc.PgPreparedStatement.executeWithFlags(PgPreparedStatement.java:155) at org.postgresql.jdbc.PgPreparedStatement.execute(PgPreparedStatement.java:144) at com.zaxxer.hikari.pool.ProxyPreparedStatement.execute(ProxyPreparedStatement.java:44) at com.zaxxer.hikari.pool.HikariProxyPreparedStatement.execute(HikariProxyPreparedStatement.java) at org.jooq.tools.jdbc.DefaultPreparedStatement.execute(DefaultPreparedStatement.java:209) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:429) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:347) ... 6 more ``` ---- See also: https://groups.google.com/forum/#!topic/jooq-user/tnMjwO-p8sM
defect
cannot use meta getschema for drop schema statement as reported on the jooq user group java schema schema from a call to meta and finding the right schema db getdslcontext dropschema schema getname execute works db getdslcontext dropschema schema execute does not work see exception below db getdslcontext execute drop schema jsl db works the name of my database is test and the schema is obviously jsl db exception in thread main org jooq exception dataaccessexception sql error syntax error at or near position at org jooq postgres debug unknown source at org jooq impl tools translate tools java at org jooq impl defaultexecutecontext sqlexception defaultexecutecontext java at org jooq impl abstractquery execute abstractquery java at jsl utilities reporting jsldatabase dropjsldbschema jsldatabase java at jsl utilities reporting jsldatabase createjsldatabase jsldatabase java at jsl utilities reporting jsldatabase createpostgresjsldatabase jsldatabase java at jsl utilities reporting jsldatabase createpostgreslocalhostjsldatabase jsldatabase java at ex queueing driverlicensebureauwithq runexperiment driverlicensebureauwithq java at ex queueing driverlicensebureauwithq main driverlicensebureauwithq java caused by org postgresql util psqlexception error syntax error at or near position at org postgresql core queryexecutorimpl receiveerrorresponse queryexecutorimpl java at org postgresql core queryexecutorimpl processresults queryexecutorimpl java at org postgresql core queryexecutorimpl execute queryexecutorimpl java at org postgresql jdbc pgstatement executeinternal pgstatement java at org postgresql jdbc pgstatement execute pgstatement java at org postgresql jdbc pgpreparedstatement executewithflags pgpreparedstatement java at org postgresql jdbc pgpreparedstatement execute pgpreparedstatement java at com zaxxer hikari pool proxypreparedstatement execute proxypreparedstatement java at com zaxxer hikari pool hikariproxypreparedstatement execute hikariproxypreparedstatement java at org jooq tools jdbc defaultpreparedstatement execute defaultpreparedstatement java at org jooq impl abstractquery execute abstractquery java at org jooq impl abstractquery execute abstractquery java more see also
1
47,971
13,067,345,387
IssuesEvent
2020-07-31 00:09:55
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
cmake - `make deploy-docs` goes to the wrong destination (Trac #1548)
Migrated from Trac cmake defect
goes to "METAPROJECT_release" should just go to "METAPROJECT" Migrated from https://code.icecube.wisc.edu/ticket/1548 ```json { "status": "closed", "changetime": "2016-04-21T18:46:34", "description": "goes to \"METAPROJECT_release\"\nshould just go to \"METAPROJECT\"", "reporter": "nega", "cc": "", "resolution": "fixed", "_ts": "1461264394191468", "component": "cmake", "summary": "cmake - `make deploy-docs` goes to the wrong destination", "priority": "normal", "keywords": "docs website", "time": "2016-02-12T22:53:11", "milestone": "", "owner": "nega", "type": "defect" } ```
1.0
cmake - `make deploy-docs` goes to the wrong destination (Trac #1548) - goes to "METAPROJECT_release" should just go to "METAPROJECT" Migrated from https://code.icecube.wisc.edu/ticket/1548 ```json { "status": "closed", "changetime": "2016-04-21T18:46:34", "description": "goes to \"METAPROJECT_release\"\nshould just go to \"METAPROJECT\"", "reporter": "nega", "cc": "", "resolution": "fixed", "_ts": "1461264394191468", "component": "cmake", "summary": "cmake - `make deploy-docs` goes to the wrong destination", "priority": "normal", "keywords": "docs website", "time": "2016-02-12T22:53:11", "milestone": "", "owner": "nega", "type": "defect" } ```
defect
cmake make deploy docs goes to the wrong destination trac goes to metaproject release should just go to metaproject migrated from json status closed changetime description goes to metaproject release nshould just go to metaproject reporter nega cc resolution fixed ts component cmake summary cmake make deploy docs goes to the wrong destination priority normal keywords docs website time milestone owner nega type defect
1
7,151
2,610,329,423
IssuesEvent
2015-02-26 19:46:08
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
closed
Mechis III Typo
auto-migrated Priority-Medium Type-Defect
``` In its description there is the following mistake: "Cost of Munificent frigates is reduced 15% due to Mechis III's massive automated fondrys." Should be: "Cost of Munificent frigates is reduced by 15% due to Mechis III's massive automated foundries." ``` ----- Original issue reported on code.google.com by `jkouzman...@gmail.com` on 14 Jun 2011 at 6:37
1.0
Mechis III Typo - ``` In its description there is the following mistake: "Cost of Munificent frigates is reduced 15% due to Mechis III's massive automated fondrys." Should be: "Cost of Munificent frigates is reduced by 15% due to Mechis III's massive automated foundries." ``` ----- Original issue reported on code.google.com by `jkouzman...@gmail.com` on 14 Jun 2011 at 6:37
defect
mechis iii typo in its description there is the following mistake cost of munificent frigates is reduced due to mechis iii s massive automated fondrys should be cost of munificent frigates is reduced by due to mechis iii s massive automated foundries original issue reported on code google com by jkouzman gmail com on jun at
1
114,354
9,702,231,347
IssuesEvent
2019-05-27 08:13:43
kyma-project/kyma
https://api.github.com/repos/kyma-project/kyma
closed
Knative Test is Failing
area/serverless bug test-failing
Knative test is failing. - [x] get pods to spin up using curl ```curl istio-ingressgateway.istio-system.svc.cluster.local -H "Host: test-service.knative-serving.kyma.local"``` ```curl test-service.knative-serving.svc.cluster.local``` - [x] request less cpu shares so that the container startup is more reliable
1.0
Knative Test is Failing - Knative test is failing. - [x] get pods to spin up using curl ```curl istio-ingressgateway.istio-system.svc.cluster.local -H "Host: test-service.knative-serving.kyma.local"``` ```curl test-service.knative-serving.svc.cluster.local``` - [x] request less cpu shares so that the container startup is more reliable
non_defect
knative test is failing knative test is failing get pods to spin up using curl curl istio ingressgateway istio system svc cluster local h host test service knative serving kyma local curl test service knative serving svc cluster local request less cpu shares so that the container startup is more reliable
0
163,543
20,363,865,049
IssuesEvent
2022-02-21 01:38:02
dreamboy9/mongo
https://api.github.com/repos/dreamboy9/mongo
opened
CVE-2022-0512 (High) detected in url-parse-1.5.1.tgz
security vulnerability
## CVE-2022-0512 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.5.1.tgz</b></p></summary> <p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p> <p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz</a></p> <p>Path to dependency file: /buildscripts/libdeps/graph_visualizer_web_stack/package.json</p> <p>Path to vulnerable library: /buildscripts/libdeps/graph_visualizer_web_stack/node_modules/url-parse/package.json</p> <p> Dependency Hierarchy: - react-scripts-4.0.3.tgz (Root Library) - webpack-dev-server-3.11.1.tgz - sockjs-client-1.5.1.tgz - :x: **url-parse-1.5.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.6. <p>Publish Date: 2022-02-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0512>CVE-2022-0512</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0512">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0512</a></p> <p>Release Date: 2022-02-14</p> <p>Fix Resolution: url-parse - 1.5.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-0512 (High) detected in url-parse-1.5.1.tgz - ## CVE-2022-0512 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.5.1.tgz</b></p></summary> <p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p> <p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.5.1.tgz</a></p> <p>Path to dependency file: /buildscripts/libdeps/graph_visualizer_web_stack/package.json</p> <p>Path to vulnerable library: /buildscripts/libdeps/graph_visualizer_web_stack/node_modules/url-parse/package.json</p> <p> Dependency Hierarchy: - react-scripts-4.0.3.tgz (Root Library) - webpack-dev-server-3.11.1.tgz - sockjs-client-1.5.1.tgz - :x: **url-parse-1.5.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.6. <p>Publish Date: 2022-02-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0512>CVE-2022-0512</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0512">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0512</a></p> <p>Release Date: 2022-02-14</p> <p>Fix Resolution: url-parse - 1.5.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in url parse tgz cve high severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file buildscripts libdeps graph visualizer web stack package json path to vulnerable library buildscripts libdeps graph visualizer web stack node modules url parse package json dependency hierarchy react scripts tgz root library webpack dev server tgz sockjs client tgz x url parse tgz vulnerable library found in base branch master vulnerability details authorization bypass through user controlled key in npm url parse prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse step up your open source security game with whitesource
0
57,872
16,112,345,801
IssuesEvent
2021-04-27 23:47:25
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
zfs bulld failure on ppc64el on Linux 5.12
Status: Triage Needed Type: Defect
<## System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Ubuntu Distribution Version | 21.10 (development) Linux Kernel | 5.12 Architecture | ppc64el ZFS Version | 2.0.4 SPL Version | 2.0.4 ### Describe the problem you're observing ERROR: modpost: GPL-incompatible module zfs.ko uses GPL-only symbol 'mmu_feature_keys' ppc64el linux commit 7613f5a66becfd0e43a0f34de8518695888f5458 "powerpc/64s/kuap: Use mmu_has_feature()" introduced a change to use mmu_has_feature instead of early_mmu_has_feature. This unfortunately means that get_user() in zfs.ko (and others) will indirectly use the inline'd functions allow_read_from_user() and indirectly mmu_has_feature() and the latter accesses a GPL'd symbol mmu_feature_keys[] which causes the build failure. I'll reach out to Michael Ellerman and Christophe Leroy on Wednesday to see if this can be fixed with the GPL'd symbol as this is rather restrictive now on get_user() callers.
1.0
zfs bulld failure on ppc64el on Linux 5.12 - <## System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Ubuntu Distribution Version | 21.10 (development) Linux Kernel | 5.12 Architecture | ppc64el ZFS Version | 2.0.4 SPL Version | 2.0.4 ### Describe the problem you're observing ERROR: modpost: GPL-incompatible module zfs.ko uses GPL-only symbol 'mmu_feature_keys' ppc64el linux commit 7613f5a66becfd0e43a0f34de8518695888f5458 "powerpc/64s/kuap: Use mmu_has_feature()" introduced a change to use mmu_has_feature instead of early_mmu_has_feature. This unfortunately means that get_user() in zfs.ko (and others) will indirectly use the inline'd functions allow_read_from_user() and indirectly mmu_has_feature() and the latter accesses a GPL'd symbol mmu_feature_keys[] which causes the build failure. I'll reach out to Michael Ellerman and Christophe Leroy on Wednesday to see if this can be fixed with the GPL'd symbol as this is rather restrictive now on get_user() callers.
defect
zfs bulld failure on on linux system information type version name distribution name ubuntu distribution version development linux kernel architecture zfs version spl version describe the problem you re observing error modpost gpl incompatible module zfs ko uses gpl only symbol mmu feature keys linux commit powerpc kuap use mmu has feature introduced a change to use mmu has feature instead of early mmu has feature this unfortunately means that get user in zfs ko and others will indirectly use the inline d functions allow read from user and indirectly mmu has feature and the latter accesses a gpl d symbol mmu feature keys which causes the build failure i ll reach out to michael ellerman and christophe leroy on wednesday to see if this can be fixed with the gpl d symbol as this is rather restrictive now on get user callers
1
75,438
25,841,408,878
IssuesEvent
2022-12-13 00:51:42
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
FE | Profile | Accessibility | Focus Area
508/Accessibility vsa authenticated-experience needs-grooming profile 508-defect-2 sprint-planning needs-refinement
## Background We received feedback from a11y team that when editing in Profile the in-line error message doesn't focus on the update button. Recommendation is to include descriptive error messages for screen readers and other users to help them determine what needs to be corrected. ## Tasks - [ ] ## Acceptance Criteria - [ ] ## Validation - Who can validate this ticket? (FE, BE, Design, PM)? - How can this work be validated? - What updates need to be made (e.g. product outline, use cases, contact center guide) ## Associated Tickets - #46731 ## How to configure this issue - Include practice area in title -- e.g., [Design], [FE], [BE] - Include project/initiative name - e.g., Auth Profile: Address Change Messaging - Add label for practice area (frontend, backend, design) - Add label for project if applicable - Assign to a Profile team member - Associate with an Epic if applicable - If creating for yourself: estimate work if enough info exists to do so
1.0
FE | Profile | Accessibility | Focus Area - ## Background We received feedback from a11y team that when editing in Profile the in-line error message doesn't focus on the update button. Recommendation is to include descriptive error messages for screen readers and other users to help them determine what needs to be corrected. ## Tasks - [ ] ## Acceptance Criteria - [ ] ## Validation - Who can validate this ticket? (FE, BE, Design, PM)? - How can this work be validated? - What updates need to be made (e.g. product outline, use cases, contact center guide) ## Associated Tickets - #46731 ## How to configure this issue - Include practice area in title -- e.g., [Design], [FE], [BE] - Include project/initiative name - e.g., Auth Profile: Address Change Messaging - Add label for practice area (frontend, backend, design) - Add label for project if applicable - Assign to a Profile team member - Associate with an Epic if applicable - If creating for yourself: estimate work if enough info exists to do so
defect
fe profile accessibility focus area background we received feedback from team that when editing in profile the in line error message doesn t focus on the update button recommendation is to include descriptive error messages for screen readers and other users to help them determine what needs to be corrected tasks acceptance criteria validation who can validate this ticket fe be design pm how can this work be validated what updates need to be made e g product outline use cases contact center guide associated tickets how to configure this issue include practice area in title e g include project initiative name e g auth profile address change messaging add label for practice area frontend backend design add label for project if applicable assign to a profile team member associate with an epic if applicable if creating for yourself estimate work if enough info exists to do so
1
27,477
5,030,442,118
IssuesEvent
2016-12-16 00:49:18
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
closed
i18n - Specified key was too long
Defect
This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: 3.3.10. * Platform and Target: Linux with MySQL version: 5.1.45p1-log - PHP Version 5.5.17 - ### What you did I run i18n.sql ```sql CREATE TABLE i18n ( id int NOT NULL auto_increment, locale varchar(6) NOT NULL, model varchar(255) NOT NULL, foreign_key int(10) NOT NULL, field varchar(255) NOT NULL, content text, PRIMARY KEY (id), UNIQUE INDEX I18N_LOCALE_FIELD(locale, model, foreign_key, field), INDEX I18N_FIELD(model, foreign_key, field) ) ``` ### What happened I got this error ```Error in query (1071): Specified key was too long; max key length is 1000 bytes``` ### What you expected to happen I expect that the i18n table should be created.
1.0
i18n - Specified key was too long - This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: 3.3.10. * Platform and Target: Linux with MySQL version: 5.1.45p1-log - PHP Version 5.5.17 - ### What you did I run i18n.sql ```sql CREATE TABLE i18n ( id int NOT NULL auto_increment, locale varchar(6) NOT NULL, model varchar(255) NOT NULL, foreign_key int(10) NOT NULL, field varchar(255) NOT NULL, content text, PRIMARY KEY (id), UNIQUE INDEX I18N_LOCALE_FIELD(locale, model, foreign_key, field), INDEX I18N_FIELD(model, foreign_key, field) ) ``` ### What happened I got this error ```Error in query (1071): Specified key was too long; max key length is 1000 bytes``` ### What you expected to happen I expect that the i18n table should be created.
defect
specified key was too long this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target linux with mysql version log php version what you did i run sql sql create table id int not null auto increment locale varchar not null model varchar not null foreign key int not null field varchar not null content text primary key id unique index locale field locale model foreign key field index field model foreign key field what happened i got this error error in query specified key was too long max key length is bytes what you expected to happen i expect that the table should be created
1
66,848
20,731,280,396
IssuesEvent
2022-03-14 09:40:54
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Poll votes different on pin message sidebar and on main room timeline
T-Defect S-Major A-Message-Pinning O-Occasional A-Polls
### Steps to reproduce 1. Pin a poll 2. Have people vote and then come back to the room later ### Outcome <img width="1138" alt="Screen Shot 2022-03-14 at 09 37 54" src="https://user-images.githubusercontent.com/769871/158145984-aad0e9d6-1f2f-4361-b450-a7fd5ea181a2.png"> ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp _No response_ ### Application version _No response_ ### Homeserver _No response_ ### Will you send logs? No
1.0
Poll votes different on pin message sidebar and on main room timeline - ### Steps to reproduce 1. Pin a poll 2. Have people vote and then come back to the room later ### Outcome <img width="1138" alt="Screen Shot 2022-03-14 at 09 37 54" src="https://user-images.githubusercontent.com/769871/158145984-aad0e9d6-1f2f-4361-b450-a7fd5ea181a2.png"> ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp _No response_ ### Application version _No response_ ### Homeserver _No response_ ### Will you send logs? No
defect
poll votes different on pin message sidebar and on main room timeline steps to reproduce pin a poll have people vote and then come back to the room later outcome img width alt screen shot at src operating system no response browser information no response url for webapp no response application version no response homeserver no response will you send logs no
1
32,518
6,817,705,903
IssuesEvent
2017-11-07 00:50:14
tpfinal-pp1/tp-final
https://api.github.com/repos/tpfinal-pp1/tp-final
closed
Ventana Matching Personas muestra Botones de edicion y demas
Defecto medio Liberado por desarrollo
En especial el de Matching inmuebles haciendo que entonces se pueda hacer un loop infinito
1.0
Ventana Matching Personas muestra Botones de edicion y demas - En especial el de Matching inmuebles haciendo que entonces se pueda hacer un loop infinito
defect
ventana matching personas muestra botones de edicion y demas en especial el de matching inmuebles haciendo que entonces se pueda hacer un loop infinito
1
51,273
12,695,945,567
IssuesEvent
2020-06-22 09:15:08
pwa-builder/PWABuilder
https://api.github.com/repos/pwa-builder/PWABuilder
opened
[Keyboard Navigation - PWA Builder - PWA Builder Generate] : Focus order not logical for upload dialogue box
A11yCT A11yMAS A11yMediumImpact Accessibility HCL- PWABuilder MAS2.4.3 Severity3 bug :bug: needs triage :mag:
**User Experience:** The user has to navigate through all elements present on the page to get the focus on the dialogue box controls. **Test Environment:** OS: Windows 10 build 19608.1006 Browser: Edge - Anaheim - Version 85.0.545.0 (Official build) dev (64-bit) URL: https://preview.pwabuilder.com/generate **Repro Steps:** 1. Open URL: https://preview.pwabuilder.com/generate in Edge Anaheim dev browser. 2. Navigate to Images Tabbed button 3. Now press tab until focus reaches Upload Button and press enter key 4. Now press tab and try to navigate to first element of the dialogue box opened 5. Observe the issue. **Actual Result:** The whole page has to be navigated to get the focus on the dialogue box opened on clicking upload button **Expected Result:** The focus should travel directly to the first element present on the dialogue box. **MAS Reference:** https://microsoft.sharepoint.com/:w:/r/teams/msenable/_layouts/15/Doc.aspx?sourcedoc=%7B684573A7-B089-4131-9B39-0009054125D3%7D&file=MAS%204.1.3%20%E2%80%93%20Status%20Messages.docx&action=default&mobileredirect=true [2.4.3 Focus Order Upload Dialogue Box.zip](https://github.com/pwa-builder/PWABuilder/files/4812297/2.4.3.Focus.Order.Upload.Dialogue.Box.zip) ![PWABuilder - Personal - Microsoft​ Edge 6_22_2020 2_40_52 PM](https://user-images.githubusercontent.com/66783558/85270366-cddb0b00-b496-11ea-9dcf-865c1facde21.png)
1.0
[Keyboard Navigation - PWA Builder - PWA Builder Generate] : Focus order not logical for upload dialogue box - **User Experience:** The user has to navigate through all elements present on the page to get the focus on the dialogue box controls. **Test Environment:** OS: Windows 10 build 19608.1006 Browser: Edge - Anaheim - Version 85.0.545.0 (Official build) dev (64-bit) URL: https://preview.pwabuilder.com/generate **Repro Steps:** 1. Open URL: https://preview.pwabuilder.com/generate in Edge Anaheim dev browser. 2. Navigate to Images Tabbed button 3. Now press tab until focus reaches Upload Button and press enter key 4. Now press tab and try to navigate to first element of the dialogue box opened 5. Observe the issue. **Actual Result:** The whole page has to be navigated to get the focus on the dialogue box opened on clicking upload button **Expected Result:** The focus should travel directly to the first element present on the dialogue box. **MAS Reference:** https://microsoft.sharepoint.com/:w:/r/teams/msenable/_layouts/15/Doc.aspx?sourcedoc=%7B684573A7-B089-4131-9B39-0009054125D3%7D&file=MAS%204.1.3%20%E2%80%93%20Status%20Messages.docx&action=default&mobileredirect=true [2.4.3 Focus Order Upload Dialogue Box.zip](https://github.com/pwa-builder/PWABuilder/files/4812297/2.4.3.Focus.Order.Upload.Dialogue.Box.zip) ![PWABuilder - Personal - Microsoft​ Edge 6_22_2020 2_40_52 PM](https://user-images.githubusercontent.com/66783558/85270366-cddb0b00-b496-11ea-9dcf-865c1facde21.png)
non_defect
focus order not logical for upload dialogue box user experience the user has to navigate through all elements present on the page to get the focus on the dialogue box controls test environment os windows build browser edge anaheim version official build dev bit url repro steps open url in edge anaheim dev browser navigate to images tabbed button now press tab until focus reaches upload button and press enter key now press tab and try to navigate to first element of the dialogue box opened observe the issue actual result the whole page has to be navigated to get the focus on the dialogue box opened on clicking upload button expected result the focus should travel directly to the first element present on the dialogue box mas reference
0
68,730
21,803,302,701
IssuesEvent
2022-05-16 08:01:42
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Starting a thread on an edited message will render a wrong thread summary (non-persistent)
T-Defect S-Minor O-Occasional A-Threads
### Steps to reproduce 1. Click on "Reply in thread" on an edited message. ### Outcome #### What did you expect? The thread summary to not be rendered until I post the reply. #### What happened instead? A thread summary is generated as soon as I clicked on "Reply in thread" displaying the contents of the root message. ![Element 5 | 🧪 Playground Non-encrypted](https://user-images.githubusercontent.com/60790/168100850-7f9cf134-7ec8-4b27-ab6d-f0e5f57456c5.png) ### Operating system macOS ### Browser information Chrome 101 ### URL for webapp develop.element.io ### Application version Element version: 479d4bf64d97-react-fbb8581f779e-js-34cfa511049e Olm version: 3.2.8 ### Homeserver matrix.org ### Will you send logs? No
1.0
Starting a thread on an edited message will render a wrong thread summary (non-persistent) - ### Steps to reproduce 1. Click on "Reply in thread" on an edited message. ### Outcome #### What did you expect? The thread summary to not be rendered until I post the reply. #### What happened instead? A thread summary is generated as soon as I clicked on "Reply in thread" displaying the contents of the root message. ![Element 5 | 🧪 Playground Non-encrypted](https://user-images.githubusercontent.com/60790/168100850-7f9cf134-7ec8-4b27-ab6d-f0e5f57456c5.png) ### Operating system macOS ### Browser information Chrome 101 ### URL for webapp develop.element.io ### Application version Element version: 479d4bf64d97-react-fbb8581f779e-js-34cfa511049e Olm version: 3.2.8 ### Homeserver matrix.org ### Will you send logs? No
defect
starting a thread on an edited message will render a wrong thread summary non persistent steps to reproduce click on reply in thread on an edited message outcome what did you expect the thread summary to not be rendered until i post the reply what happened instead a thread summary is generated as soon as i clicked on reply in thread displaying the contents of the root message operating system macos browser information chrome url for webapp develop element io application version element version react js olm version homeserver matrix org will you send logs no
1
92,195
3,868,754,545
IssuesEvent
2016-04-10 05:33:54
cs2103jan2016-t16-4j/main
https://api.github.com/repos/cs2103jan2016-t16-4j/main
closed
A user can leave out obvious fields which we can fill by default
priority.medium type.enhancement type.story
So that it is easier and faster for the user to input tasks.
1.0
A user can leave out obvious fields which we can fill by default - So that it is easier and faster for the user to input tasks.
non_defect
a user can leave out obvious fields which we can fill by default so that it is easier and faster for the user to input tasks
0
249,577
26,954,452,180
IssuesEvent
2023-02-08 14:02:09
simplycubed/terraform-google-static-assets
https://api.github.com/repos/simplycubed/terraform-google-static-assets
closed
CVE-2022-1996 (High) detected in github.com/emicklei/go-restful/v3-v3.8.0 - autoclosed
security vulnerability
## CVE-2022-1996 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/emicklei/go-restful/v3-v3.8.0</b></p></summary> <p>package for building REST-style Web Services using Go</p> <p>Library home page: <a href="https://proxy.golang.org/github.com/emicklei/go-restful/v3/@v/v3.8.0.zip">https://proxy.golang.org/github.com/emicklei/go-restful/v3/@v/v3.8.0.zip</a></p> <p> Dependency Hierarchy: - github.com/gruntwork-io/terratest-v0.40.17 (Root Library) - k8s.io/client-go-v0.24.2 - k8s.io/kube-openapi - :x: **github.com/emicklei/go-restful/v3-v3.8.0** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/simplycubed/terraform-google-static-assets/commit/e49e2f33b77657ce4ab7eac9abebafc4a1fd18ba">e49e2f33b77657ce4ab7eac9abebafc4a1fd18ba</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Authorization Bypass Through User-Controlled Key in GitHub repository emicklei/go-restful prior to v3.8.0. <p>Publish Date: 2022-06-08 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1996>CVE-2022-1996</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-1996">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-1996</a></p> <p>Release Date: 2022-06-08</p> <p>Fix Resolution: v3.8.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-1996 (High) detected in github.com/emicklei/go-restful/v3-v3.8.0 - autoclosed - ## CVE-2022-1996 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/emicklei/go-restful/v3-v3.8.0</b></p></summary> <p>package for building REST-style Web Services using Go</p> <p>Library home page: <a href="https://proxy.golang.org/github.com/emicklei/go-restful/v3/@v/v3.8.0.zip">https://proxy.golang.org/github.com/emicklei/go-restful/v3/@v/v3.8.0.zip</a></p> <p> Dependency Hierarchy: - github.com/gruntwork-io/terratest-v0.40.17 (Root Library) - k8s.io/client-go-v0.24.2 - k8s.io/kube-openapi - :x: **github.com/emicklei/go-restful/v3-v3.8.0** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/simplycubed/terraform-google-static-assets/commit/e49e2f33b77657ce4ab7eac9abebafc4a1fd18ba">e49e2f33b77657ce4ab7eac9abebafc4a1fd18ba</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Authorization Bypass Through User-Controlled Key in GitHub repository emicklei/go-restful prior to v3.8.0. <p>Publish Date: 2022-06-08 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1996>CVE-2022-1996</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-1996">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-1996</a></p> <p>Release Date: 2022-06-08</p> <p>Fix Resolution: v3.8.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in github com emicklei go restful autoclosed cve high severity vulnerability vulnerable library github com emicklei go restful package for building rest style web services using go library home page a href dependency hierarchy github com gruntwork io terratest root library io client go io kube openapi x github com emicklei go restful vulnerable library found in head commit a href found in base branch master vulnerability details authorization bypass through user controlled key in github repository emicklei go restful prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
55,158
7,964,069,987
IssuesEvent
2018-07-13 19:58:12
AzureAD/azure-activedirectory-library-for-android
https://api.github.com/repos/AzureAD/azure-activedirectory-library-for-android
closed
DEVELOPER_REDIRECTURI_INVALID
Documentation In Progress P2
I am trying the Quick Start.... ./gradlew installDebug 'hello' did not show up on the android device, but 'Fancy ADAL Test App' did. When I try to 'AQUIRETOKEN' I get: `03-09 16:23:30.725 5846 5914 E SampleTelemetry: Microsoft.ADAL.request_id:910c1e6b-4a3e-4d78-b08e-a4c5c1a54f20 03-09 16:23:30.725 5846 5914 E SampleTelemetry: Microsoft.ADAL.api_error_code:DEVELOPER_REDIRECTURI_INVALID 03-09 16:23:30.725 5846 5914 E SampleTelemetry: Microsoft.ADAL.device_id:akubWTkC8z0Ff/mWJ+plvTXGhFkhtbUi1cg5TjiQSIU= 03-09 16:23:30.725 5846 5914 E SampleTelemetry: Microsoft.ADAL.api_id:115 03-09 16:23:30.725 5846 5914 E SampleTelemetry: Microsoft.ADAL.authority_type:aad 03-09 16:23:30.725 5846 5914 E SampleTelemetry: Microsoft.ADAL.application_name:com.microsoft.aad.adal.userappwithbroker 03-09 16:23:30.725 5846 5914 E SampleTelemetry: Microsoft.ADAL.application_version:1.0 03-09 16:23:30.726 5846 5914 E SampleTelemetry: Microsoft.ADAL.authority_validation_status:yes 03-09 16:23:30.726 5846 5914 E SampleTelemetry: Microsoft.ADAL.response_time:28 03-09 16:23:30.726 5846 5914 E SampleTelemetry: Microsoft.ADAL.client_id:b92e0ba5-f86e-4411-8e18-6b5f928d968a 03-09 16:23:30.726 5846 5914 E SampleTelemetry: Microsoft.ADAL.is_successful:false 03-09 16:23:30.726 5846 5914 E SampleTelemetry: Microsoft.ADAL.correlation_id:e0ce9485-1df2-4923-961d-3c403b00d5f3 03-09 16:23:30.726 5846 5914 E SampleTelemetry: Microsoft.ADAL.prompt_behavior:Auto 03-09 16:23:30.727 5846 5913 V UserAppWithBroker.Main: MainActivity userapp:The base64 url encoded signature component of the redirect uri does not match the expected value. ` The android device is currently logged in with Outlook.....
1.0
DEVELOPER_REDIRECTURI_INVALID - I am trying the Quick Start.... ./gradlew installDebug 'hello' did not show up on the android device, but 'Fancy ADAL Test App' did. When I try to 'AQUIRETOKEN' I get: `03-09 16:23:30.725 5846 5914 E SampleTelemetry: Microsoft.ADAL.request_id:910c1e6b-4a3e-4d78-b08e-a4c5c1a54f20 03-09 16:23:30.725 5846 5914 E SampleTelemetry: Microsoft.ADAL.api_error_code:DEVELOPER_REDIRECTURI_INVALID 03-09 16:23:30.725 5846 5914 E SampleTelemetry: Microsoft.ADAL.device_id:akubWTkC8z0Ff/mWJ+plvTXGhFkhtbUi1cg5TjiQSIU= 03-09 16:23:30.725 5846 5914 E SampleTelemetry: Microsoft.ADAL.api_id:115 03-09 16:23:30.725 5846 5914 E SampleTelemetry: Microsoft.ADAL.authority_type:aad 03-09 16:23:30.725 5846 5914 E SampleTelemetry: Microsoft.ADAL.application_name:com.microsoft.aad.adal.userappwithbroker 03-09 16:23:30.725 5846 5914 E SampleTelemetry: Microsoft.ADAL.application_version:1.0 03-09 16:23:30.726 5846 5914 E SampleTelemetry: Microsoft.ADAL.authority_validation_status:yes 03-09 16:23:30.726 5846 5914 E SampleTelemetry: Microsoft.ADAL.response_time:28 03-09 16:23:30.726 5846 5914 E SampleTelemetry: Microsoft.ADAL.client_id:b92e0ba5-f86e-4411-8e18-6b5f928d968a 03-09 16:23:30.726 5846 5914 E SampleTelemetry: Microsoft.ADAL.is_successful:false 03-09 16:23:30.726 5846 5914 E SampleTelemetry: Microsoft.ADAL.correlation_id:e0ce9485-1df2-4923-961d-3c403b00d5f3 03-09 16:23:30.726 5846 5914 E SampleTelemetry: Microsoft.ADAL.prompt_behavior:Auto 03-09 16:23:30.727 5846 5913 V UserAppWithBroker.Main: MainActivity userapp:The base64 url encoded signature component of the redirect uri does not match the expected value. ` The android device is currently logged in with Outlook.....
non_defect
developer redirecturi invalid i am trying the quick start gradlew installdebug hello did not show up on the android device but fancy adal test app did when i try to aquiretoken i get e sampletelemetry microsoft adal request id e sampletelemetry microsoft adal api error code developer redirecturi invalid e sampletelemetry microsoft adal device id mwj e sampletelemetry microsoft adal api id e sampletelemetry microsoft adal authority type aad e sampletelemetry microsoft adal application name com microsoft aad adal userappwithbroker e sampletelemetry microsoft adal application version e sampletelemetry microsoft adal authority validation status yes e sampletelemetry microsoft adal response time e sampletelemetry microsoft adal client id e sampletelemetry microsoft adal is successful false e sampletelemetry microsoft adal correlation id e sampletelemetry microsoft adal prompt behavior auto v userappwithbroker main mainactivity userapp the url encoded signature component of the redirect uri does not match the expected value the android device is currently logged in with outlook
0
367,086
25,718,592,212
IssuesEvent
2022-12-07 12:08:26
icflorescu/mantine-datatable
https://api.github.com/repos/icflorescu/mantine-datatable
closed
is there a way to add an HTML attribute for every record ❓
documentation enhancement good first issue
is there a way to add an HTML tag for every record, for example I want to add data-cy="equipment-item" to every item rendered by the DataTable component so that I can get it in cypress without having to set the render function and return custom component for every column object
1.0
is there a way to add an HTML attribute for every record ❓ - is there a way to add an HTML tag for every record, for example I want to add data-cy="equipment-item" to every item rendered by the DataTable component so that I can get it in cypress without having to set the render function and return custom component for every column object
non_defect
is there a way to add an html attribute for every record ❓ is there a way to add an html tag for every record for example i want to add data cy equipment item to every item rendered by the datatable component so that i can get it in cypress without having to set the render function and return custom component for every column object
0
57,647
15,892,720,193
IssuesEvent
2021-04-11 01:37:16
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
closed
RFC: use full path by default as reference in .pot files when using I18n extract
defect
This is a (multiple allowed): * [x] feature-discussion (RFC) * CakePHP Version: 4.2.5 ### What you did ```bash bin/cake i18n extract (multiple directories) ``` ### What happened When extracting multiple folders like vendor, templates, src, etc, the references are not clear: ```gettext #: Error/error400.php:39 msgid "The requested address {0} was not found on this server." msgstr "" #: Error/ExceptionRenderer.php:340 msgid "An Internal Error Has Occurred." msgstr "" ``` With the current behavior, the references start with Error/ but one is in `ROOT/template/Error` and the other is inside `ROOT/vendor/.../src/Error` ### What you expected to happen To show paths from ROOT: Applying my changes would show them this way: ```gettext #: templates/Error/error400.php:39 msgid "The requested address {0} was not found on this server." msgstr "" #: vendor/cakephp/cakephp/src/Error/ExceptionRenderer.php:340 msgid "An Internal Error Has Occurred." msgstr "" ``` I know changing this would change a LOT of lines _.pot_ and _.po_ files first time the user uses _extract command_ but I think it's way cleaner this way. ## Note: ***I found this while writing this RFC** but I still think there may be something we could do. The default behaviour of the command is supposed to show full paths instead of relative paths. It is actually showing neither the full nor the relative ones. **So what is the expected behaviour?** I say this because there's an option called `relative-paths` that can be passed to the to the _extract command_. The DocBlock says: ```php /** * Use relative paths in the pot files rather than full path * * @var bool */ protected $_relativePaths = false; ``` When this option is true, the paths will be like this: ```gettext #: ./templates/Error/error400.php:39 msgid "The requested address {0} was not found on this server." msgstr "" #: ./vendor/cakephp/cakephp/src/Error/ExceptionRenderer.php:340 msgid "An Internal Error Has Occurred." msgstr "" ``` Which is almost the same result I want to get. I don't know if I have expressed myself correctly... **At this point I do not know if the actual behavior is the expected one**. I'm ready to send a PR which, by default, shows use the "full" paths like in the second example but since it does quite the same as `relative-paths` option does, I don't know if we should just change the description of the command and some DocBlocks to state that the default behavior does not use the full path. **What do you think?** If we do not make any changes, there will still be, in my opinion, confusion in the routes when no `relative-paths` option is passed
1.0
RFC: use full path by default as reference in .pot files when using I18n extract - This is a (multiple allowed): * [x] feature-discussion (RFC) * CakePHP Version: 4.2.5 ### What you did ```bash bin/cake i18n extract (multiple directories) ``` ### What happened When extracting multiple folders like vendor, templates, src, etc, the references are not clear: ```gettext #: Error/error400.php:39 msgid "The requested address {0} was not found on this server." msgstr "" #: Error/ExceptionRenderer.php:340 msgid "An Internal Error Has Occurred." msgstr "" ``` With the current behavior, the references start with Error/ but one is in `ROOT/template/Error` and the other is inside `ROOT/vendor/.../src/Error` ### What you expected to happen To show paths from ROOT: Applying my changes would show them this way: ```gettext #: templates/Error/error400.php:39 msgid "The requested address {0} was not found on this server." msgstr "" #: vendor/cakephp/cakephp/src/Error/ExceptionRenderer.php:340 msgid "An Internal Error Has Occurred." msgstr "" ``` I know changing this would change a LOT of lines _.pot_ and _.po_ files first time the user uses _extract command_ but I think it's way cleaner this way. ## Note: ***I found this while writing this RFC** but I still think there may be something we could do. The default behaviour of the command is supposed to show full paths instead of relative paths. It is actually showing neither the full nor the relative ones. **So what is the expected behaviour?** I say this because there's an option called `relative-paths` that can be passed to the to the _extract command_. The DocBlock says: ```php /** * Use relative paths in the pot files rather than full path * * @var bool */ protected $_relativePaths = false; ``` When this option is true, the paths will be like this: ```gettext #: ./templates/Error/error400.php:39 msgid "The requested address {0} was not found on this server." msgstr "" #: ./vendor/cakephp/cakephp/src/Error/ExceptionRenderer.php:340 msgid "An Internal Error Has Occurred." msgstr "" ``` Which is almost the same result I want to get. I don't know if I have expressed myself correctly... **At this point I do not know if the actual behavior is the expected one**. I'm ready to send a PR which, by default, shows use the "full" paths like in the second example but since it does quite the same as `relative-paths` option does, I don't know if we should just change the description of the command and some DocBlocks to state that the default behavior does not use the full path. **What do you think?** If we do not make any changes, there will still be, in my opinion, confusion in the routes when no `relative-paths` option is passed
defect
rfc use full path by default as reference in pot files when using extract this is a multiple allowed feature discussion rfc cakephp version what you did bash bin cake extract multiple directories what happened when extracting multiple folders like vendor templates src etc the references are not clear gettext error php msgid the requested address was not found on this server msgstr error exceptionrenderer php msgid an internal error has occurred msgstr with the current behavior the references start with error but one is in root template error and the other is inside root vendor src error what you expected to happen to show paths from root applying my changes would show them this way gettext templates error php msgid the requested address was not found on this server msgstr vendor cakephp cakephp src error exceptionrenderer php msgid an internal error has occurred msgstr i know changing this would change a lot of lines pot and po files first time the user uses extract command but i think it s way cleaner this way note i found this while writing this rfc but i still think there may be something we could do the default behaviour of the command is supposed to show full paths instead of relative paths it is actually showing neither the full nor the relative ones so what is the expected behaviour i say this because there s an option called relative paths that can be passed to the to the extract command the docblock says php use relative paths in the pot files rather than full path var bool protected relativepaths false when this option is true the paths will be like this gettext templates error php msgid the requested address was not found on this server msgstr vendor cakephp cakephp src error exceptionrenderer php msgid an internal error has occurred msgstr which is almost the same result i want to get i don t know if i have expressed myself correctly at this point i do not know if the actual behavior is the expected one i m ready to send a pr which by default shows use the full paths like in the second example but since it does quite the same as relative paths option does i don t know if we should just change the description of the command and some docblocks to state that the default behavior does not use the full path what do you think if we do not make any changes there will still be in my opinion confusion in the routes when no relative paths option is passed
1
81,172
30,739,702,889
IssuesEvent
2023-07-28 10:25:25
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
`disable_3pid_login` doesn't hide email field on the registration page anymore
T-Defect
### Steps to reproduce 1. Add `"disable_3pid_login": true` to `config.json` 2. Go to `element.example.com/#/register` ### Outcome #### What did you expect? Email field isn't present. #### What happened instead? It's present, even though it used to be hidden with that setting set to true. It's confusing to my users as the server isn't configured to handle email. ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp _No response_ ### Application version Element version: 1.11.36 Olm version: 3.2.14 ### Homeserver Synapse v1.86.0 ### Will you send logs? No
1.0
`disable_3pid_login` doesn't hide email field on the registration page anymore - ### Steps to reproduce 1. Add `"disable_3pid_login": true` to `config.json` 2. Go to `element.example.com/#/register` ### Outcome #### What did you expect? Email field isn't present. #### What happened instead? It's present, even though it used to be hidden with that setting set to true. It's confusing to my users as the server isn't configured to handle email. ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp _No response_ ### Application version Element version: 1.11.36 Olm version: 3.2.14 ### Homeserver Synapse v1.86.0 ### Will you send logs? No
defect
disable login doesn t hide email field on the registration page anymore steps to reproduce add disable login true to config json go to element example com register outcome what did you expect email field isn t present what happened instead it s present even though it used to be hidden with that setting set to true it s confusing to my users as the server isn t configured to handle email operating system no response browser information no response url for webapp no response application version element version olm version homeserver synapse will you send logs no
1
164,791
12,812,916,254
IssuesEvent
2020-07-04 09:37:45
aliasrobotics/RVD
https://api.github.com/repos/aliasrobotics/RVD
closed
RVD#2697: (error) Uninitialized variable
bug components software cppcheck mitigated robot component: PX4 static analysis testing triage version: v1.9.0
```yaml id: 2697 title: 'RVD#2697: (error) Uninitialized variable' type: bug description: '[Firmware/src/modules/mavlink/mavlink_main.cpp:461]: (error) Uninitialized variable: inst' cwe: None cve: None keywords: - cppcheck - static analysis - testing - triage - bug - 'version: v1.9.0' - 'robot component: PX4' - components software system: Firmware/src/modules/mavlink/mavlink_main.cpp vendor: null severity: rvss-score: 0 rvss-vector: '' severity-description: '' cvss-score: 0 cvss-vector: '' links: - https://github.com/aliasrobotics/RVD/issues/2697 flaw: phase: testing specificity: N/A architectural-location: N/A application: N/A subsystem: N/A package: N/A languages: None date-detected: 2020-06-29 (14:53) detected-by: Alias Robotics detected-by-method: testing static date-reported: 2020-06-29 (14:53) reported-by: Alias Robotics reported-by-relationship: automatic issue: https://github.com/aliasrobotics/RVD/issues/2697 reproducibility: always trace: '' reproduction: See artifacts below (if available) reproduction-image: gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_px4/-/jobs/615913193/artifacts/download exploitation: description: '' exploitation-image: '' exploitation-vector: '' exploitation-recipe: '' mitigation: description: '' pull-request: '' date-mitigation: '' ```
1.0
RVD#2697: (error) Uninitialized variable - ```yaml id: 2697 title: 'RVD#2697: (error) Uninitialized variable' type: bug description: '[Firmware/src/modules/mavlink/mavlink_main.cpp:461]: (error) Uninitialized variable: inst' cwe: None cve: None keywords: - cppcheck - static analysis - testing - triage - bug - 'version: v1.9.0' - 'robot component: PX4' - components software system: Firmware/src/modules/mavlink/mavlink_main.cpp vendor: null severity: rvss-score: 0 rvss-vector: '' severity-description: '' cvss-score: 0 cvss-vector: '' links: - https://github.com/aliasrobotics/RVD/issues/2697 flaw: phase: testing specificity: N/A architectural-location: N/A application: N/A subsystem: N/A package: N/A languages: None date-detected: 2020-06-29 (14:53) detected-by: Alias Robotics detected-by-method: testing static date-reported: 2020-06-29 (14:53) reported-by: Alias Robotics reported-by-relationship: automatic issue: https://github.com/aliasrobotics/RVD/issues/2697 reproducibility: always trace: '' reproduction: See artifacts below (if available) reproduction-image: gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_px4/-/jobs/615913193/artifacts/download exploitation: description: '' exploitation-image: '' exploitation-vector: '' exploitation-recipe: '' mitigation: description: '' pull-request: '' date-mitigation: '' ```
non_defect
rvd error uninitialized variable yaml id title rvd error uninitialized variable type bug description error uninitialized variable inst cwe none cve none keywords cppcheck static analysis testing triage bug version robot component components software system firmware src modules mavlink mavlink main cpp vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity n a architectural location n a application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace reproduction see artifacts below if available reproduction image gitlab com aliasrobotics offensive alurity pipelines active pipeline jobs artifacts download exploitation description exploitation image exploitation vector exploitation recipe mitigation description pull request date mitigation
0
111,219
4,466,901,860
IssuesEvent
2016-08-25 01:11:14
Murkantilism/skylabgame
https://api.github.com/repos/Murkantilism/skylabgame
closed
Write code to support the new animation slices
enhancement Low Priority
- [x] Re: #167 will need to play animation slices at the right times based on timing decided in that ticket - [x] Additionally, there appears to be a minor issue where the player movement does not wait for the wall flip animation to complete if you hold down boost input (even if you are out of boost). Normally there should be a slight pause for the animation to complete. - [x] Wallflip should start sooner
1.0
Write code to support the new animation slices - - [x] Re: #167 will need to play animation slices at the right times based on timing decided in that ticket - [x] Additionally, there appears to be a minor issue where the player movement does not wait for the wall flip animation to complete if you hold down boost input (even if you are out of boost). Normally there should be a slight pause for the animation to complete. - [x] Wallflip should start sooner
non_defect
write code to support the new animation slices re will need to play animation slices at the right times based on timing decided in that ticket additionally there appears to be a minor issue where the player movement does not wait for the wall flip animation to complete if you hold down boost input even if you are out of boost normally there should be a slight pause for the animation to complete wallflip should start sooner
0
655,932
21,714,621,638
IssuesEvent
2022-05-10 16:38:34
WordPress/openverse-frontend
https://api.github.com/repos/WordPress/openverse-frontend
closed
Image changes size when loading single result page
🟩 priority: low 🛠 goal: fix 🕹 aspect: interface 🔧 tech: vue 🎨 tech: css
## Description <!-- Concisely describe the bug. Compare your experience with what you expected to happen. --> <!-- For example: "I clicked the 'submit' button and instead of seeing a thank you message, I saw a blank page." --> When viewing a single image result, the image thumbnail will load first, then the fullsize image will load. This frequently causes a smaller image to load momentarily, then flashes to the larger image. Should the thumbnail not be loaded in this context? Should it be loaded at the same size as the original image so there's no size discrepancy, even if there is a quality discrepancy? ## Reproduction <!-- Provide detailed steps to reproduce the bug. --> 1. Visit https://search-staging.openverse.engineering/image/0439408c-11ae-490e-a10a-8139766f68fb 2. Refresh the page 3. Observe that the thumbnail shows up momentarily before the full image loads ## Screenshots <!-- Add screenshots to show the problem; or delete the section entirely. --> https://user-images.githubusercontent.com/10214785/165646181-e8a89b82-36df-4368-b9d9-da731ed60a85.mp4 ## Environment <!-- Please complete this, unless you are certain the problem is not environment specific. --> - Device: Desktop - OS: Ubuntu - Browser: Firefox ## Additional context <!-- Add any other context about the problem here; or delete the section entirely. --> ## Resolution <!-- Replace the [ ] with [x] to check the box. --> - [ ] 🙋 I would be interested in resolving this bug.
1.0
Image changes size when loading single result page - ## Description <!-- Concisely describe the bug. Compare your experience with what you expected to happen. --> <!-- For example: "I clicked the 'submit' button and instead of seeing a thank you message, I saw a blank page." --> When viewing a single image result, the image thumbnail will load first, then the fullsize image will load. This frequently causes a smaller image to load momentarily, then flashes to the larger image. Should the thumbnail not be loaded in this context? Should it be loaded at the same size as the original image so there's no size discrepancy, even if there is a quality discrepancy? ## Reproduction <!-- Provide detailed steps to reproduce the bug. --> 1. Visit https://search-staging.openverse.engineering/image/0439408c-11ae-490e-a10a-8139766f68fb 2. Refresh the page 3. Observe that the thumbnail shows up momentarily before the full image loads ## Screenshots <!-- Add screenshots to show the problem; or delete the section entirely. --> https://user-images.githubusercontent.com/10214785/165646181-e8a89b82-36df-4368-b9d9-da731ed60a85.mp4 ## Environment <!-- Please complete this, unless you are certain the problem is not environment specific. --> - Device: Desktop - OS: Ubuntu - Browser: Firefox ## Additional context <!-- Add any other context about the problem here; or delete the section entirely. --> ## Resolution <!-- Replace the [ ] with [x] to check the box. --> - [ ] 🙋 I would be interested in resolving this bug.
non_defect
image changes size when loading single result page description when viewing a single image result the image thumbnail will load first then the fullsize image will load this frequently causes a smaller image to load momentarily then flashes to the larger image should the thumbnail not be loaded in this context should it be loaded at the same size as the original image so there s no size discrepancy even if there is a quality discrepancy reproduction visit refresh the page observe that the thumbnail shows up momentarily before the full image loads screenshots environment device desktop os ubuntu browser firefox additional context resolution 🙋 i would be interested in resolving this bug
0
28,067
5,172,103,663
IssuesEvent
2017-01-18 12:29:18
primefaces/primeng
https://api.github.com/repos/primefaces/primeng
closed
Button with Icon: the label is not updated when used with a pipe
defect
When used with a pipe such as translate (ng2-translate) the label is not correctly updated. First display - OK ![2017-01-03_10h57_00](https://cloud.githubusercontent.com/assets/9085108/21604645/9fa4b160-d1a4-11e6-8dbb-1fb083831770.png) After language switch - KO ![2017-01-03_11h00_18](https://cloud.githubusercontent.com/assets/9085108/21604657/b4e161c2-d1a4-11e6-8ebc-4c3522c5d982.png) DOM ![2017-01-03_11h07_14](https://cloud.githubusercontent.com/assets/9085108/21604674/d3021bf6-d1a4-11e6-9276-efeec1dfc625.png) "primeng": "^1.1.1", "@angular/core": "2.2.1", "ng2-translate": "^4.0.0", I'm using the ultima-ng theme but I don't thik the problem is related to the theme.
1.0
Button with Icon: the label is not updated when used with a pipe - When used with a pipe such as translate (ng2-translate) the label is not correctly updated. First display - OK ![2017-01-03_10h57_00](https://cloud.githubusercontent.com/assets/9085108/21604645/9fa4b160-d1a4-11e6-8dbb-1fb083831770.png) After language switch - KO ![2017-01-03_11h00_18](https://cloud.githubusercontent.com/assets/9085108/21604657/b4e161c2-d1a4-11e6-8ebc-4c3522c5d982.png) DOM ![2017-01-03_11h07_14](https://cloud.githubusercontent.com/assets/9085108/21604674/d3021bf6-d1a4-11e6-9276-efeec1dfc625.png) "primeng": "^1.1.1", "@angular/core": "2.2.1", "ng2-translate": "^4.0.0", I'm using the ultima-ng theme but I don't thik the problem is related to the theme.
defect
button with icon the label is not updated when used with a pipe when used with a pipe such as translate translate the label is not correctly updated first display ok after language switch ko dom primeng angular core translate i m using the ultima ng theme but i don t thik the problem is related to the theme
1
20,154
3,298,489,285
IssuesEvent
2015-11-02 14:48:27
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
closed
Const expressions is emitted in init method
defect
The compiler should calculate const expressions and do not move it to `init` method internal static class B { internal const int Valeur1 = 1; internal const int Valeur2 = Valeur1+1; } `Valeur2` should be emitted in static block instead of `init` method
1.0
Const expressions is emitted in init method - The compiler should calculate const expressions and do not move it to `init` method internal static class B { internal const int Valeur1 = 1; internal const int Valeur2 = Valeur1+1; } `Valeur2` should be emitted in static block instead of `init` method
defect
const expressions is emitted in init method the compiler should calculate const expressions and do not move it to init method internal static class b internal const int internal const int should be emitted in static block instead of init method
1
44,576
12,263,215,393
IssuesEvent
2020-05-07 00:17:50
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
closed
zpool iostat appears to hang
Type: Defect
### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | CentOS Distribution Version | 7.7.1908 Linux Kernel | 3.10.0-1062.18.1.el7.x86_64 Architecture | x86_64 ZFS Version | 0.8.3-1 SPL Version | 0.8.3-1 <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing zpool iostat appears to hang after printing the header ``` capacity operations bandwidth pool alloc free read write read write --------------------------------------------------------------------------------------------------------------------------- ``` Actually, it's printing an "infinite" number of dash - characters under the headings because the output is a terminal and the columns is less than 42 ``` [root@fortknox ~]# stty columns 42 [root@fortknox ~]# zpool iostat capacity operations bandwidth pool alloc free read write read write ---- ----- ----- ----- ----- ----- ----- rpool 1.67G 65.3G 6 0 603K 2.36K [root@fortknox ~]# stty columns 41 [root@fortknox ~]# zpool iostat capacity operations bandwidth pool alloc free read write read write --------------------------------------------------------------------------------------------------------------------------- ``` ### Describe how to reproduce the problem Set the terminal columns to 41 or less Run "zpool iostat"
1.0
zpool iostat appears to hang - ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | CentOS Distribution Version | 7.7.1908 Linux Kernel | 3.10.0-1062.18.1.el7.x86_64 Architecture | x86_64 ZFS Version | 0.8.3-1 SPL Version | 0.8.3-1 <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing zpool iostat appears to hang after printing the header ``` capacity operations bandwidth pool alloc free read write read write --------------------------------------------------------------------------------------------------------------------------- ``` Actually, it's printing an "infinite" number of dash - characters under the headings because the output is a terminal and the columns is less than 42 ``` [root@fortknox ~]# stty columns 42 [root@fortknox ~]# zpool iostat capacity operations bandwidth pool alloc free read write read write ---- ----- ----- ----- ----- ----- ----- rpool 1.67G 65.3G 6 0 603K 2.36K [root@fortknox ~]# stty columns 41 [root@fortknox ~]# zpool iostat capacity operations bandwidth pool alloc free read write read write --------------------------------------------------------------------------------------------------------------------------- ``` ### Describe how to reproduce the problem Set the terminal columns to 41 or less Run "zpool iostat"
defect
zpool iostat appears to hang system information type version name distribution name centos distribution version linux kernel architecture zfs version spl version commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing zpool iostat appears to hang after printing the header capacity operations bandwidth pool alloc free read write read write actually it s printing an infinite number of dash characters under the headings because the output is a terminal and the columns is less than stty columns zpool iostat capacity operations bandwidth pool alloc free read write read write rpool stty columns zpool iostat capacity operations bandwidth pool alloc free read write read write describe how to reproduce the problem set the terminal columns to or less run zpool iostat
1
58,728
3,090,784,043
IssuesEvent
2015-08-26 09:06:12
UnifiedViews/Plugins-QualityAssessment
https://api.github.com/repos/UnifiedViews/Plugins-QualityAssessment
closed
Installation of QualityAssessment Plugins under debian
priority: Normal severity: enhancement
It is requested to create support for installation of DPUs as debian wheezy package.
1.0
Installation of QualityAssessment Plugins under debian - It is requested to create support for installation of DPUs as debian wheezy package.
non_defect
installation of qualityassessment plugins under debian it is requested to create support for installation of dpus as debian wheezy package
0
89,338
10,595,948,568
IssuesEvent
2019-10-09 20:07:52
ICB-DCM/AMICI
https://api.github.com/repos/ICB-DCM/AMICI
closed
Make swig generate pydoc comments from doxygen comments
HacktoberFest documentation enhancement good first issue python
- [x] Check available swig version (need >= 4.0) - [x] Add `-doxygen` - [] ~Make swig ignore `Unknown Doxygen command`s (http://www.swig.org/Doc4.0/Doxygen.html)~ See also #613
1.0
Make swig generate pydoc comments from doxygen comments - - [x] Check available swig version (need >= 4.0) - [x] Add `-doxygen` - [] ~Make swig ignore `Unknown Doxygen command`s (http://www.swig.org/Doc4.0/Doxygen.html)~ See also #613
non_defect
make swig generate pydoc comments from doxygen comments check available swig version need add doxygen make swig ignore unknown doxygen command s see also
0
31,154
6,443,897,545
IssuesEvent
2017-08-12 02:15:51
opendatakit/opendatakit
https://api.github.com/repos/opendatakit/opendatakit
closed
Google Docs is interpreting space-separated numbers as date fields
Aggregate Priority-Medium Type-Defect
Originally reported on Google Code with ID 462 ``` From a user: I have a ODK aggregate form that I am uploading to Google docs. There is multiple select sections in it with numeric underlying values, i.e. 1, 2, 3, 4 ,5 etc. Now uploading gives me in the coloumn, just two numbers, say "3 5". No problem here, sending to Google docs, google change my data to a date. For the above I get 3/5/2011. Formatting the column to text is no use either, since the original data is lost already, at least I must manually change it back to "3 5". How can I prevent this from happening, other than changing the underlying value to text? ---------------- Will see what can be done to get Google Docs to not do this. ``` Reported by `mitchellsundt` on 2011-12-05 17:10:03
1.0
Google Docs is interpreting space-separated numbers as date fields - Originally reported on Google Code with ID 462 ``` From a user: I have a ODK aggregate form that I am uploading to Google docs. There is multiple select sections in it with numeric underlying values, i.e. 1, 2, 3, 4 ,5 etc. Now uploading gives me in the coloumn, just two numbers, say "3 5". No problem here, sending to Google docs, google change my data to a date. For the above I get 3/5/2011. Formatting the column to text is no use either, since the original data is lost already, at least I must manually change it back to "3 5". How can I prevent this from happening, other than changing the underlying value to text? ---------------- Will see what can be done to get Google Docs to not do this. ``` Reported by `mitchellsundt` on 2011-12-05 17:10:03
defect
google docs is interpreting space separated numbers as date fields originally reported on google code with id from a user i have a odk aggregate form that i am uploading to google docs there is multiple select sections in it with numeric underlying values i e etc now uploading gives me in the coloumn just two numbers say no problem here sending to google docs google change my data to a date for the above i get formatting the column to text is no use either since the original data is lost already at least i must manually change it back to how can i prevent this from happening other than changing the underlying value to text will see what can be done to get google docs to not do this reported by mitchellsundt on
1
66,743
20,614,415,506
IssuesEvent
2022-03-07 11:48:09
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
opened
`<font color>` is not applied inside of `<h1>`
T-Defect
### Steps to reproduce 1. send colored text inside of a heading ### Outcome ``` { "msgtype": "m.notice", "body": "weißes ?, rotes ?, blaues ?", "format": "org.matrix.custom.html", "formatted_body": "<strong><font color=\"#ffffff\">?</font><font color=\"#ff0000\">?</font><font color=\"#00b0f0\">?</font><strong/>" } ``` renders as it should: ![image](https://user-images.githubusercontent.com/2803622/157028419-97ecb9f4-082d-4e74-98d5-a3e6c582f2c9.png) ``` { "msgtype": "m.notice", "body": "großes weißes ?, großes rotes ?, großes blaues ?", "format": "org.matrix.custom.html", "formatted_body": "<h1><font color=\"#ffffff\">?</font><font color=\"#ff0000\">?</font><font color=\"#00b0f0\">?</font><h1/>" } ``` renders neither `<h1` nor `color` ![image](https://user-images.githubusercontent.com/2803622/157028380-1054e69c-2ace-4fca-931c-6ddb2a2f5d8a.png) ``` { "body": "# ???", "msgtype": "m.text", "format": "org.matrix.custom.html", "formatted_body": "<h1>???</h1>\n" } ``` renders as ![image](https://user-images.githubusercontent.com/2803622/157028752-1ad8aabc-04bb-427e-b8a5-6b4c9d93c5b4.png) ### Your phone model _No response_ ### Operating system version 10 ### Application version and app store 1.4.2 ### Homeserver 1.53 ### Will you send logs? No
1.0
`<font color>` is not applied inside of `<h1>` - ### Steps to reproduce 1. send colored text inside of a heading ### Outcome ``` { "msgtype": "m.notice", "body": "weißes ?, rotes ?, blaues ?", "format": "org.matrix.custom.html", "formatted_body": "<strong><font color=\"#ffffff\">?</font><font color=\"#ff0000\">?</font><font color=\"#00b0f0\">?</font><strong/>" } ``` renders as it should: ![image](https://user-images.githubusercontent.com/2803622/157028419-97ecb9f4-082d-4e74-98d5-a3e6c582f2c9.png) ``` { "msgtype": "m.notice", "body": "großes weißes ?, großes rotes ?, großes blaues ?", "format": "org.matrix.custom.html", "formatted_body": "<h1><font color=\"#ffffff\">?</font><font color=\"#ff0000\">?</font><font color=\"#00b0f0\">?</font><h1/>" } ``` renders neither `<h1` nor `color` ![image](https://user-images.githubusercontent.com/2803622/157028380-1054e69c-2ace-4fca-931c-6ddb2a2f5d8a.png) ``` { "body": "# ???", "msgtype": "m.text", "format": "org.matrix.custom.html", "formatted_body": "<h1>???</h1>\n" } ``` renders as ![image](https://user-images.githubusercontent.com/2803622/157028752-1ad8aabc-04bb-427e-b8a5-6b4c9d93c5b4.png) ### Your phone model _No response_ ### Operating system version 10 ### Application version and app store 1.4.2 ### Homeserver 1.53 ### Will you send logs? No
defect
is not applied inside of steps to reproduce send colored text inside of a heading outcome msgtype m notice body weißes rotes blaues format org matrix custom html formatted body renders as it should msgtype m notice body großes weißes großes rotes großes blaues format org matrix custom html formatted body renders neither nor color body msgtype m text format org matrix custom html formatted body n renders as your phone model no response operating system version application version and app store homeserver will you send logs no
1
140,291
18,901,230,139
IssuesEvent
2021-11-16 01:17:58
tim-wsdemo/nw-prism
https://api.github.com/repos/tim-wsdemo/nw-prism
opened
CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz
security vulnerability
## CVE-2021-3918 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary> <p>JSON Schema validation and specifications</p> <p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p> <p>Path to dependency file: nw-prism/package.json</p> <p>Path to vulnerable library: nw-prism/node_modules/json-schema/package.json</p> <p> Dependency Hierarchy: - botbuilder-3.30.0.tgz (Root Library) - request-2.88.2.tgz - http-signature-1.2.0.tgz - jsprim-1.4.1.tgz - :x: **json-schema-0.2.3.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution') <p>Publish Date: 2021-11-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918>CVE-2021-3918</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"json-schema","packageVersion":"0.2.3","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"botbuilder:3.30.0;request:2.88.2;http-signature:1.2.0;jsprim:1.4.1;json-schema:0.2.3","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-3918","vulnerabilityDetails":"json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes (\u0027Prototype Pollution\u0027)","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz - ## CVE-2021-3918 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary> <p>JSON Schema validation and specifications</p> <p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p> <p>Path to dependency file: nw-prism/package.json</p> <p>Path to vulnerable library: nw-prism/node_modules/json-schema/package.json</p> <p> Dependency Hierarchy: - botbuilder-3.30.0.tgz (Root Library) - request-2.88.2.tgz - http-signature-1.2.0.tgz - jsprim-1.4.1.tgz - :x: **json-schema-0.2.3.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution') <p>Publish Date: 2021-11-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918>CVE-2021-3918</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"json-schema","packageVersion":"0.2.3","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"botbuilder:3.30.0;request:2.88.2;http-signature:1.2.0;jsprim:1.4.1;json-schema:0.2.3","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-3918","vulnerabilityDetails":"json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes (\u0027Prototype Pollution\u0027)","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_defect
cve high detected in json schema tgz cve high severity vulnerability vulnerable library json schema tgz json schema validation and specifications library home page a href path to dependency file nw prism package json path to vulnerable library nw prism node modules json schema package json dependency hierarchy botbuilder tgz root library request tgz http signature tgz jsprim tgz x json schema tgz vulnerable library found in base branch master vulnerability details json schema is vulnerable to improperly controlled modification of object prototype attributes prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree botbuilder request http signature jsprim json schema isminimumfixversionavailable false basebranches vulnerabilityidentifier cve vulnerabilitydetails json schema is vulnerable to improperly controlled modification of object prototype attributes pollution vulnerabilityurl
0
52,683
13,224,928,451
IssuesEvent
2020-08-17 20:08:11
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
new singleton implementation does not compile on RHEL4 (gcc 3.4.6)_ (Trac #151)
IceTray Migrated from Trac defect
Compilation of the trunk of offline software ends with an error (error dump: see "additional information"). According to the error message an attempt is being made by the Singleton implementation to use the I3Factory() constructor. This happens both for I3ModuleFactory and for I3ServiceFactory. The error does not occur under RHEL5 (gcc 4.1.2) <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/151">https://code.icecube.wisc.edu/projects/icecube/ticket/151</a>, reported by blaufussand owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2009-01-08T14:51:17", "_ts": "1231426277000000", "description": "Compilation of the trunk of offline software ends with an error (error\ndump: see \"additional information\"). According to the error message an\nattempt is being made by the Singleton implementation to use the\nI3Factory() constructor. This happens both for I3ModuleFactory and for\nI3ServiceFactory.\n\nThe error does not occur under RHEL5 (gcc 4.1.2)", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "time": "2008-11-19T02:08:07", "component": "IceTray", "summary": "new singleton implementation does not compile on RHEL4 (gcc 3.4.6)_", "priority": "normal", "keywords": "", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
1.0
new singleton implementation does not compile on RHEL4 (gcc 3.4.6)_ (Trac #151) - Compilation of the trunk of offline software ends with an error (error dump: see "additional information"). According to the error message an attempt is being made by the Singleton implementation to use the I3Factory() constructor. This happens both for I3ModuleFactory and for I3ServiceFactory. The error does not occur under RHEL5 (gcc 4.1.2) <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/151">https://code.icecube.wisc.edu/projects/icecube/ticket/151</a>, reported by blaufussand owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2009-01-08T14:51:17", "_ts": "1231426277000000", "description": "Compilation of the trunk of offline software ends with an error (error\ndump: see \"additional information\"). According to the error message an\nattempt is being made by the Singleton implementation to use the\nI3Factory() constructor. This happens both for I3ModuleFactory and for\nI3ServiceFactory.\n\nThe error does not occur under RHEL5 (gcc 4.1.2)", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "time": "2008-11-19T02:08:07", "component": "IceTray", "summary": "new singleton implementation does not compile on RHEL4 (gcc 3.4.6)_", "priority": "normal", "keywords": "", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
defect
new singleton implementation does not compile on gcc trac compilation of the trunk of offline software ends with an error error dump see additional information according to the error message an attempt is being made by the singleton implementation to use the constructor this happens both for and for the error does not occur under gcc migrated from json status closed changetime ts description compilation of the trunk of offline software ends with an error error ndump see additional information according to the error message an nattempt is being made by the singleton implementation to use the constructor this happens both for and for n nthe error does not occur under gcc reporter blaufuss cc resolution fixed time component icetray summary new singleton implementation does not compile on gcc priority normal keywords milestone owner troy type defect
1
31,993
6,676,593,425
IssuesEvent
2017-10-05 06:45:54
proarc/proarc
https://api.github.com/repos/proarc/proarc
closed
Povinnost vyplnění - kartografický dokument
auto-migrated Priority-Low Type-Defect
``` Prosíme opravit: 1. subject - name chybí 2. subject - name part - ve specifikaci M, v ProArcu R 3. related item - chybí 4. V poli identifier přidat pole pro identifikátory urnnbn, ccnb, isbn. Z prázdného formuláře není zřejmé, že vedle uuid mají být vyplněné i další údaje. 5. identifier - chybí pole pole pro oclc a sysno 6. record info - chybí pole description standard ``` Original issue reported on code.google.com by `daneck...@knav.cz` on 13 May 2014 at 9:52
1.0
Povinnost vyplnění - kartografický dokument - ``` Prosíme opravit: 1. subject - name chybí 2. subject - name part - ve specifikaci M, v ProArcu R 3. related item - chybí 4. V poli identifier přidat pole pro identifikátory urnnbn, ccnb, isbn. Z prázdného formuláře není zřejmé, že vedle uuid mají být vyplněné i další údaje. 5. identifier - chybí pole pole pro oclc a sysno 6. record info - chybí pole description standard ``` Original issue reported on code.google.com by `daneck...@knav.cz` on 13 May 2014 at 9:52
defect
povinnost vyplnění kartografický dokument prosíme opravit subject name chybí subject name part ve specifikaci m v proarcu r related item chybí v poli identifier přidat pole pro identifikátory urnnbn ccnb isbn z prázdného formuláře není zřejmé že vedle uuid mají být vyplněné i další údaje identifier chybí pole pole pro oclc a sysno record info chybí pole description standard original issue reported on code google com by daneck knav cz on may at
1
311,785
26,814,545,554
IssuesEvent
2023-02-02 02:36:24
t3rn/t3rn
https://api.github.com/repos/t3rn/t3rn
closed
Setup a bug bounty program
test :test_tube: chainops Stale
### Item Type Task ### Description We can up our applications' security with the help of open source hackers incentivied by monetary and/or crypto rewards. Let us bootstrap a bug bounty program that is: + hosted on hackerone + initially private (select+invite hackers, later maybe public) ### Additional Information To get an idea what a program looks like check the [hackerone directory](https://hackerone.com/directory/programs). ### Acceptance Criteria + active hackerone program ### What are the points assigned to this item? 5 ### Is the issue ready? - [X] This issue is ready
2.0
Setup a bug bounty program - ### Item Type Task ### Description We can up our applications' security with the help of open source hackers incentivied by monetary and/or crypto rewards. Let us bootstrap a bug bounty program that is: + hosted on hackerone + initially private (select+invite hackers, later maybe public) ### Additional Information To get an idea what a program looks like check the [hackerone directory](https://hackerone.com/directory/programs). ### Acceptance Criteria + active hackerone program ### What are the points assigned to this item? 5 ### Is the issue ready? - [X] This issue is ready
non_defect
setup a bug bounty program item type task description we can up our applications security with the help of open source hackers incentivied by monetary and or crypto rewards let us bootstrap a bug bounty program that is hosted on hackerone initially private select invite hackers later maybe public additional information to get an idea what a program looks like check the acceptance criteria active hackerone program what are the points assigned to this item is the issue ready this issue is ready
0
351,577
25,032,816,228
IssuesEvent
2022-11-04 13:47:21
moov-io/ach
https://api.github.com/repos/moov-io/ach
closed
ACH server API documentation broken links
bug documentation
### ACH Version all ### What were you trying to do? Reading the documentation there is a link for the REST API that goes to a 404 when trying to follow how to post a JSON file. https://moov-io.github.io/ach/create-file/#upload-a-json-representation <img width="720" alt="Screenshot 2022-11-03 at 3 42 06 PM" src="https://user-images.githubusercontent.com/32989/199847910-5d751561-7086-4b21-8408-00155c9aa85f.png"> This takes you to a link that is no longer valid. https://api.moov.io/#operation/createFile The same file errors below on the same page for creating a file https://api.moov.io/#operation/createFile ### What did you expect to see? It would be nice to have a link to the server API specification from the readme and have this broken link fixed. I think that we want people to go here. But need to confirm with @adamdecaf if this is the most up-to-date API specification https://moov-io.github.io/ach/api/#get-/files/-fileID- ### What did you see? none ### How can we reproduce the problem? visit the webpage
1.0
ACH server API documentation broken links - ### ACH Version all ### What were you trying to do? Reading the documentation there is a link for the REST API that goes to a 404 when trying to follow how to post a JSON file. https://moov-io.github.io/ach/create-file/#upload-a-json-representation <img width="720" alt="Screenshot 2022-11-03 at 3 42 06 PM" src="https://user-images.githubusercontent.com/32989/199847910-5d751561-7086-4b21-8408-00155c9aa85f.png"> This takes you to a link that is no longer valid. https://api.moov.io/#operation/createFile The same file errors below on the same page for creating a file https://api.moov.io/#operation/createFile ### What did you expect to see? It would be nice to have a link to the server API specification from the readme and have this broken link fixed. I think that we want people to go here. But need to confirm with @adamdecaf if this is the most up-to-date API specification https://moov-io.github.io/ach/api/#get-/files/-fileID- ### What did you see? none ### How can we reproduce the problem? visit the webpage
non_defect
ach server api documentation broken links ach version all what were you trying to do reading the documentation there is a link for the rest api that goes to a when trying to follow how to post a json file img width alt screenshot at pm src this takes you to a link that is no longer valid the same file errors below on the same page for creating a file what did you expect to see it would be nice to have a link to the server api specification from the readme and have this broken link fixed i think that we want people to go here but need to confirm with adamdecaf if this is the most up to date api specification what did you see none how can we reproduce the problem visit the webpage
0
43,564
11,756,062,826
IssuesEvent
2020-03-13 10:46:36
STEllAR-GROUP/hpx
https://api.github.com/repos/STEllAR-GROUP/hpx
closed
HPX hello_world_distributed crashes on qb-2 when run on 75 nodes
category: parcel transport type: defect
## Expected Behavior I ran hello_world_distributed with 75 nodes / 20 cores per node using 20 MPI instances per node and 1 HPX thread per MPI instance, and expected 1500 Hello Worlds. ## Actual Behavior Segmentation fault with no stdout at all and this stderr output {stack-trace}: 4 frames: 0x2aabb8b1b772 : ??? + 0x2aabb8b1b772 in /home/dmarce1/local/relwithdebinfo/hpx/lib64/libhpx.so.1 0x2aabb8b1a278 : hpx::termination_handler(int) + 0xd8 in /home/dmarce1/local/relwithdebinfo/hpx/lib64/libhpx.so.1 0x365b40f7e0 : ??? + 0x365b40f7e0 in /lib64/libpthread.so.0 0x2aabbb208b56 : MPIDI_CH3_iSendv + 0x76 in /home/dmarce1/local/mpi/lib/libmpi.so.12 {what}: Segmentation fault ## Steps to Reproduce the Problem Run hello_world_distributed with 75 nodes and 20 MPI instances per node and argument -t1 on QB-2. This behavior has also happened with Octo-Tiger, using 75 nodes and 20 MPI instances per node, -t1, and with 128 nodes and 1 MPI instance per node, -t20. ## Specifications HPX 1.4.0 GCC 8.3 MVAPICH2 2.2
1.0
HPX hello_world_distributed crashes on qb-2 when run on 75 nodes - ## Expected Behavior I ran hello_world_distributed with 75 nodes / 20 cores per node using 20 MPI instances per node and 1 HPX thread per MPI instance, and expected 1500 Hello Worlds. ## Actual Behavior Segmentation fault with no stdout at all and this stderr output {stack-trace}: 4 frames: 0x2aabb8b1b772 : ??? + 0x2aabb8b1b772 in /home/dmarce1/local/relwithdebinfo/hpx/lib64/libhpx.so.1 0x2aabb8b1a278 : hpx::termination_handler(int) + 0xd8 in /home/dmarce1/local/relwithdebinfo/hpx/lib64/libhpx.so.1 0x365b40f7e0 : ??? + 0x365b40f7e0 in /lib64/libpthread.so.0 0x2aabbb208b56 : MPIDI_CH3_iSendv + 0x76 in /home/dmarce1/local/mpi/lib/libmpi.so.12 {what}: Segmentation fault ## Steps to Reproduce the Problem Run hello_world_distributed with 75 nodes and 20 MPI instances per node and argument -t1 on QB-2. This behavior has also happened with Octo-Tiger, using 75 nodes and 20 MPI instances per node, -t1, and with 128 nodes and 1 MPI instance per node, -t20. ## Specifications HPX 1.4.0 GCC 8.3 MVAPICH2 2.2
defect
hpx hello world distributed crashes on qb when run on nodes expected behavior i ran hello world distributed with nodes cores per node using mpi instances per node and hpx thread per mpi instance and expected hello worlds actual behavior segmentation fault with no stdout at all and this stderr output stack trace frames in home local relwithdebinfo hpx libhpx so hpx termination handler int in home local relwithdebinfo hpx libhpx so in libpthread so mpidi isendv in home local mpi lib libmpi so what segmentation fault steps to reproduce the problem run hello world distributed with nodes and mpi instances per node and argument on qb this behavior has also happened with octo tiger using nodes and mpi instances per node and with nodes and mpi instance per node specifications hpx gcc
1
299,120
25,883,277,793
IssuesEvent
2022-12-14 12:50:45
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
DISABLED test_chain_iterable_style_dataset (__main__.TestDataLoaderPersistentWorkers)
module: dataloader triaged module: flaky-tests skipped
Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_chain_iterable_style_dataset&suite=TestDataLoaderPersistentWorkers&file=test_dataloader.py) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/7141678531). Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 red and 1 green. cc @SsnL @VitalyFedyunin @ejguan @NivekT
1.0
DISABLED test_chain_iterable_style_dataset (__main__.TestDataLoaderPersistentWorkers) - Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_chain_iterable_style_dataset&suite=TestDataLoaderPersistentWorkers&file=test_dataloader.py) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/7141678531). Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 red and 1 green. cc @SsnL @VitalyFedyunin @ejguan @NivekT
non_defect
disabled test chain iterable style dataset main testdataloaderpersistentworkers platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with red and green cc ssnl vitalyfedyunin ejguan nivekt
0
22,487
3,654,276,167
IssuesEvent
2016-02-17 11:44:38
brunoais/javadude
https://api.github.com/repos/brunoais/javadude
closed
Annotations: If superclass is serializable, add serialization version id
auto-migrated Priority-Medium Project-Annotations Type-Defect
``` If specified superclass is serializable, generated superclass should have a serialization version id. ``` Original issue reported on code.google.com by `scott%ja...@gtempaccount.com` on 3 May 2009 at 3:14
1.0
Annotations: If superclass is serializable, add serialization version id - ``` If specified superclass is serializable, generated superclass should have a serialization version id. ``` Original issue reported on code.google.com by `scott%ja...@gtempaccount.com` on 3 May 2009 at 3:14
defect
annotations if superclass is serializable add serialization version id if specified superclass is serializable generated superclass should have a serialization version id original issue reported on code google com by scott ja gtempaccount com on may at
1
10,737
2,622,182,647
IssuesEvent
2015-03-04 00:19:29
byzhang/leveldb
https://api.github.com/repos/byzhang/leveldb
closed
leveldb segfaulting when no available disk space
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. sudo mount -t tmpfs tmpfs /tmp/leveldb -o size=500k 2. compile test program (see attachment): cd build && cmake -DCMAKE_BUILD_TYPE=debug .. 3. Run test program: build/src/leveldb /tmp/leveldb What is the expected output? What do you see instead? When not enought disk space, leveldb should return error on any "put" operation. Segfaulting is not good idea: What version of the product are you using? On what operating system? Re-producible bug at versions: 1. 0b9a89f Release LevelDB 1.14 2. Release: LevelDB 1.14.0 (re-check from archive sources) Operation system: Gentoo OS, ubuntu 12.04 Please provide any additional information below. 1. Test program an be found at attachment 2. See gdb output: http://pastebin.com/BpjYrPaw ``` Original issue reported on code.google.com by `feniksgo...@gmail.com` on 2 Dec 2013 at 6:56
1.0
leveldb segfaulting when no available disk space - ``` What steps will reproduce the problem? 1. sudo mount -t tmpfs tmpfs /tmp/leveldb -o size=500k 2. compile test program (see attachment): cd build && cmake -DCMAKE_BUILD_TYPE=debug .. 3. Run test program: build/src/leveldb /tmp/leveldb What is the expected output? What do you see instead? When not enought disk space, leveldb should return error on any "put" operation. Segfaulting is not good idea: What version of the product are you using? On what operating system? Re-producible bug at versions: 1. 0b9a89f Release LevelDB 1.14 2. Release: LevelDB 1.14.0 (re-check from archive sources) Operation system: Gentoo OS, ubuntu 12.04 Please provide any additional information below. 1. Test program an be found at attachment 2. See gdb output: http://pastebin.com/BpjYrPaw ``` Original issue reported on code.google.com by `feniksgo...@gmail.com` on 2 Dec 2013 at 6:56
defect
leveldb segfaulting when no available disk space what steps will reproduce the problem sudo mount t tmpfs tmpfs tmp leveldb o size compile test program see attachment cd build cmake dcmake build type debug run test program build src leveldb tmp leveldb what is the expected output what do you see instead when not enought disk space leveldb should return error on any put operation segfaulting is not good idea what version of the product are you using on what operating system re producible bug at versions release leveldb release leveldb re check from archive sources operation system gentoo os ubuntu please provide any additional information below test program an be found at attachment see gdb output original issue reported on code google com by feniksgo gmail com on dec at
1
97,204
8,651,573,771
IssuesEvent
2018-11-27 03:50:39
humera987/FXLabs-Test-Automation
https://api.github.com/repos/humera987/FXLabs-Test-Automation
closed
projecttest16 : ApiV1RunsJobidTestsuiteTestSuiteResponsesNameGetQueryParamPagesizeEmptyValue
projecttest16
Project : projecttest16 Job : UAT Env : UAT Region : US_WEST Result : fail Status Code : 404 Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=ODhjOTJmMmUtMTIwZC00NzhhLWJkYWQtMjJjMzQ4YTI4NjAy; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Fri, 16 Nov 2018 05:40:21 GMT]} Endpoint : http://13.56.210.25/api/v1/api/v1/runs/PrElm9RK/testSuite/test-suite-responses/PrElm9RK?pageSize= Request : Response : { "timestamp" : "2018-11-16T05:40:21.987+0000", "status" : 404, "error" : "Not Found", "message" : "No message available", "path" : "/api/v1/api/v1/runs/PrElm9RK/testSuite/test-suite-responses/PrElm9RK" } Logs : Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 500] resolved-to [404 != 500] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]Assertion [@StatusCode != 200] resolved-to [404 != 200] result [Passed] --- FX Bot ---
1.0
projecttest16 : ApiV1RunsJobidTestsuiteTestSuiteResponsesNameGetQueryParamPagesizeEmptyValue - Project : projecttest16 Job : UAT Env : UAT Region : US_WEST Result : fail Status Code : 404 Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=ODhjOTJmMmUtMTIwZC00NzhhLWJkYWQtMjJjMzQ4YTI4NjAy; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Fri, 16 Nov 2018 05:40:21 GMT]} Endpoint : http://13.56.210.25/api/v1/api/v1/runs/PrElm9RK/testSuite/test-suite-responses/PrElm9RK?pageSize= Request : Response : { "timestamp" : "2018-11-16T05:40:21.987+0000", "status" : 404, "error" : "Not Found", "message" : "No message available", "path" : "/api/v1/api/v1/runs/PrElm9RK/testSuite/test-suite-responses/PrElm9RK" } Logs : Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 500] resolved-to [404 != 500] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]Assertion [@StatusCode != 200] resolved-to [404 != 200] result [Passed] --- FX Bot ---
non_defect
project job uat env uat region us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options set cookie content type transfer encoding date endpoint request response timestamp status error not found message no message available path api api runs testsuite test suite responses logs assertion resolved to result assertion resolved to result assertion resolved to result assertion resolved to result fx bot
0
122,242
12,146,882,392
IssuesEvent
2020-04-24 12:02:11
google/conscrypt
https://api.github.com/repos/google/conscrypt
closed
FIPS Issue
documentation
I want to use this library for AES/GCM, EC (P-256,384) and RSA (2048,3072,4096). Is Conscrypt using FIPS 140-2 validated BoringSSL implementation?
1.0
FIPS Issue - I want to use this library for AES/GCM, EC (P-256,384) and RSA (2048,3072,4096). Is Conscrypt using FIPS 140-2 validated BoringSSL implementation?
non_defect
fips issue i want to use this library for aes gcm ec p and rsa is conscrypt using fips validated boringssl implementation
0
250,995
21,408,937,750
IssuesEvent
2022-04-22 02:10:29
pombase/canto
https://api.github.com/repos/pombase/canto
closed
The phenotype is requested for the wrong allele of a multi gene phenotype
bug next needs testing interactions
Here wee1 is the 'rescued phenotye" <img width="902" alt="Screenshot 2022-03-29 at 08 58 26" src="https://user-images.githubusercontent.com/7359272/160562996-ee22c681-2a7c-4b87-9daf-1a6b6b3c5359.png"> and here cdc5... is the rescued phenotype <img width="677" alt="Screenshot 2022-03-29 at 08 55 56" src="https://user-images.githubusercontent.com/7359272/160563025-ef00f53b-eb7e-4e94-a5b4-b70f7221a3f1.png">
1.0
The phenotype is requested for the wrong allele of a multi gene phenotype - Here wee1 is the 'rescued phenotye" <img width="902" alt="Screenshot 2022-03-29 at 08 58 26" src="https://user-images.githubusercontent.com/7359272/160562996-ee22c681-2a7c-4b87-9daf-1a6b6b3c5359.png"> and here cdc5... is the rescued phenotype <img width="677" alt="Screenshot 2022-03-29 at 08 55 56" src="https://user-images.githubusercontent.com/7359272/160563025-ef00f53b-eb7e-4e94-a5b4-b70f7221a3f1.png">
non_defect
the phenotype is requested for the wrong allele of a multi gene phenotype here is the rescued phenotye img width alt screenshot at src and here is the rescued phenotype img width alt screenshot at src
0