Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
13,163
15,590,183,061
IssuesEvent
2021-03-18 09:02:43
CNPMNC-KDH/TKB
https://api.github.com/repos/CNPMNC-KDH/TKB
reopened
Student schedule app
DuongThienKhoi NguyenThanhDuy Processing...... VuDuyVietHoang
Xây dựng ứng dụng quản lý thời khóa biểu sinh viên giúp người dùng quản lý thời gian học tập tốt hơn. - [x] Hoàn thành nghiên cứu ứng dụng. - [ ] Hoàn thành kham khảo giao diện template. - [ ] Hoàn thành thống nhất chủ đề giao diện. - [ ] Hoàn thành xây dựng hệ cơ sở dữ liệu. - [ ] Hoàn thành log in, sign in, forgot. - [ ] Hoàn thành alarm. - [ ] Hoàn thành thời khóa biểu. - [ ] Hoàn thành xây dựng ứng dụng. - [ ] Chạy thử ứng dụng trên nhiều nền tảng. - [ ] Kiểm thử và sửa lỗi. - [ ] Phát hành sản phẩm lên các nền tảng cửa hàng trực tuyến.
1.0
Student schedule app - Xây dựng ứng dụng quản lý thời khóa biểu sinh viên giúp người dùng quản lý thời gian học tập tốt hơn. - [x] Hoàn thành nghiên cứu ứng dụng. - [ ] Hoàn thành kham khảo giao diện template. - [ ] Hoàn thành thống nhất chủ đề giao diện. - [ ] Hoàn thành xây dựng hệ cơ sở dữ liệu. - [ ] Hoàn thành log in, sign in, forgot. - [ ] Hoàn thành alarm. - [ ] Hoàn thành thời khóa biểu. - [ ] Hoàn thành xây dựng ứng dụng. - [ ] Chạy thử ứng dụng trên nhiều nền tảng. - [ ] Kiểm thử và sửa lỗi. - [ ] Phát hành sản phẩm lên các nền tảng cửa hàng trực tuyến.
process
student schedule app xây dựng ứng dụng quản lý thời khóa biểu sinh viên giúp người dùng quản lý thời gian học tập tốt hơn hoàn thành nghiên cứu ứng dụng hoàn thành kham khảo giao diện template hoàn thành thống nhất chủ đề giao diện hoàn thành xây dựng hệ cơ sở dữ liệu hoàn thành log in sign in forgot hoàn thành alarm hoàn thành thời khóa biểu hoàn thành xây dựng ứng dụng chạy thử ứng dụng trên nhiều nền tảng kiểm thử và sửa lỗi phát hành sản phẩm lên các nền tảng cửa hàng trực tuyến
1
13,949
16,723,743,167
IssuesEvent
2021-06-10 10:24:23
hochschule-darmstadt/openartbrowser
https://api.github.com/repos/hochschule-darmstadt/openartbrowser
opened
Refactor the data handling during the crawling process
etl process feature refactoring
**Reason (Why?)** The crawler is rather slow, the current runtime is roghly 12 hours. This is in part due to very poorly implemented array handling. One example can be found in the [script for adding youtube videos to items](https://github.com/hochschule-darmstadt/openartbrowser/blob/staging/etl/data_enhancement/add_youtube_videos.py) from line 143 - 150: ``` # Set video attribute entries_added_count = 0 for entity in entities: if entity[ID] in videos: entries_added_count += 1 entity[VIDEOS] = videos[entity[ID]] ``` Here youtube videos get added to entities, e.g. artworks or movements. As evident the whole entities array ( for artworks currently ~500000) is processed to add a few youtube videos (~300 videos for artworks). This is just one example, similar array handling is done at multiple points in the etl process due to the entities being lists of dicts (so individual adressing by the QID is not possible) **Solution (What?)** This is bad design, especially with an increasing number of entities. Therefore the etl process should be refactored using, for example, the [pandas](https://pypi.org/project/pandas/) library. Pandas, or similar libraries are specifically designed to handle and manipulate big data and offer fast and reliable functionality for array processing and saving and loading json files. **Relation to other Issues** The issue arose after #452 where a lot of new entities were added **Acceptance criteria** If done correctly the crawl should be a lot quicker, more stable and less resource hungry as the current one.
1.0
Refactor the data handling during the crawling process - **Reason (Why?)** The crawler is rather slow, the current runtime is roghly 12 hours. This is in part due to very poorly implemented array handling. One example can be found in the [script for adding youtube videos to items](https://github.com/hochschule-darmstadt/openartbrowser/blob/staging/etl/data_enhancement/add_youtube_videos.py) from line 143 - 150: ``` # Set video attribute entries_added_count = 0 for entity in entities: if entity[ID] in videos: entries_added_count += 1 entity[VIDEOS] = videos[entity[ID]] ``` Here youtube videos get added to entities, e.g. artworks or movements. As evident the whole entities array ( for artworks currently ~500000) is processed to add a few youtube videos (~300 videos for artworks). This is just one example, similar array handling is done at multiple points in the etl process due to the entities being lists of dicts (so individual adressing by the QID is not possible) **Solution (What?)** This is bad design, especially with an increasing number of entities. Therefore the etl process should be refactored using, for example, the [pandas](https://pypi.org/project/pandas/) library. Pandas, or similar libraries are specifically designed to handle and manipulate big data and offer fast and reliable functionality for array processing and saving and loading json files. **Relation to other Issues** The issue arose after #452 where a lot of new entities were added **Acceptance criteria** If done correctly the crawl should be a lot quicker, more stable and less resource hungry as the current one.
process
refactor the data handling during the crawling process reason why the crawler is rather slow the current runtime is roghly hours this is in part due to very poorly implemented array handling one example can be found in the from line set video attribute entries added count for entity in entities if entity in videos entries added count entity videos here youtube videos get added to entities e g artworks or movements as evident the whole entities array for artworks currently is processed to add a few youtube videos videos for artworks this is just one example similar array handling is done at multiple points in the etl process due to the entities being lists of dicts so individual adressing by the qid is not possible solution what this is bad design especially with an increasing number of entities therefore the etl process should be refactored using for example the library pandas or similar libraries are specifically designed to handle and manipulate big data and offer fast and reliable functionality for array processing and saving and loading json files relation to other issues the issue arose after where a lot of new entities were added acceptance criteria if done correctly the crawl should be a lot quicker more stable and less resource hungry as the current one
1
26,239
19,786,240,780
IssuesEvent
2022-01-18 07:11:28
happy-travel/agent-app-project
https://api.github.com/repos/happy-travel/agent-app-project
closed
Development a basic supplier service infrastructure
backend infrastructure suppliers sunpu
It is necessary to develop the basic infrastructure of the provider service (database, logging connection, secrets from Vault, Swagger, etc.)
1.0
Development a basic supplier service infrastructure - It is necessary to develop the basic infrastructure of the provider service (database, logging connection, secrets from Vault, Swagger, etc.)
non_process
development a basic supplier service infrastructure it is necessary to develop the basic infrastructure of the provider service database logging connection secrets from vault swagger etc
0
1,849
4,647,825,974
IssuesEvent
2016-10-01 18:26:18
opentrials/opentrials
https://api.github.com/repos/opentrials/opentrials
closed
Remove 0-valued identifiers
4. Ready for Review Processors
Some of our sources have invalid identifiers like `NCT00000000`, `U0000-0000-0000`, `ISRCTN00000000` and others. These end up messing with our deduplication process, because the records end up having a mix of valid and invalid identifiers like: ```python # Record 1 # https://www.clinicaltrialsregister.eu/ctr-search/trial/2005-001059-39/PT { "nct": "NCT00202878", "euctr": "EUCTR2005-001059-39", "isrctn": "ISRCTN00000000" } # Record 2 # https://www.clinicaltrialsregister.eu/ctr-search/trial/2014-000042-30/IT { "nct": "NCT02300558", "who": "U0000-0000-0000", "euctr": "EUCTR2014-000042-30", "isrctn": "ISRCTN00000000" } ``` Our deduplication process will consider these 2 records as being the same, because they share the identifier `ISRCTN00000000`. So we'll end up thinking that trials `NCT02300558` and `NCT00202878` are the same, which is wrong. This is an issue with the source data, so we should write a report and ask them to fix. However, in the meantime, we can do a simple check for the identifiers: if all their numbers are `0`, consider it as invalid and don't use it. From a brief research, I could only find this issue with trial on EUCTR and HRA research summaries (e.g. http://www.hra.nhs.uk/news/research-summaries/tacrolimus-and-glucose-metabolism-in-renal-transplantation-v-1-1/)
1.0
Remove 0-valued identifiers - Some of our sources have invalid identifiers like `NCT00000000`, `U0000-0000-0000`, `ISRCTN00000000` and others. These end up messing with our deduplication process, because the records end up having a mix of valid and invalid identifiers like: ```python # Record 1 # https://www.clinicaltrialsregister.eu/ctr-search/trial/2005-001059-39/PT { "nct": "NCT00202878", "euctr": "EUCTR2005-001059-39", "isrctn": "ISRCTN00000000" } # Record 2 # https://www.clinicaltrialsregister.eu/ctr-search/trial/2014-000042-30/IT { "nct": "NCT02300558", "who": "U0000-0000-0000", "euctr": "EUCTR2014-000042-30", "isrctn": "ISRCTN00000000" } ``` Our deduplication process will consider these 2 records as being the same, because they share the identifier `ISRCTN00000000`. So we'll end up thinking that trials `NCT02300558` and `NCT00202878` are the same, which is wrong. This is an issue with the source data, so we should write a report and ask them to fix. However, in the meantime, we can do a simple check for the identifiers: if all their numbers are `0`, consider it as invalid and don't use it. From a brief research, I could only find this issue with trial on EUCTR and HRA research summaries (e.g. http://www.hra.nhs.uk/news/research-summaries/tacrolimus-and-glucose-metabolism-in-renal-transplantation-v-1-1/)
process
remove valued identifiers some of our sources have invalid identifiers like and others these end up messing with our deduplication process because the records end up having a mix of valid and invalid identifiers like python record nct euctr isrctn record nct who euctr isrctn our deduplication process will consider these records as being the same because they share the identifier so we ll end up thinking that trials and are the same which is wrong this is an issue with the source data so we should write a report and ask them to fix however in the meantime we can do a simple check for the identifiers if all their numbers are consider it as invalid and don t use it from a brief research i could only find this issue with trial on euctr and hra research summaries e g
1
512,651
14,906,335,471
IssuesEvent
2021-01-22 00:20:19
lemonsaurus/blackbox
https://api.github.com/repos/lemonsaurus/blackbox
closed
Storage handler: Google Drive
area: back-end priority: critical status: WIP
A storage handler that allows us to upload stuff to Google Drive, using the GDrive API.
1.0
Storage handler: Google Drive - A storage handler that allows us to upload stuff to Google Drive, using the GDrive API.
non_process
storage handler google drive a storage handler that allows us to upload stuff to google drive using the gdrive api
0
215,837
7,298,525,600
IssuesEvent
2018-02-26 17:22:48
manskou/Nerfus-Metera-issues
https://api.github.com/repos/manskou/Nerfus-Metera-issues
opened
Loud percussive sound that plays even when phone is muted
low priority software bug
Possibly an engine bug, happened three times in the same run before and whilst fighting Fitas boss.
1.0
Loud percussive sound that plays even when phone is muted - Possibly an engine bug, happened three times in the same run before and whilst fighting Fitas boss.
non_process
loud percussive sound that plays even when phone is muted possibly an engine bug happened three times in the same run before and whilst fighting fitas boss
0
681
3,152,709,583
IssuesEvent
2015-09-16 14:59:24
rg3/youtube-dl
https://api.github.com/repos/rg3/youtube-dl
closed
Using ':' in -o breaks ffmpeg
postprocessors
I was trying to download a full playlist from youtube, and I wanted the videos numbered so I could watch in order, so I used: youtube-dl --playlist-reverse -f bestvideo+171/bestvideo+140 -o "%(autonumber)s: %(title)s-%(id)s.%(ext)s" https://www.youtube.com/playlist?list=PL26B36130624FCAF5 The unfortunate side effect is that ffmpeg believes that the filename is specifying a protocol. I could just use a different character, but I was able to fix it by modifying FFmpegPostProcessor.run_ffmpeg_multiple_files and simply prepending 'file:' to each input_path and the out_path.
1.0
Using ':' in -o breaks ffmpeg - I was trying to download a full playlist from youtube, and I wanted the videos numbered so I could watch in order, so I used: youtube-dl --playlist-reverse -f bestvideo+171/bestvideo+140 -o "%(autonumber)s: %(title)s-%(id)s.%(ext)s" https://www.youtube.com/playlist?list=PL26B36130624FCAF5 The unfortunate side effect is that ffmpeg believes that the filename is specifying a protocol. I could just use a different character, but I was able to fix it by modifying FFmpegPostProcessor.run_ffmpeg_multiple_files and simply prepending 'file:' to each input_path and the out_path.
process
using in o breaks ffmpeg i was trying to download a full playlist from youtube and i wanted the videos numbered so i could watch in order so i used youtube dl playlist reverse f bestvideo bestvideo o autonumber s title s id s ext s the unfortunate side effect is that ffmpeg believes that the filename is specifying a protocol i could just use a different character but i was able to fix it by modifying ffmpegpostprocessor run ffmpeg multiple files and simply prepending file to each input path and the out path
1
22,693
31,996,544,481
IssuesEvent
2023-09-21 09:32:43
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
[Mirror]
P2 type: process more data needed team-OSS mirror request
### Please list the URLs of the archives you'd like to mirror: I have a test project, it is configured with maven and TestNG, it works perfectly fine with it. I want to use Bazel to run the same project. I have created WORKSPACE and BUILD files for each directory, which builds perfectly fine. I want to run the test using the 'testng_regressions.xml' in this direcotry: src/test/resources/testrunners/testng_regressions.xml This is the BUILD file for 'testng_regressions.xml', ``` load("@rules_java//java:defs.bzl", "java_test") java_library( name = "main_project_files", srcs = glob(['src/main/java/**/*.java']), runtime_deps = [ "//Web/src/main/java/io/package/com/constants:app_constants", "//Web/src/main/java/io/package/com/factory:playwright_factory", "//Web/src/main/java/io/package/com/listeners:extent_report_listener", "//Web/src/main/java/io/package/com/pages:login_page", "//Web/src/main/java/io/package/com/steps:login_steps", ], ) java_library( name = "test_project_files", srcs = glob(['src/test/java/**/*.java']), runtime_deps = [":main_project_files", "//Web/src/test/java/io/package/com/base:base_test", "//Web/src/test/java/io/package/com/tests:project_test_files"], ) # java_test( name = "testng_regressions", size = "small", timeout = "short", tags = ["regression"], runtime_deps = [ ":main_project_files", ":test_project_files" ], use_testrunner = False, main_class='org.testng.TestNG', args=['-testjar','libtest_project_files.jar','-verbose','2'], ) ``` This is the content of the testNG file ``` <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE suite SYSTEM "http://testng.org/testng-1.0.dtd"> <suite name="Open Cart Test Regression PW Suite" thread-count="5" parallel="tests" verbose="4"> <listeners> <listener class-name="io.christdoes.com.listeners.ExtentReportListener" /> </listeners> <test name="Open Cart Home Page Test_chrome"> <parameter name="browser" value="chrome" /> <parameter name="language" value="en" /> <classes> <class name="io.christdoes.com.tests.HomePageTest" /> </classes> </test> <test name="Open Cart Login Page Test_safari"> <parameter name="browser" value="safari" /> <parameter name="language" value="nl" /> <classes> <class name="io.christdoes.com.tests.HomePageTest" /> </classes> </test> </suite> ``` When I run the test, it always give a non-detailed error ``` FAIL: //Web/src/test/resources/testrunners:testng_regressions (see /private/var/tmp/_bazel_christdoes/44565daf5d57f20e3e02fd7160ffadab/execroot/__main__/bazel-out/darwin-fastbuild/testlogs/Web/src/test/resources/testrunners/testng_regressions/test.log) INFO: Elapsed time: 91.557s, Critical Path: 31.65s INFO: 17 processes: 3 internal, 6 darwin-sandbox, 8 worker. INFO: Build completed, 1 test FAILED, 17 total actions //Web/src/test/resources/testrunners:testng_regressions FAILED in 2.1s /private/var/tmp/_bazel_user/44565daf5d57f20e3e02fd7160ffadab/execroot/__main__/bazel-out/darwin-fastbuild/testlogs/Web/src/test/resources/testrunners/testng_regressions/test.log ``` Executed 1 out of 1 test: 1 fails locally.
1.0
[Mirror] - ### Please list the URLs of the archives you'd like to mirror: I have a test project, it is configured with maven and TestNG, it works perfectly fine with it. I want to use Bazel to run the same project. I have created WORKSPACE and BUILD files for each directory, which builds perfectly fine. I want to run the test using the 'testng_regressions.xml' in this direcotry: src/test/resources/testrunners/testng_regressions.xml This is the BUILD file for 'testng_regressions.xml', ``` load("@rules_java//java:defs.bzl", "java_test") java_library( name = "main_project_files", srcs = glob(['src/main/java/**/*.java']), runtime_deps = [ "//Web/src/main/java/io/package/com/constants:app_constants", "//Web/src/main/java/io/package/com/factory:playwright_factory", "//Web/src/main/java/io/package/com/listeners:extent_report_listener", "//Web/src/main/java/io/package/com/pages:login_page", "//Web/src/main/java/io/package/com/steps:login_steps", ], ) java_library( name = "test_project_files", srcs = glob(['src/test/java/**/*.java']), runtime_deps = [":main_project_files", "//Web/src/test/java/io/package/com/base:base_test", "//Web/src/test/java/io/package/com/tests:project_test_files"], ) # java_test( name = "testng_regressions", size = "small", timeout = "short", tags = ["regression"], runtime_deps = [ ":main_project_files", ":test_project_files" ], use_testrunner = False, main_class='org.testng.TestNG', args=['-testjar','libtest_project_files.jar','-verbose','2'], ) ``` This is the content of the testNG file ``` <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE suite SYSTEM "http://testng.org/testng-1.0.dtd"> <suite name="Open Cart Test Regression PW Suite" thread-count="5" parallel="tests" verbose="4"> <listeners> <listener class-name="io.christdoes.com.listeners.ExtentReportListener" /> </listeners> <test name="Open Cart Home Page Test_chrome"> <parameter name="browser" value="chrome" /> <parameter name="language" value="en" /> <classes> <class name="io.christdoes.com.tests.HomePageTest" /> </classes> </test> <test name="Open Cart Login Page Test_safari"> <parameter name="browser" value="safari" /> <parameter name="language" value="nl" /> <classes> <class name="io.christdoes.com.tests.HomePageTest" /> </classes> </test> </suite> ``` When I run the test, it always give a non-detailed error ``` FAIL: //Web/src/test/resources/testrunners:testng_regressions (see /private/var/tmp/_bazel_christdoes/44565daf5d57f20e3e02fd7160ffadab/execroot/__main__/bazel-out/darwin-fastbuild/testlogs/Web/src/test/resources/testrunners/testng_regressions/test.log) INFO: Elapsed time: 91.557s, Critical Path: 31.65s INFO: 17 processes: 3 internal, 6 darwin-sandbox, 8 worker. INFO: Build completed, 1 test FAILED, 17 total actions //Web/src/test/resources/testrunners:testng_regressions FAILED in 2.1s /private/var/tmp/_bazel_user/44565daf5d57f20e3e02fd7160ffadab/execroot/__main__/bazel-out/darwin-fastbuild/testlogs/Web/src/test/resources/testrunners/testng_regressions/test.log ``` Executed 1 out of 1 test: 1 fails locally.
process
please list the urls of the archives you d like to mirror i have a test project it is configured with maven and testng it works perfectly fine with it i want to use bazel to run the same project i have created workspace and build files for each directory which builds perfectly fine i want to run the test using the testng regressions xml in this direcotry src test resources testrunners testng regressions xml this is the build file for testng regressions xml load rules java java defs bzl java test java library name main project files srcs glob runtime deps web src main java io package com constants app constants web src main java io package com factory playwright factory web src main java io package com listeners extent report listener web src main java io package com pages login page web src main java io package com steps login steps java library name test project files srcs glob runtime deps main project files web src test java io package com base base test web src test java io package com tests project test files java test name testng regressions size small timeout short tags runtime deps main project files test project files use testrunner false main class org testng testng args this is the content of the testng file doctype suite system suite name open cart test regression pw suite thread count parallel tests verbose when i run the test it always give a non detailed error fail web src test resources testrunners testng regressions see private var tmp bazel christdoes execroot main bazel out darwin fastbuild testlogs web src test resources testrunners testng regressions test log info elapsed time critical path info processes internal darwin sandbox worker info build completed test failed total actions web src test resources testrunners testng regressions failed in private var tmp bazel user execroot main bazel out darwin fastbuild testlogs web src test resources testrunners testng regressions test log executed out of test fails locally
1
20,404
27,064,126,977
IssuesEvent
2023-02-13 22:23:51
googleapis/google-cloud-node
https://api.github.com/repos/googleapis/google-cloud-node
closed
update issue template
type: process
let's update issue template to make sure people describe the api they're referring to
1.0
update issue template - let's update issue template to make sure people describe the api they're referring to
process
update issue template let s update issue template to make sure people describe the api they re referring to
1
4,080
7,031,435,623
IssuesEvent
2017-12-26 17:39:35
nodejs/node
https://api.github.com/repos/nodejs/node
closed
NeedImmediateCallbackSetter terminates after setting global.process
process v8.x
<!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify affected core module name If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able. --> * **Version**: 8.9.3 (LTS) * **Platform**: Windows 10 Professional 64 bit * **Subsystem**: process <!-- Enter your issue details below this comment. --> This causes node 8.9.3 to crash: ```js > global.process = { __proto__: global.process, pid: 123456 } process { pid: 123456 } > process._needImmediateCallback = true true > FATAL ERROR: v8::ToLocalChecked Empty MaybeLocal. ``` I came across this problem when executing unit tests of an npm package which intentionally replaces the `process` object. The problem seems to be caused by `CheckImmediate` which is called by `NeedImmediateCallbackSetter` via libuv: https://github.com/nodejs/node/blob/ab46b8ebaa13da05513976ec2f33949aef2fae94/src/node.cc#L379-L384 `MakeCallback` fails because `callback_v` is not a function: https://github.com/nodejs/node/blob/ab46b8ebaa13da05513976ec2f33949aef2fae94/src/node.cc#L1528-L1539 node 9 is not affected.
1.0
NeedImmediateCallbackSetter terminates after setting global.process - <!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify affected core module name If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able. --> * **Version**: 8.9.3 (LTS) * **Platform**: Windows 10 Professional 64 bit * **Subsystem**: process <!-- Enter your issue details below this comment. --> This causes node 8.9.3 to crash: ```js > global.process = { __proto__: global.process, pid: 123456 } process { pid: 123456 } > process._needImmediateCallback = true true > FATAL ERROR: v8::ToLocalChecked Empty MaybeLocal. ``` I came across this problem when executing unit tests of an npm package which intentionally replaces the `process` object. The problem seems to be caused by `CheckImmediate` which is called by `NeedImmediateCallbackSetter` via libuv: https://github.com/nodejs/node/blob/ab46b8ebaa13da05513976ec2f33949aef2fae94/src/node.cc#L379-L384 `MakeCallback` fails because `callback_v` is not a function: https://github.com/nodejs/node/blob/ab46b8ebaa13da05513976ec2f33949aef2fae94/src/node.cc#L1528-L1539 node 9 is not affected.
process
needimmediatecallbacksetter terminates after setting global process thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify affected core module name if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able version lts platform windows professional bit subsystem process this causes node to crash js global process proto global process pid process pid process needimmediatecallback true true fatal error tolocalchecked empty maybelocal i came across this problem when executing unit tests of an npm package which intentionally replaces the process object the problem seems to be caused by checkimmediate which is called by needimmediatecallbacksetter via libuv makecallback fails because callback v is not a function node is not affected
1
149,803
11,924,268,850
IssuesEvent
2020-04-01 09:15:00
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
opened
An error pops up when creating one folder
:beetle: regression :gear: files 🧪 testing
**Storage Explorer Version:** 1.13.0 **Build**: [20200401.4](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3600038) **Branch**: rel/1.13.0 **Platform/OS**: Windows 10/ Linux Ubuntu 16.04/ macOS High Sierra **Architecture**: ia32/x64 **Regression From:** Previous release(1.12.0) **Steps to reproduce:** 1. Expand one storage account -> File Shares. 2. Create a new file share -> Create one new folder. 3. Check the result. **Expect Experience:** Succeed to create one folder. **Actual Experience:** An error pops up. ![image](https://user-images.githubusercontent.com/54055206/78118765-0b04a380-743a-11ea-8219-a23ca142ce9b.png)
1.0
An error pops up when creating one folder - **Storage Explorer Version:** 1.13.0 **Build**: [20200401.4](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3600038) **Branch**: rel/1.13.0 **Platform/OS**: Windows 10/ Linux Ubuntu 16.04/ macOS High Sierra **Architecture**: ia32/x64 **Regression From:** Previous release(1.12.0) **Steps to reproduce:** 1. Expand one storage account -> File Shares. 2. Create a new file share -> Create one new folder. 3. Check the result. **Expect Experience:** Succeed to create one folder. **Actual Experience:** An error pops up. ![image](https://user-images.githubusercontent.com/54055206/78118765-0b04a380-743a-11ea-8219-a23ca142ce9b.png)
non_process
an error pops up when creating one folder storage explorer version build branch rel platform os windows linux ubuntu macos high sierra architecture regression from previous release steps to reproduce expand one storage account file shares create a new file share create one new folder check the result expect experience succeed to create one folder actual experience an error pops up
0
4,343
7,246,143,770
IssuesEvent
2018-02-14 20:34:22
hardvolk/foodie-journal
https://api.github.com/repos/hardvolk/foodie-journal
closed
Subir home-component a un repo para que se haga merge con el login
In process
subir home completo para checar que el login funcione correctamente en conjunto y poder subirlo a dev
1.0
Subir home-component a un repo para que se haga merge con el login - subir home completo para checar que el login funcione correctamente en conjunto y poder subirlo a dev
process
subir home component a un repo para que se haga merge con el login subir home completo para checar que el login funcione correctamente en conjunto y poder subirlo a dev
1
12,280
14,790,580,557
IssuesEvent
2021-01-12 12:15:32
yeti-switch/yeti-web
https://api.github.com/repos/yeti-switch/yeti-web
closed
pgq processors prometheus support
PGQ Processor
pgq processors should report to **prometheus-collector** daemon counters for processed events and errors
1.0
pgq processors prometheus support - pgq processors should report to **prometheus-collector** daemon counters for processed events and errors
process
pgq processors prometheus support pgq processors should report to prometheus collector daemon counters for processed events and errors
1
11,374
14,216,503,365
IssuesEvent
2020-11-17 09:04:54
prisma/migrate
https://api.github.com/repos/prisma/migrate
opened
Detect failure from insufficient privileges in introspection (and migrate?)
kind/improvement process/candidate topic: introspect-migrate
Example from https://github.com/prisma/migrate/issues/647 run `npx prisma introspect` ``` Introspecting based on datasource defined in prisma/schema.prisma … Oops, an unexpected error occured! [libs/sql-schema-describer/src/mysql.rs:156:53] table columns not found ``` The current error is not super helpful right now.
1.0
Detect failure from insufficient privileges in introspection (and migrate?) - Example from https://github.com/prisma/migrate/issues/647 run `npx prisma introspect` ``` Introspecting based on datasource defined in prisma/schema.prisma … Oops, an unexpected error occured! [libs/sql-schema-describer/src/mysql.rs:156:53] table columns not found ``` The current error is not super helpful right now.
process
detect failure from insufficient privileges in introspection and migrate example from run npx prisma introspect introspecting based on datasource defined in prisma schema prisma … oops an unexpected error occured table columns not found the current error is not super helpful right now
1
21,231
28,321,617,951
IssuesEvent
2023-04-11 02:00:11
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Tue, 11 Apr 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### Exploring Data Geometry for Continual Learning - **Authors:** Zhi Gao, Chen Xu, Feng Li, Yunde Jia, Mehrtash Harandi, Yuwei Wu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.03931 - **Pdf link:** https://arxiv.org/pdf/2304.03931 - **Abstract** Continual learning aims to efficiently learn from a non-stationary stream of data while avoiding forgetting the knowledge of old data. In many practical applications, data complies with non-Euclidean geometry. As such, the commonly used Euclidean space cannot gracefully capture non-Euclidean geometric structures of data, leading to inferior results. In this paper, we study continual learning from a novel perspective by exploring data geometry for the non-stationary stream of data. Our method dynamically expands the geometry of the underlying space to match growing geometric structures induced by new data, and prevents forgetting by keeping geometric structures of old data into account. In doing so, making use of the mixed curvature space, we propose an incremental search scheme, through which the growing geometric structures are encoded. Then, we introduce an angular-regularization loss and a neighbor-robustness loss to train the model, capable of penalizing the change of global geometric structures and local geometric structures. Experiments show that our method achieves better performance than baseline methods designed in Euclidean space. ### Monocular 3D Human Pose Estimation for Sports Broadcasts using Partial Sports Field Registration - **Authors:** Tobias Baumgartner, Stefanie Klatt - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.04437 - **Pdf link:** https://arxiv.org/pdf/2304.04437 - **Abstract** The filming of sporting events projects and flattens the movement of athletes in the world onto a 2D broadcast image. The pixel locations of joints in these images can be detected with high validity. Recovering the actual 3D movement of the limbs (kinematics) of the athletes requires lifting these 2D pixel locations back into a third dimension, implying a certain scene geometry. The well-known line markings of sports fields allow for the calibration of the camera and for determining the actual geometry of the scene. Close-up shots of athletes are required to extract detailed kinematics, which in turn obfuscates the pertinent field markers for camera calibration. We suggest partial sports field registration, which determines a set of scene-consistent camera calibrations up to a single degree of freedom. Through joint optimization of 3D pose estimation and camera calibration, we demonstrate the successful extraction of 3D running kinematics on a 400m track. In this work, we combine advances in 2D human pose estimation and camera calibration via partial sports field registration to demonstrate an avenue for collecting valid large-scale kinematic datasets. We generate a synthetic dataset of more than 10k images in Unreal Engine 5 with different viewpoints, running styles, and body types, to show the limitations of existing monocular 3D HPE methods. Synthetic data and code are available at https://github.com/tobibaum/PartialSportsFieldReg_3DHPE. ### Event-based Camera Tracker by $\nabla$t NeRF - **Authors:** Mana Masuda, Yusuke Sekikawa, Hideo Saito - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.04559 - **Pdf link:** https://arxiv.org/pdf/2304.04559 - **Abstract** When a camera travels across a 3D world, only a fraction of pixel value changes; an event-based camera observes the change as sparse events. How can we utilize sparse events for efficient recovery of the camera pose? We show that we can recover the camera pose by minimizing the error between sparse events and the temporal gradient of the scene represented as a neural radiance field (NeRF). To enable the computation of the temporal gradient of the scene, we augment NeRF's camera pose as a time function. When the input pose to the NeRF coincides with the actual pose, the output of the temporal gradient of NeRF equals the observed intensity changes on the event's points. Using this principle, we propose an event-based camera pose tracking framework called TeGRA which realizes the pose update by using the sparse event's observation. To the best of our knowledge, this is the first camera pose estimation algorithm using the scene's implicit representation and the sparse intensity change from events. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Feature Representation Learning with Adaptive Displacement Generation and Transformer Fusion for Micro-Expression Recognition - **Authors:** Zhijun Zhai, Jianhui Zhao, Chengjiang Long, Wenju Xu, Shuangjiang He, Huijuan Zhao - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.04420 - **Pdf link:** https://arxiv.org/pdf/2304.04420 - **Abstract** Micro-expressions are spontaneous, rapid and subtle facial movements that can neither be forged nor suppressed. They are very important nonverbal communication clues, but are transient and of low intensity thus difficult to recognize. Recently deep learning based methods have been developed for micro-expression (ME) recognition using feature extraction and fusion techniques, however, targeted feature learning and efficient feature fusion still lack further study according to the ME characteristics. To address these issues, we propose a novel framework Feature Representation Learning with adaptive Displacement Generation and Transformer fusion (FRL-DGT), in which a convolutional Displacement Generation Module (DGM) with self-supervised learning is used to extract dynamic features from onset/apex frames targeted to the subsequent ME recognition task, and a well-designed Transformer Fusion mechanism composed of three Transformer-based fusion modules (local, global fusions based on AU regions and full-face fusion) is applied to extract the multi-level informative features after DGM for the final ME prediction. The extensive experiments with solid leave-one-subject-out (LOSO) evaluation results have demonstrated the superiority of our proposed FRL-DGT to state-of-the-art methods. ### Grouped Knowledge Distillation for Deep Face Recognition - **Authors:** Weisong Zhao, Xiangyu Zhu, Kaiwen Guo, Xiao-Yu Zhang, Zhen Lei - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.04462 - **Pdf link:** https://arxiv.org/pdf/2304.04462 - **Abstract** Compared with the feature-based distillation methods, logits distillation can liberalize the requirements of consistent feature dimension between teacher and student networks, while the performance is deemed inferior in face recognition. One major challenge is that the light-weight student network has difficulty fitting the target logits due to its low model capacity, which is attributed to the significant number of identities in face recognition. Therefore, we seek to probe the target logits to extract the primary knowledge related to face identity, and discard the others, to make the distillation more achievable for the student network. Specifically, there is a tail group with near-zero values in the prediction, containing minor knowledge for distillation. To provide a clear perspective of its impact, we first partition the logits into two groups, i.e., Primary Group and Secondary Group, according to the cumulative probability of the softened prediction. Then, we reorganize the Knowledge Distillation (KD) loss of grouped logits into three parts, i.e., Primary-KD, Secondary-KD, and Binary-KD. Primary-KD refers to distilling the primary knowledge from the teacher, Secondary-KD aims to refine minor knowledge but increases the difficulty of distillation, and Binary-KD ensures the consistency of knowledge distribution between teacher and student. We experimentally found that (1) Primary-KD and Binary-KD are indispensable for KD, and (2) Secondary-KD is the culprit restricting KD at the bottleneck. Therefore, we propose a Grouped Knowledge Distillation (GKD) that retains the Primary-KD and Binary-KD but omits Secondary-KD in the ultimate KD loss calculation. Extensive experimental results on popular face recognition benchmarks demonstrate the superiority of proposed GKD over state-of-the-art methods. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### Neural Residual Radiance Fields for Streamably Free-Viewpoint Videos - **Authors:** Liao Wang, Qiang Hu, Qihan He, Ziyu Wang, Jingyi Yu, Tinne Tuytelaars, Lan Xu, Minye Wu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.04452 - **Pdf link:** https://arxiv.org/pdf/2304.04452 - **Abstract** The success of the Neural Radiance Fields (NeRFs) for modeling and free-view rendering static objects has inspired numerous attempts on dynamic scenes. Current techniques that utilize neural rendering for facilitating free-view videos (FVVs) are restricted to either offline rendering or are capable of processing only brief sequences with minimal motion. In this paper, we present a novel technique, Residual Radiance Field or ReRF, as a highly compact neural representation to achieve real-time FVV rendering on long-duration dynamic scenes. ReRF explicitly models the residual information between adjacent timestamps in the spatial-temporal feature space, with a global coordinate-based tiny MLP as the feature decoder. Specifically, ReRF employs a compact motion grid along with a residual feature grid to exploit inter-frame feature similarities. We show such a strategy can handle large motions without sacrificing quality. We further present a sequential training scheme to maintain the smoothness and the sparsity of the motion/residual grids. Based on ReRF, we design a special FVV codec that achieves three orders of magnitudes compression rate and provides a companion ReRF player to support online streaming of long-duration FVVs of dynamic scenes. Extensive experiments demonstrate the effectiveness of ReRF for compactly representing dynamic radiance fields, enabling an unprecedented free-viewpoint viewing experience in speed and quality. ### Are Visual Recognition Models Robust to Image Compression? - **Authors:** João Maria Janeiro, Stanislav Frolov, Alaaeldin El-Nouby, Jakob Verbeek - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2304.04518 - **Pdf link:** https://arxiv.org/pdf/2304.04518 - **Abstract** Reducing the data footprint of visual content via image compression is essential to reduce storage requirements, but also to reduce the bandwidth and latency requirements for transmission. In particular, the use of compressed images allows for faster transfer of data, and faster response times for visual recognition in edge devices that rely on cloud-based services. In this paper, we first analyze the impact of image compression using traditional codecs, as well as recent state-of-the-art neural compression approaches, on three visual recognition tasks: image classification, object detection, and semantic segmentation. We consider a wide range of compression levels, ranging from 0.1 to 2 bits-per-pixel (bpp). We find that for all three tasks, the recognition ability is significantly impacted when using strong compression. For example, for segmentation mIoU is reduced from 44.5 to 30.5 mIoU when compressing to 0.1 bpp using the best compression model we evaluated. Second, we test to what extent this performance drop can be ascribed to a loss of relevant information in the compressed image, or to a lack of generalization of visual recognition models to images with compression artefacts. We find that to a large extent the performance loss is due to the latter: by finetuning the recognition models on compressed training images, most of the performance loss is recovered. For example, bringing segmentation accuracy back up to 42 mIoU, i.e. recovering 82% of the original drop in accuracy. ## Keyword: RAW ### GANHead: Towards Generative Animatable Neural Head Avatars - **Authors:** Sijing Wu, Yichao Yan, Yunhao Li, Yuhao Cheng, Wenhan Zhu, Ke Gao, Xiaobo Li, Guangtao Zhai - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.03950 - **Pdf link:** https://arxiv.org/pdf/2304.03950 - **Abstract** To bring digital avatars into people's lives, it is highly demanded to efficiently generate complete, realistic, and animatable head avatars. This task is challenging, and it is difficult for existing methods to satisfy all the requirements at once. To achieve these goals, we propose GANHead (Generative Animatable Neural Head Avatar), a novel generative head model that takes advantages of both the fine-grained control over the explicit expression parameters and the realistic rendering results of implicit representations. Specifically, GANHead represents coarse geometry, fine-gained details and texture via three networks in canonical space to obtain the ability to generate complete and realistic head avatars. To achieve flexible animation, we define the deformation filed by standard linear blend skinning (LBS), with the learned continuous pose and expression bases and LBS weights. This allows the avatars to be directly animated by FLAME parameters and generalize well to unseen poses and expressions. Compared to state-of-the-art (SOTA) methods, GANHead achieves superior performance on head avatar generation and raw scan fitting. ### Unsupervised Sampling Promoting for Stochastic Human Trajectory Prediction - **Authors:** Guangyi Chen, Zhenhao Chen, Shunxing Fan, Kun Zhang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.04298 - **Pdf link:** https://arxiv.org/pdf/2304.04298 - **Abstract** The indeterminate nature of human motion requires trajectory prediction systems to use a probabilistic model to formulate the multi-modality phenomenon and infer a finite set of future trajectories. However, the inference processes of most existing methods rely on Monte Carlo random sampling, which is insufficient to cover the realistic paths with finite samples, due to the long tail effect of the predicted distribution. To promote the sampling process of stochastic prediction, we propose a novel method, called BOsampler, to adaptively mine potential paths with Bayesian optimization in an unsupervised manner, as a sequential design strategy in which new prediction is dependent on the previously drawn samples. Specifically, we model the trajectory sampling as a Gaussian process and construct an acquisition function to measure the potential sampling value. This acquisition function applies the original distribution as prior and encourages exploring paths in the long-tail region. This sampling method can be integrated with existing stochastic predictive models without retraining. Experimental results on various baseline methods demonstrate the effectiveness of our method. ### Federated Incremental Semantic Segmentation - **Authors:** Jiahua Dong, Duzhen Zhang, Yang Cong, Wei Cong, Henghui Ding, Dengxin Dai - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.04620 - **Pdf link:** https://arxiv.org/pdf/2304.04620 - **Abstract** Federated learning-based semantic segmentation (FSS) has drawn widespread attention via decentralized training on local clients. However, most FSS models assume categories are fixed in advance, thus heavily undergoing forgetting on old categories in practical applications where local clients receive new categories incrementally while have no memory storage to access old classes. Moreover, new clients collecting novel classes may join in the global training of FSS, which further exacerbates catastrophic forgetting. To surmount the above challenges, we propose a Forgetting-Balanced Learning (FBL) model to address heterogeneous forgetting on old classes from both intra-client and inter-client aspects. Specifically, under the guidance of pseudo labels generated via adaptive class-balanced pseudo labeling, we develop a forgetting-balanced semantic compensation loss and a forgetting-balanced relation consistency loss to rectify intra-client heterogeneous forgetting of old categories with background shift. It performs balanced gradient propagation and relation consistency distillation within local clients. Moreover, to tackle heterogeneous forgetting from inter-client aspect, we propose a task transition monitor. It can identify new classes under privacy protection and store the latest old global model for relation distillation. Qualitative experiments reveal large improvement of our model against comparison methods. The code is available at https://github.com/JiahuaDong/FISS. ### Do We Train on Test Data? The Impact of Near-Duplicates on License Plate Recognition - **Authors:** Rayson Laroca, Valter Estevam, Alceu S. Britto Jr., Rodrigo Minetto, David Menotti - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.04653 - **Pdf link:** https://arxiv.org/pdf/2304.04653 - **Abstract** This work draws attention to the large fraction of near-duplicates in the training and test sets of datasets widely adopted in License Plate Recognition (LPR) research. These duplicates refer to images that, although different, show the same license plate. Our experiments, conducted on the two most popular datasets in the field, show a substantial decrease in recognition rate when six well-known models are trained and tested under fair splits, that is, in the absence of duplicates in the training and test sets. Moreover, in one of the datasets, the ranking of models changed considerably when they were trained and tested under duplicate-free splits. These findings suggest that such duplicates have significantly biased the evaluation and development of deep learning-based models for LPR. The list of near-duplicates we have found and proposals for fair splits are publicly available for further research at https://raysonlaroca.github.io/supp/lpr-train-on-test/ ## Keyword: raw image There is no result
2.0
New submissions for Tue, 11 Apr 23 - ## Keyword: events ### Exploring Data Geometry for Continual Learning - **Authors:** Zhi Gao, Chen Xu, Feng Li, Yunde Jia, Mehrtash Harandi, Yuwei Wu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.03931 - **Pdf link:** https://arxiv.org/pdf/2304.03931 - **Abstract** Continual learning aims to efficiently learn from a non-stationary stream of data while avoiding forgetting the knowledge of old data. In many practical applications, data complies with non-Euclidean geometry. As such, the commonly used Euclidean space cannot gracefully capture non-Euclidean geometric structures of data, leading to inferior results. In this paper, we study continual learning from a novel perspective by exploring data geometry for the non-stationary stream of data. Our method dynamically expands the geometry of the underlying space to match growing geometric structures induced by new data, and prevents forgetting by keeping geometric structures of old data into account. In doing so, making use of the mixed curvature space, we propose an incremental search scheme, through which the growing geometric structures are encoded. Then, we introduce an angular-regularization loss and a neighbor-robustness loss to train the model, capable of penalizing the change of global geometric structures and local geometric structures. Experiments show that our method achieves better performance than baseline methods designed in Euclidean space. ### Monocular 3D Human Pose Estimation for Sports Broadcasts using Partial Sports Field Registration - **Authors:** Tobias Baumgartner, Stefanie Klatt - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.04437 - **Pdf link:** https://arxiv.org/pdf/2304.04437 - **Abstract** The filming of sporting events projects and flattens the movement of athletes in the world onto a 2D broadcast image. The pixel locations of joints in these images can be detected with high validity. Recovering the actual 3D movement of the limbs (kinematics) of the athletes requires lifting these 2D pixel locations back into a third dimension, implying a certain scene geometry. The well-known line markings of sports fields allow for the calibration of the camera and for determining the actual geometry of the scene. Close-up shots of athletes are required to extract detailed kinematics, which in turn obfuscates the pertinent field markers for camera calibration. We suggest partial sports field registration, which determines a set of scene-consistent camera calibrations up to a single degree of freedom. Through joint optimization of 3D pose estimation and camera calibration, we demonstrate the successful extraction of 3D running kinematics on a 400m track. In this work, we combine advances in 2D human pose estimation and camera calibration via partial sports field registration to demonstrate an avenue for collecting valid large-scale kinematic datasets. We generate a synthetic dataset of more than 10k images in Unreal Engine 5 with different viewpoints, running styles, and body types, to show the limitations of existing monocular 3D HPE methods. Synthetic data and code are available at https://github.com/tobibaum/PartialSportsFieldReg_3DHPE. ### Event-based Camera Tracker by $\nabla$t NeRF - **Authors:** Mana Masuda, Yusuke Sekikawa, Hideo Saito - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.04559 - **Pdf link:** https://arxiv.org/pdf/2304.04559 - **Abstract** When a camera travels across a 3D world, only a fraction of pixel value changes; an event-based camera observes the change as sparse events. How can we utilize sparse events for efficient recovery of the camera pose? We show that we can recover the camera pose by minimizing the error between sparse events and the temporal gradient of the scene represented as a neural radiance field (NeRF). To enable the computation of the temporal gradient of the scene, we augment NeRF's camera pose as a time function. When the input pose to the NeRF coincides with the actual pose, the output of the temporal gradient of NeRF equals the observed intensity changes on the event's points. Using this principle, we propose an event-based camera pose tracking framework called TeGRA which realizes the pose update by using the sparse event's observation. To the best of our knowledge, this is the first camera pose estimation algorithm using the scene's implicit representation and the sparse intensity change from events. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Feature Representation Learning with Adaptive Displacement Generation and Transformer Fusion for Micro-Expression Recognition - **Authors:** Zhijun Zhai, Jianhui Zhao, Chengjiang Long, Wenju Xu, Shuangjiang He, Huijuan Zhao - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.04420 - **Pdf link:** https://arxiv.org/pdf/2304.04420 - **Abstract** Micro-expressions are spontaneous, rapid and subtle facial movements that can neither be forged nor suppressed. They are very important nonverbal communication clues, but are transient and of low intensity thus difficult to recognize. Recently deep learning based methods have been developed for micro-expression (ME) recognition using feature extraction and fusion techniques, however, targeted feature learning and efficient feature fusion still lack further study according to the ME characteristics. To address these issues, we propose a novel framework Feature Representation Learning with adaptive Displacement Generation and Transformer fusion (FRL-DGT), in which a convolutional Displacement Generation Module (DGM) with self-supervised learning is used to extract dynamic features from onset/apex frames targeted to the subsequent ME recognition task, and a well-designed Transformer Fusion mechanism composed of three Transformer-based fusion modules (local, global fusions based on AU regions and full-face fusion) is applied to extract the multi-level informative features after DGM for the final ME prediction. The extensive experiments with solid leave-one-subject-out (LOSO) evaluation results have demonstrated the superiority of our proposed FRL-DGT to state-of-the-art methods. ### Grouped Knowledge Distillation for Deep Face Recognition - **Authors:** Weisong Zhao, Xiangyu Zhu, Kaiwen Guo, Xiao-Yu Zhang, Zhen Lei - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.04462 - **Pdf link:** https://arxiv.org/pdf/2304.04462 - **Abstract** Compared with the feature-based distillation methods, logits distillation can liberalize the requirements of consistent feature dimension between teacher and student networks, while the performance is deemed inferior in face recognition. One major challenge is that the light-weight student network has difficulty fitting the target logits due to its low model capacity, which is attributed to the significant number of identities in face recognition. Therefore, we seek to probe the target logits to extract the primary knowledge related to face identity, and discard the others, to make the distillation more achievable for the student network. Specifically, there is a tail group with near-zero values in the prediction, containing minor knowledge for distillation. To provide a clear perspective of its impact, we first partition the logits into two groups, i.e., Primary Group and Secondary Group, according to the cumulative probability of the softened prediction. Then, we reorganize the Knowledge Distillation (KD) loss of grouped logits into three parts, i.e., Primary-KD, Secondary-KD, and Binary-KD. Primary-KD refers to distilling the primary knowledge from the teacher, Secondary-KD aims to refine minor knowledge but increases the difficulty of distillation, and Binary-KD ensures the consistency of knowledge distribution between teacher and student. We experimentally found that (1) Primary-KD and Binary-KD are indispensable for KD, and (2) Secondary-KD is the culprit restricting KD at the bottleneck. Therefore, we propose a Grouped Knowledge Distillation (GKD) that retains the Primary-KD and Binary-KD but omits Secondary-KD in the ultimate KD loss calculation. Extensive experimental results on popular face recognition benchmarks demonstrate the superiority of proposed GKD over state-of-the-art methods. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### Neural Residual Radiance Fields for Streamably Free-Viewpoint Videos - **Authors:** Liao Wang, Qiang Hu, Qihan He, Ziyu Wang, Jingyi Yu, Tinne Tuytelaars, Lan Xu, Minye Wu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.04452 - **Pdf link:** https://arxiv.org/pdf/2304.04452 - **Abstract** The success of the Neural Radiance Fields (NeRFs) for modeling and free-view rendering static objects has inspired numerous attempts on dynamic scenes. Current techniques that utilize neural rendering for facilitating free-view videos (FVVs) are restricted to either offline rendering or are capable of processing only brief sequences with minimal motion. In this paper, we present a novel technique, Residual Radiance Field or ReRF, as a highly compact neural representation to achieve real-time FVV rendering on long-duration dynamic scenes. ReRF explicitly models the residual information between adjacent timestamps in the spatial-temporal feature space, with a global coordinate-based tiny MLP as the feature decoder. Specifically, ReRF employs a compact motion grid along with a residual feature grid to exploit inter-frame feature similarities. We show such a strategy can handle large motions without sacrificing quality. We further present a sequential training scheme to maintain the smoothness and the sparsity of the motion/residual grids. Based on ReRF, we design a special FVV codec that achieves three orders of magnitudes compression rate and provides a companion ReRF player to support online streaming of long-duration FVVs of dynamic scenes. Extensive experiments demonstrate the effectiveness of ReRF for compactly representing dynamic radiance fields, enabling an unprecedented free-viewpoint viewing experience in speed and quality. ### Are Visual Recognition Models Robust to Image Compression? - **Authors:** João Maria Janeiro, Stanislav Frolov, Alaaeldin El-Nouby, Jakob Verbeek - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2304.04518 - **Pdf link:** https://arxiv.org/pdf/2304.04518 - **Abstract** Reducing the data footprint of visual content via image compression is essential to reduce storage requirements, but also to reduce the bandwidth and latency requirements for transmission. In particular, the use of compressed images allows for faster transfer of data, and faster response times for visual recognition in edge devices that rely on cloud-based services. In this paper, we first analyze the impact of image compression using traditional codecs, as well as recent state-of-the-art neural compression approaches, on three visual recognition tasks: image classification, object detection, and semantic segmentation. We consider a wide range of compression levels, ranging from 0.1 to 2 bits-per-pixel (bpp). We find that for all three tasks, the recognition ability is significantly impacted when using strong compression. For example, for segmentation mIoU is reduced from 44.5 to 30.5 mIoU when compressing to 0.1 bpp using the best compression model we evaluated. Second, we test to what extent this performance drop can be ascribed to a loss of relevant information in the compressed image, or to a lack of generalization of visual recognition models to images with compression artefacts. We find that to a large extent the performance loss is due to the latter: by finetuning the recognition models on compressed training images, most of the performance loss is recovered. For example, bringing segmentation accuracy back up to 42 mIoU, i.e. recovering 82% of the original drop in accuracy. ## Keyword: RAW ### GANHead: Towards Generative Animatable Neural Head Avatars - **Authors:** Sijing Wu, Yichao Yan, Yunhao Li, Yuhao Cheng, Wenhan Zhu, Ke Gao, Xiaobo Li, Guangtao Zhai - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.03950 - **Pdf link:** https://arxiv.org/pdf/2304.03950 - **Abstract** To bring digital avatars into people's lives, it is highly demanded to efficiently generate complete, realistic, and animatable head avatars. This task is challenging, and it is difficult for existing methods to satisfy all the requirements at once. To achieve these goals, we propose GANHead (Generative Animatable Neural Head Avatar), a novel generative head model that takes advantages of both the fine-grained control over the explicit expression parameters and the realistic rendering results of implicit representations. Specifically, GANHead represents coarse geometry, fine-gained details and texture via three networks in canonical space to obtain the ability to generate complete and realistic head avatars. To achieve flexible animation, we define the deformation filed by standard linear blend skinning (LBS), with the learned continuous pose and expression bases and LBS weights. This allows the avatars to be directly animated by FLAME parameters and generalize well to unseen poses and expressions. Compared to state-of-the-art (SOTA) methods, GANHead achieves superior performance on head avatar generation and raw scan fitting. ### Unsupervised Sampling Promoting for Stochastic Human Trajectory Prediction - **Authors:** Guangyi Chen, Zhenhao Chen, Shunxing Fan, Kun Zhang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.04298 - **Pdf link:** https://arxiv.org/pdf/2304.04298 - **Abstract** The indeterminate nature of human motion requires trajectory prediction systems to use a probabilistic model to formulate the multi-modality phenomenon and infer a finite set of future trajectories. However, the inference processes of most existing methods rely on Monte Carlo random sampling, which is insufficient to cover the realistic paths with finite samples, due to the long tail effect of the predicted distribution. To promote the sampling process of stochastic prediction, we propose a novel method, called BOsampler, to adaptively mine potential paths with Bayesian optimization in an unsupervised manner, as a sequential design strategy in which new prediction is dependent on the previously drawn samples. Specifically, we model the trajectory sampling as a Gaussian process and construct an acquisition function to measure the potential sampling value. This acquisition function applies the original distribution as prior and encourages exploring paths in the long-tail region. This sampling method can be integrated with existing stochastic predictive models without retraining. Experimental results on various baseline methods demonstrate the effectiveness of our method. ### Federated Incremental Semantic Segmentation - **Authors:** Jiahua Dong, Duzhen Zhang, Yang Cong, Wei Cong, Henghui Ding, Dengxin Dai - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.04620 - **Pdf link:** https://arxiv.org/pdf/2304.04620 - **Abstract** Federated learning-based semantic segmentation (FSS) has drawn widespread attention via decentralized training on local clients. However, most FSS models assume categories are fixed in advance, thus heavily undergoing forgetting on old categories in practical applications where local clients receive new categories incrementally while have no memory storage to access old classes. Moreover, new clients collecting novel classes may join in the global training of FSS, which further exacerbates catastrophic forgetting. To surmount the above challenges, we propose a Forgetting-Balanced Learning (FBL) model to address heterogeneous forgetting on old classes from both intra-client and inter-client aspects. Specifically, under the guidance of pseudo labels generated via adaptive class-balanced pseudo labeling, we develop a forgetting-balanced semantic compensation loss and a forgetting-balanced relation consistency loss to rectify intra-client heterogeneous forgetting of old categories with background shift. It performs balanced gradient propagation and relation consistency distillation within local clients. Moreover, to tackle heterogeneous forgetting from inter-client aspect, we propose a task transition monitor. It can identify new classes under privacy protection and store the latest old global model for relation distillation. Qualitative experiments reveal large improvement of our model against comparison methods. The code is available at https://github.com/JiahuaDong/FISS. ### Do We Train on Test Data? The Impact of Near-Duplicates on License Plate Recognition - **Authors:** Rayson Laroca, Valter Estevam, Alceu S. Britto Jr., Rodrigo Minetto, David Menotti - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.04653 - **Pdf link:** https://arxiv.org/pdf/2304.04653 - **Abstract** This work draws attention to the large fraction of near-duplicates in the training and test sets of datasets widely adopted in License Plate Recognition (LPR) research. These duplicates refer to images that, although different, show the same license plate. Our experiments, conducted on the two most popular datasets in the field, show a substantial decrease in recognition rate when six well-known models are trained and tested under fair splits, that is, in the absence of duplicates in the training and test sets. Moreover, in one of the datasets, the ranking of models changed considerably when they were trained and tested under duplicate-free splits. These findings suggest that such duplicates have significantly biased the evaluation and development of deep learning-based models for LPR. The list of near-duplicates we have found and proposals for fair splits are publicly available for further research at https://raysonlaroca.github.io/supp/lpr-train-on-test/ ## Keyword: raw image There is no result
process
new submissions for tue apr keyword events exploring data geometry for continual learning authors zhi gao chen xu feng li yunde jia mehrtash harandi yuwei wu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract continual learning aims to efficiently learn from a non stationary stream of data while avoiding forgetting the knowledge of old data in many practical applications data complies with non euclidean geometry as such the commonly used euclidean space cannot gracefully capture non euclidean geometric structures of data leading to inferior results in this paper we study continual learning from a novel perspective by exploring data geometry for the non stationary stream of data our method dynamically expands the geometry of the underlying space to match growing geometric structures induced by new data and prevents forgetting by keeping geometric structures of old data into account in doing so making use of the mixed curvature space we propose an incremental search scheme through which the growing geometric structures are encoded then we introduce an angular regularization loss and a neighbor robustness loss to train the model capable of penalizing the change of global geometric structures and local geometric structures experiments show that our method achieves better performance than baseline methods designed in euclidean space monocular human pose estimation for sports broadcasts using partial sports field registration authors tobias baumgartner stefanie klatt subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract the filming of sporting events projects and flattens the movement of athletes in the world onto a broadcast image the pixel locations of joints in these images can be detected with high validity recovering the actual movement of the limbs kinematics of the athletes requires lifting these pixel locations back into a third dimension implying a certain scene geometry the well known line markings of sports fields allow for the calibration of the camera and for determining the actual geometry of the scene close up shots of athletes are required to extract detailed kinematics which in turn obfuscates the pertinent field markers for camera calibration we suggest partial sports field registration which determines a set of scene consistent camera calibrations up to a single degree of freedom through joint optimization of pose estimation and camera calibration we demonstrate the successful extraction of running kinematics on a track in this work we combine advances in human pose estimation and camera calibration via partial sports field registration to demonstrate an avenue for collecting valid large scale kinematic datasets we generate a synthetic dataset of more than images in unreal engine with different viewpoints running styles and body types to show the limitations of existing monocular hpe methods synthetic data and code are available at event based camera tracker by nabla t nerf authors mana masuda yusuke sekikawa hideo saito subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract when a camera travels across a world only a fraction of pixel value changes an event based camera observes the change as sparse events how can we utilize sparse events for efficient recovery of the camera pose we show that we can recover the camera pose by minimizing the error between sparse events and the temporal gradient of the scene represented as a neural radiance field nerf to enable the computation of the temporal gradient of the scene we augment nerf s camera pose as a time function when the input pose to the nerf coincides with the actual pose the output of the temporal gradient of nerf equals the observed intensity changes on the event s points using this principle we propose an event based camera pose tracking framework called tegra which realizes the pose update by using the sparse event s observation to the best of our knowledge this is the first camera pose estimation algorithm using the scene s implicit representation and the sparse intensity change from events keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp feature representation learning with adaptive displacement generation and transformer fusion for micro expression recognition authors zhijun zhai jianhui zhao chengjiang long wenju xu shuangjiang he huijuan zhao subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract micro expressions are spontaneous rapid and subtle facial movements that can neither be forged nor suppressed they are very important nonverbal communication clues but are transient and of low intensity thus difficult to recognize recently deep learning based methods have been developed for micro expression me recognition using feature extraction and fusion techniques however targeted feature learning and efficient feature fusion still lack further study according to the me characteristics to address these issues we propose a novel framework feature representation learning with adaptive displacement generation and transformer fusion frl dgt in which a convolutional displacement generation module dgm with self supervised learning is used to extract dynamic features from onset apex frames targeted to the subsequent me recognition task and a well designed transformer fusion mechanism composed of three transformer based fusion modules local global fusions based on au regions and full face fusion is applied to extract the multi level informative features after dgm for the final me prediction the extensive experiments with solid leave one subject out loso evaluation results have demonstrated the superiority of our proposed frl dgt to state of the art methods grouped knowledge distillation for deep face recognition authors weisong zhao xiangyu zhu kaiwen guo xiao yu zhang zhen lei subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract compared with the feature based distillation methods logits distillation can liberalize the requirements of consistent feature dimension between teacher and student networks while the performance is deemed inferior in face recognition one major challenge is that the light weight student network has difficulty fitting the target logits due to its low model capacity which is attributed to the significant number of identities in face recognition therefore we seek to probe the target logits to extract the primary knowledge related to face identity and discard the others to make the distillation more achievable for the student network specifically there is a tail group with near zero values in the prediction containing minor knowledge for distillation to provide a clear perspective of its impact we first partition the logits into two groups i e primary group and secondary group according to the cumulative probability of the softened prediction then we reorganize the knowledge distillation kd loss of grouped logits into three parts i e primary kd secondary kd and binary kd primary kd refers to distilling the primary knowledge from the teacher secondary kd aims to refine minor knowledge but increases the difficulty of distillation and binary kd ensures the consistency of knowledge distribution between teacher and student we experimentally found that primary kd and binary kd are indispensable for kd and secondary kd is the culprit restricting kd at the bottleneck therefore we propose a grouped knowledge distillation gkd that retains the primary kd and binary kd but omits secondary kd in the ultimate kd loss calculation extensive experimental results on popular face recognition benchmarks demonstrate the superiority of proposed gkd over state of the art methods keyword image signal processing there is no result keyword image signal process there is no result keyword compression neural residual radiance fields for streamably free viewpoint videos authors liao wang qiang hu qihan he ziyu wang jingyi yu tinne tuytelaars lan xu minye wu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract the success of the neural radiance fields nerfs for modeling and free view rendering static objects has inspired numerous attempts on dynamic scenes current techniques that utilize neural rendering for facilitating free view videos fvvs are restricted to either offline rendering or are capable of processing only brief sequences with minimal motion in this paper we present a novel technique residual radiance field or rerf as a highly compact neural representation to achieve real time fvv rendering on long duration dynamic scenes rerf explicitly models the residual information between adjacent timestamps in the spatial temporal feature space with a global coordinate based tiny mlp as the feature decoder specifically rerf employs a compact motion grid along with a residual feature grid to exploit inter frame feature similarities we show such a strategy can handle large motions without sacrificing quality we further present a sequential training scheme to maintain the smoothness and the sparsity of the motion residual grids based on rerf we design a special fvv codec that achieves three orders of magnitudes compression rate and provides a companion rerf player to support online streaming of long duration fvvs of dynamic scenes extensive experiments demonstrate the effectiveness of rerf for compactly representing dynamic radiance fields enabling an unprecedented free viewpoint viewing experience in speed and quality are visual recognition models robust to image compression authors joão maria janeiro stanislav frolov alaaeldin el nouby jakob verbeek subjects computer vision and pattern recognition cs cv artificial intelligence cs ai machine learning cs lg arxiv link pdf link abstract reducing the data footprint of visual content via image compression is essential to reduce storage requirements but also to reduce the bandwidth and latency requirements for transmission in particular the use of compressed images allows for faster transfer of data and faster response times for visual recognition in edge devices that rely on cloud based services in this paper we first analyze the impact of image compression using traditional codecs as well as recent state of the art neural compression approaches on three visual recognition tasks image classification object detection and semantic segmentation we consider a wide range of compression levels ranging from to bits per pixel bpp we find that for all three tasks the recognition ability is significantly impacted when using strong compression for example for segmentation miou is reduced from to miou when compressing to bpp using the best compression model we evaluated second we test to what extent this performance drop can be ascribed to a loss of relevant information in the compressed image or to a lack of generalization of visual recognition models to images with compression artefacts we find that to a large extent the performance loss is due to the latter by finetuning the recognition models on compressed training images most of the performance loss is recovered for example bringing segmentation accuracy back up to miou i e recovering of the original drop in accuracy keyword raw ganhead towards generative animatable neural head avatars authors sijing wu yichao yan yunhao li yuhao cheng wenhan zhu ke gao xiaobo li guangtao zhai subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract to bring digital avatars into people s lives it is highly demanded to efficiently generate complete realistic and animatable head avatars this task is challenging and it is difficult for existing methods to satisfy all the requirements at once to achieve these goals we propose ganhead generative animatable neural head avatar a novel generative head model that takes advantages of both the fine grained control over the explicit expression parameters and the realistic rendering results of implicit representations specifically ganhead represents coarse geometry fine gained details and texture via three networks in canonical space to obtain the ability to generate complete and realistic head avatars to achieve flexible animation we define the deformation filed by standard linear blend skinning lbs with the learned continuous pose and expression bases and lbs weights this allows the avatars to be directly animated by flame parameters and generalize well to unseen poses and expressions compared to state of the art sota methods ganhead achieves superior performance on head avatar generation and raw scan fitting unsupervised sampling promoting for stochastic human trajectory prediction authors guangyi chen zhenhao chen shunxing fan kun zhang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract the indeterminate nature of human motion requires trajectory prediction systems to use a probabilistic model to formulate the multi modality phenomenon and infer a finite set of future trajectories however the inference processes of most existing methods rely on monte carlo random sampling which is insufficient to cover the realistic paths with finite samples due to the long tail effect of the predicted distribution to promote the sampling process of stochastic prediction we propose a novel method called bosampler to adaptively mine potential paths with bayesian optimization in an unsupervised manner as a sequential design strategy in which new prediction is dependent on the previously drawn samples specifically we model the trajectory sampling as a gaussian process and construct an acquisition function to measure the potential sampling value this acquisition function applies the original distribution as prior and encourages exploring paths in the long tail region this sampling method can be integrated with existing stochastic predictive models without retraining experimental results on various baseline methods demonstrate the effectiveness of our method federated incremental semantic segmentation authors jiahua dong duzhen zhang yang cong wei cong henghui ding dengxin dai subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract federated learning based semantic segmentation fss has drawn widespread attention via decentralized training on local clients however most fss models assume categories are fixed in advance thus heavily undergoing forgetting on old categories in practical applications where local clients receive new categories incrementally while have no memory storage to access old classes moreover new clients collecting novel classes may join in the global training of fss which further exacerbates catastrophic forgetting to surmount the above challenges we propose a forgetting balanced learning fbl model to address heterogeneous forgetting on old classes from both intra client and inter client aspects specifically under the guidance of pseudo labels generated via adaptive class balanced pseudo labeling we develop a forgetting balanced semantic compensation loss and a forgetting balanced relation consistency loss to rectify intra client heterogeneous forgetting of old categories with background shift it performs balanced gradient propagation and relation consistency distillation within local clients moreover to tackle heterogeneous forgetting from inter client aspect we propose a task transition monitor it can identify new classes under privacy protection and store the latest old global model for relation distillation qualitative experiments reveal large improvement of our model against comparison methods the code is available at do we train on test data the impact of near duplicates on license plate recognition authors rayson laroca valter estevam alceu s britto jr rodrigo minetto david menotti subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract this work draws attention to the large fraction of near duplicates in the training and test sets of datasets widely adopted in license plate recognition lpr research these duplicates refer to images that although different show the same license plate our experiments conducted on the two most popular datasets in the field show a substantial decrease in recognition rate when six well known models are trained and tested under fair splits that is in the absence of duplicates in the training and test sets moreover in one of the datasets the ranking of models changed considerably when they were trained and tested under duplicate free splits these findings suggest that such duplicates have significantly biased the evaluation and development of deep learning based models for lpr the list of near duplicates we have found and proposals for fair splits are publicly available for further research at keyword raw image there is no result
1
9,886
12,888,950,020
IssuesEvent
2020-07-13 13:48:56
pystatgen/sgkit
https://api.github.com/repos/pystatgen/sgkit
closed
Mypy/Typing check
process + tools
Current master doesn't pass mypy check: ``` (sgkit-dev) ➜ sgkit git:(master) ✗ mypy --strict . setup.py:2: error: Skipping analyzing 'setuptools': found module but no type hints or library stubs sgkit/utils.py:4: error: Function is missing a type annotation sgkit/tests/test_utils.py:2: error: Skipping analyzing 'pytest': found module but no type hints or library stubs sgkit/tests/test_utils.py:2: note: See https://mypy.readthedocs.io/en/latest/running_mypy.html#missing-imports sgkit/tests/test_utils.py:9: error: Call to untyped function "check_array_like" in typed context sgkit/tests/test_utils.py:12: error: Call to untyped function "check_array_like" in typed context sgkit/tests/test_utils.py:14: error: Call to untyped function "check_array_like" in typed context sgkit/tests/test_utils.py:16: error: Call to untyped function "check_array_like" in typed context sgkit/tests/test_utils.py:18: error: Call to untyped function "check_array_like" in typed context sgkit/tests/test_utils.py:20: error: Call to untyped function "check_array_like" in typed context sgkit/api.py:55: error: Call to untyped function "check_array_like" in typed context sgkit/api.py:56: error: Call to untyped function "check_array_like" in typed context sgkit/api.py:57: error: Call to untyped function "check_array_like" in typed context sgkit/api.py:58: error: Call to untyped function "check_array_like" in typed context sgkit/api.py:59: error: Call to untyped function "check_array_like" in typed context sgkit/api.py:72: error: Call to untyped function "check_array_like" in typed context sgkit/api.py:73: error: Unsupported target for indexed assignment ("Mapping[Hashable, Any]") sgkit/api.py:78: error: Call to untyped function "check_array_like" in typed context sgkit/api.py:79: error: Unsupported target for indexed assignment ("Mapping[Hashable, Any]") sgkit/api.py:81: error: Argument "attrs" to "Dataset" has incompatible type "Dict[str, List[str]]"; expected "Mapping[Hashable, Any]" Found 19 errors in 4 files (checked 7 source files) ``` It goes without saying that typing provides huge benefits (especially over time), there are also many typing limitations in the python (data) ecosystem. Currently our build and pre-commit don't run mypy check, there isn't much code in this repo, but there is a bunch coming in soon. So some questions about this: * do we want to enable mypy check for both build and pre-commit right now (I'm +1) * we likely won't be able to type everything, so we need to agree which parts need to be typed and which part can be omitted
1.0
Mypy/Typing check - Current master doesn't pass mypy check: ``` (sgkit-dev) ➜ sgkit git:(master) ✗ mypy --strict . setup.py:2: error: Skipping analyzing 'setuptools': found module but no type hints or library stubs sgkit/utils.py:4: error: Function is missing a type annotation sgkit/tests/test_utils.py:2: error: Skipping analyzing 'pytest': found module but no type hints or library stubs sgkit/tests/test_utils.py:2: note: See https://mypy.readthedocs.io/en/latest/running_mypy.html#missing-imports sgkit/tests/test_utils.py:9: error: Call to untyped function "check_array_like" in typed context sgkit/tests/test_utils.py:12: error: Call to untyped function "check_array_like" in typed context sgkit/tests/test_utils.py:14: error: Call to untyped function "check_array_like" in typed context sgkit/tests/test_utils.py:16: error: Call to untyped function "check_array_like" in typed context sgkit/tests/test_utils.py:18: error: Call to untyped function "check_array_like" in typed context sgkit/tests/test_utils.py:20: error: Call to untyped function "check_array_like" in typed context sgkit/api.py:55: error: Call to untyped function "check_array_like" in typed context sgkit/api.py:56: error: Call to untyped function "check_array_like" in typed context sgkit/api.py:57: error: Call to untyped function "check_array_like" in typed context sgkit/api.py:58: error: Call to untyped function "check_array_like" in typed context sgkit/api.py:59: error: Call to untyped function "check_array_like" in typed context sgkit/api.py:72: error: Call to untyped function "check_array_like" in typed context sgkit/api.py:73: error: Unsupported target for indexed assignment ("Mapping[Hashable, Any]") sgkit/api.py:78: error: Call to untyped function "check_array_like" in typed context sgkit/api.py:79: error: Unsupported target for indexed assignment ("Mapping[Hashable, Any]") sgkit/api.py:81: error: Argument "attrs" to "Dataset" has incompatible type "Dict[str, List[str]]"; expected "Mapping[Hashable, Any]" Found 19 errors in 4 files (checked 7 source files) ``` It goes without saying that typing provides huge benefits (especially over time), there are also many typing limitations in the python (data) ecosystem. Currently our build and pre-commit don't run mypy check, there isn't much code in this repo, but there is a bunch coming in soon. So some questions about this: * do we want to enable mypy check for both build and pre-commit right now (I'm +1) * we likely won't be able to type everything, so we need to agree which parts need to be typed and which part can be omitted
process
mypy typing check current master doesn t pass mypy check sgkit dev ➜ sgkit git master ✗ mypy strict setup py error skipping analyzing setuptools found module but no type hints or library stubs sgkit utils py error function is missing a type annotation sgkit tests test utils py error skipping analyzing pytest found module but no type hints or library stubs sgkit tests test utils py note see sgkit tests test utils py error call to untyped function check array like in typed context sgkit tests test utils py error call to untyped function check array like in typed context sgkit tests test utils py error call to untyped function check array like in typed context sgkit tests test utils py error call to untyped function check array like in typed context sgkit tests test utils py error call to untyped function check array like in typed context sgkit tests test utils py error call to untyped function check array like in typed context sgkit api py error call to untyped function check array like in typed context sgkit api py error call to untyped function check array like in typed context sgkit api py error call to untyped function check array like in typed context sgkit api py error call to untyped function check array like in typed context sgkit api py error call to untyped function check array like in typed context sgkit api py error call to untyped function check array like in typed context sgkit api py error unsupported target for indexed assignment mapping sgkit api py error call to untyped function check array like in typed context sgkit api py error unsupported target for indexed assignment mapping sgkit api py error argument attrs to dataset has incompatible type dict expected mapping found errors in files checked source files it goes without saying that typing provides huge benefits especially over time there are also many typing limitations in the python data ecosystem currently our build and pre commit don t run mypy check there isn t much code in this repo but there is a bunch coming in soon so some questions about this do we want to enable mypy check for both build and pre commit right now i m we likely won t be able to type everything so we need to agree which parts need to be typed and which part can be omitted
1
1,728
4,388,041,791
IssuesEvent
2016-08-08 17:38:42
kerubistan/kerub
https://api.github.com/repos/kerubistan/kerub
opened
get rid of the abstraction layer
component:data processing enhancement
Some interfaces, like * Hypervisor * PowerManager are just trying to hide the details of the power management protocol. This can be done (and actually is done in storage) in the planner, so for now these interfaces can go away.
1.0
get rid of the abstraction layer - Some interfaces, like * Hypervisor * PowerManager are just trying to hide the details of the power management protocol. This can be done (and actually is done in storage) in the planner, so for now these interfaces can go away.
process
get rid of the abstraction layer some interfaces like hypervisor powermanager are just trying to hide the details of the power management protocol this can be done and actually is done in storage in the planner so for now these interfaces can go away
1
80,002
15,326,038,066
IssuesEvent
2021-02-26 02:41:40
Azure/azure-sdk-for-go
https://api.github.com/repos/Azure/azure-sdk-for-go
closed
Migrate from satori/go.uuid to gofrs/uuid
Mgmt codegen enhancement
We are planning to replace the `github.com/satori/go.uuid` package with `github.com/gofrs/uuid` to resolve its [potential security issues](https://github.com/satori/go.uuid/issues/73) in a major version at the end of this month. In major version release v53.0.0, you will have to replace all the import statements of the `uuid` in order to avoid compile errors. Thank you!
1.0
Migrate from satori/go.uuid to gofrs/uuid - We are planning to replace the `github.com/satori/go.uuid` package with `github.com/gofrs/uuid` to resolve its [potential security issues](https://github.com/satori/go.uuid/issues/73) in a major version at the end of this month. In major version release v53.0.0, you will have to replace all the import statements of the `uuid` in order to avoid compile errors. Thank you!
non_process
migrate from satori go uuid to gofrs uuid we are planning to replace the github com satori go uuid package with github com gofrs uuid to resolve its in a major version at the end of this month in major version release you will have to replace all the import statements of the uuid in order to avoid compile errors thank you
0
34,625
9,421,830,636
IssuesEvent
2019-04-11 07:55:33
commercialhaskell/stack
https://api.github.com/repos/commercialhaskell/stack
closed
Don't work `stack script --resolver lts-9.3 --package ghc --package ghc-paths`
component: build component: script interpreter further investigation required
Failure bolow command. ```bash $ stack script Main.hs --resolver lts-9.3 --package ghc --package ghc-paths ``` But, success below command. ```bash $ stack script Main.hs --resolver lts-9.3 --package ghc-paths $ stack script Main.hs --resolver lts-9.3 --package ghc --package ghc-paths ``` ### Steps to reproduce 1. Clean dir `rm -rf ~/.stack` 2. Create `Main.hs` ```haskell module Main where main :: IO () main = undefined ``` 3. Run `stack script Main.hs --resolver lts-9.3 --package ghc --package ghc-paths` ### Expected I expected that displayed to result (`undefined`) after build & run. ### Actual #### Failure case Not found `ghc` package error. ```bash $ stack script test.hs --resolver lts-9.3 --package ghc --package ghc-paths --verbose Version 1.5.1, Git revision 2b0392468a03174a7c5d317291c62452d771b055 x86_64 hpack-0.18.1 2017-09-05 18:13:21.093922: [debug] Ignoring config files @(Stack/Config.hs:915:13) 2017-09-05 18:13:21.094564: [debug] Using resolver: lts-9.3 specified on command line @(Stack/Config.hs:537:7) 2017-09-05 18:13:21.094643: [info] Using resolver: lts-9.3 specified on command line @(Stack/Config.hs:628:17) 2017-09-05 18:13:21.094830: [debug] Decoding build plan from: /home/foo/.stack/build-plan/lts-9.3.yaml @(Stack/Snapshot.hs:151:5) 2017-09-05 18:13:22.319973: [debug] Run process: /sbin/ldconfig -p @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.321808: [debug] Process finished in 1ms: /sbin/ldconfig -p @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.322099: [debug] Run process: /usr/bin/gcc -v @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.323198: [debug] Process finished in 1ms: /usr/bin/gcc -v @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.323300: [debug] PIE enabled @(Stack/Setup.hs:585:17) 2017-09-05 18:13:22.323427: [debug] Found shared library libtinfo.so.5 in 'ldconfig -p' output @(Stack/Setup.hs:561:29) 2017-09-05 18:13:22.323589: [debug] Did not find shared library libtinfo.so.6 @(Stack/Setup.hs:575:38) 2017-09-05 18:13:22.323663: [debug] Did not find shared library libncursesw.so.6 @(Stack/Setup.hs:575:38) 2017-09-05 18:13:22.323725: [debug] Found shared library libgmp.so.10 in 'ldconfig -p' output @(Stack/Setup.hs:561:29) 2017-09-05 18:13:22.323795: [debug] Did not find shared library libgmp.so.3 @(Stack/Setup.hs:575:38) 2017-09-05 18:13:22.323846: [debug] Using standard GHC build @(Stack/Setup.hs:613:9) 2017-09-05 18:13:22.324144: [debug] Getting global package database location @(Stack/GhcPkg.hs:46:5) 2017-09-05 18:13:22.324321: [debug] Asking GHC for its version @(Stack/Setup/Installed.hs:98:13) 2017-09-05 18:13:22.324438: [debug] Getting Cabal package version @(Stack/GhcPkg.hs:185:5) 2017-09-05 18:13:22.324509: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db list --global @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.324932: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc --numeric-version @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.325906: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db field --simple-output Cabal version @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.342697: [debug] Process finished in 18ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db list --global @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.352001: [debug] Process finished in 25ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db field --simple-output Cabal version @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.373818: [debug] Process finished in 47ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc --numeric-version @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.374015: [debug] GHC version is: ghc-8.0.2 @(Stack/Setup/Installed.hs:102:13) 2017-09-05 18:13:22.374136: [debug] Resolving package entries @(Stack/Setup.hs:252:5) 2017-09-05 18:13:22.374362: [debug] Trying to decode /home/foo/.stack/loaded-snapshot-cache/x86_64-linux/ghc-8.0.2/lts-9.3.cache @(Data/Store/VersionTagged.hs:66:5) 2017-09-05 18:13:22.400517: [debug] Success decoding /home/foo/.stack/loaded-snapshot-cache/x86_64-linux/ghc-8.0.2/lts-9.3.cache @(Data/Store/VersionTagged.hs:70:13) 2017-09-05 18:13:22.401300: [debug] Starting to execute command inside EnvConfig @(Stack/Runners.hs:168:18) 2017-09-05 18:13:22.401519: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg list --simple-output @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.419844: [debug] Process finished in 18ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg list --simple-output @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.420053: [debug] Missing packages, performing installation @(Stack/Script.hs:84:21) 2017-09-05 18:13:22.420143: [debug] Parsing the targets @(Stack/Build/Target.hs:473:3) 2017-09-05 18:13:22.460835: [debug] Finding out which packages are already installed @(Stack/Build/Installed.hs:60:5) 2017-09-05 18:13:22.461683: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --global --no-user-package-db dump --expand-pkgroot @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.487589: [debug] Process finished in 25ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --global --no-user-package-db dump --expand-pkgroot @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.488454: [debug] Ignoring package haskeline due to wanting version 0.7.4.0 instead of 0.7.3.0 @(Stack/Build/Installed.hs:190:5) 2017-09-05 18:13:22.488553: [debug] Ignoring package terminfo due to wanting version 0.4.1.0 instead of 0.4.0.2 @(Stack/Build/Installed.hs:190:5) 2017-09-05 18:13:22.488626: [debug] Ignoring package xhtml due to wanting version 3000.2.2 instead of 3000.2.1 @(Stack/Build/Installed.hs:190:5) 2017-09-05 18:13:22.488895: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb dump --expand-pkgroot @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.507026: [debug] Process finished in 17ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb dump --expand-pkgroot @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.507375: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/script/lts-9.3/.stack-work/install/x86_64-linux/lts-9.3/8.0.2/pkgdb dump --expand-pkgroot @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.525772: [debug] Process finished in 18ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/script/lts-9.3/.stack-work/install/x86_64-linux/lts-9.3/8.0.2/pkgdb dump --expand-pkgroot @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.526398: [debug] Constructing the build plan @(Stack/Build/ConstructPlan.hs:179:5) 2017-09-05 18:13:22.527063: [debug] Trying to decode /home/foo/.stack/indices/Hackage/01-index.cache @(Data/Store/VersionTagged.hs:66:5) 2017-09-05 18:13:22.650981: [debug] Success decoding /home/foo/.stack/indices/Hackage/01-index.cache @(Data/Store/VersionTagged.hs:70:13) 2017-09-05 18:13:22.653824: [info] Didn't see ghc-8.0.2 in your package indices. Updating and trying again. @(Stack/Fetch.hs:357:33) 2017-09-05 18:13:22.655279: [info] Selected mirror https://s3.amazonaws.com/hackage.fpcomplete.com/ @(Stack/PackageIndex.hs:307:24) 2017-09-05 18:13:22.659140: [info] Downloading timestamp @(Stack/PackageIndex.hs:307:24) 2017-09-05 18:13:24.080026: [info] No updates to your package index were found @(Stack/PackageIndex.hs:362:14) Update complete 2017-09-05 18:13:27.790829: [debug] Trying to decode /home/foo/.stack/indices/Hackage/01-index.cache @(Data/Store/VersionTagged.hs:66:5) 2017-09-05 18:13:27.990528: [debug] Success decoding /home/foo/.stack/indices/Hackage/01-index.cache @(Data/Store/VersionTagged.hs:70:13) The following package identifiers were not found in your indices: ghc-8.0.2 Possible candidates: ghc-8.2.1. ``` #### Success case step 1. ```bash stack script test.hs --resolver lts-9.3 --package ghc-paths --verbose Version 1.5.1, Git revision 2b0392468a03174a7c5d317291c62452d771b055 x86_64 hpack-0.18.1 2017-09-05 18:15:02.972926: [debug] Ignoring config files @(Stack/Config.hs:915:13) 2017-09-05 18:15:02.973516: [debug] Using resolver: lts-9.3 specified on command line @(Stack/Config.hs:537:7) 2017-09-05 18:15:02.973588: [info] Using resolver: lts-9.3 specified on command line @(Stack/Config.hs:628:17) 2017-09-05 18:15:02.973655: [debug] Decoding build plan from: /home/foo/.stack/build-plan/lts-9.3.yaml @(Stack/Snapshot.hs:151:5) 2017-09-05 18:15:04.180276: [debug] Run process: /sbin/ldconfig -p @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.182259: [debug] Process finished in 1ms: /sbin/ldconfig -p @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.182457: [debug] Run process: /usr/bin/gcc -v @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.183728: [debug] Process finished in 1ms: /usr/bin/gcc -v @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.183838: [debug] PIE enabled @(Stack/Setup.hs:585:17) 2017-09-05 18:15:04.183976: [debug] Found shared library libtinfo.so.5 in 'ldconfig -p' output @(Stack/Setup.hs:561:29) 2017-09-05 18:15:04.184180: [debug] Did not find shared library libtinfo.so.6 @(Stack/Setup.hs:575:38) 2017-09-05 18:15:04.184270: [debug] Did not find shared library libncursesw.so.6 @(Stack/Setup.hs:575:38) 2017-09-05 18:15:04.184354: [debug] Found shared library libgmp.so.10 in 'ldconfig -p' output @(Stack/Setup.hs:561:29) 2017-09-05 18:15:04.184446: [debug] Did not find shared library libgmp.so.3 @(Stack/Setup.hs:575:38) 2017-09-05 18:15:04.184524: [debug] Using standard GHC build @(Stack/Setup.hs:613:9) 2017-09-05 18:15:04.184852: [debug] Getting global package database location @(Stack/GhcPkg.hs:46:5) 2017-09-05 18:15:04.184990: [debug] Asking GHC for its version @(Stack/Setup/Installed.hs:98:13) 2017-09-05 18:15:04.185118: [debug] Getting Cabal package version @(Stack/GhcPkg.hs:185:5) 2017-09-05 18:15:04.185189: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db list --global @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.185575: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc --numeric-version @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.189497: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db field --simple-output Cabal version @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.208970: [debug] Process finished in 22ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db list --global @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.231413: [debug] Process finished in 29ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db field --simple-output Cabal version @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.231731: [debug] Process finished in 46ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc --numeric-version @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.231846: [debug] GHC version is: ghc-8.0.2 @(Stack/Setup/Installed.hs:102:13) 2017-09-05 18:15:04.231992: [debug] Resolving package entries @(Stack/Setup.hs:252:5) 2017-09-05 18:15:04.232213: [debug] Trying to decode /home/foo/.stack/loaded-snapshot-cache/x86_64-linux/ghc-8.0.2/lts-9.3.cache @(Data/Store/VersionTagged.hs:66:5) 2017-09-05 18:15:04.258664: [debug] Success decoding /home/foo/.stack/loaded-snapshot-cache/x86_64-linux/ghc-8.0.2/lts-9.3.cache @(Data/Store/VersionTagged.hs:70:13) 2017-09-05 18:15:04.259391: [debug] Starting to execute command inside EnvConfig @(Stack/Runners.hs:168:18) 2017-09-05 18:15:04.259593: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg list --simple-output @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.278507: [debug] Process finished in 18ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg list --simple-output @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.278649: [debug] Missing packages, performing installation @(Stack/Script.hs:84:21) 2017-09-05 18:15:04.278764: [debug] Parsing the targets @(Stack/Build/Target.hs:473:3) 2017-09-05 18:15:04.310866: [debug] Finding out which packages are already installed @(Stack/Build/Installed.hs:60:5) 2017-09-05 18:15:04.311614: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --global --no-user-package-db dump --expand-pkgroot @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.346677: [debug] Process finished in 34ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --global --no-user-package-db dump --expand-pkgroot @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.347377: [debug] Ignoring package haskeline due to wanting version 0.7.4.0 instead of 0.7.3.0 @(Stack/Build/Installed.hs:190:5) 2017-09-05 18:15:04.347468: [debug] Ignoring package terminfo due to wanting version 0.4.1.0 instead of 0.4.0.2 @(Stack/Build/Installed.hs:190:5) 2017-09-05 18:15:04.347535: [debug] Ignoring package xhtml due to wanting version 3000.2.2 instead of 3000.2.1 @(Stack/Build/Installed.hs:190:5) 2017-09-05 18:15:04.347748: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb dump --expand-pkgroot @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.368349: [debug] Process finished in 20ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb dump --expand-pkgroot @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.368581: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/script/lts-9.3/.stack-work/install/x86_64-linux/lts-9.3/8.0.2/pkgdb dump --expand-pkgroot @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.385803: [debug] Process finished in 17ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/script/lts-9.3/.stack-work/install/x86_64-linux/lts-9.3/8.0.2/pkgdb dump --expand-pkgroot @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.386433: [debug] Constructing the build plan @(Stack/Build/ConstructPlan.hs:179:5) 2017-09-05 18:15:04.387302: [debug] Trying to decode /home/foo/.stack/indices/Hackage/01-index.cache @(Data/Store/VersionTagged.hs:66:5) 2017-09-05 18:15:04.535856: [debug] Success decoding /home/foo/.stack/indices/Hackage/01-index.cache @(Data/Store/VersionTagged.hs:70:13) 2017-09-05 18:15:04.540159: [debug] Checking if we are going to build multiple executables with the same name @(Stack/Build.hs:177:5) 2017-09-05 18:15:04.540262: [debug] Executing the build plan @(Stack/Build/Execute.hs:487:5) 2017-09-05 18:15:04.558429: [debug] Creating process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc -rtsopts -threaded -clear-package-db -global-package-db -hide-all-packages -package base -main-is StackSetupShim.mainOverride -package Cabal-1.24.2.0 /home/foo/.stack/setup-exe-src/setup-mPHDZzAJ.hs /home/foo/.stack/setup-exe-src/setup-shim-mPHDZzAJ.hs -o /home/foo/.stack/setup-exe-cache/x86_64-linux/tmp-Cabal-simple_mPHDZzAJ_1.24.2.0_ghc-8.0.2 @(System/Process/Log.hs:22:3) [1 of 2] Compiling Main ( /home/foo/.stack/setup-exe-src/setup-mPHDZzAJ.hs, /home/foo/.stack/setup-exe-src/setup-mPHDZzAJ.o ) [2 of 2] Compiling StackSetupShim ( /home/foo/.stack/setup-exe-src/setup-shim-mPHDZzAJ.hs, /home/foo/.stack/setup-exe-src/setup-shim-mPHDZzAJ.o ) Linking /home/foo/.stack/setup-exe-cache/x86_64-linux/tmp-Cabal-simple_mPHDZzAJ_1.24.2.0_ghc-8.0.2 ... 2017-09-05 18:15:08.144445: [debug] Getting global package database location @(Stack/GhcPkg.hs:46:5) 2017-09-05 18:15:08.144627: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db list --global @(System/Process/Log.hs:37:3) 2017-09-05 18:15:08.163161: [debug] Process finished in 18ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db list --global @(System/Process/Log.hs:44:3) 2017-09-05 18:15:08.163848: [debug] Exception ignored when attempting to load /home/foo/.stack/precompiled/x86_64-linux/ghc-8.0.2/1.24.2.0/ghc-paths-0.1.0.9@sha256:d3f3470c7bd13b765891fb56b28d809cb7aeda0a78050679ae6f29b6705c46bf,656/SWXaDPn62wF2mi5c-wlXeEG1Td-bFCxHP1uj741nlPo=: /home/foo/.stack/precompiled/x86_64-linux/ghc-8.0.2/1.24.2.0/ghc-paths-0.1.0.9@sha256:d3f3470c7bd13b765891fb56b28d809cb7aeda0a78050679ae6f29b6705c46bf,656/SWXaDPn62wF2mi5c-wlXeEG1Td-bFCxHP1uj741nlPo=: openBinaryFile: does not exist (No such file or directory) @(Data/Store/VersionTagged.hs:84:9) 2017-09-05 18:15:08.164431: [debug] Downloading /hackage.fpcomplete.com/package/ghc-paths-0.1.0.9.tar.gz @(Network/HTTP/Download/Verified.hs:226:9) 2017-09-05 18:15:09.756666: [info] ghc-paths-0.1.0.9: download @(Stack/Fetch.hs:525:32) 2017-09-05 18:15:09.759156: [debug] Exception ignored when attempting to load /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/stack-config-cache: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/stack-config-cache: openBinaryFile: does not exist (No such file or directory) @(Data/Store/VersionTagged.hs:84:9) 2017-09-05 18:15:09.759328: [debug] Exception ignored when attempting to load /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/stack-cabal-mod: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/stack-cabal-mod: openBinaryFile: does not exist (No such file or directory) @(Data/Store/VersionTagged.hs:84:9) 2017-09-05 18:15:09.759562: [info] ghc-paths-0.1.0.9: configure @(Stack/Build/Execute.hs:838:23) 2017-09-05 18:15:09.760713: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-8.0.2 --make -odir /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup -hidir /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup -i -i. -clear-package-db -global-package-db -package-db=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb -package-db=/home/foo/.stack/script/lts-9.3/.stack-work/install/x86_64-linux/lts-9.3/8.0.2/pkgdb -hide-all-packages -package-id=Cabal-1.24.2.0 -package-id=base-4.9.1.0 -package-id=directory-1.3.0.0 -optP-include -optP/tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup_macros.h /tmp/stack31298/ghc-paths-0.1.0.9/Setup.hs /home/foo/.stack/setup-exe-src/setup-shim-mPHDZzAJ.hs -main-is StackSetupShim.mainOverride -o /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup -threaded @(System/Process/Log.hs:37:3) 2017-09-05 18:15:13.436617: [debug] Process finished in 3675ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-8.0.2 --make -odir /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup -hidir /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup -i -i. -clear-package-db -global-package-db -package-db=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb -package-db=/home/foo/.stack/script/lts-9.3/.stack-work/install/x86_64-linux/lts-9.3/8.0.2/pkgdb -hide-all-packages -package-id=Cabal-1.24.2.0 -package-id=base-4.9.1.0 -package-id=directory-1.3.0.0 -optP-include -optP/tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup_macros.h /tmp/stack31298/ghc-paths-0.1.0.9/Setup.hs /home/foo/.stack/setup-exe-src/setup-shim-mPHDZzAJ.hs -main-is StackSetupShim.mainOverride -o /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup -threaded @(System/Process/Log.hs:44:3) 2017-09-05 18:15:13.437039: [debug] Run process: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup --builddir=.stack-work/dist/x86_64-linux/Cabal-1.24.2.0 configure --with-ghc=/home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc --with-ghc-pkg=/home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --package-db=clear --package-db=global --package-db=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb --libdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/lib --bindir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/bin --datadir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/share --libexecdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/libexec --sysconfdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/etc --docdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/doc/ghc-paths-0.1.0.9 --htmldir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/doc/ghc-paths-0.1.0.9 --haddockdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/doc/ghc-paths-0.1.0.9 --dependency=Cabal=Cabal-1.24.2.0 --dependency=base=base-4.9.1.0 --dependency=directory=directory-1.3.0.0 @(System/Process/Log.hs:37:3) 2017-09-05 18:15:14.003558: [debug] Process finished in 566ms: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup --builddir=.stack-work/dist/x86_64-linux/Cabal-1.24.2.0 configure --with-ghc=/home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc --with-ghc-pkg=/home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --package-db=clear --package-db=global --package-db=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb --libdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/lib --bindir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/bin --datadir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/share --libexecdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/libexec --sysconfdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/etc --docdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/doc/ghc-paths-0.1.0.9 --htmldir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/doc/ghc-paths-0.1.0.9 --haddockdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/doc/ghc-paths-0.1.0.9 --dependency=Cabal=Cabal-1.24.2.0 --dependency=base=base-4.9.1.0 --dependency=directory=directory-1.3.0.0 @(System/Process/Log.hs:44:3) 2017-09-05 18:15:14.003874: [debug] Encoding /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/stack-config-cache @(Data/Store/VersionTagged.hs:48:5) 2017-09-05 18:15:14.004236: [debug] Finished writing /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/stack-config-cache @(Data/Store/VersionTagged.hs:53:5) 2017-09-05 18:15:14.004341: [debug] Encoding /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/stack-cabal-mod @(Data/Store/VersionTagged.hs:48:5) 2017-09-05 18:15:14.004649: [debug] Finished writing /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/stack-cabal-mod @(Data/Store/VersionTagged.hs:53:5) 2017-09-05 18:15:14.004799: [info] ghc-paths-0.1.0.9: build @(Stack/Build/Execute.hs:838:23) 2017-09-05 18:15:14.005078: [debug] Run process: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup --builddir=.stack-work/dist/x86_64-linux/Cabal-1.24.2.0 build --ghc-options " -ddump-hi -ddump-to-file" @(System/Process/Log.hs:37:3) 2017-09-05 18:15:14.470439: [debug] Process finished in 465ms: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup --builddir=.stack-work/dist/x86_64-linux/Cabal-1.24.2.0 build --ghc-options " -ddump-hi -ddump-to-file" @(System/Process/Log.hs:44:3) 2017-09-05 18:15:14.470627: [info] ghc-paths-0.1.0.9: copy/register @(Stack/Build/Execute.hs:838:23) 2017-09-05 18:15:14.470949: [debug] Run process: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup --builddir=.stack-work/dist/x86_64-linux/Cabal-1.24.2.0 copy @(System/Process/Log.hs:37:3) 2017-09-05 18:15:14.485856: [debug] Process finished in 14ms: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup --builddir=.stack-work/dist/x86_64-linux/Cabal-1.24.2.0 copy @(System/Process/Log.hs:44:3) 2017-09-05 18:15:14.486332: [debug] Run process: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup --builddir=.stack-work/dist/x86_64-linux/Cabal-1.24.2.0 register @(System/Process/Log.hs:37:3) 2017-09-05 18:15:14.606404: [debug] Process finished in 119ms: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup --builddir=.stack-work/dist/x86_64-linux/Cabal-1.24.2.0 register @(System/Process/Log.hs:44:3) 2017-09-05 18:15:14.606903: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb describe --simple-output ghc-paths --expand-pkgroot @(System/Process/Log.hs:37:3) 2017-09-05 18:15:14.627097: [debug] Process finished in 19ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb describe --simple-output ghc-paths --expand-pkgroot @(System/Process/Log.hs:44:3) 2017-09-05 18:15:14.627936: [debug] Encoding /home/foo/.stack/precompiled/x86_64-linux/ghc-8.0.2/1.24.2.0/ghc-paths-0.1.0.9@sha256:d3f3470c7bd13b765891fb56b28d809cb7aeda0a78050679ae6f29b6705c46bf,656/SWXaDPn62wF2mi5c-wlXeEG1Td-bFCxHP1uj741nlPo= @(Data/Store/VersionTagged.hs:48:5) 2017-09-05 18:15:14.628291: [debug] Finished writing /home/foo/.stack/precompiled/x86_64-linux/ghc-8.0.2/1.24.2.0/ghc-paths-0.1.0.9@sha256:d3f3470c7bd13b765891fb56b28d809cb7aeda0a78050679ae6f29b6705c46bf,656/SWXaDPn62wF2mi5c-wlXeEG1Td-bFCxHP1uj741nlPo= @(Data/Store/VersionTagged.hs:53:5) 2017-09-05 18:15:14.631289: [debug] Encoding /home/foo/.stack/script/lts-9.3/.stack-work/install/x86_64-linux/lts-9.3/8.0.2/flag-cache/ghc-paths-0.1.0.9-AhaDlGOsRAepox069XzG @(Data/Store/VersionTagged.hs:48:5) 2017-09-05 18:15:14.631680: [debug] Finished writing /home/foo/.stack/script/lts-9.3/.stack-work/install/x86_64-linux/lts-9.3/8.0.2/flag-cache/ghc-paths-0.1.0.9-AhaDlGOsRAepox069XzG @(Data/Store/VersionTagged.hs:53:5) 2017-09-05 18:15:14.631958: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/runghc -hide-all-packages -packagebase -packageghc-paths /home/foo/test.hs @(System/Process/Log.hs:37:3) test.hs: Prelude.undefined CallStack (from HasCallStack): error, called at libraries/base/GHC/Err.hs:79:14 in base:GHC.Err undefined, called at /home/foo/test.hs:4:8 in main:Main ``` step 2 ```bash $ stack script test.hs --resolver lts-9.3 --package ghc --package ghc-paths --verbose Version 1.5.1, Git revision 2b0392468a03174a7c5d317291c62452d771b055 x86_64 hpack-0.18.1 2017-09-05 18:21:34.453142: [debug] Ignoring config files @(Stack/Config.hs:915:13) 2017-09-05 18:21:34.453729: [debug] Using resolver: lts-9.3 specified on command line @(Stack/Config.hs:537:7) 2017-09-05 18:21:34.453809: [info] Using resolver: lts-9.3 specified on command line @(Stack/Config.hs:628:17) 2017-09-05 18:21:34.454004: [debug] Decoding build plan from: /home/foo/.stack/build-plan/lts-9.3.yaml @(Stack/Snapshot.hs:151:5) 2017-09-05 18:21:35.689984: [debug] Run process: /sbin/ldconfig -p @(System/Process/Log.hs:37:3) 2017-09-05 18:21:35.691721: [debug] Process finished in 1ms: /sbin/ldconfig -p @(System/Process/Log.hs:44:3) 2017-09-05 18:21:35.691915: [debug] Run process: /usr/bin/gcc -v @(System/Process/Log.hs:37:3) 2017-09-05 18:21:35.693237: [debug] Process finished in 1ms: /usr/bin/gcc -v @(System/Process/Log.hs:44:3) 2017-09-05 18:21:35.693446: [debug] PIE enabled @(Stack/Setup.hs:585:17) 2017-09-05 18:21:35.693588: [debug] Found shared library libtinfo.so.5 in 'ldconfig -p' output @(Stack/Setup.hs:561:29) 2017-09-05 18:21:35.693751: [debug] Did not find shared library libtinfo.so.6 @(Stack/Setup.hs:575:38) 2017-09-05 18:21:35.693829: [debug] Did not find shared library libncursesw.so.6 @(Stack/Setup.hs:575:38) 2017-09-05 18:21:35.693889: [debug] Found shared library libgmp.so.10 in 'ldconfig -p' output @(Stack/Setup.hs:561:29) 2017-09-05 18:21:35.693963: [debug] Did not find shared library libgmp.so.3 @(Stack/Setup.hs:575:38) 2017-09-05 18:21:35.694014: [debug] Using standard GHC build @(Stack/Setup.hs:613:9) 2017-09-05 18:21:35.694326: [debug] Asking GHC for its version @(Stack/Setup/Installed.hs:98:13) 2017-09-05 18:21:35.694490: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc --numeric-version @(System/Process/Log.hs:37:3) 2017-09-05 18:21:35.694868: [debug] Getting Cabal package version @(Stack/GhcPkg.hs:185:5) 2017-09-05 18:21:35.695774: [debug] Getting global package database location @(Stack/GhcPkg.hs:46:5) 2017-09-05 18:21:35.695850: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db field --simple-output Cabal version @(System/Process/Log.hs:37:3) 2017-09-05 18:21:35.696084: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db list --global @(System/Process/Log.hs:37:3) 2017-09-05 18:21:35.714060: [debug] Process finished in 18ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db field --simple-output Cabal version @(System/Process/Log.hs:44:3) 2017-09-05 18:21:35.727648: [debug] Process finished in 32ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc --numeric-version @(System/Process/Log.hs:44:3) 2017-09-05 18:21:35.727766: [debug] GHC version is: ghc-8.0.2 @(Stack/Setup/Installed.hs:102:13) 2017-09-05 18:21:35.731071: [debug] Process finished in 17ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db list --global @(System/Process/Log.hs:44:3) 2017-09-05 18:21:35.731359: [debug] Resolving package entries @(Stack/Setup.hs:252:5) 2017-09-05 18:21:35.731510: [debug] Trying to decode /home/foo/.stack/loaded-snapshot-cache/x86_64-linux/ghc-8.0.2/lts-9.3.cache @(Data/Store/VersionTagged.hs:66:5) 2017-09-05 18:21:35.757467: [debug] Success decoding /home/foo/.stack/loaded-snapshot-cache/x86_64-linux/ghc-8.0.2/lts-9.3.cache @(Data/Store/VersionTagged.hs:70:13) 2017-09-05 18:21:35.758334: [debug] Starting to execute command inside EnvConfig @(Stack/Runners.hs:168:18) 2017-09-05 18:21:35.758541: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg list --simple-output @(System/Process/Log.hs:37:3) 2017-09-05 18:21:35.777333: [debug] Process finished in 18ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg list --simple-output @(System/Process/Log.hs:44:3) 2017-09-05 18:21:35.777506: [debug] All packages already installed @(Stack/Script.hs:82:22) 2017-09-05 18:21:35.777688: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/runghc -hide-all-packages -packagebase -packageghc -packageghc-paths /home/foo/test.hs @(System/Process/Log.hs:37:3) test.hs: Prelude.undefined CallStack (from HasCallStack): error, called at libraries/base/GHC/Err.hs:79:14 in base:GHC.Err undefined, called at /home/foo/test.hs:4:8 in main:Main ``` ### Stack version ``` $ stack --version Version 1.5.1, Git revision 2b0392468a03174a7c5d317291c62452d771b055 x86_64 hpack-0.18.1 ``` ### Method of installation * Official binary, downloaded from stackage.org or fpcomplete's package repository
1.0
Don't work `stack script --resolver lts-9.3 --package ghc --package ghc-paths` - Failure bolow command. ```bash $ stack script Main.hs --resolver lts-9.3 --package ghc --package ghc-paths ``` But, success below command. ```bash $ stack script Main.hs --resolver lts-9.3 --package ghc-paths $ stack script Main.hs --resolver lts-9.3 --package ghc --package ghc-paths ``` ### Steps to reproduce 1. Clean dir `rm -rf ~/.stack` 2. Create `Main.hs` ```haskell module Main where main :: IO () main = undefined ``` 3. Run `stack script Main.hs --resolver lts-9.3 --package ghc --package ghc-paths` ### Expected I expected that displayed to result (`undefined`) after build & run. ### Actual #### Failure case Not found `ghc` package error. ```bash $ stack script test.hs --resolver lts-9.3 --package ghc --package ghc-paths --verbose Version 1.5.1, Git revision 2b0392468a03174a7c5d317291c62452d771b055 x86_64 hpack-0.18.1 2017-09-05 18:13:21.093922: [debug] Ignoring config files @(Stack/Config.hs:915:13) 2017-09-05 18:13:21.094564: [debug] Using resolver: lts-9.3 specified on command line @(Stack/Config.hs:537:7) 2017-09-05 18:13:21.094643: [info] Using resolver: lts-9.3 specified on command line @(Stack/Config.hs:628:17) 2017-09-05 18:13:21.094830: [debug] Decoding build plan from: /home/foo/.stack/build-plan/lts-9.3.yaml @(Stack/Snapshot.hs:151:5) 2017-09-05 18:13:22.319973: [debug] Run process: /sbin/ldconfig -p @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.321808: [debug] Process finished in 1ms: /sbin/ldconfig -p @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.322099: [debug] Run process: /usr/bin/gcc -v @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.323198: [debug] Process finished in 1ms: /usr/bin/gcc -v @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.323300: [debug] PIE enabled @(Stack/Setup.hs:585:17) 2017-09-05 18:13:22.323427: [debug] Found shared library libtinfo.so.5 in 'ldconfig -p' output @(Stack/Setup.hs:561:29) 2017-09-05 18:13:22.323589: [debug] Did not find shared library libtinfo.so.6 @(Stack/Setup.hs:575:38) 2017-09-05 18:13:22.323663: [debug] Did not find shared library libncursesw.so.6 @(Stack/Setup.hs:575:38) 2017-09-05 18:13:22.323725: [debug] Found shared library libgmp.so.10 in 'ldconfig -p' output @(Stack/Setup.hs:561:29) 2017-09-05 18:13:22.323795: [debug] Did not find shared library libgmp.so.3 @(Stack/Setup.hs:575:38) 2017-09-05 18:13:22.323846: [debug] Using standard GHC build @(Stack/Setup.hs:613:9) 2017-09-05 18:13:22.324144: [debug] Getting global package database location @(Stack/GhcPkg.hs:46:5) 2017-09-05 18:13:22.324321: [debug] Asking GHC for its version @(Stack/Setup/Installed.hs:98:13) 2017-09-05 18:13:22.324438: [debug] Getting Cabal package version @(Stack/GhcPkg.hs:185:5) 2017-09-05 18:13:22.324509: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db list --global @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.324932: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc --numeric-version @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.325906: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db field --simple-output Cabal version @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.342697: [debug] Process finished in 18ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db list --global @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.352001: [debug] Process finished in 25ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db field --simple-output Cabal version @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.373818: [debug] Process finished in 47ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc --numeric-version @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.374015: [debug] GHC version is: ghc-8.0.2 @(Stack/Setup/Installed.hs:102:13) 2017-09-05 18:13:22.374136: [debug] Resolving package entries @(Stack/Setup.hs:252:5) 2017-09-05 18:13:22.374362: [debug] Trying to decode /home/foo/.stack/loaded-snapshot-cache/x86_64-linux/ghc-8.0.2/lts-9.3.cache @(Data/Store/VersionTagged.hs:66:5) 2017-09-05 18:13:22.400517: [debug] Success decoding /home/foo/.stack/loaded-snapshot-cache/x86_64-linux/ghc-8.0.2/lts-9.3.cache @(Data/Store/VersionTagged.hs:70:13) 2017-09-05 18:13:22.401300: [debug] Starting to execute command inside EnvConfig @(Stack/Runners.hs:168:18) 2017-09-05 18:13:22.401519: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg list --simple-output @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.419844: [debug] Process finished in 18ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg list --simple-output @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.420053: [debug] Missing packages, performing installation @(Stack/Script.hs:84:21) 2017-09-05 18:13:22.420143: [debug] Parsing the targets @(Stack/Build/Target.hs:473:3) 2017-09-05 18:13:22.460835: [debug] Finding out which packages are already installed @(Stack/Build/Installed.hs:60:5) 2017-09-05 18:13:22.461683: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --global --no-user-package-db dump --expand-pkgroot @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.487589: [debug] Process finished in 25ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --global --no-user-package-db dump --expand-pkgroot @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.488454: [debug] Ignoring package haskeline due to wanting version 0.7.4.0 instead of 0.7.3.0 @(Stack/Build/Installed.hs:190:5) 2017-09-05 18:13:22.488553: [debug] Ignoring package terminfo due to wanting version 0.4.1.0 instead of 0.4.0.2 @(Stack/Build/Installed.hs:190:5) 2017-09-05 18:13:22.488626: [debug] Ignoring package xhtml due to wanting version 3000.2.2 instead of 3000.2.1 @(Stack/Build/Installed.hs:190:5) 2017-09-05 18:13:22.488895: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb dump --expand-pkgroot @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.507026: [debug] Process finished in 17ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb dump --expand-pkgroot @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.507375: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/script/lts-9.3/.stack-work/install/x86_64-linux/lts-9.3/8.0.2/pkgdb dump --expand-pkgroot @(System/Process/Log.hs:37:3) 2017-09-05 18:13:22.525772: [debug] Process finished in 18ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/script/lts-9.3/.stack-work/install/x86_64-linux/lts-9.3/8.0.2/pkgdb dump --expand-pkgroot @(System/Process/Log.hs:44:3) 2017-09-05 18:13:22.526398: [debug] Constructing the build plan @(Stack/Build/ConstructPlan.hs:179:5) 2017-09-05 18:13:22.527063: [debug] Trying to decode /home/foo/.stack/indices/Hackage/01-index.cache @(Data/Store/VersionTagged.hs:66:5) 2017-09-05 18:13:22.650981: [debug] Success decoding /home/foo/.stack/indices/Hackage/01-index.cache @(Data/Store/VersionTagged.hs:70:13) 2017-09-05 18:13:22.653824: [info] Didn't see ghc-8.0.2 in your package indices. Updating and trying again. @(Stack/Fetch.hs:357:33) 2017-09-05 18:13:22.655279: [info] Selected mirror https://s3.amazonaws.com/hackage.fpcomplete.com/ @(Stack/PackageIndex.hs:307:24) 2017-09-05 18:13:22.659140: [info] Downloading timestamp @(Stack/PackageIndex.hs:307:24) 2017-09-05 18:13:24.080026: [info] No updates to your package index were found @(Stack/PackageIndex.hs:362:14) Update complete 2017-09-05 18:13:27.790829: [debug] Trying to decode /home/foo/.stack/indices/Hackage/01-index.cache @(Data/Store/VersionTagged.hs:66:5) 2017-09-05 18:13:27.990528: [debug] Success decoding /home/foo/.stack/indices/Hackage/01-index.cache @(Data/Store/VersionTagged.hs:70:13) The following package identifiers were not found in your indices: ghc-8.0.2 Possible candidates: ghc-8.2.1. ``` #### Success case step 1. ```bash stack script test.hs --resolver lts-9.3 --package ghc-paths --verbose Version 1.5.1, Git revision 2b0392468a03174a7c5d317291c62452d771b055 x86_64 hpack-0.18.1 2017-09-05 18:15:02.972926: [debug] Ignoring config files @(Stack/Config.hs:915:13) 2017-09-05 18:15:02.973516: [debug] Using resolver: lts-9.3 specified on command line @(Stack/Config.hs:537:7) 2017-09-05 18:15:02.973588: [info] Using resolver: lts-9.3 specified on command line @(Stack/Config.hs:628:17) 2017-09-05 18:15:02.973655: [debug] Decoding build plan from: /home/foo/.stack/build-plan/lts-9.3.yaml @(Stack/Snapshot.hs:151:5) 2017-09-05 18:15:04.180276: [debug] Run process: /sbin/ldconfig -p @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.182259: [debug] Process finished in 1ms: /sbin/ldconfig -p @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.182457: [debug] Run process: /usr/bin/gcc -v @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.183728: [debug] Process finished in 1ms: /usr/bin/gcc -v @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.183838: [debug] PIE enabled @(Stack/Setup.hs:585:17) 2017-09-05 18:15:04.183976: [debug] Found shared library libtinfo.so.5 in 'ldconfig -p' output @(Stack/Setup.hs:561:29) 2017-09-05 18:15:04.184180: [debug] Did not find shared library libtinfo.so.6 @(Stack/Setup.hs:575:38) 2017-09-05 18:15:04.184270: [debug] Did not find shared library libncursesw.so.6 @(Stack/Setup.hs:575:38) 2017-09-05 18:15:04.184354: [debug] Found shared library libgmp.so.10 in 'ldconfig -p' output @(Stack/Setup.hs:561:29) 2017-09-05 18:15:04.184446: [debug] Did not find shared library libgmp.so.3 @(Stack/Setup.hs:575:38) 2017-09-05 18:15:04.184524: [debug] Using standard GHC build @(Stack/Setup.hs:613:9) 2017-09-05 18:15:04.184852: [debug] Getting global package database location @(Stack/GhcPkg.hs:46:5) 2017-09-05 18:15:04.184990: [debug] Asking GHC for its version @(Stack/Setup/Installed.hs:98:13) 2017-09-05 18:15:04.185118: [debug] Getting Cabal package version @(Stack/GhcPkg.hs:185:5) 2017-09-05 18:15:04.185189: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db list --global @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.185575: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc --numeric-version @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.189497: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db field --simple-output Cabal version @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.208970: [debug] Process finished in 22ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db list --global @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.231413: [debug] Process finished in 29ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db field --simple-output Cabal version @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.231731: [debug] Process finished in 46ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc --numeric-version @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.231846: [debug] GHC version is: ghc-8.0.2 @(Stack/Setup/Installed.hs:102:13) 2017-09-05 18:15:04.231992: [debug] Resolving package entries @(Stack/Setup.hs:252:5) 2017-09-05 18:15:04.232213: [debug] Trying to decode /home/foo/.stack/loaded-snapshot-cache/x86_64-linux/ghc-8.0.2/lts-9.3.cache @(Data/Store/VersionTagged.hs:66:5) 2017-09-05 18:15:04.258664: [debug] Success decoding /home/foo/.stack/loaded-snapshot-cache/x86_64-linux/ghc-8.0.2/lts-9.3.cache @(Data/Store/VersionTagged.hs:70:13) 2017-09-05 18:15:04.259391: [debug] Starting to execute command inside EnvConfig @(Stack/Runners.hs:168:18) 2017-09-05 18:15:04.259593: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg list --simple-output @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.278507: [debug] Process finished in 18ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg list --simple-output @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.278649: [debug] Missing packages, performing installation @(Stack/Script.hs:84:21) 2017-09-05 18:15:04.278764: [debug] Parsing the targets @(Stack/Build/Target.hs:473:3) 2017-09-05 18:15:04.310866: [debug] Finding out which packages are already installed @(Stack/Build/Installed.hs:60:5) 2017-09-05 18:15:04.311614: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --global --no-user-package-db dump --expand-pkgroot @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.346677: [debug] Process finished in 34ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --global --no-user-package-db dump --expand-pkgroot @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.347377: [debug] Ignoring package haskeline due to wanting version 0.7.4.0 instead of 0.7.3.0 @(Stack/Build/Installed.hs:190:5) 2017-09-05 18:15:04.347468: [debug] Ignoring package terminfo due to wanting version 0.4.1.0 instead of 0.4.0.2 @(Stack/Build/Installed.hs:190:5) 2017-09-05 18:15:04.347535: [debug] Ignoring package xhtml due to wanting version 3000.2.2 instead of 3000.2.1 @(Stack/Build/Installed.hs:190:5) 2017-09-05 18:15:04.347748: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb dump --expand-pkgroot @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.368349: [debug] Process finished in 20ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb dump --expand-pkgroot @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.368581: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/script/lts-9.3/.stack-work/install/x86_64-linux/lts-9.3/8.0.2/pkgdb dump --expand-pkgroot @(System/Process/Log.hs:37:3) 2017-09-05 18:15:04.385803: [debug] Process finished in 17ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/script/lts-9.3/.stack-work/install/x86_64-linux/lts-9.3/8.0.2/pkgdb dump --expand-pkgroot @(System/Process/Log.hs:44:3) 2017-09-05 18:15:04.386433: [debug] Constructing the build plan @(Stack/Build/ConstructPlan.hs:179:5) 2017-09-05 18:15:04.387302: [debug] Trying to decode /home/foo/.stack/indices/Hackage/01-index.cache @(Data/Store/VersionTagged.hs:66:5) 2017-09-05 18:15:04.535856: [debug] Success decoding /home/foo/.stack/indices/Hackage/01-index.cache @(Data/Store/VersionTagged.hs:70:13) 2017-09-05 18:15:04.540159: [debug] Checking if we are going to build multiple executables with the same name @(Stack/Build.hs:177:5) 2017-09-05 18:15:04.540262: [debug] Executing the build plan @(Stack/Build/Execute.hs:487:5) 2017-09-05 18:15:04.558429: [debug] Creating process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc -rtsopts -threaded -clear-package-db -global-package-db -hide-all-packages -package base -main-is StackSetupShim.mainOverride -package Cabal-1.24.2.0 /home/foo/.stack/setup-exe-src/setup-mPHDZzAJ.hs /home/foo/.stack/setup-exe-src/setup-shim-mPHDZzAJ.hs -o /home/foo/.stack/setup-exe-cache/x86_64-linux/tmp-Cabal-simple_mPHDZzAJ_1.24.2.0_ghc-8.0.2 @(System/Process/Log.hs:22:3) [1 of 2] Compiling Main ( /home/foo/.stack/setup-exe-src/setup-mPHDZzAJ.hs, /home/foo/.stack/setup-exe-src/setup-mPHDZzAJ.o ) [2 of 2] Compiling StackSetupShim ( /home/foo/.stack/setup-exe-src/setup-shim-mPHDZzAJ.hs, /home/foo/.stack/setup-exe-src/setup-shim-mPHDZzAJ.o ) Linking /home/foo/.stack/setup-exe-cache/x86_64-linux/tmp-Cabal-simple_mPHDZzAJ_1.24.2.0_ghc-8.0.2 ... 2017-09-05 18:15:08.144445: [debug] Getting global package database location @(Stack/GhcPkg.hs:46:5) 2017-09-05 18:15:08.144627: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db list --global @(System/Process/Log.hs:37:3) 2017-09-05 18:15:08.163161: [debug] Process finished in 18ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db list --global @(System/Process/Log.hs:44:3) 2017-09-05 18:15:08.163848: [debug] Exception ignored when attempting to load /home/foo/.stack/precompiled/x86_64-linux/ghc-8.0.2/1.24.2.0/ghc-paths-0.1.0.9@sha256:d3f3470c7bd13b765891fb56b28d809cb7aeda0a78050679ae6f29b6705c46bf,656/SWXaDPn62wF2mi5c-wlXeEG1Td-bFCxHP1uj741nlPo=: /home/foo/.stack/precompiled/x86_64-linux/ghc-8.0.2/1.24.2.0/ghc-paths-0.1.0.9@sha256:d3f3470c7bd13b765891fb56b28d809cb7aeda0a78050679ae6f29b6705c46bf,656/SWXaDPn62wF2mi5c-wlXeEG1Td-bFCxHP1uj741nlPo=: openBinaryFile: does not exist (No such file or directory) @(Data/Store/VersionTagged.hs:84:9) 2017-09-05 18:15:08.164431: [debug] Downloading /hackage.fpcomplete.com/package/ghc-paths-0.1.0.9.tar.gz @(Network/HTTP/Download/Verified.hs:226:9) 2017-09-05 18:15:09.756666: [info] ghc-paths-0.1.0.9: download @(Stack/Fetch.hs:525:32) 2017-09-05 18:15:09.759156: [debug] Exception ignored when attempting to load /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/stack-config-cache: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/stack-config-cache: openBinaryFile: does not exist (No such file or directory) @(Data/Store/VersionTagged.hs:84:9) 2017-09-05 18:15:09.759328: [debug] Exception ignored when attempting to load /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/stack-cabal-mod: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/stack-cabal-mod: openBinaryFile: does not exist (No such file or directory) @(Data/Store/VersionTagged.hs:84:9) 2017-09-05 18:15:09.759562: [info] ghc-paths-0.1.0.9: configure @(Stack/Build/Execute.hs:838:23) 2017-09-05 18:15:09.760713: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-8.0.2 --make -odir /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup -hidir /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup -i -i. -clear-package-db -global-package-db -package-db=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb -package-db=/home/foo/.stack/script/lts-9.3/.stack-work/install/x86_64-linux/lts-9.3/8.0.2/pkgdb -hide-all-packages -package-id=Cabal-1.24.2.0 -package-id=base-4.9.1.0 -package-id=directory-1.3.0.0 -optP-include -optP/tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup_macros.h /tmp/stack31298/ghc-paths-0.1.0.9/Setup.hs /home/foo/.stack/setup-exe-src/setup-shim-mPHDZzAJ.hs -main-is StackSetupShim.mainOverride -o /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup -threaded @(System/Process/Log.hs:37:3) 2017-09-05 18:15:13.436617: [debug] Process finished in 3675ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-8.0.2 --make -odir /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup -hidir /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup -i -i. -clear-package-db -global-package-db -package-db=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb -package-db=/home/foo/.stack/script/lts-9.3/.stack-work/install/x86_64-linux/lts-9.3/8.0.2/pkgdb -hide-all-packages -package-id=Cabal-1.24.2.0 -package-id=base-4.9.1.0 -package-id=directory-1.3.0.0 -optP-include -optP/tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup_macros.h /tmp/stack31298/ghc-paths-0.1.0.9/Setup.hs /home/foo/.stack/setup-exe-src/setup-shim-mPHDZzAJ.hs -main-is StackSetupShim.mainOverride -o /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup -threaded @(System/Process/Log.hs:44:3) 2017-09-05 18:15:13.437039: [debug] Run process: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup --builddir=.stack-work/dist/x86_64-linux/Cabal-1.24.2.0 configure --with-ghc=/home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc --with-ghc-pkg=/home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --package-db=clear --package-db=global --package-db=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb --libdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/lib --bindir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/bin --datadir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/share --libexecdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/libexec --sysconfdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/etc --docdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/doc/ghc-paths-0.1.0.9 --htmldir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/doc/ghc-paths-0.1.0.9 --haddockdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/doc/ghc-paths-0.1.0.9 --dependency=Cabal=Cabal-1.24.2.0 --dependency=base=base-4.9.1.0 --dependency=directory=directory-1.3.0.0 @(System/Process/Log.hs:37:3) 2017-09-05 18:15:14.003558: [debug] Process finished in 566ms: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup --builddir=.stack-work/dist/x86_64-linux/Cabal-1.24.2.0 configure --with-ghc=/home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc --with-ghc-pkg=/home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --package-db=clear --package-db=global --package-db=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb --libdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/lib --bindir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/bin --datadir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/share --libexecdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/libexec --sysconfdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/etc --docdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/doc/ghc-paths-0.1.0.9 --htmldir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/doc/ghc-paths-0.1.0.9 --haddockdir=/home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/doc/ghc-paths-0.1.0.9 --dependency=Cabal=Cabal-1.24.2.0 --dependency=base=base-4.9.1.0 --dependency=directory=directory-1.3.0.0 @(System/Process/Log.hs:44:3) 2017-09-05 18:15:14.003874: [debug] Encoding /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/stack-config-cache @(Data/Store/VersionTagged.hs:48:5) 2017-09-05 18:15:14.004236: [debug] Finished writing /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/stack-config-cache @(Data/Store/VersionTagged.hs:53:5) 2017-09-05 18:15:14.004341: [debug] Encoding /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/stack-cabal-mod @(Data/Store/VersionTagged.hs:48:5) 2017-09-05 18:15:14.004649: [debug] Finished writing /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/stack-cabal-mod @(Data/Store/VersionTagged.hs:53:5) 2017-09-05 18:15:14.004799: [info] ghc-paths-0.1.0.9: build @(Stack/Build/Execute.hs:838:23) 2017-09-05 18:15:14.005078: [debug] Run process: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup --builddir=.stack-work/dist/x86_64-linux/Cabal-1.24.2.0 build --ghc-options " -ddump-hi -ddump-to-file" @(System/Process/Log.hs:37:3) 2017-09-05 18:15:14.470439: [debug] Process finished in 465ms: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup --builddir=.stack-work/dist/x86_64-linux/Cabal-1.24.2.0 build --ghc-options " -ddump-hi -ddump-to-file" @(System/Process/Log.hs:44:3) 2017-09-05 18:15:14.470627: [info] ghc-paths-0.1.0.9: copy/register @(Stack/Build/Execute.hs:838:23) 2017-09-05 18:15:14.470949: [debug] Run process: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup --builddir=.stack-work/dist/x86_64-linux/Cabal-1.24.2.0 copy @(System/Process/Log.hs:37:3) 2017-09-05 18:15:14.485856: [debug] Process finished in 14ms: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup --builddir=.stack-work/dist/x86_64-linux/Cabal-1.24.2.0 copy @(System/Process/Log.hs:44:3) 2017-09-05 18:15:14.486332: [debug] Run process: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup --builddir=.stack-work/dist/x86_64-linux/Cabal-1.24.2.0 register @(System/Process/Log.hs:37:3) 2017-09-05 18:15:14.606404: [debug] Process finished in 119ms: /tmp/stack31298/ghc-paths-0.1.0.9/.stack-work/dist/x86_64-linux/Cabal-1.24.2.0/setup/setup --builddir=.stack-work/dist/x86_64-linux/Cabal-1.24.2.0 register @(System/Process/Log.hs:44:3) 2017-09-05 18:15:14.606903: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb describe --simple-output ghc-paths --expand-pkgroot @(System/Process/Log.hs:37:3) 2017-09-05 18:15:14.627097: [debug] Process finished in 19ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --user --no-user-package-db --package-db /home/foo/.stack/snapshots/x86_64-linux/lts-9.3/8.0.2/pkgdb describe --simple-output ghc-paths --expand-pkgroot @(System/Process/Log.hs:44:3) 2017-09-05 18:15:14.627936: [debug] Encoding /home/foo/.stack/precompiled/x86_64-linux/ghc-8.0.2/1.24.2.0/ghc-paths-0.1.0.9@sha256:d3f3470c7bd13b765891fb56b28d809cb7aeda0a78050679ae6f29b6705c46bf,656/SWXaDPn62wF2mi5c-wlXeEG1Td-bFCxHP1uj741nlPo= @(Data/Store/VersionTagged.hs:48:5) 2017-09-05 18:15:14.628291: [debug] Finished writing /home/foo/.stack/precompiled/x86_64-linux/ghc-8.0.2/1.24.2.0/ghc-paths-0.1.0.9@sha256:d3f3470c7bd13b765891fb56b28d809cb7aeda0a78050679ae6f29b6705c46bf,656/SWXaDPn62wF2mi5c-wlXeEG1Td-bFCxHP1uj741nlPo= @(Data/Store/VersionTagged.hs:53:5) 2017-09-05 18:15:14.631289: [debug] Encoding /home/foo/.stack/script/lts-9.3/.stack-work/install/x86_64-linux/lts-9.3/8.0.2/flag-cache/ghc-paths-0.1.0.9-AhaDlGOsRAepox069XzG @(Data/Store/VersionTagged.hs:48:5) 2017-09-05 18:15:14.631680: [debug] Finished writing /home/foo/.stack/script/lts-9.3/.stack-work/install/x86_64-linux/lts-9.3/8.0.2/flag-cache/ghc-paths-0.1.0.9-AhaDlGOsRAepox069XzG @(Data/Store/VersionTagged.hs:53:5) 2017-09-05 18:15:14.631958: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/runghc -hide-all-packages -packagebase -packageghc-paths /home/foo/test.hs @(System/Process/Log.hs:37:3) test.hs: Prelude.undefined CallStack (from HasCallStack): error, called at libraries/base/GHC/Err.hs:79:14 in base:GHC.Err undefined, called at /home/foo/test.hs:4:8 in main:Main ``` step 2 ```bash $ stack script test.hs --resolver lts-9.3 --package ghc --package ghc-paths --verbose Version 1.5.1, Git revision 2b0392468a03174a7c5d317291c62452d771b055 x86_64 hpack-0.18.1 2017-09-05 18:21:34.453142: [debug] Ignoring config files @(Stack/Config.hs:915:13) 2017-09-05 18:21:34.453729: [debug] Using resolver: lts-9.3 specified on command line @(Stack/Config.hs:537:7) 2017-09-05 18:21:34.453809: [info] Using resolver: lts-9.3 specified on command line @(Stack/Config.hs:628:17) 2017-09-05 18:21:34.454004: [debug] Decoding build plan from: /home/foo/.stack/build-plan/lts-9.3.yaml @(Stack/Snapshot.hs:151:5) 2017-09-05 18:21:35.689984: [debug] Run process: /sbin/ldconfig -p @(System/Process/Log.hs:37:3) 2017-09-05 18:21:35.691721: [debug] Process finished in 1ms: /sbin/ldconfig -p @(System/Process/Log.hs:44:3) 2017-09-05 18:21:35.691915: [debug] Run process: /usr/bin/gcc -v @(System/Process/Log.hs:37:3) 2017-09-05 18:21:35.693237: [debug] Process finished in 1ms: /usr/bin/gcc -v @(System/Process/Log.hs:44:3) 2017-09-05 18:21:35.693446: [debug] PIE enabled @(Stack/Setup.hs:585:17) 2017-09-05 18:21:35.693588: [debug] Found shared library libtinfo.so.5 in 'ldconfig -p' output @(Stack/Setup.hs:561:29) 2017-09-05 18:21:35.693751: [debug] Did not find shared library libtinfo.so.6 @(Stack/Setup.hs:575:38) 2017-09-05 18:21:35.693829: [debug] Did not find shared library libncursesw.so.6 @(Stack/Setup.hs:575:38) 2017-09-05 18:21:35.693889: [debug] Found shared library libgmp.so.10 in 'ldconfig -p' output @(Stack/Setup.hs:561:29) 2017-09-05 18:21:35.693963: [debug] Did not find shared library libgmp.so.3 @(Stack/Setup.hs:575:38) 2017-09-05 18:21:35.694014: [debug] Using standard GHC build @(Stack/Setup.hs:613:9) 2017-09-05 18:21:35.694326: [debug] Asking GHC for its version @(Stack/Setup/Installed.hs:98:13) 2017-09-05 18:21:35.694490: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc --numeric-version @(System/Process/Log.hs:37:3) 2017-09-05 18:21:35.694868: [debug] Getting Cabal package version @(Stack/GhcPkg.hs:185:5) 2017-09-05 18:21:35.695774: [debug] Getting global package database location @(Stack/GhcPkg.hs:46:5) 2017-09-05 18:21:35.695850: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db field --simple-output Cabal version @(System/Process/Log.hs:37:3) 2017-09-05 18:21:35.696084: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db list --global @(System/Process/Log.hs:37:3) 2017-09-05 18:21:35.714060: [debug] Process finished in 18ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db field --simple-output Cabal version @(System/Process/Log.hs:44:3) 2017-09-05 18:21:35.727648: [debug] Process finished in 32ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc --numeric-version @(System/Process/Log.hs:44:3) 2017-09-05 18:21:35.727766: [debug] GHC version is: ghc-8.0.2 @(Stack/Setup/Installed.hs:102:13) 2017-09-05 18:21:35.731071: [debug] Process finished in 17ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg --no-user-package-db list --global @(System/Process/Log.hs:44:3) 2017-09-05 18:21:35.731359: [debug] Resolving package entries @(Stack/Setup.hs:252:5) 2017-09-05 18:21:35.731510: [debug] Trying to decode /home/foo/.stack/loaded-snapshot-cache/x86_64-linux/ghc-8.0.2/lts-9.3.cache @(Data/Store/VersionTagged.hs:66:5) 2017-09-05 18:21:35.757467: [debug] Success decoding /home/foo/.stack/loaded-snapshot-cache/x86_64-linux/ghc-8.0.2/lts-9.3.cache @(Data/Store/VersionTagged.hs:70:13) 2017-09-05 18:21:35.758334: [debug] Starting to execute command inside EnvConfig @(Stack/Runners.hs:168:18) 2017-09-05 18:21:35.758541: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg list --simple-output @(System/Process/Log.hs:37:3) 2017-09-05 18:21:35.777333: [debug] Process finished in 18ms: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/ghc-pkg list --simple-output @(System/Process/Log.hs:44:3) 2017-09-05 18:21:35.777506: [debug] All packages already installed @(Stack/Script.hs:82:22) 2017-09-05 18:21:35.777688: [debug] Run process: /home/foo/.stack/programs/x86_64-linux/ghc-8.0.2/bin/runghc -hide-all-packages -packagebase -packageghc -packageghc-paths /home/foo/test.hs @(System/Process/Log.hs:37:3) test.hs: Prelude.undefined CallStack (from HasCallStack): error, called at libraries/base/GHC/Err.hs:79:14 in base:GHC.Err undefined, called at /home/foo/test.hs:4:8 in main:Main ``` ### Stack version ``` $ stack --version Version 1.5.1, Git revision 2b0392468a03174a7c5d317291c62452d771b055 x86_64 hpack-0.18.1 ``` ### Method of installation * Official binary, downloaded from stackage.org or fpcomplete's package repository
non_process
don t work stack script resolver lts package ghc package ghc paths failure bolow command bash stack script main hs resolver lts package ghc package ghc paths but success below command bash stack script main hs resolver lts package ghc paths stack script main hs resolver lts package ghc package ghc paths steps to reproduce clean dir rm rf stack create main hs haskell module main where main io main undefined run stack script main hs resolver lts package ghc package ghc paths expected i expected that displayed to result undefined after build run actual failure case not found ghc package error bash stack script test hs resolver lts package ghc package ghc paths verbose version git revision hpack ignoring config files stack config hs using resolver lts specified on command line stack config hs using resolver lts specified on command line stack config hs decoding build plan from home foo stack build plan lts yaml stack snapshot hs run process sbin ldconfig p system process log hs process finished in sbin ldconfig p system process log hs run process usr bin gcc v system process log hs process finished in usr bin gcc v system process log hs pie enabled stack setup hs found shared library libtinfo so in ldconfig p output stack setup hs did not find shared library libtinfo so stack setup hs did not find shared library libncursesw so stack setup hs found shared library libgmp so in ldconfig p output stack setup hs did not find shared library libgmp so stack setup hs using standard ghc build stack setup hs getting global package database location stack ghcpkg hs asking ghc for its version stack setup installed hs getting cabal package version stack ghcpkg hs run process home foo stack programs linux ghc bin ghc pkg no user package db list global system process log hs run process home foo stack programs linux ghc bin ghc numeric version system process log hs run process home foo stack programs linux ghc bin ghc pkg no user package db field simple output cabal version system process log hs process finished in home foo stack programs linux ghc bin ghc pkg no user package db list global system process log hs process finished in home foo stack programs linux ghc bin ghc pkg no user package db field simple output cabal version system process log hs process finished in home foo stack programs linux ghc bin ghc numeric version system process log hs ghc version is ghc stack setup installed hs resolving package entries stack setup hs trying to decode home foo stack loaded snapshot cache linux ghc lts cache data store versiontagged hs success decoding home foo stack loaded snapshot cache linux ghc lts cache data store versiontagged hs starting to execute command inside envconfig stack runners hs run process home foo stack programs linux ghc bin ghc pkg list simple output system process log hs process finished in home foo stack programs linux ghc bin ghc pkg list simple output system process log hs missing packages performing installation stack script hs parsing the targets stack build target hs finding out which packages are already installed stack build installed hs run process home foo stack programs linux ghc bin ghc pkg global no user package db dump expand pkgroot system process log hs process finished in home foo stack programs linux ghc bin ghc pkg global no user package db dump expand pkgroot system process log hs ignoring package haskeline due to wanting version instead of stack build installed hs ignoring package terminfo due to wanting version instead of stack build installed hs ignoring package xhtml due to wanting version instead of stack build installed hs run process home foo stack programs linux ghc bin ghc pkg user no user package db package db home foo stack snapshots linux lts pkgdb dump expand pkgroot system process log hs process finished in home foo stack programs linux ghc bin ghc pkg user no user package db package db home foo stack snapshots linux lts pkgdb dump expand pkgroot system process log hs run process home foo stack programs linux ghc bin ghc pkg user no user package db package db home foo stack script lts stack work install linux lts pkgdb dump expand pkgroot system process log hs process finished in home foo stack programs linux ghc bin ghc pkg user no user package db package db home foo stack script lts stack work install linux lts pkgdb dump expand pkgroot system process log hs constructing the build plan stack build constructplan hs trying to decode home foo stack indices hackage index cache data store versiontagged hs success decoding home foo stack indices hackage index cache data store versiontagged hs didn t see ghc in your package indices updating and trying again stack fetch hs selected mirror stack packageindex hs downloading timestamp stack packageindex hs no updates to your package index were found stack packageindex hs update complete trying to decode home foo stack indices hackage index cache data store versiontagged hs success decoding home foo stack indices hackage index cache data store versiontagged hs the following package identifiers were not found in your indices ghc possible candidates ghc success case step bash stack script test hs resolver lts package ghc paths verbose version git revision hpack ignoring config files stack config hs using resolver lts specified on command line stack config hs using resolver lts specified on command line stack config hs decoding build plan from home foo stack build plan lts yaml stack snapshot hs run process sbin ldconfig p system process log hs process finished in sbin ldconfig p system process log hs run process usr bin gcc v system process log hs process finished in usr bin gcc v system process log hs pie enabled stack setup hs found shared library libtinfo so in ldconfig p output stack setup hs did not find shared library libtinfo so stack setup hs did not find shared library libncursesw so stack setup hs found shared library libgmp so in ldconfig p output stack setup hs did not find shared library libgmp so stack setup hs using standard ghc build stack setup hs getting global package database location stack ghcpkg hs asking ghc for its version stack setup installed hs getting cabal package version stack ghcpkg hs run process home foo stack programs linux ghc bin ghc pkg no user package db list global system process log hs run process home foo stack programs linux ghc bin ghc numeric version system process log hs run process home foo stack programs linux ghc bin ghc pkg no user package db field simple output cabal version system process log hs process finished in home foo stack programs linux ghc bin ghc pkg no user package db list global system process log hs process finished in home foo stack programs linux ghc bin ghc pkg no user package db field simple output cabal version system process log hs process finished in home foo stack programs linux ghc bin ghc numeric version system process log hs ghc version is ghc stack setup installed hs resolving package entries stack setup hs trying to decode home foo stack loaded snapshot cache linux ghc lts cache data store versiontagged hs success decoding home foo stack loaded snapshot cache linux ghc lts cache data store versiontagged hs starting to execute command inside envconfig stack runners hs run process home foo stack programs linux ghc bin ghc pkg list simple output system process log hs process finished in home foo stack programs linux ghc bin ghc pkg list simple output system process log hs missing packages performing installation stack script hs parsing the targets stack build target hs finding out which packages are already installed stack build installed hs run process home foo stack programs linux ghc bin ghc pkg global no user package db dump expand pkgroot system process log hs process finished in home foo stack programs linux ghc bin ghc pkg global no user package db dump expand pkgroot system process log hs ignoring package haskeline due to wanting version instead of stack build installed hs ignoring package terminfo due to wanting version instead of stack build installed hs ignoring package xhtml due to wanting version instead of stack build installed hs run process home foo stack programs linux ghc bin ghc pkg user no user package db package db home foo stack snapshots linux lts pkgdb dump expand pkgroot system process log hs process finished in home foo stack programs linux ghc bin ghc pkg user no user package db package db home foo stack snapshots linux lts pkgdb dump expand pkgroot system process log hs run process home foo stack programs linux ghc bin ghc pkg user no user package db package db home foo stack script lts stack work install linux lts pkgdb dump expand pkgroot system process log hs process finished in home foo stack programs linux ghc bin ghc pkg user no user package db package db home foo stack script lts stack work install linux lts pkgdb dump expand pkgroot system process log hs constructing the build plan stack build constructplan hs trying to decode home foo stack indices hackage index cache data store versiontagged hs success decoding home foo stack indices hackage index cache data store versiontagged hs checking if we are going to build multiple executables with the same name stack build hs executing the build plan stack build execute hs creating process home foo stack programs linux ghc bin ghc rtsopts threaded clear package db global package db hide all packages package base main is stacksetupshim mainoverride package cabal home foo stack setup exe src setup mphdzzaj hs home foo stack setup exe src setup shim mphdzzaj hs o home foo stack setup exe cache linux tmp cabal simple mphdzzaj ghc system process log hs compiling main home foo stack setup exe src setup mphdzzaj hs home foo stack setup exe src setup mphdzzaj o compiling stacksetupshim home foo stack setup exe src setup shim mphdzzaj hs home foo stack setup exe src setup shim mphdzzaj o linking home foo stack setup exe cache linux tmp cabal simple mphdzzaj ghc getting global package database location stack ghcpkg hs run process home foo stack programs linux ghc bin ghc pkg no user package db list global system process log hs process finished in home foo stack programs linux ghc bin ghc pkg no user package db list global system process log hs exception ignored when attempting to load home foo stack precompiled linux ghc ghc paths home foo stack precompiled linux ghc ghc paths openbinaryfile does not exist no such file or directory data store versiontagged hs downloading hackage fpcomplete com package ghc paths tar gz network http download verified hs ghc paths download stack fetch hs exception ignored when attempting to load tmp ghc paths stack work dist linux cabal stack config cache tmp ghc paths stack work dist linux cabal stack config cache openbinaryfile does not exist no such file or directory data store versiontagged hs exception ignored when attempting to load tmp ghc paths stack work dist linux cabal stack cabal mod tmp ghc paths stack work dist linux cabal stack cabal mod openbinaryfile does not exist no such file or directory data store versiontagged hs ghc paths configure stack build execute hs run process home foo stack programs linux ghc bin ghc make odir tmp ghc paths stack work dist linux cabal setup hidir tmp ghc paths stack work dist linux cabal setup i i clear package db global package db package db home foo stack snapshots linux lts pkgdb package db home foo stack script lts stack work install linux lts pkgdb hide all packages package id cabal package id base package id directory optp include optp tmp ghc paths stack work dist linux cabal setup setup macros h tmp ghc paths setup hs home foo stack setup exe src setup shim mphdzzaj hs main is stacksetupshim mainoverride o tmp ghc paths stack work dist linux cabal setup setup threaded system process log hs process finished in home foo stack programs linux ghc bin ghc make odir tmp ghc paths stack work dist linux cabal setup hidir tmp ghc paths stack work dist linux cabal setup i i clear package db global package db package db home foo stack snapshots linux lts pkgdb package db home foo stack script lts stack work install linux lts pkgdb hide all packages package id cabal package id base package id directory optp include optp tmp ghc paths stack work dist linux cabal setup setup macros h tmp ghc paths setup hs home foo stack setup exe src setup shim mphdzzaj hs main is stacksetupshim mainoverride o tmp ghc paths stack work dist linux cabal setup setup threaded system process log hs run process tmp ghc paths stack work dist linux cabal setup setup builddir stack work dist linux cabal configure with ghc home foo stack programs linux ghc bin ghc with ghc pkg home foo stack programs linux ghc bin ghc pkg user package db clear package db global package db home foo stack snapshots linux lts pkgdb libdir home foo stack snapshots linux lts lib bindir home foo stack snapshots linux lts bin datadir home foo stack snapshots linux lts share libexecdir home foo stack snapshots linux lts libexec sysconfdir home foo stack snapshots linux lts etc docdir home foo stack snapshots linux lts doc ghc paths htmldir home foo stack snapshots linux lts doc ghc paths haddockdir home foo stack snapshots linux lts doc ghc paths dependency cabal cabal dependency base base dependency directory directory system process log hs process finished in tmp ghc paths stack work dist linux cabal setup setup builddir stack work dist linux cabal configure with ghc home foo stack programs linux ghc bin ghc with ghc pkg home foo stack programs linux ghc bin ghc pkg user package db clear package db global package db home foo stack snapshots linux lts pkgdb libdir home foo stack snapshots linux lts lib bindir home foo stack snapshots linux lts bin datadir home foo stack snapshots linux lts share libexecdir home foo stack snapshots linux lts libexec sysconfdir home foo stack snapshots linux lts etc docdir home foo stack snapshots linux lts doc ghc paths htmldir home foo stack snapshots linux lts doc ghc paths haddockdir home foo stack snapshots linux lts doc ghc paths dependency cabal cabal dependency base base dependency directory directory system process log hs encoding tmp ghc paths stack work dist linux cabal stack config cache data store versiontagged hs finished writing tmp ghc paths stack work dist linux cabal stack config cache data store versiontagged hs encoding tmp ghc paths stack work dist linux cabal stack cabal mod data store versiontagged hs finished writing tmp ghc paths stack work dist linux cabal stack cabal mod data store versiontagged hs ghc paths build stack build execute hs run process tmp ghc paths stack work dist linux cabal setup setup builddir stack work dist linux cabal build ghc options ddump hi ddump to file system process log hs process finished in tmp ghc paths stack work dist linux cabal setup setup builddir stack work dist linux cabal build ghc options ddump hi ddump to file system process log hs ghc paths copy register stack build execute hs run process tmp ghc paths stack work dist linux cabal setup setup builddir stack work dist linux cabal copy system process log hs process finished in tmp ghc paths stack work dist linux cabal setup setup builddir stack work dist linux cabal copy system process log hs run process tmp ghc paths stack work dist linux cabal setup setup builddir stack work dist linux cabal register system process log hs process finished in tmp ghc paths stack work dist linux cabal setup setup builddir stack work dist linux cabal register system process log hs run process home foo stack programs linux ghc bin ghc pkg user no user package db package db home foo stack snapshots linux lts pkgdb describe simple output ghc paths expand pkgroot system process log hs process finished in home foo stack programs linux ghc bin ghc pkg user no user package db package db home foo stack snapshots linux lts pkgdb describe simple output ghc paths expand pkgroot system process log hs encoding home foo stack precompiled linux ghc ghc paths data store versiontagged hs finished writing home foo stack precompiled linux ghc ghc paths data store versiontagged hs encoding home foo stack script lts stack work install linux lts flag cache ghc paths data store versiontagged hs finished writing home foo stack script lts stack work install linux lts flag cache ghc paths data store versiontagged hs run process home foo stack programs linux ghc bin runghc hide all packages packagebase packageghc paths home foo test hs system process log hs test hs prelude undefined callstack from hascallstack error called at libraries base ghc err hs in base ghc err undefined called at home foo test hs in main main step bash stack script test hs resolver lts package ghc package ghc paths verbose version git revision hpack ignoring config files stack config hs using resolver lts specified on command line stack config hs using resolver lts specified on command line stack config hs decoding build plan from home foo stack build plan lts yaml stack snapshot hs run process sbin ldconfig p system process log hs process finished in sbin ldconfig p system process log hs run process usr bin gcc v system process log hs process finished in usr bin gcc v system process log hs pie enabled stack setup hs found shared library libtinfo so in ldconfig p output stack setup hs did not find shared library libtinfo so stack setup hs did not find shared library libncursesw so stack setup hs found shared library libgmp so in ldconfig p output stack setup hs did not find shared library libgmp so stack setup hs using standard ghc build stack setup hs asking ghc for its version stack setup installed hs run process home foo stack programs linux ghc bin ghc numeric version system process log hs getting cabal package version stack ghcpkg hs getting global package database location stack ghcpkg hs run process home foo stack programs linux ghc bin ghc pkg no user package db field simple output cabal version system process log hs run process home foo stack programs linux ghc bin ghc pkg no user package db list global system process log hs process finished in home foo stack programs linux ghc bin ghc pkg no user package db field simple output cabal version system process log hs process finished in home foo stack programs linux ghc bin ghc numeric version system process log hs ghc version is ghc stack setup installed hs process finished in home foo stack programs linux ghc bin ghc pkg no user package db list global system process log hs resolving package entries stack setup hs trying to decode home foo stack loaded snapshot cache linux ghc lts cache data store versiontagged hs success decoding home foo stack loaded snapshot cache linux ghc lts cache data store versiontagged hs starting to execute command inside envconfig stack runners hs run process home foo stack programs linux ghc bin ghc pkg list simple output system process log hs process finished in home foo stack programs linux ghc bin ghc pkg list simple output system process log hs all packages already installed stack script hs run process home foo stack programs linux ghc bin runghc hide all packages packagebase packageghc packageghc paths home foo test hs system process log hs test hs prelude undefined callstack from hascallstack error called at libraries base ghc err hs in base ghc err undefined called at home foo test hs in main main stack version stack version version git revision hpack method of installation official binary downloaded from stackage org or fpcomplete s package repository
0
101,158
16,493,344,728
IssuesEvent
2021-05-25 07:37:28
benchmarkdebricked/angular
https://api.github.com/repos/benchmarkdebricked/angular
opened
CVE-2020-7645 (High) detected in chrome-launcher-0.10.5.tgz
security vulnerability
## CVE-2020-7645 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>chrome-launcher-0.10.5.tgz</b></p></summary> <p>Launch latest Chrome with the Devtools Protocol port open</p> <p>Library home page: <a href="https://registry.npmjs.org/chrome-launcher/-/chrome-launcher-0.10.5.tgz">https://registry.npmjs.org/chrome-launcher/-/chrome-launcher-0.10.5.tgz</a></p> <p>Path to dependency file: angular/aio/package.json</p> <p>Path to vulnerable library: angular/aio/node_modules/chrome-launcher</p> <p> Dependency Hierarchy: - :x: **chrome-launcher-0.10.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/benchmarkdebricked/angular/commit/69ad0ced6b2b3ebe8f9d4d3b9dda14051592bbdf">69ad0ced6b2b3ebe8f9d4d3b9dda14051592bbdf</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of chrome-launcher allow execution of arbitrary commands, by controlling the $HOME environment variable in Linux operating systems. <p>Publish Date: 2020-05-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7645>CVE-2020-7645</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7645 (High) detected in chrome-launcher-0.10.5.tgz - ## CVE-2020-7645 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>chrome-launcher-0.10.5.tgz</b></p></summary> <p>Launch latest Chrome with the Devtools Protocol port open</p> <p>Library home page: <a href="https://registry.npmjs.org/chrome-launcher/-/chrome-launcher-0.10.5.tgz">https://registry.npmjs.org/chrome-launcher/-/chrome-launcher-0.10.5.tgz</a></p> <p>Path to dependency file: angular/aio/package.json</p> <p>Path to vulnerable library: angular/aio/node_modules/chrome-launcher</p> <p> Dependency Hierarchy: - :x: **chrome-launcher-0.10.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/benchmarkdebricked/angular/commit/69ad0ced6b2b3ebe8f9d4d3b9dda14051592bbdf">69ad0ced6b2b3ebe8f9d4d3b9dda14051592bbdf</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of chrome-launcher allow execution of arbitrary commands, by controlling the $HOME environment variable in Linux operating systems. <p>Publish Date: 2020-05-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7645>CVE-2020-7645</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in chrome launcher tgz cve high severity vulnerability vulnerable library chrome launcher tgz launch latest chrome with the devtools protocol port open library home page a href path to dependency file angular aio package json path to vulnerable library angular aio node modules chrome launcher dependency hierarchy x chrome launcher tgz vulnerable library found in head commit a href vulnerability details all versions of chrome launcher allow execution of arbitrary commands by controlling the home environment variable in linux operating systems publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href step up your open source security game with whitesource
0
13,765
16,524,669,475
IssuesEvent
2021-05-26 18:26:15
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
Publish macOS/arm64 binaries
kind/process lifecycle/rotten priority/important-longterm
NOTE: No released version of Go supports macOS/arm64 as of this date, but v1.16 will. https://github.com/golang/go/issues/38485 has some details on how to get started with making builds from the go trunk source code. We may have to do something about code-signing on macOS/arm64 Related: #9224 We'll want to make sure that the documentation points to the appropriate URL's as well.
1.0
Publish macOS/arm64 binaries - NOTE: No released version of Go supports macOS/arm64 as of this date, but v1.16 will. https://github.com/golang/go/issues/38485 has some details on how to get started with making builds from the go trunk source code. We may have to do something about code-signing on macOS/arm64 Related: #9224 We'll want to make sure that the documentation points to the appropriate URL's as well.
process
publish macos binaries note no released version of go supports macos as of this date but will has some details on how to get started with making builds from the go trunk source code we may have to do something about code signing on macos related we ll want to make sure that the documentation points to the appropriate url s as well
1
697,686
23,948,969,183
IssuesEvent
2022-09-12 09:49:06
rathena/rathena
https://api.github.com/repos/rathena/rathena
closed
[RE] Gunslinger job change reward
component:script priority:low mode:renewal
* **rAthena Hash**: 78cda50e3b5cfb3d053c7d08181345418e5a83a1 * **Client Date**: N/A * **Server Mode**: Renewal * **Description of Issue**: * Result: Missing renewal reward items from changing Novice to Gunslinger, preventing gunslinger changing Novice Revolver into other guns by talking to Sharp Snake's Fang in academy * Expected Result: Renewal reward items need to be updated to follow official * How to Reproduce: Check script * Official Information: https://irowiki.org/wiki/First_Expanded_Job_Change_Guides#Gunslinger I don't have Gunslinger in iRO to test, but the reward should be same as wiki. Reward items: 1 Novice Rifle [3], 2 Cartridge and 1 Silver Cartridge * **Modifications that may affect results**: N/A
1.0
[RE] Gunslinger job change reward - * **rAthena Hash**: 78cda50e3b5cfb3d053c7d08181345418e5a83a1 * **Client Date**: N/A * **Server Mode**: Renewal * **Description of Issue**: * Result: Missing renewal reward items from changing Novice to Gunslinger, preventing gunslinger changing Novice Revolver into other guns by talking to Sharp Snake's Fang in academy * Expected Result: Renewal reward items need to be updated to follow official * How to Reproduce: Check script * Official Information: https://irowiki.org/wiki/First_Expanded_Job_Change_Guides#Gunslinger I don't have Gunslinger in iRO to test, but the reward should be same as wiki. Reward items: 1 Novice Rifle [3], 2 Cartridge and 1 Silver Cartridge * **Modifications that may affect results**: N/A
non_process
gunslinger job change reward rathena hash client date n a server mode renewal description of issue result missing renewal reward items from changing novice to gunslinger preventing gunslinger changing novice revolver into other guns by talking to sharp snake s fang in academy expected result renewal reward items need to be updated to follow official how to reproduce check script official information i don t have gunslinger in iro to test but the reward should be same as wiki reward items novice rifle cartridge and silver cartridge modifications that may affect results n a
0
9,425
12,417,733,467
IssuesEvent
2020-05-22 21:31:08
CDLUC3/mrt-doc
https://api.github.com/repos/CDLUC3/mrt-doc
closed
Investigate methods to automate aspects of collection creation
Tools & Processes
### Summary Given the upcoming reduction in cost per/TB for Merritt storage, UCB has expressed interest in creating multiple, new collections. "Multiple" may mean as many as 100 or more collections. We should investigate what aspects of Merritt collection creation could be automated. ### Tasks - [ ] Refresh documentation regarding collection creation - [ ] Investigate which steps might be easily automated in the near term, with the end goal of eventually automating the entire process.
1.0
Investigate methods to automate aspects of collection creation - ### Summary Given the upcoming reduction in cost per/TB for Merritt storage, UCB has expressed interest in creating multiple, new collections. "Multiple" may mean as many as 100 or more collections. We should investigate what aspects of Merritt collection creation could be automated. ### Tasks - [ ] Refresh documentation regarding collection creation - [ ] Investigate which steps might be easily automated in the near term, with the end goal of eventually automating the entire process.
process
investigate methods to automate aspects of collection creation summary given the upcoming reduction in cost per tb for merritt storage ucb has expressed interest in creating multiple new collections multiple may mean as many as or more collections we should investigate what aspects of merritt collection creation could be automated tasks refresh documentation regarding collection creation investigate which steps might be easily automated in the near term with the end goal of eventually automating the entire process
1
184,013
6,700,139,846
IssuesEvent
2017-10-11 02:38:41
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www3.lenovo.com - site is not usable
browser-firefox-mobile-tablet priority-important status-worksforme
<!-- @browser: Firefox Mobile (Tablet) 58.0 --> <!-- @ua_header: Mozilla/5.0 (Android 7.0; Tablet; rv:58.0) Gecko/58.0 Firefox/58.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www3.lenovo.com/jp/ja/notebooks/thinkpad/t-series/T470s/p/20HGCTO1WWJAJP9/customize **Browser / Version**: Firefox Mobile (Tablet) 58.0 **Operating System**: Android 7.0 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: customization page doesn't open **Steps to Reproduce**: By clicking any of the thinkpad products to buy. The pages don't open in Firefox but do with Chrome. (Linux desktop has the same issue) [![Screenshot Description](https://webcompat.com/uploads/2017/10/13811d5f-a07d-4b2a-8f49-ff78f8b05057-thumb.jpg)](https://webcompat.com/uploads/2017/10/13811d5f-a07d-4b2a-8f49-ff78f8b05057.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www3.lenovo.com - site is not usable - <!-- @browser: Firefox Mobile (Tablet) 58.0 --> <!-- @ua_header: Mozilla/5.0 (Android 7.0; Tablet; rv:58.0) Gecko/58.0 Firefox/58.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www3.lenovo.com/jp/ja/notebooks/thinkpad/t-series/T470s/p/20HGCTO1WWJAJP9/customize **Browser / Version**: Firefox Mobile (Tablet) 58.0 **Operating System**: Android 7.0 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: customization page doesn't open **Steps to Reproduce**: By clicking any of the thinkpad products to buy. The pages don't open in Firefox but do with Chrome. (Linux desktop has the same issue) [![Screenshot Description](https://webcompat.com/uploads/2017/10/13811d5f-a07d-4b2a-8f49-ff78f8b05057-thumb.jpg)](https://webcompat.com/uploads/2017/10/13811d5f-a07d-4b2a-8f49-ff78f8b05057.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
lenovo com site is not usable url browser version firefox mobile tablet operating system android tested another browser yes problem type site is not usable description customization page doesn t open steps to reproduce by clicking any of the thinkpad products to buy the pages don t open in firefox but do with chrome linux desktop has the same issue from with ❤️
0
433,929
12,512,447,795
IssuesEvent
2020-06-02 22:50:09
googleapis/gax-nodejs
https://api.github.com/repos/googleapis/gax-nodejs
closed
Add more information on "Retry total timeout exceeded before any response was received" error.
priority: p2 type: feature request
Hi, `gax-nodejs` team, I'd like to propose to add some more information on error message for ["Retry total timeout error"](https://github.com/googleapis/gax-nodejs/blob/4b9a35e509c5f617de3e2212cfa66aa75192f98a/src/normalCalls/retries.ts#L82-L89) so that GCP users can easily debug which GCP library and which API calls are hitting "retry total timeout" error. (I've already contacted GCP Support Team, they created an [issue on public issue tracker](https://issuetracker.google.com/issues/150856656), and suggested me to create an issue here). ## Is your feature request related to a problem? Please describe. (background) We are running Google Cloud Function with node10, and we occasionally hit the "Retry total timeout exceeded before any response was received" error. <details> <summary>FYI my piece of yarn.lock related to google-gax</summary> ``` "@google-cloud/datastore@^4.5.2": version "4.5.2" resolved "https://registry.yarnpkg.com/@google-cloud/datastore/-/datastore-4.5.2.tgz#b1f9300ca2de91bcfd5783f54354411ece1446bc" integrity sha512-dILePgHiA14/6l47XrsTjQqGhZVtznrqg3Q3ent5jOhL+0QGGokY+Y3OtYTQ1yCAfUg/ZcwjLY1s2MBbrHqnog== dependencies: "@google-cloud/projectify" "^1.0.0" "@google-cloud/promisify" "^1.0.0" "@grpc/grpc-js" "0.6.11" "@types/duplexify" "^3.6.0" "@types/long" "^4.0.0" arrify "^2.0.1" concat-stream "^2.0.0" extend "^3.0.2" google-auth-library "^5.0.0" google-gax "^1.7.5" is "^3.3.0" split-array-stream "^2.0.0" stream-events "^1.0.5" "@google-cloud/logging@^7.0.0": version "7.2.3" resolved "https://registry.yarnpkg.com/@google-cloud/logging/-/logging-7.2.3.tgz#499f04bb57c607e28c426566853e632eab13b8cf" integrity sha512-MLAlYVBihCs0e581n9VUYOPJcrSpwSdL7KAjy2wgQidmRD9aWjvg97F22JMCtvrUiFBmAjiJVtn1JyNIXlv0Yw== dependencies: "@google-cloud/common" "^2.2.2" "@google-cloud/paginator" "^2.0.0" "@google-cloud/projectify" "^1.0.0" "@google-cloud/promisify" "^1.0.0" "@opencensus/propagation-stackdriver" "0.0.20" arrify "^2.0.0" dot-prop "^5.1.0" eventid "^1.0.0" extend "^3.0.2" gcp-metadata "^3.1.0" google-auth-library "^5.2.2" google-gax "^1.11.0" is "^3.3.0" on-finished "^2.3.0" pumpify "^2.0.0" snakecase-keys "^3.0.0" stream-events "^1.0.4" through2 "^3.0.0" type-fest "^0.11.0" "@google-cloud/pubsub@^1.5.0": version "1.5.0" resolved "https://registry.yarnpkg.com/@google-cloud/pubsub/-/pubsub-1.5.0.tgz#0ce5a60a90c87af9e9dcd28433a0da74c223007b" integrity sha512-SCuNClo/xDGjYmciExxVjX78WPY1Ul6MK5Qn5eX2tsILqxXpf5Lan+XU/jnL53pirAIFgcxt8I2CWibSdSqRww== dependencies: "@google-cloud/paginator" "^2.0.0" "@google-cloud/precise-date" "^1.0.0" "@google-cloud/projectify" "^1.0.0" "@google-cloud/promisify" "^1.0.0" "@types/duplexify" "^3.6.0" "@types/long" "^4.0.0" arrify "^2.0.0" async-each "^1.0.1" extend "^3.0.2" google-auth-library "^5.5.0" google-gax "^1.7.5" is-stream-ended "^0.1.4" lodash.snakecase "^4.1.1" p-defer "^3.0.0" protobufjs "^6.8.1" "@google-cloud/vision@^1.7.2": version "1.9.0" resolved "https://registry.yarnpkg.com/@google-cloud/vision/-/vision-1.9.0.tgz#3dd9708a2d87c5a2ab46c104f2a320ee99cd51fb" integrity sha512-6ka0MggI8u4WgGJe1wAIdxIAKE9gRp1CQ5ADXNM+JmSUniEZfbeNwOG8twFCxyWBlUqC+cQJaG4Amd6SLXzk/w== dependencies: "@google-cloud/promisify" "^1.0.0" google-gax "^1.7.5" is "^3.2.1" gaxios@^2.0.1, gaxios@^2.1.0: version "2.3.1" resolved "https://registry.yarnpkg.com/gaxios/-/gaxios-2.3.1.tgz#52bd832b5d6a252072783b9afd9742bde835b2f4" integrity sha512-DQOesWEx59/bm63lTX0uHDDXpGTW9oKqNsoigwCoRe2lOb5rFqxzHjLTa6aqEBecLcz69dHLw7rbS068z1fvIQ== dependencies: abort-controller "^3.0.0" extend "^3.0.2" https-proxy-agent "^5.0.0" is-stream "^2.0.0" node-fetch "^2.3.0" gcp-metadata@^3.0.0, gcp-metadata@^3.1.0, gcp-metadata@^3.4.0: version "3.4.0" resolved "https://registry.yarnpkg.com/gcp-metadata/-/gcp-metadata-3.4.0.tgz#dfcbb6abe2c262c86c8eb01e4729d46b44e21484" integrity sha512-fizmBtCXHp8b7FZuzbgKaixO8DzsSYoEVmMgZIna7x8t6cfBF3eqirODWYxVbgmasA5qudCAKiszfB7yVwroIQ== dependencies: gaxios "^2.1.0" json-bigint "^0.3.0" google-auth-library@^5.0.0, google-auth-library@^5.2.2, google-auth-library@^5.5.0, google-auth-library@^5.6.1, google-auth-library@^5.9.0: version "5.10.1" resolved "https://registry.yarnpkg.com/google-auth-library/-/google-auth-library-5.10.1.tgz#504ec75487ad140e68dd577c21affa363c87ddff" integrity sha512-rOlaok5vlpV9rSiUu5EpR0vVpc+PhN62oF4RyX/6++DG1VsaulAFEMlDYBLjJDDPI6OcNOCGAKy9UVB/3NIDXg== dependencies: arrify "^2.0.0" base64-js "^1.3.0" ecdsa-sig-formatter "^1.0.11" fast-text-encoding "^1.0.0" gaxios "^2.1.0" gcp-metadata "^3.4.0" gtoken "^4.1.0" jws "^4.0.0" lru-cache "^5.0.0" google-gax@^1.11.0, google-gax@^1.7.5: version "1.14.2" resolved "https://registry.yarnpkg.com/google-gax/-/google-gax-1.14.2.tgz#ce4f9a42c1bc2ca4a4ed8e8cc70c6f7a3548b790" integrity sha512-Nde+FdqALbV3QgMA4KlkxOHfrj9busnZ3EECwy/1gDJm9vhKGwDLWzErqRU5g80OoGSAMgyY7DWIfqz7ina4Jw== dependencies: "@grpc/grpc-js" "^0.6.18" "@grpc/proto-loader" "^0.5.1" "@types/fs-extra" "^8.0.1" "@types/long" "^4.0.0" abort-controller "^3.0.0" duplexify "^3.6.0" google-auth-library "^5.0.0" is-stream-ended "^0.1.4" lodash.at "^4.6.0" lodash.has "^4.5.2" node-fetch "^2.6.0" protobufjs "^6.8.8" retry-request "^4.0.0" semver "^6.0.0" walkdir "^0.4.0" googleapis-common@^3.2.0: version "3.2.1" resolved "https://registry.yarnpkg.com/googleapis-common/-/googleapis-common-3.2.1.tgz#1c81dfce74876948fb1e68a2a99f842da6123877" integrity sha512-lPAJXYpZLvY4AZp57RWj1eUXS2D5LxS3y6W0n9Jbl6eKOWmyLgz99f84XVCxVnyq8DcgV/ZAxt1lpFEzGZkVvQ== dependencies: extend "^3.0.2" gaxios "^2.0.1" google-auth-library "^5.6.1" qs "^6.7.0" url-template "^2.0.8" uuid "^3.3.2" gtoken@^4.1.0: version "4.1.4" resolved "https://registry.yarnpkg.com/gtoken/-/gtoken-4.1.4.tgz#925ff1e7df3aaada06611d30ea2d2abf60fcd6a7" integrity sha512-VxirzD0SWoFUo5p8RDP8Jt2AGyOmyYcT/pOUgDKJCK+iSw0TMqwrVfY37RXTNmoKwrzmDHSk0GMT9FsgVmnVSA== dependencies: gaxios "^2.1.0" google-p12-pem "^2.0.0" jws "^4.0.0" mime "^2.2.0" ``` </details> I believe this error happens when google cloud library (technically its underlying library, "gax-nodejs") tries to access GCP API and didn't get a response from the server within the deadline here https://github.com/googleapis/gax-nodejs/blob/4b9a35e509c5f617de3e2212cfa66aa75192f98a/src/normalCalls/retries.ts#L82-L89 I tried to debug which API call and which library emit this error message, but I couldn't because the error message and stack trace don't provide any information other than 'Retry total timeout exceeded before any response was received'. Here's the full stack-trace when we hit the error. ``` Error: Retry total timeout exceeded before any response was received at repeat (/srv/functions/node_modules/google-gax/build/src/normalCalls/retries.js:65:31) at Timeout.setTimeout [as _onTimeout] (/srv/functions/node_modules/google-gax/build/src/normalCalls/retries.js:100:25) at ontimeout (timers.js:436:11) at tryOnTimeout (timers.js:300:5) at listOnTimeout (timers.js:263:5) at Timer.processTimers (timers.js:223:10) ``` ## Describe the solution you'd like Therefore, I'd like to propose that gax shows some more information on DEADLINE_EXCEEDED error so that the GCP users can easily debug and where's the root cause of this issue. For example: - Which library, or API didn't respond within the deadline. - How long the API call took. - What is the retry settings? Unfortunately, since I'm not familiar with GRPC and GCP internal endpoints, I'm not sure is it possible. ### Additional context I guess an informative error message would be helpful for a lot of GCP users because as far as I know, there're quite a few people who are suffering from this issue: - https://github.com/googleapis/nodejs-pubsub/issues/770 - https://github.com/googleapis/google-auth-library-nodejs/issues/283 Anyway, thank you so much for maintaining this fundamental library for GCP, you always save my days :)
1.0
Add more information on "Retry total timeout exceeded before any response was received" error. - Hi, `gax-nodejs` team, I'd like to propose to add some more information on error message for ["Retry total timeout error"](https://github.com/googleapis/gax-nodejs/blob/4b9a35e509c5f617de3e2212cfa66aa75192f98a/src/normalCalls/retries.ts#L82-L89) so that GCP users can easily debug which GCP library and which API calls are hitting "retry total timeout" error. (I've already contacted GCP Support Team, they created an [issue on public issue tracker](https://issuetracker.google.com/issues/150856656), and suggested me to create an issue here). ## Is your feature request related to a problem? Please describe. (background) We are running Google Cloud Function with node10, and we occasionally hit the "Retry total timeout exceeded before any response was received" error. <details> <summary>FYI my piece of yarn.lock related to google-gax</summary> ``` "@google-cloud/datastore@^4.5.2": version "4.5.2" resolved "https://registry.yarnpkg.com/@google-cloud/datastore/-/datastore-4.5.2.tgz#b1f9300ca2de91bcfd5783f54354411ece1446bc" integrity sha512-dILePgHiA14/6l47XrsTjQqGhZVtznrqg3Q3ent5jOhL+0QGGokY+Y3OtYTQ1yCAfUg/ZcwjLY1s2MBbrHqnog== dependencies: "@google-cloud/projectify" "^1.0.0" "@google-cloud/promisify" "^1.0.0" "@grpc/grpc-js" "0.6.11" "@types/duplexify" "^3.6.0" "@types/long" "^4.0.0" arrify "^2.0.1" concat-stream "^2.0.0" extend "^3.0.2" google-auth-library "^5.0.0" google-gax "^1.7.5" is "^3.3.0" split-array-stream "^2.0.0" stream-events "^1.0.5" "@google-cloud/logging@^7.0.0": version "7.2.3" resolved "https://registry.yarnpkg.com/@google-cloud/logging/-/logging-7.2.3.tgz#499f04bb57c607e28c426566853e632eab13b8cf" integrity sha512-MLAlYVBihCs0e581n9VUYOPJcrSpwSdL7KAjy2wgQidmRD9aWjvg97F22JMCtvrUiFBmAjiJVtn1JyNIXlv0Yw== dependencies: "@google-cloud/common" "^2.2.2" "@google-cloud/paginator" "^2.0.0" "@google-cloud/projectify" "^1.0.0" "@google-cloud/promisify" "^1.0.0" "@opencensus/propagation-stackdriver" "0.0.20" arrify "^2.0.0" dot-prop "^5.1.0" eventid "^1.0.0" extend "^3.0.2" gcp-metadata "^3.1.0" google-auth-library "^5.2.2" google-gax "^1.11.0" is "^3.3.0" on-finished "^2.3.0" pumpify "^2.0.0" snakecase-keys "^3.0.0" stream-events "^1.0.4" through2 "^3.0.0" type-fest "^0.11.0" "@google-cloud/pubsub@^1.5.0": version "1.5.0" resolved "https://registry.yarnpkg.com/@google-cloud/pubsub/-/pubsub-1.5.0.tgz#0ce5a60a90c87af9e9dcd28433a0da74c223007b" integrity sha512-SCuNClo/xDGjYmciExxVjX78WPY1Ul6MK5Qn5eX2tsILqxXpf5Lan+XU/jnL53pirAIFgcxt8I2CWibSdSqRww== dependencies: "@google-cloud/paginator" "^2.0.0" "@google-cloud/precise-date" "^1.0.0" "@google-cloud/projectify" "^1.0.0" "@google-cloud/promisify" "^1.0.0" "@types/duplexify" "^3.6.0" "@types/long" "^4.0.0" arrify "^2.0.0" async-each "^1.0.1" extend "^3.0.2" google-auth-library "^5.5.0" google-gax "^1.7.5" is-stream-ended "^0.1.4" lodash.snakecase "^4.1.1" p-defer "^3.0.0" protobufjs "^6.8.1" "@google-cloud/vision@^1.7.2": version "1.9.0" resolved "https://registry.yarnpkg.com/@google-cloud/vision/-/vision-1.9.0.tgz#3dd9708a2d87c5a2ab46c104f2a320ee99cd51fb" integrity sha512-6ka0MggI8u4WgGJe1wAIdxIAKE9gRp1CQ5ADXNM+JmSUniEZfbeNwOG8twFCxyWBlUqC+cQJaG4Amd6SLXzk/w== dependencies: "@google-cloud/promisify" "^1.0.0" google-gax "^1.7.5" is "^3.2.1" gaxios@^2.0.1, gaxios@^2.1.0: version "2.3.1" resolved "https://registry.yarnpkg.com/gaxios/-/gaxios-2.3.1.tgz#52bd832b5d6a252072783b9afd9742bde835b2f4" integrity sha512-DQOesWEx59/bm63lTX0uHDDXpGTW9oKqNsoigwCoRe2lOb5rFqxzHjLTa6aqEBecLcz69dHLw7rbS068z1fvIQ== dependencies: abort-controller "^3.0.0" extend "^3.0.2" https-proxy-agent "^5.0.0" is-stream "^2.0.0" node-fetch "^2.3.0" gcp-metadata@^3.0.0, gcp-metadata@^3.1.0, gcp-metadata@^3.4.0: version "3.4.0" resolved "https://registry.yarnpkg.com/gcp-metadata/-/gcp-metadata-3.4.0.tgz#dfcbb6abe2c262c86c8eb01e4729d46b44e21484" integrity sha512-fizmBtCXHp8b7FZuzbgKaixO8DzsSYoEVmMgZIna7x8t6cfBF3eqirODWYxVbgmasA5qudCAKiszfB7yVwroIQ== dependencies: gaxios "^2.1.0" json-bigint "^0.3.0" google-auth-library@^5.0.0, google-auth-library@^5.2.2, google-auth-library@^5.5.0, google-auth-library@^5.6.1, google-auth-library@^5.9.0: version "5.10.1" resolved "https://registry.yarnpkg.com/google-auth-library/-/google-auth-library-5.10.1.tgz#504ec75487ad140e68dd577c21affa363c87ddff" integrity sha512-rOlaok5vlpV9rSiUu5EpR0vVpc+PhN62oF4RyX/6++DG1VsaulAFEMlDYBLjJDDPI6OcNOCGAKy9UVB/3NIDXg== dependencies: arrify "^2.0.0" base64-js "^1.3.0" ecdsa-sig-formatter "^1.0.11" fast-text-encoding "^1.0.0" gaxios "^2.1.0" gcp-metadata "^3.4.0" gtoken "^4.1.0" jws "^4.0.0" lru-cache "^5.0.0" google-gax@^1.11.0, google-gax@^1.7.5: version "1.14.2" resolved "https://registry.yarnpkg.com/google-gax/-/google-gax-1.14.2.tgz#ce4f9a42c1bc2ca4a4ed8e8cc70c6f7a3548b790" integrity sha512-Nde+FdqALbV3QgMA4KlkxOHfrj9busnZ3EECwy/1gDJm9vhKGwDLWzErqRU5g80OoGSAMgyY7DWIfqz7ina4Jw== dependencies: "@grpc/grpc-js" "^0.6.18" "@grpc/proto-loader" "^0.5.1" "@types/fs-extra" "^8.0.1" "@types/long" "^4.0.0" abort-controller "^3.0.0" duplexify "^3.6.0" google-auth-library "^5.0.0" is-stream-ended "^0.1.4" lodash.at "^4.6.0" lodash.has "^4.5.2" node-fetch "^2.6.0" protobufjs "^6.8.8" retry-request "^4.0.0" semver "^6.0.0" walkdir "^0.4.0" googleapis-common@^3.2.0: version "3.2.1" resolved "https://registry.yarnpkg.com/googleapis-common/-/googleapis-common-3.2.1.tgz#1c81dfce74876948fb1e68a2a99f842da6123877" integrity sha512-lPAJXYpZLvY4AZp57RWj1eUXS2D5LxS3y6W0n9Jbl6eKOWmyLgz99f84XVCxVnyq8DcgV/ZAxt1lpFEzGZkVvQ== dependencies: extend "^3.0.2" gaxios "^2.0.1" google-auth-library "^5.6.1" qs "^6.7.0" url-template "^2.0.8" uuid "^3.3.2" gtoken@^4.1.0: version "4.1.4" resolved "https://registry.yarnpkg.com/gtoken/-/gtoken-4.1.4.tgz#925ff1e7df3aaada06611d30ea2d2abf60fcd6a7" integrity sha512-VxirzD0SWoFUo5p8RDP8Jt2AGyOmyYcT/pOUgDKJCK+iSw0TMqwrVfY37RXTNmoKwrzmDHSk0GMT9FsgVmnVSA== dependencies: gaxios "^2.1.0" google-p12-pem "^2.0.0" jws "^4.0.0" mime "^2.2.0" ``` </details> I believe this error happens when google cloud library (technically its underlying library, "gax-nodejs") tries to access GCP API and didn't get a response from the server within the deadline here https://github.com/googleapis/gax-nodejs/blob/4b9a35e509c5f617de3e2212cfa66aa75192f98a/src/normalCalls/retries.ts#L82-L89 I tried to debug which API call and which library emit this error message, but I couldn't because the error message and stack trace don't provide any information other than 'Retry total timeout exceeded before any response was received'. Here's the full stack-trace when we hit the error. ``` Error: Retry total timeout exceeded before any response was received at repeat (/srv/functions/node_modules/google-gax/build/src/normalCalls/retries.js:65:31) at Timeout.setTimeout [as _onTimeout] (/srv/functions/node_modules/google-gax/build/src/normalCalls/retries.js:100:25) at ontimeout (timers.js:436:11) at tryOnTimeout (timers.js:300:5) at listOnTimeout (timers.js:263:5) at Timer.processTimers (timers.js:223:10) ``` ## Describe the solution you'd like Therefore, I'd like to propose that gax shows some more information on DEADLINE_EXCEEDED error so that the GCP users can easily debug and where's the root cause of this issue. For example: - Which library, or API didn't respond within the deadline. - How long the API call took. - What is the retry settings? Unfortunately, since I'm not familiar with GRPC and GCP internal endpoints, I'm not sure is it possible. ### Additional context I guess an informative error message would be helpful for a lot of GCP users because as far as I know, there're quite a few people who are suffering from this issue: - https://github.com/googleapis/nodejs-pubsub/issues/770 - https://github.com/googleapis/google-auth-library-nodejs/issues/283 Anyway, thank you so much for maintaining this fundamental library for GCP, you always save my days :)
non_process
add more information on retry total timeout exceeded before any response was received error hi gax nodejs team i d like to propose to add some more information on error message for so that gcp users can easily debug which gcp library and which api calls are hitting retry total timeout error i ve already contacted gcp support team they created an and suggested me to create an issue here is your feature request related to a problem please describe background we are running google cloud function with and we occasionally hit the retry total timeout exceeded before any response was received error fyi my piece of yarn lock related to google gax google cloud datastore version resolved integrity dependencies google cloud projectify google cloud promisify grpc grpc js types duplexify types long arrify concat stream extend google auth library google gax is split array stream stream events google cloud logging version resolved integrity dependencies google cloud common google cloud paginator google cloud projectify google cloud promisify opencensus propagation stackdriver arrify dot prop eventid extend gcp metadata google auth library google gax is on finished pumpify snakecase keys stream events type fest google cloud pubsub version resolved integrity scunclo xu dependencies google cloud paginator google cloud precise date google cloud projectify google cloud promisify types duplexify types long arrify async each extend google auth library google gax is stream ended lodash snakecase p defer protobufjs google cloud vision version resolved integrity w dependencies google cloud promisify google gax is gaxios gaxios version resolved integrity dependencies abort controller extend https proxy agent is stream node fetch gcp metadata gcp metadata gcp metadata version resolved integrity dependencies gaxios json bigint google auth library google auth library google auth library google auth library google auth library version resolved integrity dependencies arrify js ecdsa sig formatter fast text encoding gaxios gcp metadata gtoken jws lru cache google gax google gax version resolved integrity nde dependencies grpc grpc js grpc proto loader types fs extra types long abort controller duplexify google auth library is stream ended lodash at lodash has node fetch protobufjs retry request semver walkdir googleapis common version resolved integrity dependencies extend gaxios google auth library qs url template uuid gtoken version resolved integrity pougdkjck dependencies gaxios google pem jws mime i believe this error happens when google cloud library technically its underlying library gax nodejs tries to access gcp api and didn t get a response from the server within the deadline here i tried to debug which api call and which library emit this error message but i couldn t because the error message and stack trace don t provide any information other than retry total timeout exceeded before any response was received here s the full stack trace when we hit the error error retry total timeout exceeded before any response was received at repeat srv functions node modules google gax build src normalcalls retries js at timeout settimeout srv functions node modules google gax build src normalcalls retries js at ontimeout timers js at tryontimeout timers js at listontimeout timers js at timer processtimers timers js describe the solution you d like therefore i d like to propose that gax shows some more information on deadline exceeded error so that the gcp users can easily debug and where s the root cause of this issue for example which library or api didn t respond within the deadline how long the api call took what is the retry settings unfortunately since i m not familiar with grpc and gcp internal endpoints i m not sure is it possible additional context i guess an informative error message would be helpful for a lot of gcp users because as far as i know there re quite a few people who are suffering from this issue anyway thank you so much for maintaining this fundamental library for gcp you always save my days
0
18,040
24,051,687,955
IssuesEvent
2022-09-16 13:21:44
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Taxon constraint: U12-type spliceosomal complex (GO:0005689)
RNA processes taxon constraints protein complex
Please provide as much information as you can: * **GO term ID label:** U12-type spliceosomal complex (GO:0005689) * **Request to add a taxon constraint:** * * **only in taxon: ** * * **never in taxon: ** Never in yeast * **Request to remove a taxon constraint:** Please specify * **Supporting evidence if available (e.g PMID):** Repeating this analysis for all genomes in the IAOD (except S. cerevisiae, S. pombe and C. elegans, as they lack U12-type introns) r https://academic.oup.com/nar/article/48/13/7066/5850313
1.0
Taxon constraint: U12-type spliceosomal complex (GO:0005689) - Please provide as much information as you can: * **GO term ID label:** U12-type spliceosomal complex (GO:0005689) * **Request to add a taxon constraint:** * * **only in taxon: ** * * **never in taxon: ** Never in yeast * **Request to remove a taxon constraint:** Please specify * **Supporting evidence if available (e.g PMID):** Repeating this analysis for all genomes in the IAOD (except S. cerevisiae, S. pombe and C. elegans, as they lack U12-type introns) r https://academic.oup.com/nar/article/48/13/7066/5850313
process
taxon constraint type spliceosomal complex go please provide as much information as you can go term id label type spliceosomal complex go request to add a taxon constraint only in taxon never in taxon never in yeast request to remove a taxon constraint please specify supporting evidence if available e g pmid repeating this analysis for all genomes in the iaod except s cerevisiae s pombe and c elegans as they lack type introns r
1
13,955
16,737,259,191
IssuesEvent
2021-06-11 04:30:38
largecats/blog
https://api.github.com/repos/largecats/blog
opened
Stream Processing 101
/blog/2021/06/11/stream-processing-101/ Gitalk
https://largecats.github.io/blog/2021/06/11/stream-processing-101/ Background Batch Processing Stream Processing Concepts State State + Time Time Processing ...
1.0
Stream Processing 101 - https://largecats.github.io/blog/2021/06/11/stream-processing-101/ Background Batch Processing Stream Processing Concepts State State + Time Time Processing ...
process
stream processing background batch processing stream processing concepts state state time time processing
1
11,936
14,706,606,781
IssuesEvent
2021-01-04 20:08:42
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Prisma not applying limit when take specified with id cursor
bug/2-confirmed kind/bug process/candidate team/client topic: pagination
<!-- Thanks for helping us improve Prisma! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by setting the `DEBUG="*"` environment variable and enabling additional logging output in Prisma Client. Learn more about writing proper bug reports here: https://pris.ly/d/bug-reports --> ## Bug description I was noticing that when paginating through one of my larger tables with prisma, the first page was always loading quickly but subsequent ones were taking many seconds. My queries look like this: ```typescript function getPaginationArgs(cursor: string | undefined) { return { take: 10, cursor: cursor ? { id: cursor } : undefined, skip: cursor ? 1 : undefined, orderBy: { createdAt: "desc" }, } as const; } const page1 = await prisma.page.findMany(getPaginationArgs(undefined)); const cursor = page1[page1.length - 1].id; const page2 = await prisma.page.findMany(getPaginationArgs(cursor)); ``` I went in and looked at the queries prisma was issuing: First page: ```sql SELECT "prisma_test_schema_1"."page"."id", "prisma_test_schema_1"."page"."created_at", "prisma_test_schema_1"."page"."url" FROM "prisma_test_schema_1"."page" WHERE 1=1 ORDER BY "prisma_test_schema_1"."page"."created_at" DESC LIMIT $1 OFFSET $2 ``` Second Page: ```sql SELECT "prisma_test_schema_1"."page"."id", "prisma_test_schema_1"."page"."created_at", "prisma_test_schema_1"."page"."url" FROM "prisma_test_schema_1"."page", (SELECT "prisma_test_schema_1"."page"."created_at" FROM "prisma_test_schema_1"."page" WHERE ("prisma_test_schema_1"."page"."id") = ($1)) AS "order_cmp" WHERE "prisma_test_schema_1"."page"."created_at" <= "order_cmp"."created_at" ORDER BY "prisma_test_schema_1"."page"."created_at" DESC OFFSET $2 ``` The big difference here is that `LIMIT` is present in the first query but not the second. ## How to reproduce https://github.com/TLadd/prisma-unique-composite-key-query-bug/blob/master/README.md Follow instructions in README for a reproducible example. Basically just the above code snippet. ## Expected behavior I would expect the `LIMIT` to be applied to the second query as well. ## Prisma information ``` generator client { provider = "prisma-client-js" binaryTargets = ["native", "debian-openssl-1.1.x"] } datasource db { provider = "postgresql" url = env("DATABASE_URL") } model Page { id String @id @default(dbgenerated()) createdAt DateTime @default(now()) @map("created_at") url String @unique @@map("page") } ``` ## Environment & setup <!-- In which environment does the problem occur --> - OS: Mac OS - Database: PostgreSQL - Node.js version: v12.16.2 - Prisma version: 2.12.1
1.0
Prisma not applying limit when take specified with id cursor - <!-- Thanks for helping us improve Prisma! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by setting the `DEBUG="*"` environment variable and enabling additional logging output in Prisma Client. Learn more about writing proper bug reports here: https://pris.ly/d/bug-reports --> ## Bug description I was noticing that when paginating through one of my larger tables with prisma, the first page was always loading quickly but subsequent ones were taking many seconds. My queries look like this: ```typescript function getPaginationArgs(cursor: string | undefined) { return { take: 10, cursor: cursor ? { id: cursor } : undefined, skip: cursor ? 1 : undefined, orderBy: { createdAt: "desc" }, } as const; } const page1 = await prisma.page.findMany(getPaginationArgs(undefined)); const cursor = page1[page1.length - 1].id; const page2 = await prisma.page.findMany(getPaginationArgs(cursor)); ``` I went in and looked at the queries prisma was issuing: First page: ```sql SELECT "prisma_test_schema_1"."page"."id", "prisma_test_schema_1"."page"."created_at", "prisma_test_schema_1"."page"."url" FROM "prisma_test_schema_1"."page" WHERE 1=1 ORDER BY "prisma_test_schema_1"."page"."created_at" DESC LIMIT $1 OFFSET $2 ``` Second Page: ```sql SELECT "prisma_test_schema_1"."page"."id", "prisma_test_schema_1"."page"."created_at", "prisma_test_schema_1"."page"."url" FROM "prisma_test_schema_1"."page", (SELECT "prisma_test_schema_1"."page"."created_at" FROM "prisma_test_schema_1"."page" WHERE ("prisma_test_schema_1"."page"."id") = ($1)) AS "order_cmp" WHERE "prisma_test_schema_1"."page"."created_at" <= "order_cmp"."created_at" ORDER BY "prisma_test_schema_1"."page"."created_at" DESC OFFSET $2 ``` The big difference here is that `LIMIT` is present in the first query but not the second. ## How to reproduce https://github.com/TLadd/prisma-unique-composite-key-query-bug/blob/master/README.md Follow instructions in README for a reproducible example. Basically just the above code snippet. ## Expected behavior I would expect the `LIMIT` to be applied to the second query as well. ## Prisma information ``` generator client { provider = "prisma-client-js" binaryTargets = ["native", "debian-openssl-1.1.x"] } datasource db { provider = "postgresql" url = env("DATABASE_URL") } model Page { id String @id @default(dbgenerated()) createdAt DateTime @default(now()) @map("created_at") url String @unique @@map("page") } ``` ## Environment & setup <!-- In which environment does the problem occur --> - OS: Mac OS - Database: PostgreSQL - Node.js version: v12.16.2 - Prisma version: 2.12.1
process
prisma not applying limit when take specified with id cursor thanks for helping us improve prisma 🙏 please follow the sections in the template and provide as much information as possible about your problem e g by setting the debug environment variable and enabling additional logging output in prisma client learn more about writing proper bug reports here bug description i was noticing that when paginating through one of my larger tables with prisma the first page was always loading quickly but subsequent ones were taking many seconds my queries look like this typescript function getpaginationargs cursor string undefined return take cursor cursor id cursor undefined skip cursor undefined orderby createdat desc as const const await prisma page findmany getpaginationargs undefined const cursor id const await prisma page findmany getpaginationargs cursor i went in and looked at the queries prisma was issuing first page sql select prisma test schema page id prisma test schema page created at prisma test schema page url from prisma test schema page where order by prisma test schema page created at desc limit offset second page sql select prisma test schema page id prisma test schema page created at prisma test schema page url from prisma test schema page select prisma test schema page created at from prisma test schema page where prisma test schema page id as order cmp where prisma test schema page created at order cmp created at order by prisma test schema page created at desc offset the big difference here is that limit is present in the first query but not the second how to reproduce follow instructions in readme for a reproducible example basically just the above code snippet expected behavior i would expect the limit to be applied to the second query as well prisma information generator client provider prisma client js binarytargets datasource db provider postgresql url env database url model page id string id default dbgenerated createdat datetime default now map created at url string unique map page environment setup os mac os database postgresql node js version prisma version
1
6,612
9,696,035,326
IssuesEvent
2019-05-25 03:11:44
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
topicsetref generates duplicate ID
bug preprocess priority/low stale
Converted from [SourceForge issue 3512608](http://sourceforge.net/support/tracker.php?aid=3512608), submitted by tmakita If I use topicset & topisetref in the bookmap, DITA-OT generates duplicate id in the result bookmap. This causes error in the formatting phase. The test data is quite simple. ``` xml <topicset id="a1" navtitle="Topicset"> <topicref> <topicref href="taskbook/task_preface.dita"></topicref> </topicref> </topicset> <topicsetref href="#a1" navtitle="Topicsetref"></topicsetref> ``` In this case the id value "a1" remains both topicset defined side topicref and referenced side of topicref. Please fix this bug. I attached test data and result bookmap file generated in the temp directory.
1.0
topicsetref generates duplicate ID - Converted from [SourceForge issue 3512608](http://sourceforge.net/support/tracker.php?aid=3512608), submitted by tmakita If I use topicset & topisetref in the bookmap, DITA-OT generates duplicate id in the result bookmap. This causes error in the formatting phase. The test data is quite simple. ``` xml <topicset id="a1" navtitle="Topicset"> <topicref> <topicref href="taskbook/task_preface.dita"></topicref> </topicref> </topicset> <topicsetref href="#a1" navtitle="Topicsetref"></topicsetref> ``` In this case the id value "a1" remains both topicset defined side topicref and referenced side of topicref. Please fix this bug. I attached test data and result bookmap file generated in the temp directory.
process
topicsetref generates duplicate id converted from submitted by tmakita if i use topicset topisetref in the bookmap dita ot generates duplicate id in the result bookmap this causes error in the formatting phase the test data is quite simple xml in this case the id value remains both topicset defined side topicref and referenced side of topicref please fix this bug i attached test data and result bookmap file generated in the temp directory
1
220,601
24,565,295,900
IssuesEvent
2022-10-13 02:02:02
flamencist/CopyCssSelector
https://api.github.com/repos/flamencist/CopyCssSelector
closed
CVE-2021-35065 (High) detected in glob-parent-5.1.1.tgz - autoclosed
security vulnerability
## CVE-2021-35065 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-5.1.1.tgz</b></p></summary> <p>Extract the non-magic parent path from a glob string.</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - karma-4.4.1.tgz (Root Library) - chokidar-3.4.0.tgz - :x: **glob-parent-5.1.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS) <p>Publish Date: 2021-06-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p> <p>Release Date: 2021-06-22</p> <p>Fix Resolution: glob-parent - 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-35065 (High) detected in glob-parent-5.1.1.tgz - autoclosed - ## CVE-2021-35065 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-5.1.1.tgz</b></p></summary> <p>Extract the non-magic parent path from a glob string.</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - karma-4.4.1.tgz (Root Library) - chokidar-3.4.0.tgz - :x: **glob-parent-5.1.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS) <p>Publish Date: 2021-06-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p> <p>Release Date: 2021-06-22</p> <p>Fix Resolution: glob-parent - 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in glob parent tgz autoclosed cve high severity vulnerability vulnerable library glob parent tgz extract the non magic parent path from a glob string library home page a href path to dependency file package json path to vulnerable library node modules glob parent package json dependency hierarchy karma tgz root library chokidar tgz x glob parent tgz vulnerable library vulnerability details the package glob parent before are vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent step up your open source security game with mend
0
184,170
6,706,319,762
IssuesEvent
2017-10-12 06:24:57
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.glassdoor.com - design is broken
browser-firefox priority-important status-needstriage type-stylo
<!-- @browser: Firefox 58.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:58.0) Gecko/20100101 Firefox/58.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://www.glassdoor.com/member/profile/index.htm **Browser / Version**: Firefox 58.0 **Operating System**: Windows 10 **Tested Another Browser**: No **Problem type**: Design is broken **Description**: Broken layout. **Steps to Reproduce**: layout.css.servo.enabled: true [![Screenshot Description](https://webcompat.com/uploads/2017/10/09a2969d-86f6-44b1-97ab-7c396f3f17f0-thumb.jpg)](https://webcompat.com/uploads/2017/10/09a2969d-86f6-44b1-97ab-7c396f3f17f0.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.glassdoor.com - design is broken - <!-- @browser: Firefox 58.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:58.0) Gecko/20100101 Firefox/58.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://www.glassdoor.com/member/profile/index.htm **Browser / Version**: Firefox 58.0 **Operating System**: Windows 10 **Tested Another Browser**: No **Problem type**: Design is broken **Description**: Broken layout. **Steps to Reproduce**: layout.css.servo.enabled: true [![Screenshot Description](https://webcompat.com/uploads/2017/10/09a2969d-86f6-44b1-97ab-7c396f3f17f0-thumb.jpg)](https://webcompat.com/uploads/2017/10/09a2969d-86f6-44b1-97ab-7c396f3f17f0.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
design is broken url browser version firefox operating system windows tested another browser no problem type design is broken description broken layout steps to reproduce layout css servo enabled true from with ❤️
0
326,213
9,948,811,118
IssuesEvent
2019-07-04 09:46:30
chrisly-bear/PEBRApp
https://api.github.com/repos/chrisly-bear/PEBRApp
opened
Changes from Email
high priority
Various changes which were communicated via email (18. June 2019) are summarized in this issue: **New Patient Screen**: - [ ] Ask the sticker number question only for eligible patients, i.e., put the sticker number question _after_ the consent question. Otherwise we will give out sticker numbers to those not eligible. We want to spare stickers (and minimize confusion). **Patient Screen**: - [ ] For the notifications section, replace the word "disabled" with "not wished" if, at time of assessment, the patient had a phone number but did not select a given notification option. If, at time of assessment, the patient did not have a phone, write "no phone available at the time of assessment". **Preference Assessment Screen**: - [ ] Some changes were made in the codebook. Implement those changes.
1.0
Changes from Email - Various changes which were communicated via email (18. June 2019) are summarized in this issue: **New Patient Screen**: - [ ] Ask the sticker number question only for eligible patients, i.e., put the sticker number question _after_ the consent question. Otherwise we will give out sticker numbers to those not eligible. We want to spare stickers (and minimize confusion). **Patient Screen**: - [ ] For the notifications section, replace the word "disabled" with "not wished" if, at time of assessment, the patient had a phone number but did not select a given notification option. If, at time of assessment, the patient did not have a phone, write "no phone available at the time of assessment". **Preference Assessment Screen**: - [ ] Some changes were made in the codebook. Implement those changes.
non_process
changes from email various changes which were communicated via email june are summarized in this issue new patient screen ask the sticker number question only for eligible patients i e put the sticker number question after the consent question otherwise we will give out sticker numbers to those not eligible we want to spare stickers and minimize confusion patient screen for the notifications section replace the word disabled with not wished if at time of assessment the patient had a phone number but did not select a given notification option if at time of assessment the patient did not have a phone write no phone available at the time of assessment preference assessment screen some changes were made in the codebook implement those changes
0
6,873
10,012,332,389
IssuesEvent
2019-07-15 12:58:11
Jeffail/benthos
https://api.github.com/repos/Jeffail/benthos
closed
Storing response codes from HTTP requests in metadata
enhancement outputs processors
You can kind of hack yourself a response code check when you use an `http` processor: ```yaml - http: parallel: true request: url: "TODO" verb: GET - for_each: - conditional: condition: check_interpolation: value: ${!error} condition: text: operator: contains arg: "(404)" ``` However, this is pretty nasty and prone to long term problems. It's also exclusive to the `http` processor, but it would also be nice to store response codes from the `http_client` output and perhaps more output destinations. This would allow us to build DLQ output destinations based on the _cause_ of the failure at our chosen output.
1.0
Storing response codes from HTTP requests in metadata - You can kind of hack yourself a response code check when you use an `http` processor: ```yaml - http: parallel: true request: url: "TODO" verb: GET - for_each: - conditional: condition: check_interpolation: value: ${!error} condition: text: operator: contains arg: "(404)" ``` However, this is pretty nasty and prone to long term problems. It's also exclusive to the `http` processor, but it would also be nice to store response codes from the `http_client` output and perhaps more output destinations. This would allow us to build DLQ output destinations based on the _cause_ of the failure at our chosen output.
process
storing response codes from http requests in metadata you can kind of hack yourself a response code check when you use an http processor yaml http parallel true request url todo verb get for each conditional condition check interpolation value error condition text operator contains arg however this is pretty nasty and prone to long term problems it s also exclusive to the http processor but it would also be nice to store response codes from the http client output and perhaps more output destinations this would allow us to build dlq output destinations based on the cause of the failure at our chosen output
1
22,637
31,885,275,718
IssuesEvent
2023-09-16 21:50:06
winter-telescope/mirar
https://api.github.com/repos/winter-telescope/mirar
closed
[FEATURE] run source photometry on many sources without image subtraction
enhancement processors
**Is your feature request related to a problem? Please describe.** I'm always frustrated when I want to perform PSF photometry on all of the sources in my image, but the `SourceBatch`/`SourcePSFPhotometry` infrastructure requires that I run ZOGY first. It seems that the only way to generate a source table with several sources in it is to use `source_detector.py` which runs on a ZOGY difference image. **Describe the solution you'd like** I would like to have a class or function that will turn all the sources in a sextractor table into a candidate table, which I can then hand to the `Source{PSF/Aperture}Photometry` processors. **Describe alternatives you've considered** I've also considered shifting gears and setting up image subtraction first, to abide by the commonly expected workflow. **Additional context** I know that @robertdstein is restructuring sources in an upcoming PR, but the priority there is for single sources (to get pushed to Fritz) rather than a list of sources. I mentioned the ZOGY prereq before in the context of `source_exporter.py,` so it's possible your current work will accidentally fix this as well.
1.0
[FEATURE] run source photometry on many sources without image subtraction - **Is your feature request related to a problem? Please describe.** I'm always frustrated when I want to perform PSF photometry on all of the sources in my image, but the `SourceBatch`/`SourcePSFPhotometry` infrastructure requires that I run ZOGY first. It seems that the only way to generate a source table with several sources in it is to use `source_detector.py` which runs on a ZOGY difference image. **Describe the solution you'd like** I would like to have a class or function that will turn all the sources in a sextractor table into a candidate table, which I can then hand to the `Source{PSF/Aperture}Photometry` processors. **Describe alternatives you've considered** I've also considered shifting gears and setting up image subtraction first, to abide by the commonly expected workflow. **Additional context** I know that @robertdstein is restructuring sources in an upcoming PR, but the priority there is for single sources (to get pushed to Fritz) rather than a list of sources. I mentioned the ZOGY prereq before in the context of `source_exporter.py,` so it's possible your current work will accidentally fix this as well.
process
run source photometry on many sources without image subtraction is your feature request related to a problem please describe i m always frustrated when i want to perform psf photometry on all of the sources in my image but the sourcebatch sourcepsfphotometry infrastructure requires that i run zogy first it seems that the only way to generate a source table with several sources in it is to use source detector py which runs on a zogy difference image describe the solution you d like i would like to have a class or function that will turn all the sources in a sextractor table into a candidate table which i can then hand to the source psf aperture photometry processors describe alternatives you ve considered i ve also considered shifting gears and setting up image subtraction first to abide by the commonly expected workflow additional context i know that robertdstein is restructuring sources in an upcoming pr but the priority there is for single sources to get pushed to fritz rather than a list of sources i mentioned the zogy prereq before in the context of source exporter py so it s possible your current work will accidentally fix this as well
1
127,594
5,037,050,513
IssuesEvent
2016-12-17 12:34:59
bounswe/bounswe2016group12
https://api.github.com/repos/bounswe/bounswe2016group12
closed
Android - FollowTopic , Get profile data implementation
priority
Profile data-> followed topics, commented topics , voted comments and relations.
1.0
Android - FollowTopic , Get profile data implementation - Profile data-> followed topics, commented topics , voted comments and relations.
non_process
android followtopic get profile data implementation profile data followed topics commented topics voted comments and relations
0
8,829
8,418,472,958
IssuesEvent
2018-10-15 00:33:16
Kaushalop/vote-app-blockchain
https://api.github.com/repos/Kaushalop/vote-app-blockchain
opened
Microservices exists for operations on the collection votes
backend-services
This collection will basically focus on storing information regards to the casted vote email. Each vote will have the voterId, partyId, electionType (enum of PVP, SENATOR) basically preventing voting for two parties standing for PVP(President, Vice-President) elections. 1/ Create a vote 2/ Delete a vote 3/ Get a vote 4/ Get all votes with filters on the data Model the data according to the rules and Issue #4 rules.
1.0
Microservices exists for operations on the collection votes - This collection will basically focus on storing information regards to the casted vote email. Each vote will have the voterId, partyId, electionType (enum of PVP, SENATOR) basically preventing voting for two parties standing for PVP(President, Vice-President) elections. 1/ Create a vote 2/ Delete a vote 3/ Get a vote 4/ Get all votes with filters on the data Model the data according to the rules and Issue #4 rules.
non_process
microservices exists for operations on the collection votes this collection will basically focus on storing information regards to the casted vote email each vote will have the voterid partyid electiontype enum of pvp senator basically preventing voting for two parties standing for pvp president vice president elections create a vote delete a vote get a vote get all votes with filters on the data model the data according to the rules and issue rules
0
37,137
4,779,464,564
IssuesEvent
2016-10-27 22:35:47
elegantthemes/Divi-Beta
https://api.github.com/repos/elegantthemes/Divi-Beta
closed
Footer is broken on No Result Fond pages
BUG DESIGN SIGNOFF QUALITY ASSURED READY FOR REVIEW
### Problem: Here is the demo page where you can clearly see the problem in the footer: https://www.elegantthemes.com/preview/Divi/?s=%21%238787dfff Screenshot: ![1](https://cloud.githubusercontent.com/assets/3352130/19404619/34773068-9278-11e6-9dcc-154ae2b11c8e.png)
1.0
Footer is broken on No Result Fond pages - ### Problem: Here is the demo page where you can clearly see the problem in the footer: https://www.elegantthemes.com/preview/Divi/?s=%21%238787dfff Screenshot: ![1](https://cloud.githubusercontent.com/assets/3352130/19404619/34773068-9278-11e6-9dcc-154ae2b11c8e.png)
non_process
footer is broken on no result fond pages problem here is the demo page where you can clearly see the problem in the footer screenshot
0
3,912
6,827,172,203
IssuesEvent
2017-11-08 16:15:56
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
opened
The href change of the iframe location works wrong
AREA: client SYSTEM: sandbox SYSTEM: URL processing TYPE: bug
When cross domain iframe changes location, it leave proxy port as is, but parent window location and new iframe location may be the same domain. Related with https://github.com/DevExpress/testcafe/issues/1922
1.0
The href change of the iframe location works wrong - When cross domain iframe changes location, it leave proxy port as is, but parent window location and new iframe location may be the same domain. Related with https://github.com/DevExpress/testcafe/issues/1922
process
the href change of the iframe location works wrong when cross domain iframe changes location it leave proxy port as is but parent window location and new iframe location may be the same domain related with
1
10,055
7,884,886,970
IssuesEvent
2018-06-27 10:33:07
ltb-project/self-service-password
https://api.github.com/repos/ltb-project/self-service-password
closed
reduce info released in error messages
enhancement security
Do you think it's too much information to report `User ... not found` and `Mail ... does not match for user ...`? Similar to other authentication things, I suggest it might be better to reveal less, perhaps just `There was a problem ...`. Otherwise, I can use this mechanism to verify that usernames exist. (Currently on version 1.0, intending to upgrade soon.)
True
reduce info released in error messages - Do you think it's too much information to report `User ... not found` and `Mail ... does not match for user ...`? Similar to other authentication things, I suggest it might be better to reveal less, perhaps just `There was a problem ...`. Otherwise, I can use this mechanism to verify that usernames exist. (Currently on version 1.0, intending to upgrade soon.)
non_process
reduce info released in error messages do you think it s too much information to report user not found and mail does not match for user similar to other authentication things i suggest it might be better to reveal less perhaps just there was a problem otherwise i can use this mechanism to verify that usernames exist currently on version intending to upgrade soon
0
21,626
30,025,443,765
IssuesEvent
2023-06-27 05:33:51
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
pih 1.45 has 2 GuardDog issues
guarddog typosquatting silent-process-execution
https://pypi.org/project/pih https://inspector.pypi.io/project/pih ```{ "dependency": "pih", "version": "1.45", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pip, pid", "silent-process-execution": [ { "location": "pih-1.45/pih/tools.py:719", "code": " result = subprocess.run(command, stdin=subprocess.DEVNULL, stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL)", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpqlmtkein/pih" } }```
1.0
pih 1.45 has 2 GuardDog issues - https://pypi.org/project/pih https://inspector.pypi.io/project/pih ```{ "dependency": "pih", "version": "1.45", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pip, pid", "silent-process-execution": [ { "location": "pih-1.45/pih/tools.py:719", "code": " result = subprocess.run(command, stdin=subprocess.DEVNULL, stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL)", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpqlmtkein/pih" } }```
process
pih has guarddog issues dependency pih version result issues errors results typosquatting this package closely ressembles the following package names and might be a typosquatting attempt pip pid silent process execution location pih pih tools py code result subprocess run command stdin subprocess devnull stdout subprocess devnull stderr subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp tmpqlmtkein pih
1
237,760
19,672,722,649
IssuesEvent
2022-01-11 09:11:19
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: X-Pack Detection Engine API Integration Tests.x-pack/test/detection_engine_api_integration/security_and_spaces/tests/create_threat_matching·ts - detection engine api security and spaces enabled create_threat_matching tests with auditbeat data should be able to execute and get 10 signals when doing a specific query
failed-test Team: SecuritySolution Team: CTI
A test failed on a tracked branch ``` Error: expected undefined to be truthy at Assertion.assert (/opt/local-ssd/buildkite/builds/kb-n2-4-661597c297fc7312/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:100:11) at Assertion.ok (/opt/local-ssd/buildkite/builds/kb-n2-4-661597c297fc7312/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:122:8) at Function.ok (/opt/local-ssd/buildkite/builds/kb-n2-4-661597c297fc7312/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:531:15) at Context.<anonymous> (test/detection_engine_api_integration/security_and_spaces/tests/create_threat_matching.ts:172:43) at runMicrotasks (<anonymous>) at processTicksAndRejections (node:internal/process/task_queues:96:5) at Object.apply (/opt/local-ssd/buildkite/builds/kb-n2-4-661597c297fc7312/elastic/kibana-hourly/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` First failure: [CI Build - 7.17](https://buildkite.com/elastic/kibana-hourly/builds/7382#7ba149a6-f9a6-490d-a09a-934d88467a13) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Detection Engine API Integration Tests.x-pack/test/detection_engine_api_integration/security_and_spaces/tests/create_threat_matching·ts","test.name":"detection engine api security and spaces enabled create_threat_matching tests with auditbeat data should be able to execute and get 10 signals when doing a specific query","test.failCount":2}} -->
1.0
Failing test: X-Pack Detection Engine API Integration Tests.x-pack/test/detection_engine_api_integration/security_and_spaces/tests/create_threat_matching·ts - detection engine api security and spaces enabled create_threat_matching tests with auditbeat data should be able to execute and get 10 signals when doing a specific query - A test failed on a tracked branch ``` Error: expected undefined to be truthy at Assertion.assert (/opt/local-ssd/buildkite/builds/kb-n2-4-661597c297fc7312/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:100:11) at Assertion.ok (/opt/local-ssd/buildkite/builds/kb-n2-4-661597c297fc7312/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:122:8) at Function.ok (/opt/local-ssd/buildkite/builds/kb-n2-4-661597c297fc7312/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:531:15) at Context.<anonymous> (test/detection_engine_api_integration/security_and_spaces/tests/create_threat_matching.ts:172:43) at runMicrotasks (<anonymous>) at processTicksAndRejections (node:internal/process/task_queues:96:5) at Object.apply (/opt/local-ssd/buildkite/builds/kb-n2-4-661597c297fc7312/elastic/kibana-hourly/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` First failure: [CI Build - 7.17](https://buildkite.com/elastic/kibana-hourly/builds/7382#7ba149a6-f9a6-490d-a09a-934d88467a13) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Detection Engine API Integration Tests.x-pack/test/detection_engine_api_integration/security_and_spaces/tests/create_threat_matching·ts","test.name":"detection engine api security and spaces enabled create_threat_matching tests with auditbeat data should be able to execute and get 10 signals when doing a specific query","test.failCount":2}} -->
non_process
failing test x pack detection engine api integration tests x pack test detection engine api integration security and spaces tests create threat matching·ts detection engine api security and spaces enabled create threat matching tests with auditbeat data should be able to execute and get signals when doing a specific query a test failed on a tracked branch error expected undefined to be truthy at assertion assert opt local ssd buildkite builds kb elastic kibana hourly kibana node modules kbn expect expect js at assertion ok opt local ssd buildkite builds kb elastic kibana hourly kibana node modules kbn expect expect js at function ok opt local ssd buildkite builds kb elastic kibana hourly kibana node modules kbn expect expect js at context test detection engine api integration security and spaces tests create threat matching ts at runmicrotasks at processticksandrejections node internal process task queues at object apply opt local ssd buildkite builds kb elastic kibana hourly kibana node modules kbn test target node functional test runner lib mocha wrap function js first failure
0
2,419
5,201,200,124
IssuesEvent
2017-01-24 03:20:49
rubberduck-vba/Rubberduck
https://api.github.com/repos/rubberduck-vba/Rubberduck
closed
FunctionReturnValueNotUsedInspection inserts extra `End Sub`
bug parse-tree-preprocessing parse-tree-processing
I'm still working on finding the minimal example for this, but the when I ran the code from https://github.com/peterennis/aegit/blob/master/aerc/src/basGDIPlus.bas (commit bbfdc5a) through FunctionReturnValueNotUsedInspection quick fix, I ended up with this at the bottom of the Sub: ``` VB #If Win64 Then 'Help function to get a OLE-Picture from Windows-Bitmap-Handle Private Sub BitmapToPicture(ByVal hBmp As LongPtr, Optional ByRef bIsIcon As Boolean = False) Dim TPicConv As PICTDESC Dim UID As GUID With TPicConv If bIsIcon Then .cbSizeOfStruct = 16 .PicType = 3 'PicType Icon Else .cbSizeOfStruct = Len(TPicConv) .PicType = 1 'PicType Bitmap End If .hImage = hBmp End With CLSIDFromString StrPtr(GUID_IPicture), UID OleCreatePictureIndirect TPicConv, UID, True, BitmapToPicture End Sub End Sub #Else Private Function BitmapToPicture(ByVal hBmp As Long, Optional bIsIcon As Boolean = False) As StdPicture Dim TPicConv As PICTDESC, UID As GUID With TPicConv If bIsIcon Then .cbSizeOfStruct = 16 .PicType = 3 'PicType Icon Else .cbSizeOfStruct = Len(TPicConv) .PicType = 1 'PicType Bitmap End If .hImage = hBmp End With CLSIDFromString StrPtr(GUID_IPicture), UID OleCreatePictureIndirect TPicConv, UID, True, BitmapToPicture End Function #End If ``` This might(?) be related to #2021 (note that it exhibits that behaviour also), but I doubt it - this looks more like an issue with keeping track of the insertion point.
2.0
FunctionReturnValueNotUsedInspection inserts extra `End Sub` - I'm still working on finding the minimal example for this, but the when I ran the code from https://github.com/peterennis/aegit/blob/master/aerc/src/basGDIPlus.bas (commit bbfdc5a) through FunctionReturnValueNotUsedInspection quick fix, I ended up with this at the bottom of the Sub: ``` VB #If Win64 Then 'Help function to get a OLE-Picture from Windows-Bitmap-Handle Private Sub BitmapToPicture(ByVal hBmp As LongPtr, Optional ByRef bIsIcon As Boolean = False) Dim TPicConv As PICTDESC Dim UID As GUID With TPicConv If bIsIcon Then .cbSizeOfStruct = 16 .PicType = 3 'PicType Icon Else .cbSizeOfStruct = Len(TPicConv) .PicType = 1 'PicType Bitmap End If .hImage = hBmp End With CLSIDFromString StrPtr(GUID_IPicture), UID OleCreatePictureIndirect TPicConv, UID, True, BitmapToPicture End Sub End Sub #Else Private Function BitmapToPicture(ByVal hBmp As Long, Optional bIsIcon As Boolean = False) As StdPicture Dim TPicConv As PICTDESC, UID As GUID With TPicConv If bIsIcon Then .cbSizeOfStruct = 16 .PicType = 3 'PicType Icon Else .cbSizeOfStruct = Len(TPicConv) .PicType = 1 'PicType Bitmap End If .hImage = hBmp End With CLSIDFromString StrPtr(GUID_IPicture), UID OleCreatePictureIndirect TPicConv, UID, True, BitmapToPicture End Function #End If ``` This might(?) be related to #2021 (note that it exhibits that behaviour also), but I doubt it - this looks more like an issue with keeping track of the insertion point.
process
functionreturnvaluenotusedinspection inserts extra end sub i m still working on finding the minimal example for this but the when i ran the code from commit through functionreturnvaluenotusedinspection quick fix i ended up with this at the bottom of the sub vb if then help function to get a ole picture from windows bitmap handle private sub bitmaptopicture byval hbmp as longptr optional byref bisicon as boolean false dim tpicconv as pictdesc dim uid as guid with tpicconv if bisicon then cbsizeofstruct pictype pictype icon else cbsizeofstruct len tpicconv pictype pictype bitmap end if himage hbmp end with clsidfromstring strptr guid ipicture uid olecreatepictureindirect tpicconv uid true bitmaptopicture end sub end sub else private function bitmaptopicture byval hbmp as long optional bisicon as boolean false as stdpicture dim tpicconv as pictdesc uid as guid with tpicconv if bisicon then cbsizeofstruct pictype pictype icon else cbsizeofstruct len tpicconv pictype pictype bitmap end if himage hbmp end with clsidfromstring strptr guid ipicture uid olecreatepictureindirect tpicconv uid true bitmaptopicture end function end if this might be related to note that it exhibits that behaviour also but i doubt it this looks more like an issue with keeping track of the insertion point
1
278,660
21,091,707,189
IssuesEvent
2022-04-04 06:10:27
awse2050/Side-Project-1
https://api.github.com/repos/awse2050/Side-Project-1
closed
Service, Presentation Layer 추가 필요
documentation duplicate
#1 * 추가적인 복습사항 1. AttributeConverter Interface 2. @TestMethodOrder
1.0
Service, Presentation Layer 추가 필요 - #1 * 추가적인 복습사항 1. AttributeConverter Interface 2. @TestMethodOrder
non_process
service presentation layer 추가 필요 추가적인 복습사항 attributeconverter interface testmethodorder
0
189,267
6,795,848,022
IssuesEvent
2017-11-01 17:00:37
ArctosDB/arctos
https://api.github.com/repos/ArctosDB/arctos
closed
How to bulk modify or delete "Other identifiers"?
Function-DataEntry/Bulkloading Priority-Normal
I'd like to get rid of a particular identifier from a group of specimens. For example http://arctos.database.museum/guid/MSB:Herp:98676 has both ID Type="collector number" Prefix=CLL0065 ID Type="collector number" Prefix=CLL ID Number=65 Is there a way to delete the first record (where "collector number"="CLL0065") for a suite of specimens? I have a file with matching MSB:Herp and collector numbers to delete. thanks, Tom
1.0
How to bulk modify or delete "Other identifiers"? - I'd like to get rid of a particular identifier from a group of specimens. For example http://arctos.database.museum/guid/MSB:Herp:98676 has both ID Type="collector number" Prefix=CLL0065 ID Type="collector number" Prefix=CLL ID Number=65 Is there a way to delete the first record (where "collector number"="CLL0065") for a suite of specimens? I have a file with matching MSB:Herp and collector numbers to delete. thanks, Tom
non_process
how to bulk modify or delete other identifiers i d like to get rid of a particular identifier from a group of specimens for example has both id type collector number prefix id type collector number prefix cll id number is there a way to delete the first record where collector number for a suite of specimens i have a file with matching msb herp and collector numbers to delete thanks tom
0
73,638
24,730,033,549
IssuesEvent
2022-10-20 16:46:37
STEllAR-GROUP/hpx
https://api.github.com/repos/STEllAR-GROUP/hpx
closed
Can't call nullary callables wrapped with `hpx::unwrapping`
type: defect compiler: gcc category: utilities
## Expected Behavior `hpx::unwrapping` with a nullary lambda should be callable, e.g. `hpx::unwrapping([]{})()` (even though the unwrapping isn't needed in this particular case, it's useful for generic contexts). Note that this isn't something we rely on, but we happened to bump into this accidentally. I thought you might be interested. ## Actual Behavior It fails to compile. Note that simply creating the `unwrapping` callable with `hpx::unwrapping([]{})` compiles just fine. Calling it does not work. ## Steps to Reproduce the Problem Attempt to compile the following: ``` #include <hpx/unwrap.hpp> int main() { hpx::unwrapping([]{})(); } ``` ## Specifications ... Please describe your environment - HPX Version: latest master (https://github.com/STEllAR-GROUP/hpx/commit/1e4b87375d907e88c57c50e90a26179689f46485) - Platform (compiler, OS): GCC 11.3.0
1.0
Can't call nullary callables wrapped with `hpx::unwrapping` - ## Expected Behavior `hpx::unwrapping` with a nullary lambda should be callable, e.g. `hpx::unwrapping([]{})()` (even though the unwrapping isn't needed in this particular case, it's useful for generic contexts). Note that this isn't something we rely on, but we happened to bump into this accidentally. I thought you might be interested. ## Actual Behavior It fails to compile. Note that simply creating the `unwrapping` callable with `hpx::unwrapping([]{})` compiles just fine. Calling it does not work. ## Steps to Reproduce the Problem Attempt to compile the following: ``` #include <hpx/unwrap.hpp> int main() { hpx::unwrapping([]{})(); } ``` ## Specifications ... Please describe your environment - HPX Version: latest master (https://github.com/STEllAR-GROUP/hpx/commit/1e4b87375d907e88c57c50e90a26179689f46485) - Platform (compiler, OS): GCC 11.3.0
non_process
can t call nullary callables wrapped with hpx unwrapping expected behavior hpx unwrapping with a nullary lambda should be callable e g hpx unwrapping even though the unwrapping isn t needed in this particular case it s useful for generic contexts note that this isn t something we rely on but we happened to bump into this accidentally i thought you might be interested actual behavior it fails to compile note that simply creating the unwrapping callable with hpx unwrapping compiles just fine calling it does not work steps to reproduce the problem attempt to compile the following include int main hpx unwrapping specifications please describe your environment hpx version latest master platform compiler os gcc
0
5,878
8,700,793,685
IssuesEvent
2018-12-05 09:45:20
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Process.ProcessorAffinity returns a 32-bit value on ARM64
area-System.Diagnostics.Process
When we run the following code on ARM64 machine with 48 cores (Ubuntu) without setting the CPU affinity in explicit way: ```cs Console.WriteLine(System.Diagnostics.Process.GetCurrentProcess().ProcessorAffinity); ``` We get `00000000FFFFFFFF` which is `2^32 - 1` while it should be `2^48 - 1` I don't know if this is specific to ARM64 or Ubuntu or 64 bit in general. I just don't have an access to a machine with more than 32 core to test. @AndyAyersMS have hit this issue when he was benchmarking .NET Core 3.0 on ARM
1.0
Process.ProcessorAffinity returns a 32-bit value on ARM64 - When we run the following code on ARM64 machine with 48 cores (Ubuntu) without setting the CPU affinity in explicit way: ```cs Console.WriteLine(System.Diagnostics.Process.GetCurrentProcess().ProcessorAffinity); ``` We get `00000000FFFFFFFF` which is `2^32 - 1` while it should be `2^48 - 1` I don't know if this is specific to ARM64 or Ubuntu or 64 bit in general. I just don't have an access to a machine with more than 32 core to test. @AndyAyersMS have hit this issue when he was benchmarking .NET Core 3.0 on ARM
process
process processoraffinity returns a bit value on when we run the following code on machine with cores ubuntu without setting the cpu affinity in explicit way cs console writeline system diagnostics process getcurrentprocess processoraffinity we get which is while it should be i don t know if this is specific to or ubuntu or bit in general i just don t have an access to a machine with more than core to test andyayersms have hit this issue when he was benchmarking net core on arm
1
12,771
15,149,026,533
IssuesEvent
2021-02-11 11:28:11
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Can't create table (errno 150 "Foreign key constraint is incorrectly formed")
bug/2-confirmed kind/bug process/candidate team/migrations tech/engines
When I execute the `prisma migrate up --experimental` command, I get an error like this: ![image](https://user-images.githubusercontent.com/25415068/83087217-1acdec00-a067-11ea-993d-3fb6934bc9fc.png) The report id are 6340. If you and more informations, please make me know :D. Thank you for the tools!!
1.0
Can't create table (errno 150 "Foreign key constraint is incorrectly formed") - When I execute the `prisma migrate up --experimental` command, I get an error like this: ![image](https://user-images.githubusercontent.com/25415068/83087217-1acdec00-a067-11ea-993d-3fb6934bc9fc.png) The report id are 6340. If you and more informations, please make me know :D. Thank you for the tools!!
process
can t create table errno foreign key constraint is incorrectly formed when i execute the prisma migrate up experimental command i get an error like this the report id are if you and more informations please make me know d thank you for the tools
1
3,644
6,677,466,219
IssuesEvent
2017-10-05 10:34:22
rogerthat-platform/rogerthat-backend
https://api.github.com/repos/rogerthat-platform/rogerthat-backend
closed
We need a userDataUpdated capi call
process_duplicate
Now, when the userData is updated, the full key-value store is pushed to the android/ios clients. We need a userDataUpdated capi call such that a smaller change-set is pushed to the clients. The clients should also store the userData in a new table instead of saving the json string in the friend table.
1.0
We need a userDataUpdated capi call - Now, when the userData is updated, the full key-value store is pushed to the android/ios clients. We need a userDataUpdated capi call such that a smaller change-set is pushed to the clients. The clients should also store the userData in a new table instead of saving the json string in the friend table.
process
we need a userdataupdated capi call now when the userdata is updated the full key value store is pushed to the android ios clients we need a userdataupdated capi call such that a smaller change set is pushed to the clients the clients should also store the userdata in a new table instead of saving the json string in the friend table
1
2,049
4,859,047,442
IssuesEvent
2016-11-13 13:09:41
MedFlyt/medflyt_webapp
https://api.github.com/repos/MedFlyt/medflyt_webapp
opened
In the registration process, add a field (menu picker) to choose the company type
Registration Process
Registration: Please choose the type of company: Home health care agency Home health care agency (Private paid) Managed long time care staffing agency Nursing home
1.0
In the registration process, add a field (menu picker) to choose the company type - Registration: Please choose the type of company: Home health care agency Home health care agency (Private paid) Managed long time care staffing agency Nursing home
process
in the registration process add a field menu picker to choose the company type registration please choose the type of company home health care agency home health care agency private paid managed long time care staffing agency nursing home
1
37,397
8,286,451,110
IssuesEvent
2018-09-19 04:54:55
MicrosoftDocs/live-share
https://api.github.com/repos/MicrosoftDocs/live-share
closed
[VS Code] Guest's *OnSave settings can inadvertedly impact the host
area: co-edit community feedback requested vscode
I was collaborating with a friend on a C++ project, and my being present enabled format-on-save. It was enabled in my settings, and not explicitly set in hers. Overriding it in the project settings stopped it. It seems like client settings shouldn't influence the host's settings, and ideally settings like `editor.formatOnSave` should defer to the host's settings whether or not they're per-project.
1.0
[VS Code] Guest's *OnSave settings can inadvertedly impact the host - I was collaborating with a friend on a C++ project, and my being present enabled format-on-save. It was enabled in my settings, and not explicitly set in hers. Overriding it in the project settings stopped it. It seems like client settings shouldn't influence the host's settings, and ideally settings like `editor.formatOnSave` should defer to the host's settings whether or not they're per-project.
non_process
guest s onsave settings can inadvertedly impact the host i was collaborating with a friend on a c project and my being present enabled format on save it was enabled in my settings and not explicitly set in hers overriding it in the project settings stopped it it seems like client settings shouldn t influence the host s settings and ideally settings like editor formatonsave should defer to the host s settings whether or not they re per project
0
9,216
12,248,370,826
IssuesEvent
2020-05-05 17:23:02
googleapis/gcp-metadata
https://api.github.com/repos/googleapis/gcp-metadata
closed
Simplify fastFailMetadataRequest() logic
type: process
To address a bug with DNS lookup in some GCP environments, we introduced logic that attempts both the IP address and DNS lookup of `metadata.google.internal.`. This logic has been working well, but is a little crufty: * it seems like we could pull some of the logic used to build up constants into the global scope into a helper. * it feels like the `replace` logic, used to create the `secondary` URL lookup could also be simplified. Now that we've also introduced the `GCE_METADATA_IP` (in https://github.com/googleapis/gcp-metadata/pull/346) environment variable, it seems like a refactor is needed.
1.0
Simplify fastFailMetadataRequest() logic - To address a bug with DNS lookup in some GCP environments, we introduced logic that attempts both the IP address and DNS lookup of `metadata.google.internal.`. This logic has been working well, but is a little crufty: * it seems like we could pull some of the logic used to build up constants into the global scope into a helper. * it feels like the `replace` logic, used to create the `secondary` URL lookup could also be simplified. Now that we've also introduced the `GCE_METADATA_IP` (in https://github.com/googleapis/gcp-metadata/pull/346) environment variable, it seems like a refactor is needed.
process
simplify fastfailmetadatarequest logic to address a bug with dns lookup in some gcp environments we introduced logic that attempts both the ip address and dns lookup of metadata google internal this logic has been working well but is a little crufty it seems like we could pull some of the logic used to build up constants into the global scope into a helper it feels like the replace logic used to create the secondary url lookup could also be simplified now that we ve also introduced the gce metadata ip in environment variable it seems like a refactor is needed
1
437,201
30,591,702,914
IssuesEvent
2023-07-21 17:38:56
MrXcitement/devc-cobol
https://api.github.com/repos/MrXcitement/devc-cobol
opened
Readme needs usage information
documentation
Add to the Usage section the steps on how to use the devcontainer image in a vscode project.
1.0
Readme needs usage information - Add to the Usage section the steps on how to use the devcontainer image in a vscode project.
non_process
readme needs usage information add to the usage section the steps on how to use the devcontainer image in a vscode project
0
119,404
25,518,835,648
IssuesEvent
2022-11-28 18:37:17
gmdavef/example-java-maven
https://api.github.com/repos/gmdavef/example-java-maven
opened
CVE: 2022-23305 found in Apache Log4j - Version: 1.2.15 [JAVA]
Severity: High Veracode Dependency Scanning
Veracode Software Composition Analysis =============================== Attribute | Details | --- | --- | Library | Apache Log4j Description | Apache Log4j 1.2 Language | JAVA Vulnerability | SQL Injection Vulnerability description | JDBCAppender in Log4j is vulnerable to SQL injection attacks. An attacker is able to execute arbitrary SQL commands via entering crafted strings into input fields and headers where the values to be inserted are converters from `PatternLayout` CVE | 2022-23305 CVSS score | 6.8 Vulnerability present in version/s | 1.1.3-1.2.17 Found library version/s | 1.2.15 Vulnerability fixed in version | Library latest version | 1.2.17 Fix | No fix is released. Users should upgrade to Log4j 2 or remove usage of the JDBCAppender from their configurations Links: - https://sca.analysiscenter.veracode.com/vulnerability-database/libraries/163?version=1.2.15 - https://sca.analysiscenter.veracode.com/vulnerability-database/vulnerabilities/33766 - Patch:
1.0
CVE: 2022-23305 found in Apache Log4j - Version: 1.2.15 [JAVA] - Veracode Software Composition Analysis =============================== Attribute | Details | --- | --- | Library | Apache Log4j Description | Apache Log4j 1.2 Language | JAVA Vulnerability | SQL Injection Vulnerability description | JDBCAppender in Log4j is vulnerable to SQL injection attacks. An attacker is able to execute arbitrary SQL commands via entering crafted strings into input fields and headers where the values to be inserted are converters from `PatternLayout` CVE | 2022-23305 CVSS score | 6.8 Vulnerability present in version/s | 1.1.3-1.2.17 Found library version/s | 1.2.15 Vulnerability fixed in version | Library latest version | 1.2.17 Fix | No fix is released. Users should upgrade to Log4j 2 or remove usage of the JDBCAppender from their configurations Links: - https://sca.analysiscenter.veracode.com/vulnerability-database/libraries/163?version=1.2.15 - https://sca.analysiscenter.veracode.com/vulnerability-database/vulnerabilities/33766 - Patch:
non_process
cve found in apache version veracode software composition analysis attribute details library apache description apache language java vulnerability sql injection vulnerability description jdbcappender in is vulnerable to sql injection attacks an attacker is able to execute arbitrary sql commands via entering crafted strings into input fields and headers where the values to be inserted are converters from patternlayout cve cvss score vulnerability present in version s found library version s vulnerability fixed in version library latest version fix no fix is released users should upgrade to or remove usage of the jdbcappender from their configurations links patch
0
115,501
14,789,710,026
IssuesEvent
2021-01-12 10:57:16
lafranceinsoumise/actionpopulaire.fr
https://api.github.com/repos/lafranceinsoumise/actionpopulaire.fr
closed
Ajouter une fonctionnalité de création d'un événement à partir d'un événement Facebook
Amélioration help wanted javascript python webdesign
## Objectif Dans le but de faciliter pour les utilisateurs la création d'événements sur la plateforme pour les utilisateurs qui souhaitent de toute façon créer le même événement sur Facebook, il faudrait permettre de pré-remplir le formulaire de création d'événement à partir de l'événement Facebook. ## Compétences nécessaires * javascript * html * Django * Un peu de Python ?
1.0
Ajouter une fonctionnalité de création d'un événement à partir d'un événement Facebook - ## Objectif Dans le but de faciliter pour les utilisateurs la création d'événements sur la plateforme pour les utilisateurs qui souhaitent de toute façon créer le même événement sur Facebook, il faudrait permettre de pré-remplir le formulaire de création d'événement à partir de l'événement Facebook. ## Compétences nécessaires * javascript * html * Django * Un peu de Python ?
non_process
ajouter une fonctionnalité de création d un événement à partir d un événement facebook objectif dans le but de faciliter pour les utilisateurs la création d événements sur la plateforme pour les utilisateurs qui souhaitent de toute façon créer le même événement sur facebook il faudrait permettre de pré remplir le formulaire de création d événement à partir de l événement facebook compétences nécessaires javascript html django un peu de python
0
23,379
10,881,877,133
IssuesEvent
2019-11-17 20:35:59
MirkoV1987/P5-PHPBlog
https://api.github.com/repos/MirkoV1987/P5-PHPBlog
opened
CVE-2019-6286 (Medium) detected in node-sass-v4.11.0
security vulnerability
## CVE-2019-6286 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.11.0</b></p></summary> <p> <p>:rainbow: Node.js bindings to libsass</p> <p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/MirkoV1987/P5-PHPBlog/commit/272c4b6213ea4cc72fed4b0e3ea3c3ceaccf5ba0">272c4b6213ea4cc72fed4b0e3ea3c3ceaccf5ba0</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (66)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/expand.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/expand.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/factory.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/operators.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/boolean.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/util.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/value.h - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/emitter.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/callback_bridge.h - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/file.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/sass.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/operation.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/operators.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/constants.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/error_handling.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/custom_importer_bridge.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/parser.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/constants.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/list.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/cssize.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/functions.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/util.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/custom_function_bridge.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/custom_importer_bridge.h - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/bind.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/eval.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/inspect.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/backtrace.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/extend.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_context_wrapper.h - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/sass_value_wrapper.h - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/error_handling.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/parser.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/debugger.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/emitter.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/number.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/color.h - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/sass_values.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/ast.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/output.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/check_nesting.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/null.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/functions.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/cssize.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/prelexer.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/ast.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/to_c.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/to_value.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/inspect.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/color.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/values.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_context_wrapper.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/list.h - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/check_nesting.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/map.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/to_value.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/context.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/binding.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/string.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/sass_context.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/prelexer.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/context.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/boolean.h - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/eval.cpp </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::skip_over_scopes in prelexer.hpp when called from Sass::Parser::parse_import(), a similar issue to CVE-2018-11693. <p>Publish Date: 2019-01-14 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6286>CVE-2019-6286</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6286">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6286</a></p> <p>Release Date: 2019-08-06</p> <p>Fix Resolution: 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-6286 (Medium) detected in node-sass-v4.11.0 - ## CVE-2019-6286 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.11.0</b></p></summary> <p> <p>:rainbow: Node.js bindings to libsass</p> <p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/MirkoV1987/P5-PHPBlog/commit/272c4b6213ea4cc72fed4b0e3ea3c3ceaccf5ba0">272c4b6213ea4cc72fed4b0e3ea3c3ceaccf5ba0</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (66)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/expand.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/expand.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/factory.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/operators.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/boolean.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/util.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/value.h - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/emitter.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/callback_bridge.h - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/file.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/sass.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/operation.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/operators.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/constants.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/error_handling.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/custom_importer_bridge.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/parser.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/constants.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/list.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/cssize.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/functions.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/util.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/custom_function_bridge.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/custom_importer_bridge.h - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/bind.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/eval.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/inspect.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/backtrace.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/extend.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_context_wrapper.h - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/sass_value_wrapper.h - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/error_handling.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/parser.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/debugger.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/emitter.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/number.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/color.h - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/sass_values.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/ast.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/output.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/check_nesting.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/null.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/functions.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/cssize.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/prelexer.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/ast.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/to_c.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/to_value.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/inspect.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/color.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/values.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_context_wrapper.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/list.h - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/check_nesting.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/map.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/to_value.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/context.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/binding.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/string.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/sass_context.cpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/prelexer.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/context.hpp - /P5-PHPBlog/Public/node_modules/node-sass/src/sass_types/boolean.h - /P5-PHPBlog/Public/node_modules/node-sass/src/libsass/src/eval.cpp </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::skip_over_scopes in prelexer.hpp when called from Sass::Parser::parse_import(), a similar issue to CVE-2018-11693. <p>Publish Date: 2019-01-14 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6286>CVE-2019-6286</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6286">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6286</a></p> <p>Release Date: 2019-08-06</p> <p>Fix Resolution: 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in node sass cve medium severity vulnerability vulnerable library node rainbow node js bindings to libsass library home page a href found in head commit a href library source files the source files were matched to this source library based on a best effort match source libraries are selected from a list of probable public libraries phpblog public node modules node sass src libsass src expand hpp phpblog public node modules node sass src libsass src expand cpp phpblog public node modules node sass src sass types factory cpp phpblog public node modules node sass src libsass src operators cpp phpblog public node modules node sass src sass types boolean cpp phpblog public node modules node sass src libsass src util hpp phpblog public node modules node sass src sass types value h phpblog public node modules node sass src libsass src emitter hpp phpblog public node modules node sass src callback bridge h phpblog public node modules node sass src libsass src file cpp phpblog public node modules node sass src libsass src sass cpp phpblog public node modules node sass src libsass src operation hpp phpblog public node modules node sass src libsass src operators hpp phpblog public node modules node sass src libsass src constants hpp phpblog public node modules node sass src libsass src error handling hpp phpblog public node modules node sass src custom importer bridge cpp phpblog public node modules node sass src libsass src parser hpp phpblog public node modules node sass src libsass src constants cpp phpblog public node modules node sass src sass types list cpp phpblog public node modules node sass src libsass src cssize cpp phpblog public node modules node sass src libsass src functions hpp phpblog public node modules node sass src libsass src util cpp phpblog public node modules node sass src custom function bridge cpp phpblog public node modules node sass src custom importer bridge h phpblog public node modules node sass src libsass src bind cpp phpblog public node modules node sass src libsass src eval hpp phpblog public node modules node sass src libsass src inspect cpp phpblog public node modules node sass src libsass src backtrace cpp phpblog public node modules node sass src libsass src extend cpp phpblog public node modules node sass src sass context wrapper h phpblog public node modules node sass src sass types sass value wrapper h phpblog public node modules node sass src libsass src error handling cpp phpblog public node modules node sass src libsass src parser cpp phpblog public node modules node sass src libsass src debugger hpp phpblog public node modules node sass src libsass src emitter cpp phpblog public node modules node sass src sass types number cpp phpblog public node modules node sass src sass types color h phpblog public node modules node sass src libsass src sass values cpp phpblog public node modules node sass src libsass src ast hpp phpblog public node modules node sass src libsass src output cpp phpblog public node modules node sass src libsass src check nesting cpp phpblog public node modules node sass src sass types null cpp phpblog public node modules node sass src libsass src ast def macros hpp phpblog public node modules node sass src libsass src functions cpp phpblog public node modules node sass src libsass src cssize hpp phpblog public node modules node sass src libsass src prelexer cpp phpblog public node modules node sass src libsass src ast cpp phpblog public node modules node sass src libsass src to c cpp phpblog public node modules node sass src libsass src to value hpp phpblog public node modules node sass src libsass src ast fwd decl hpp phpblog public node modules node sass src libsass src inspect hpp phpblog public node modules node sass src sass types color cpp phpblog public node modules node sass src libsass src values cpp phpblog public node modules node sass src sass context wrapper cpp phpblog public node modules node sass src sass types list h phpblog public node modules node sass src libsass src check nesting hpp phpblog public node modules node sass src sass types map cpp phpblog public node modules node sass src libsass src to value cpp phpblog public node modules node sass src libsass src context cpp phpblog public node modules node sass src binding cpp phpblog public node modules node sass src sass types string cpp phpblog public node modules node sass src libsass src sass context cpp phpblog public node modules node sass src libsass src prelexer hpp phpblog public node modules node sass src libsass src context hpp phpblog public node modules node sass src sass types boolean h phpblog public node modules node sass src libsass src eval cpp vulnerability details in libsass a heap based buffer over read exists in sass prelexer skip over scopes in prelexer hpp when called from sass parser parse import a similar issue to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
10,542
13,312,529,749
IssuesEvent
2020-08-26 09:52:07
e4exp/paper_manager_abstract
https://api.github.com/repos/e4exp/paper_manager_abstract
opened
Top-down Tree Structured Decoding with Syntactic Connections for Neural Machine Translation and Parsing
2018 Natural Language Processing Recurrent Neural Network Tree Structure _read_later
* https://arxiv.org/abs/1809.01854 * EMNLP 2018 ニューラル機械翻訳(NMT)システムにおける構文を意識した復号化の追加には、効果的な木構造ニューラルネットワーク、構文を意識した注意モデル、文構造に敏感な言語生成モデルが必要である。 我々は、Alvarez-MelisとJaakola(2017)によって最初に提案されたDRNN(Doubly-Recurrent Neural Networks)と呼ばれるトップダウン型の木構造ニューラルネットワークを利用して、シーケンシャルエンコーダと、構文を意識した注意モデルで補強された木構造デコーディングを組み合わせたSeq2DRNNと呼ばれるNMTモデルを作成する。 依存性解析モデルを使用する構文ベースのNMTへの以前のアプローチとは異なり、我々の手法は翻訳に有用な情報を提供すると主張する構成要素解析を使用しています。 さらに、文の構文構造を利用して、木構造デコーダニューラルネットワーク(Seq2DRNN+SynC)に新しい接続を追加します。 我々は、我々のNMTモデルをシーケンシャルモデルや最新の構文ベースのNMTモデルと比較し、我々のモデルがより良い順序付けでより流暢な翻訳を生成することを示している。 我々のモデルは翻訳と構成要素の解析を同時に行うことができるので、他のニューラル解析モデルと比較して解析精度を比較しています。
1.0
Top-down Tree Structured Decoding with Syntactic Connections for Neural Machine Translation and Parsing - * https://arxiv.org/abs/1809.01854 * EMNLP 2018 ニューラル機械翻訳(NMT)システムにおける構文を意識した復号化の追加には、効果的な木構造ニューラルネットワーク、構文を意識した注意モデル、文構造に敏感な言語生成モデルが必要である。 我々は、Alvarez-MelisとJaakola(2017)によって最初に提案されたDRNN(Doubly-Recurrent Neural Networks)と呼ばれるトップダウン型の木構造ニューラルネットワークを利用して、シーケンシャルエンコーダと、構文を意識した注意モデルで補強された木構造デコーディングを組み合わせたSeq2DRNNと呼ばれるNMTモデルを作成する。 依存性解析モデルを使用する構文ベースのNMTへの以前のアプローチとは異なり、我々の手法は翻訳に有用な情報を提供すると主張する構成要素解析を使用しています。 さらに、文の構文構造を利用して、木構造デコーダニューラルネットワーク(Seq2DRNN+SynC)に新しい接続を追加します。 我々は、我々のNMTモデルをシーケンシャルモデルや最新の構文ベースのNMTモデルと比較し、我々のモデルがより良い順序付けでより流暢な翻訳を生成することを示している。 我々のモデルは翻訳と構成要素の解析を同時に行うことができるので、他のニューラル解析モデルと比較して解析精度を比較しています。
process
top down tree structured decoding with syntactic connections for neural machine translation and parsing emnlp ニューラル機械翻訳 nmt システムにおける構文を意識した復号化の追加には、効果的な木構造ニューラルネットワーク、構文を意識した注意モデル、文構造に敏感な言語生成モデルが必要である。 我々は、alvarez melisとjaakola( )によって最初に提案されたdrnn(doubly recurrent neural networks)と呼ばれるトップダウン型の木構造ニューラルネットワークを利用して、シーケンシャルエンコーダと、 。 依存性解析モデルを使用する構文ベースのnmtへの以前のアプローチとは異なり、我々の手法は翻訳に有用な情報を提供すると主張する構成要素解析を使用しています。 さらに、文の構文構造を利用して、木構造デコーダニューラルネットワーク( sync)に新しい接続を追加します。 我々は、我々のnmtモデルをシーケンシャルモデルや最新の構文ベースのnmtモデルと比較し、我々のモデルがより良い順序付けでより流暢な翻訳を生成することを示している。 我々のモデルは翻訳と構成要素の解析を同時に行うことができるので、他のニューラル解析モデルと比較して解析精度を比較しています。
1
382,800
11,320,085,777
IssuesEvent
2020-01-21 02:33:14
C4C-TC-MSS/Sprint_Demo
https://api.github.com/repos/C4C-TC-MSS/Sprint_Demo
closed
Add next steps content at the end of the app
enhancement high priority
Add content around what potential applicants need to do next after receiving certificate information.
1.0
Add next steps content at the end of the app - Add content around what potential applicants need to do next after receiving certificate information.
non_process
add next steps content at the end of the app add content around what potential applicants need to do next after receiving certificate information
0
17,192
22,771,350,165
IssuesEvent
2022-07-08 10:16:42
googleapis/python-spanner
https://api.github.com/repos/googleapis/python-spanner
closed
Add a py.typed file for submodule
api: spanner type: process
To be in accordance with [PEP 561](https://www.python.org/dev/peps/pep-0561/) we should add a py.typed file to google/cloud/datastore/py.typed The file can be empty but does need to be present for some type checkers. Also, for this to work, google/cloud/spanner.py needs to be restructured to google/cloud/spanner/__init__.py See https://github.com/googleapis/python-firestore/issues/447 for additional context
1.0
Add a py.typed file for submodule - To be in accordance with [PEP 561](https://www.python.org/dev/peps/pep-0561/) we should add a py.typed file to google/cloud/datastore/py.typed The file can be empty but does need to be present for some type checkers. Also, for this to work, google/cloud/spanner.py needs to be restructured to google/cloud/spanner/__init__.py See https://github.com/googleapis/python-firestore/issues/447 for additional context
process
add a py typed file for submodule to be in accordance with we should add a py typed file to google cloud datastore py typed the file can be empty but does need to be present for some type checkers also for this to work google cloud spanner py needs to be restructured to google cloud spanner init py see for additional context
1
169,818
26,865,174,269
IssuesEvent
2023-02-03 22:37:10
Timurkazan99/fr-map
https://api.github.com/repos/Timurkazan99/fr-map
opened
Базовый дизайн
Design Priority: Medium
**Описание:** Спроектировать базовый дизайн пользовательского интерфйса **Задачи** - [ ] Дизайн основной страницы - [ ] Дизайн компонента окна информации - [ ] Дизайн поисковой строки - [ ] Дизайн страницы "О проекте"
1.0
Базовый дизайн - **Описание:** Спроектировать базовый дизайн пользовательского интерфйса **Задачи** - [ ] Дизайн основной страницы - [ ] Дизайн компонента окна информации - [ ] Дизайн поисковой строки - [ ] Дизайн страницы "О проекте"
non_process
базовый дизайн описание спроектировать базовый дизайн пользовательского интерфйса задачи дизайн основной страницы дизайн компонента окна информации дизайн поисковой строки дизайн страницы о проекте
0
26,895
6,812,693,486
IssuesEvent
2017-11-06 05:05:07
BTDF/DeploymentFramework
https://api.github.com/repos/BTDF/DeploymentFramework
closed
Add back in "Gac This" Tools menu item
bug CodePlexMigrationInitiated Impact: Low Installer Release 5.0
Add back in "Gac This" Tools menu item as "BTDF - GAC Selected Project Output". #### This work item was migrated from CodePlex CodePlex work item ID: '3658' Assigned to: 'tfabraham' Vote count: '0'
1.0
Add back in "Gac This" Tools menu item - Add back in "Gac This" Tools menu item as "BTDF - GAC Selected Project Output". #### This work item was migrated from CodePlex CodePlex work item ID: '3658' Assigned to: 'tfabraham' Vote count: '0'
non_process
add back in gac this tools menu item add back in gac this tools menu item as btdf gac selected project output this work item was migrated from codeplex codeplex work item id assigned to tfabraham vote count
0
686,965
23,510,332,668
IssuesEvent
2022-08-18 15:56:14
emory-libraries/ezpaarse-platforms
https://api.github.com/repos/emory-libraries/ezpaarse-platforms
closed
GlobalData
Add Parser High Priority Stakeholder Priority
### Example:star::star: : https://medical-globaldata-com.proxy.library.emory.edu/HomePage/Home https://login-globaldata-com.proxy.library.emory.edu/Home https://pharma-globaldata-com.proxy.library.emory.edu/ https://technology-globaldata-com.proxy.library.emory.edu/HomePage/Home Note that folks must register (once) to use Global Data via the Global Data portal prior to use, and then they can click on “IP access” afterwards. ### Priority: High ### Subscriber (Library): Woodruff ### ezPAARSE Analysis: http://ang.couperin.org/platforms/to_be_completed/
2.0
GlobalData - ### Example:star::star: : https://medical-globaldata-com.proxy.library.emory.edu/HomePage/Home https://login-globaldata-com.proxy.library.emory.edu/Home https://pharma-globaldata-com.proxy.library.emory.edu/ https://technology-globaldata-com.proxy.library.emory.edu/HomePage/Home Note that folks must register (once) to use Global Data via the Global Data portal prior to use, and then they can click on “IP access” afterwards. ### Priority: High ### Subscriber (Library): Woodruff ### ezPAARSE Analysis: http://ang.couperin.org/platforms/to_be_completed/
non_process
globaldata example star star note that folks must register once to use global data via the global data portal prior to use and then they can click on “ip access” afterwards priority high subscriber library woodruff ezpaarse analysis
0
316,383
9,646,709,564
IssuesEvent
2019-05-17 12:06:46
geosolutions-it/MapStore2
https://api.github.com/repos/geosolutions-it/MapStore2
closed
After login on public map you can not save
Accepted Priority: High bug pending review review
### Description If you open an existing map and login, you can not see the "Save" button, only "Save as". This because `mapInfo` are not reloaded. ### In case of Bug (otherwise remove this paragraph) *Browser Affected* any *Steps to reproduce* - Open a map - login as owner or admin - Open burger menu *Expected Result* - I have save button *Current Result* - save button is missing (only save as). If you refresh or go back and than open again the map, the save button disappears. ### Other useful information (optional): This issue is related to #1993 , probably the same cause ( `mapInfo` not reloaded ).
1.0
After login on public map you can not save - ### Description If you open an existing map and login, you can not see the "Save" button, only "Save as". This because `mapInfo` are not reloaded. ### In case of Bug (otherwise remove this paragraph) *Browser Affected* any *Steps to reproduce* - Open a map - login as owner or admin - Open burger menu *Expected Result* - I have save button *Current Result* - save button is missing (only save as). If you refresh or go back and than open again the map, the save button disappears. ### Other useful information (optional): This issue is related to #1993 , probably the same cause ( `mapInfo` not reloaded ).
non_process
after login on public map you can not save description if you open an existing map and login you can not see the save button only save as this because mapinfo are not reloaded in case of bug otherwise remove this paragraph browser affected any steps to reproduce open a map login as owner or admin open burger menu expected result i have save button current result save button is missing only save as if you refresh or go back and than open again the map the save button disappears other useful information optional this issue is related to probably the same cause mapinfo not reloaded
0
330,180
24,249,635,375
IssuesEvent
2022-09-27 13:20:53
usnistgov/OSCAL
https://api.github.com/repos/usnistgov/OSCAL
closed
Provide guidance around relative URI handling for href instances
enhancement User Story Scope: Metaschema Scope: Documentation
# User Story: As an OSCAL developer, I would like to have consistent guidance around relative URI handling within href in OSCAL. ## Goals: Update the OSCAL docs references with consistent guidance, including [/profile/metadata/link/@href](https://github.com/usnistgov/OSCAL/blob/main/docs/content/reference/latest/profile/xml-reference.md?plain=1#L659) and [/profile/back-matter/resource/rlink/@href](https://github.com/usnistgov/OSCAL/blob/main/docs/content/reference/latest/profile/xml-reference.md?plain=1#L8788). - [x] Ensure each instance of model reference documentation clearly indicates if an href requires a URI reference or an abolute URI. This issue builds on the work to review the OSCAL models (#1066 #1331). ## Dependencies: none ## Acceptance Criteria - [x] The guidance addresses when relative URIs are supported and not. - [x] All [OSCAL website](https://pages.nist.gov/OSCAL) and readme documentation affected by the changes in this issue have been updated. Changes to the OSCAL website can be made in the docs/content directory of your branch. - [x] A Pull Request (PR) is submitted that fully addresses the goals of this User Story. This issue is referenced in the PR. - [x] The CI-CD build process runs without any reported errors on the PR. This can be confirmed by reviewing that all checks have passed in the PR.
1.0
Provide guidance around relative URI handling for href instances - # User Story: As an OSCAL developer, I would like to have consistent guidance around relative URI handling within href in OSCAL. ## Goals: Update the OSCAL docs references with consistent guidance, including [/profile/metadata/link/@href](https://github.com/usnistgov/OSCAL/blob/main/docs/content/reference/latest/profile/xml-reference.md?plain=1#L659) and [/profile/back-matter/resource/rlink/@href](https://github.com/usnistgov/OSCAL/blob/main/docs/content/reference/latest/profile/xml-reference.md?plain=1#L8788). - [x] Ensure each instance of model reference documentation clearly indicates if an href requires a URI reference or an abolute URI. This issue builds on the work to review the OSCAL models (#1066 #1331). ## Dependencies: none ## Acceptance Criteria - [x] The guidance addresses when relative URIs are supported and not. - [x] All [OSCAL website](https://pages.nist.gov/OSCAL) and readme documentation affected by the changes in this issue have been updated. Changes to the OSCAL website can be made in the docs/content directory of your branch. - [x] A Pull Request (PR) is submitted that fully addresses the goals of this User Story. This issue is referenced in the PR. - [x] The CI-CD build process runs without any reported errors on the PR. This can be confirmed by reviewing that all checks have passed in the PR.
non_process
provide guidance around relative uri handling for href instances user story as an oscal developer i would like to have consistent guidance around relative uri handling within href in oscal goals update the oscal docs references with consistent guidance including and ensure each instance of model reference documentation clearly indicates if an href requires a uri reference or an abolute uri this issue builds on the work to review the oscal models dependencies none acceptance criteria the guidance addresses when relative uris are supported and not all and readme documentation affected by the changes in this issue have been updated changes to the oscal website can be made in the docs content directory of your branch a pull request pr is submitted that fully addresses the goals of this user story this issue is referenced in the pr the ci cd build process runs without any reported errors on the pr this can be confirmed by reviewing that all checks have passed in the pr
0
69,914
3,316,304,685
IssuesEvent
2015-11-06 16:21:19
TeselaGen/Peony-Issue-Tracking
https://api.github.com/repos/TeselaGen/Peony-Issue-Tracking
opened
Can't import or create a valid FASTA AA file
Customer: DAS Phase I Milestone #4 - Oracle Rewrite Priority: High Status: In Progress Type: Bug
_From @mfero on October 18, 2015 5:29_ Can't import or create a valid FASTA AA file. Should be able to create in a simple text box editor for direct typing or cut/paste. >gi|129295|sp|P01013|OVAX_CHICK GENE X PROTEIN (OVALBUMIN-RELATED) QIKDLLVSSSTDLDTTLVLVNAIYFKGMWKTAFNAEDTREMPFHVTKQESKPVQMMCMNNSFNVAT PAEKMKILELPFASGDLSMLVLLPDEVSDLERIE TINFEKLTEWTNPNTMEKRRVKVYLPQMKIEEK NLTSVLMALGMTDLFIPSANLTGISSAESLKISQ VHGAFMELSEDGIEMAGSTGVIEDIKHSPESEQ RADHPFLFLIKHNPTNTIVYFGRYWSP _Copied from original issue: TeselaGen/ve#1438_
1.0
Can't import or create a valid FASTA AA file - _From @mfero on October 18, 2015 5:29_ Can't import or create a valid FASTA AA file. Should be able to create in a simple text box editor for direct typing or cut/paste. >gi|129295|sp|P01013|OVAX_CHICK GENE X PROTEIN (OVALBUMIN-RELATED) QIKDLLVSSSTDLDTTLVLVNAIYFKGMWKTAFNAEDTREMPFHVTKQESKPVQMMCMNNSFNVAT PAEKMKILELPFASGDLSMLVLLPDEVSDLERIE TINFEKLTEWTNPNTMEKRRVKVYLPQMKIEEK NLTSVLMALGMTDLFIPSANLTGISSAESLKISQ VHGAFMELSEDGIEMAGSTGVIEDIKHSPESEQ RADHPFLFLIKHNPTNTIVYFGRYWSP _Copied from original issue: TeselaGen/ve#1438_
non_process
can t import or create a valid fasta aa file from mfero on october can t import or create a valid fasta aa file should be able to create in a simple text box editor for direct typing or cut paste gi sp ovax chick gene x protein ovalbumin related qikdllvssstdldttlvlvnaiyfkgmwktafnaedtrempfhvtkqeskpvqmmcmnnsfnvat paekmkilelpfasgdlsmlvllpdevsdlerie tinfekltewtnpntmekrrvkvylpqmkieek nltsvlmalgmtdlfipsanltgissaeslkisq vhgafmelsedgiemagstgviedikhspeseq radhpflflikhnptntivyfgrywsp copied from original issue teselagen ve
0
88,822
15,820,495,765
IssuesEvent
2021-04-05 19:03:41
dmyers87/tika
https://api.github.com/repos/dmyers87/tika
closed
CVE-2019-17573 (Medium) detected in cxf-rt-transports-http-3.3.2.jar - autoclosed
security vulnerability
## CVE-2019-17573 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cxf-rt-transports-http-3.3.2.jar</b></p></summary> <p>Apache CXF Runtime HTTP Transport</p> <p>Library home page: <a href="http://cxf.apache.org">http://cxf.apache.org</a></p> <p>Path to dependency file: tika/tika-langdetect/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/cxf/cxf-rt-transports-http/3.3.2/cxf-rt-transports-http-3.3.2.jar,/home/wss-scanner/.m2/repository/org/apache/cxf/cxf-rt-transports-http/3.3.2/cxf-rt-transports-http-3.3.2.jar</p> <p> Dependency Hierarchy: - cxf-rt-rs-client-3.3.2.jar (Root Library) - :x: **cxf-rt-transports-http-3.3.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dmyers87/tika/commit/b0634f6d9bc18cc79f623715d40c9e8ed98924fc">b0634f6d9bc18cc79f623715d40c9e8ed98924fc</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> By default, Apache CXF creates a /services page containing a listing of the available endpoint names and addresses. This webpage is vulnerable to a reflected Cross-Site Scripting (XSS) attack, which allows a malicious actor to inject javascript into the web page. Please note that the attack exploits a feature which is not typically not present in modern browsers, who remove dot segments before sending the request. However, Mobile applications may be vulnerable. <p>Publish Date: 2020-01-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17573>CVE-2019-17573</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17573">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17573</a></p> <p>Release Date: 2020-01-16</p> <p>Fix Resolution: org.apache.cxf:cxf-rt-transports-http:3.2.12,3.3.5</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.cxf","packageName":"cxf-rt-transports-http","packageVersion":"3.3.2","packageFilePaths":["/tika-langdetect/pom.xml","/tika-parsers/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.cxf:cxf-rt-rs-client:3.3.2;org.apache.cxf:cxf-rt-transports-http:3.3.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.cxf:cxf-rt-transports-http:3.2.12,3.3.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-17573","vulnerabilityDetails":"By default, Apache CXF creates a /services page containing a listing of the available endpoint names and addresses. This webpage is vulnerable to a reflected Cross-Site Scripting (XSS) attack, which allows a malicious actor to inject javascript into the web page. Please note that the attack exploits a feature which is not typically not present in modern browsers, who remove dot segments before sending the request. However, Mobile applications may be vulnerable.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17573","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-17573 (Medium) detected in cxf-rt-transports-http-3.3.2.jar - autoclosed - ## CVE-2019-17573 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cxf-rt-transports-http-3.3.2.jar</b></p></summary> <p>Apache CXF Runtime HTTP Transport</p> <p>Library home page: <a href="http://cxf.apache.org">http://cxf.apache.org</a></p> <p>Path to dependency file: tika/tika-langdetect/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/cxf/cxf-rt-transports-http/3.3.2/cxf-rt-transports-http-3.3.2.jar,/home/wss-scanner/.m2/repository/org/apache/cxf/cxf-rt-transports-http/3.3.2/cxf-rt-transports-http-3.3.2.jar</p> <p> Dependency Hierarchy: - cxf-rt-rs-client-3.3.2.jar (Root Library) - :x: **cxf-rt-transports-http-3.3.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dmyers87/tika/commit/b0634f6d9bc18cc79f623715d40c9e8ed98924fc">b0634f6d9bc18cc79f623715d40c9e8ed98924fc</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> By default, Apache CXF creates a /services page containing a listing of the available endpoint names and addresses. This webpage is vulnerable to a reflected Cross-Site Scripting (XSS) attack, which allows a malicious actor to inject javascript into the web page. Please note that the attack exploits a feature which is not typically not present in modern browsers, who remove dot segments before sending the request. However, Mobile applications may be vulnerable. <p>Publish Date: 2020-01-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17573>CVE-2019-17573</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17573">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17573</a></p> <p>Release Date: 2020-01-16</p> <p>Fix Resolution: org.apache.cxf:cxf-rt-transports-http:3.2.12,3.3.5</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.cxf","packageName":"cxf-rt-transports-http","packageVersion":"3.3.2","packageFilePaths":["/tika-langdetect/pom.xml","/tika-parsers/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.cxf:cxf-rt-rs-client:3.3.2;org.apache.cxf:cxf-rt-transports-http:3.3.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.cxf:cxf-rt-transports-http:3.2.12,3.3.5"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-17573","vulnerabilityDetails":"By default, Apache CXF creates a /services page containing a listing of the available endpoint names and addresses. This webpage is vulnerable to a reflected Cross-Site Scripting (XSS) attack, which allows a malicious actor to inject javascript into the web page. Please note that the attack exploits a feature which is not typically not present in modern browsers, who remove dot segments before sending the request. However, Mobile applications may be vulnerable.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17573","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in cxf rt transports http jar autoclosed cve medium severity vulnerability vulnerable library cxf rt transports http jar apache cxf runtime http transport library home page a href path to dependency file tika tika langdetect pom xml path to vulnerable library home wss scanner repository org apache cxf cxf rt transports http cxf rt transports http jar home wss scanner repository org apache cxf cxf rt transports http cxf rt transports http jar dependency hierarchy cxf rt rs client jar root library x cxf rt transports http jar vulnerable library found in head commit a href found in base branch master vulnerability details by default apache cxf creates a services page containing a listing of the available endpoint names and addresses this webpage is vulnerable to a reflected cross site scripting xss attack which allows a malicious actor to inject javascript into the web page please note that the attack exploits a feature which is not typically not present in modern browsers who remove dot segments before sending the request however mobile applications may be vulnerable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache cxf cxf rt transports http isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org apache cxf cxf rt rs client org apache cxf cxf rt transports http isminimumfixversionavailable true minimumfixversion org apache cxf cxf rt transports http basebranches vulnerabilityidentifier cve vulnerabilitydetails by default apache cxf creates a services page containing a listing of the available endpoint names and addresses this webpage is vulnerable to a reflected cross site scripting xss attack which allows a malicious actor to inject javascript into the web page please note that the attack exploits a feature which is not typically not present in modern browsers who remove dot segments before sending the request however mobile applications may be vulnerable vulnerabilityurl
0
1,841
20,387,881,072
IssuesEvent
2022-02-22 09:03:45
FoundationDB/fdb-kubernetes-operator
https://api.github.com/repos/FoundationDB/fdb-kubernetes-operator
closed
Skipping instances to be removed when generating the initial cluster file
good first issue reliability
If we have an instance that fails to come up when we first create a cluster, we may want to spin up a new instance to replace it. This causes problems with the step to generate the initial cluster file, because it requires that all processes have IPs assigned. I think we should ignore instances that are marked for removal when generating the initial cluster file.
True
Skipping instances to be removed when generating the initial cluster file - If we have an instance that fails to come up when we first create a cluster, we may want to spin up a new instance to replace it. This causes problems with the step to generate the initial cluster file, because it requires that all processes have IPs assigned. I think we should ignore instances that are marked for removal when generating the initial cluster file.
non_process
skipping instances to be removed when generating the initial cluster file if we have an instance that fails to come up when we first create a cluster we may want to spin up a new instance to replace it this causes problems with the step to generate the initial cluster file because it requires that all processes have ips assigned i think we should ignore instances that are marked for removal when generating the initial cluster file
0
11,747
14,583,351,135
IssuesEvent
2020-12-18 13:52:27
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
Participant manager has checkbox in unexpected location with unclear purpose
Feature request P2 Participant manager Process: Tested QA Process: Tested dev UI UX
The checkbox in the screenshot attached is in a location that does not make its purpose very clear. I am guessing that checking that box means that I've selected all of the records on the page - but since the individual records do not have check boxes I can't really confirm that is the case. ![Screenshot 2020-12-15 at 6 41 23 PM](https://user-images.githubusercontent.com/35972680/102286018-6366cd00-3f05-11eb-90c3-02a188d15a7e.png)
2.0
Participant manager has checkbox in unexpected location with unclear purpose - The checkbox in the screenshot attached is in a location that does not make its purpose very clear. I am guessing that checking that box means that I've selected all of the records on the page - but since the individual records do not have check boxes I can't really confirm that is the case. ![Screenshot 2020-12-15 at 6 41 23 PM](https://user-images.githubusercontent.com/35972680/102286018-6366cd00-3f05-11eb-90c3-02a188d15a7e.png)
process
participant manager has checkbox in unexpected location with unclear purpose the checkbox in the screenshot attached is in a location that does not make its purpose very clear i am guessing that checking that box means that i ve selected all of the records on the page but since the individual records do not have check boxes i can t really confirm that is the case
1
697,750
23,951,882,150
IssuesEvent
2022-09-12 12:13:54
FactorioAntigrief/FactorioAntigrief
https://api.github.com/repos/FactorioAntigrief/FactorioAntigrief
closed
Add an option to replace the APIURL when generating the banlist
scope:community-bot priority:high type:feature
Add the option of replacing the APIURL when generating the banlist, so that the backend can be contacted through a localhost domain, but still show up as another (factoriobans.club) in the exported JSON file.
1.0
Add an option to replace the APIURL when generating the banlist - Add the option of replacing the APIURL when generating the banlist, so that the backend can be contacted through a localhost domain, but still show up as another (factoriobans.club) in the exported JSON file.
non_process
add an option to replace the apiurl when generating the banlist add the option of replacing the apiurl when generating the banlist so that the backend can be contacted through a localhost domain but still show up as another factoriobans club in the exported json file
0
502,234
14,542,779,347
IssuesEvent
2020-12-15 16:06:12
magento/magento2
https://api.github.com/repos/magento/magento2
opened
[Issue] Spellings mistake
Component: Checkout Priority: P4 Severity: S4
This issue is automatically created based on existing pull request: magento/magento2#31276: Spellings mistake --------- <!--- Thank you for contributing to Magento. To help us process this pull request we recommend that you add the following information: - Summary of the pull request, - Issue(s) related to the changes made, - Manual testing scenarios Fields marked with (*) are required. Please don't remove the template. --> <!--- Please provide a general summary of the Pull Request in the Title above --> ### Description (*) <!--- Please provide a description of the changes proposed in the pull request. Letting us know what has changed and why it needed changing will help us validate this pull request. --> ### Related Pull Requests <!-- related pull request placeholder --> ### Fixed Issues (if relevant) <!--- If relevant, please provide a list of fixed issues in the format magento/magento2#<issue_number>. There could be 1 or more issues linked here and it will help us find some more information about the reasoning behind this change. --> 1. Fixes magento/magento2#<issue_number> ### Manual testing scenarios (*) <!--- Please provide a set of unambiguous steps to test the proposed code change. Giving us manual testing scenarios will help with the processing and validation process. --> 1. ... 2. ... ### Questions or comments <!--- If relevant, here you can ask questions or provide comments on your pull request for the reviewer For example if you need assistance with writing tests or would like some feedback on one of your development ideas --> ### Contribution checklist (*) - [ ] Pull request has a meaningful description of its purpose - [ ] All commits are accompanied by meaningful commit messages - [ ] All new or changed code is covered with unit/integration tests (if applicable) - [ ] All automated tests passed successfully (all builds are green)
1.0
[Issue] Spellings mistake - This issue is automatically created based on existing pull request: magento/magento2#31276: Spellings mistake --------- <!--- Thank you for contributing to Magento. To help us process this pull request we recommend that you add the following information: - Summary of the pull request, - Issue(s) related to the changes made, - Manual testing scenarios Fields marked with (*) are required. Please don't remove the template. --> <!--- Please provide a general summary of the Pull Request in the Title above --> ### Description (*) <!--- Please provide a description of the changes proposed in the pull request. Letting us know what has changed and why it needed changing will help us validate this pull request. --> ### Related Pull Requests <!-- related pull request placeholder --> ### Fixed Issues (if relevant) <!--- If relevant, please provide a list of fixed issues in the format magento/magento2#<issue_number>. There could be 1 or more issues linked here and it will help us find some more information about the reasoning behind this change. --> 1. Fixes magento/magento2#<issue_number> ### Manual testing scenarios (*) <!--- Please provide a set of unambiguous steps to test the proposed code change. Giving us manual testing scenarios will help with the processing and validation process. --> 1. ... 2. ... ### Questions or comments <!--- If relevant, here you can ask questions or provide comments on your pull request for the reviewer For example if you need assistance with writing tests or would like some feedback on one of your development ideas --> ### Contribution checklist (*) - [ ] Pull request has a meaningful description of its purpose - [ ] All commits are accompanied by meaningful commit messages - [ ] All new or changed code is covered with unit/integration tests (if applicable) - [ ] All automated tests passed successfully (all builds are green)
non_process
spellings mistake this issue is automatically created based on existing pull request magento spellings mistake thank you for contributing to magento to help us process this pull request we recommend that you add the following information summary of the pull request issue s related to the changes made manual testing scenarios fields marked with are required please don t remove the template description please provide a description of the changes proposed in the pull request letting us know what has changed and why it needed changing will help us validate this pull request related pull requests fixed issues if relevant if relevant please provide a list of fixed issues in the format magento there could be or more issues linked here and it will help us find some more information about the reasoning behind this change fixes magento manual testing scenarios please provide a set of unambiguous steps to test the proposed code change giving us manual testing scenarios will help with the processing and validation process questions or comments if relevant here you can ask questions or provide comments on your pull request for the reviewer for example if you need assistance with writing tests or would like some feedback on one of your development ideas contribution checklist pull request has a meaningful description of its purpose all commits are accompanied by meaningful commit messages all new or changed code is covered with unit integration tests if applicable all automated tests passed successfully all builds are green
0
16,917
2,615,126,502
IssuesEvent
2015-03-01 05:54:55
chrsmith/google-api-java-client
https://api.github.com/repos/chrsmith/google-api-java-client
closed
Update GoogleClient to use rootUrl and servicePath
auto-migrated Milestone-Version1.10.0 Priority-Medium Type-Enhancement
``` External references, such as a standards document, or specification? http://codereview.appspot.com/6258043/ Java environments (e.g. Java 6, Android 2.3, App Engine, or All)? All Please describe the feature requested. Update GoogleClient to use rootUrl and servicePath from JsonHttpClient. ``` Original issue reported on code.google.com by `rmis...@google.com` on 29 May 2012 at 1:40
1.0
Update GoogleClient to use rootUrl and servicePath - ``` External references, such as a standards document, or specification? http://codereview.appspot.com/6258043/ Java environments (e.g. Java 6, Android 2.3, App Engine, or All)? All Please describe the feature requested. Update GoogleClient to use rootUrl and servicePath from JsonHttpClient. ``` Original issue reported on code.google.com by `rmis...@google.com` on 29 May 2012 at 1:40
non_process
update googleclient to use rooturl and servicepath external references such as a standards document or specification java environments e g java android app engine or all all please describe the feature requested update googleclient to use rooturl and servicepath from jsonhttpclient original issue reported on code google com by rmis google com on may at
0
336,362
30,190,174,414
IssuesEvent
2023-07-04 14:47:10
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: cdc/cloud-sink-gcs/rangefeed=true failed
C-test-failure O-robot A-cdc O-roachtest release-blocker T-cdc branch-release-23.1
roachtest.cdc/cloud-sink-gcs/rangefeed=true [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/10756610?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/10756610?buildTab=artifacts#/cdc/cloud-sink-gcs/rangefeed=true) on release-23.1 @ [e12e85479312972b551677203849d29aeb38ad5f](https://github.com/cockroachdb/cockroach/commits/e12e85479312972b551677203849d29aeb38ad5f): ``` (cdc.go:411).newChangefeed: failed to create changefeed: pq: failed to create google cloud client: dialing: google: could not find default credentials. See https://developers.google.com/accounts/docs/application-default-credentials for more information. (cluster.go:2247).Run: output in run_083655.788592718_n4_workload-run-tpcc-wa: ./workload run tpcc --warehouses=50 --duration=30m {pgurl:1-3} returned: context canceled (cdc.go:315).Close: error shutting down prometheus/grafana: context canceled test artifacts and logs in: /artifacts/cdc/cloud-sink-gcs/rangefeed=true/run_1 ``` <p>Parameters: <code>ROACHTEST_arch=amd64</code> , <code>ROACHTEST_cloud=aws</code> , <code>ROACHTEST_cpu=16</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=true</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/cdc <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*cdc/cloud-sink-gcs/rangefeed=true.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-29325
2.0
roachtest: cdc/cloud-sink-gcs/rangefeed=true failed - roachtest.cdc/cloud-sink-gcs/rangefeed=true [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/10756610?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/10756610?buildTab=artifacts#/cdc/cloud-sink-gcs/rangefeed=true) on release-23.1 @ [e12e85479312972b551677203849d29aeb38ad5f](https://github.com/cockroachdb/cockroach/commits/e12e85479312972b551677203849d29aeb38ad5f): ``` (cdc.go:411).newChangefeed: failed to create changefeed: pq: failed to create google cloud client: dialing: google: could not find default credentials. See https://developers.google.com/accounts/docs/application-default-credentials for more information. (cluster.go:2247).Run: output in run_083655.788592718_n4_workload-run-tpcc-wa: ./workload run tpcc --warehouses=50 --duration=30m {pgurl:1-3} returned: context canceled (cdc.go:315).Close: error shutting down prometheus/grafana: context canceled test artifacts and logs in: /artifacts/cdc/cloud-sink-gcs/rangefeed=true/run_1 ``` <p>Parameters: <code>ROACHTEST_arch=amd64</code> , <code>ROACHTEST_cloud=aws</code> , <code>ROACHTEST_cpu=16</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=true</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/cdc <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*cdc/cloud-sink-gcs/rangefeed=true.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-29325
non_process
roachtest cdc cloud sink gcs rangefeed true failed roachtest cdc cloud sink gcs rangefeed true with on release cdc go newchangefeed failed to create changefeed pq failed to create google cloud client dialing google could not find default credentials see for more information cluster go run output in run workload run tpcc wa workload run tpcc warehouses duration pgurl returned context canceled cdc go close error shutting down prometheus grafana context canceled test artifacts and logs in artifacts cdc cloud sink gcs rangefeed true run parameters roachtest arch roachtest cloud aws roachtest cpu roachtest encrypted false roachtest fs roachtest localssd true roachtest ssd help see see cc cockroachdb cdc jira issue crdb
0
397,239
11,725,595,946
IssuesEvent
2020-03-10 13:14:20
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.price.com.hk - see bug description
browser-firefox engine-gecko form-v2-experiment os-linux priority-important
<!-- @browser: Firefox 73.0 --> <!-- @ua_header: Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:73.0) Gecko/20100101 Firefox/73.0 --> <!-- @reported_with: --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/49913 --> <!-- @extra_labels: form-v2-experiment --> **URL**: https://www.price.com.hk/search.php?g=A&q=s10e **Browser / Version**: Firefox 73.0 **Operating System**: Ubuntu **Tested Another Browser**: Yes **Problem type**: Something else **Description**: contents shown on right side only and truncated **Steps to Reproduce**: the page appears normal in Chromium <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/3/db2760ba-d1cb-420c-9cf2-2c5f8c71be78.jpg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.price.com.hk - see bug description - <!-- @browser: Firefox 73.0 --> <!-- @ua_header: Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:73.0) Gecko/20100101 Firefox/73.0 --> <!-- @reported_with: --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/49913 --> <!-- @extra_labels: form-v2-experiment --> **URL**: https://www.price.com.hk/search.php?g=A&q=s10e **Browser / Version**: Firefox 73.0 **Operating System**: Ubuntu **Tested Another Browser**: Yes **Problem type**: Something else **Description**: contents shown on right side only and truncated **Steps to Reproduce**: the page appears normal in Chromium <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/3/db2760ba-d1cb-420c-9cf2-2c5f8c71be78.jpg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
see bug description url browser version firefox operating system ubuntu tested another browser yes problem type something else description contents shown on right side only and truncated steps to reproduce the page appears normal in chromium view the screenshot img alt screenshot src browser configuration none from with ❤️
0
10,697
13,492,884,190
IssuesEvent
2020-09-11 18:44:50
googleapis/nodejs-logging
https://api.github.com/repos/googleapis/nodejs-logging
closed
flaky samples-test
api: logging type: process
The samples-test pass but looking at the details of the test run, it actually throws error: https://source.cloud.google.com/results/invocations/4aad0f84-b684-41db-ac9e-e520c0515ea2/targets/cloud-devrel%2Fclient-libraries%2Fnodejs%2Fpresubmit%2Fgoogleapis%2Fnodejs-logging%2Fnode10%2Fsamples-test/log And same thing happens for several previous PRs.
1.0
flaky samples-test - The samples-test pass but looking at the details of the test run, it actually throws error: https://source.cloud.google.com/results/invocations/4aad0f84-b684-41db-ac9e-e520c0515ea2/targets/cloud-devrel%2Fclient-libraries%2Fnodejs%2Fpresubmit%2Fgoogleapis%2Fnodejs-logging%2Fnode10%2Fsamples-test/log And same thing happens for several previous PRs.
process
flaky samples test the samples test pass but looking at the details of the test run it actually throws error and same thing happens for several previous prs
1
4,961
7,803,433,839
IssuesEvent
2018-06-10 23:54:10
exercism/cli
https://api.github.com/repos/exercism/cli
closed
Homebrew bash/zsh completion
pinned release-process
I noticed that the bash/zsh completion files aren't included in the install tar files, and am unsure what the generation process looks like. If not included in the tar files, have you considered leveraging the [Homebrew resource calls](https://github.com/Homebrew/brew/blob/master/docs/Formula-Cookbook.md#specifying-gems-python-modules-go-projects-etc-as-dependencies) the the cli website and installing that way? Also happy to put this over in the Homebrew repo if more appropriate, but figured I would start here since the files didn't seem to be in the tars. https://github.com/Homebrew/homebrew-core/blob/c53141dcda9d1a062eba89f77a5c02f09f6b5bd2/Formula/exercism.rb Thanks, jon
1.0
Homebrew bash/zsh completion - I noticed that the bash/zsh completion files aren't included in the install tar files, and am unsure what the generation process looks like. If not included in the tar files, have you considered leveraging the [Homebrew resource calls](https://github.com/Homebrew/brew/blob/master/docs/Formula-Cookbook.md#specifying-gems-python-modules-go-projects-etc-as-dependencies) the the cli website and installing that way? Also happy to put this over in the Homebrew repo if more appropriate, but figured I would start here since the files didn't seem to be in the tars. https://github.com/Homebrew/homebrew-core/blob/c53141dcda9d1a062eba89f77a5c02f09f6b5bd2/Formula/exercism.rb Thanks, jon
process
homebrew bash zsh completion i noticed that the bash zsh completion files aren t included in the install tar files and am unsure what the generation process looks like if not included in the tar files have you considered leveraging the the the cli website and installing that way also happy to put this over in the homebrew repo if more appropriate but figured i would start here since the files didn t seem to be in the tars thanks jon
1
16,620
21,678,126,856
IssuesEvent
2022-05-09 01:23:14
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add The Mesmerizer
suggested title in process
Please add as much of the following info as you can: Title: The Mesmerizer Type (film/tv show): police procedural drama Film or show in which it appears: The Boys Is the parent film/show streaming anywhere? Yes - Amazon Prime About when in the parent film/show does it appear? Ep. 1x06: "The Innocents." -> 25:08 - 25:28 Actual footage of the film/show can be seen (yes/no)? Yes Cast: Mesmer as Lt. Howser Catchphrase: "There are no secrets from me." Factoid: "The Mesmerizer" was a critic's choice TV procedural that ran for three seasons in the 1990's.
1.0
Add The Mesmerizer - Please add as much of the following info as you can: Title: The Mesmerizer Type (film/tv show): police procedural drama Film or show in which it appears: The Boys Is the parent film/show streaming anywhere? Yes - Amazon Prime About when in the parent film/show does it appear? Ep. 1x06: "The Innocents." -> 25:08 - 25:28 Actual footage of the film/show can be seen (yes/no)? Yes Cast: Mesmer as Lt. Howser Catchphrase: "There are no secrets from me." Factoid: "The Mesmerizer" was a critic's choice TV procedural that ran for three seasons in the 1990's.
process
add the mesmerizer please add as much of the following info as you can title the mesmerizer type film tv show police procedural drama film or show in which it appears the boys is the parent film show streaming anywhere yes amazon prime about when in the parent film show does it appear ep the innocents actual footage of the film show can be seen yes no yes cast mesmer as lt howser catchphrase there are no secrets from me factoid the mesmerizer was a critic s choice tv procedural that ran for three seasons in the s
1
21,278
28,442,550,421
IssuesEvent
2023-04-16 04:03:01
cse442-at-ub/project_s23-team-infinity
https://api.github.com/repos/cse442-at-ub/project_s23-team-infinity
closed
Create password hashing and verifying feature for login and register page
Processing Task Sprint 3
**Task test** *Test 1* 1) Go to https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442ad/signup. 2) Create an account with account example@buffalo.edu and password "password" 3) Click the Sign up button (click the word "Sign up") and verify that the pop-up says "Account Created". 4) Click on the "Login" button on the below the "Sign up" button to be redirected to the login page. 5) Enter account example@buffalo.edu and password "password" 6) Click the "Login" button and verify that you are redirected to the homepage. 7) Go to https://www-student.cse.buffalo.edu/tools/db/phpmyadmin/ to check if the hash has been applied to your account's password. 8) Make sure your server choice is "oceanus". 9) Enter the Username: "duncenzh" and Password: "123" in the required fields (without quotation) and click "go". 10) On the left of the site click on "cse442_2023_spring_team_ad_db" and click "Users" in the "Table". 11) Verify that on the very bottom of the list from the "Users" table that the password saved is not your password.
1.0
Create password hashing and verifying feature for login and register page - **Task test** *Test 1* 1) Go to https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442ad/signup. 2) Create an account with account example@buffalo.edu and password "password" 3) Click the Sign up button (click the word "Sign up") and verify that the pop-up says "Account Created". 4) Click on the "Login" button on the below the "Sign up" button to be redirected to the login page. 5) Enter account example@buffalo.edu and password "password" 6) Click the "Login" button and verify that you are redirected to the homepage. 7) Go to https://www-student.cse.buffalo.edu/tools/db/phpmyadmin/ to check if the hash has been applied to your account's password. 8) Make sure your server choice is "oceanus". 9) Enter the Username: "duncenzh" and Password: "123" in the required fields (without quotation) and click "go". 10) On the left of the site click on "cse442_2023_spring_team_ad_db" and click "Users" in the "Table". 11) Verify that on the very bottom of the list from the "Users" table that the password saved is not your password.
process
create password hashing and verifying feature for login and register page task test test go to create an account with account example buffalo edu and password password click the sign up button click the word sign up and verify that the pop up says account created click on the login button on the below the sign up button to be redirected to the login page enter account example buffalo edu and password password click the login button and verify that you are redirected to the homepage go to to check if the hash has been applied to your account s password make sure your server choice is oceanus enter the username duncenzh and password in the required fields without quotation and click go on the left of the site click on spring team ad db and click users in the table verify that on the very bottom of the list from the users table that the password saved is not your password
1
6,077
8,922,895,435
IssuesEvent
2019-01-21 14:16:16
linnovate/root
https://api.github.com/repos/linnovate/root
opened
in documents, updates field gets stuck after selecting a document from template
Process bug
go to documents select a folder that has an office with a doc template in it click on select from template select the file that appears in the window write some updates so the updates field will move downwards try to scroll the updates field to see the add button again the updates field is stuck and cant move downwards anymore
1.0
in documents, updates field gets stuck after selecting a document from template - go to documents select a folder that has an office with a doc template in it click on select from template select the file that appears in the window write some updates so the updates field will move downwards try to scroll the updates field to see the add button again the updates field is stuck and cant move downwards anymore
process
in documents updates field gets stuck after selecting a document from template go to documents select a folder that has an office with a doc template in it click on select from template select the file that appears in the window write some updates so the updates field will move downwards try to scroll the updates field to see the add button again the updates field is stuck and cant move downwards anymore
1
6,939
10,110,543,815
IssuesEvent
2019-07-30 10:30:57
bisq-network/bisq
https://api.github.com/repos/bisq-network/bisq
closed
Error at MakerCreateAndSignContract on v1.0.1 not allowing trade to complete.
in:trade-process was:dropped
I made an offer to buy BSQ with BTC under the updated v1.0.1 client and my offer is currently erroring out any time someone accepts my offer, not allowing me to complete the trade. I noticed one other user in the community forum had posted about the same issue. Here is the log file from the most recent error in my offer: Apr-22 03:15:00.488 [JavaFX Application Thread] INFO b.c.o.OpenOfferManager: OfferAvailabilityResponse arrived at peer: offerId=FTMEB-57b4d1bc-2db5-436b-99c1-261af50278b2-101; uid=921c1323-d23a-4532-9d23-a00657bcff94 Apr-22 03:15:04.125 [JavaFX Application Thread] INFO b.c.t.TradeManager: Received PayDepositRequest from wx4he67epfl34npg.onion:9999 with tradeId FTMEB-57b4d1bc-2db5-436b-99c1-261af50278b2-101 and uid 1a1299ee-92a1-4246-9c3c-25d3e3006d18 Apr-22 03:15:04.126 [JavaFX Application Thread] INFO b.c.n.a.TradeEvents: We got a new trade. id=FTMEB-57b4d1bc-2db5-436b-99c1-261af50278b2-101 Apr-22 03:15:04.126 [JavaFX Application Thread] INFO b.c.t.TaskRunner: Run task: MakerProcessPayDepositRequest Apr-22 03:15:04.126 [JavaFX Application Thread] INFO b.c.offer.Offer: Price at take-offer time: id=FTMEB, currency=BSQ, takersPrice=22350, makersPrice=22350, deviation=0.0% Apr-22 03:15:04.126 [JavaFX Application Thread] INFO b.c.t.TaskRunner: Run task: CheckIfPeerIsBanned Apr-22 03:15:04.126 [JavaFX Application Thread] INFO b.c.t.TaskRunner: Run task: MakerVerifyTakerAccount Apr-22 03:15:04.126 [JavaFX Application Thread] INFO b.c.t.TaskRunner: Run task: VerifyPeersAccountAgeWitness Apr-22 03:15:04.126 [JavaFX Application Thread] INFO b.c.t.TaskRunner: Run task: MakerVerifyTakerFeePayment Apr-22 03:15:04.127 [JavaFX Application Thread] INFO b.c.t.TaskRunner: Run task: MakerCreateAndSignContract **Apr-22 03:15:04.129 [JavaFX Application Thread] ERROR b.c.t.Task: An error occurred at task: MakerCreateAndSignContract Exception message: makerPaymentAccountPayload must not be null Apr-22 03:15:04.129 [JavaFX Application Thread] ERROR b.c.t.TaskRunner: Task failed: MakerCreateAndSignContract / errorMessage: An error occurred at task: MakerCreateAndSignContract Exception message: makerPaymentAccountPayload must not be null** **Apr-22 03:15:04.139 [JavaFX Application Thread] ERROR b.c.t.p.TradeProtocol: An error occurred at task: MakerCreateAndSignContract Exception message: makerPaymentAccountPayload must not be null** Apr-22 03:15:04.139 [JavaFX Application Thread] INFO b.c.t.p.TradeProtocol: Send AckMessage for PayDepositRequest to peer wx4he67epfl34npg.onion:9999. tradeId=FTMEB-57b4d1bc-2db5-436b-99c1-261af50278b2-101, sourceUid=1a1299ee-92a1-4246-9c3c-25d3e3006d18 Apr-22 03:15:04.140 [JavaFX Application Thread] WARN b.n.p.P2PService: We don't have the peer in our persisted peers so we don't know his capabilities. We decide to not sent the msg. peersNodeAddress=wx4he67epfl34npg.onion:9999 Apr-22 03:15:04.141 [JavaFX Application Thread] ERROR b.c.t.p.TradeProtocol: AckMessage for PayDepositRequest failed. Peer wx4he67epfl34npg.onion:9999. tradeId=FTMEB-57b4d1bc-2db5-436b-99c1-261af50278b2-101, sourceUid=1a1299ee-92a1-4246-9c3c-25d3e3006d18, errorMessage=We did not send the EncryptedMailboxMessage because the peer does not support the capability. Apr-22 03:15:04.141 [JavaFX Application Thread] WARN b.c.t.p.TradeProtocol: cleanupTradableOnFault tradeState=PREPARATION Apr-22 03:15:13.101 [JavaFX Application Thread] INFO b.c.p.p.PriceFeedService: request from provider http://xc3nh4juf2hshy7e.onion/ 61.011 sec. after last request Apr-22 03:15:14.393 [JavaFX Application Thread] INFO b.c.p.p.PriceFeedService: Market price for currency BSQ is not provided by the provider http://xc3nh4juf2hshy7e.onion/. That is expected for currencies not listed at providers. Apr-22 03:15:14.393 [JavaFX Application Thread] INFO b.c.p.p.PriceFeedService: Received new MarketPrice(currencyCode=BSQ, price=2.65E-4, timestampSec=0, isExternallyProvidedPrice=false) from provider http://xc3nh4juf2hshy7e.onion/ after 1.291 sec.
1.0
Error at MakerCreateAndSignContract on v1.0.1 not allowing trade to complete. - I made an offer to buy BSQ with BTC under the updated v1.0.1 client and my offer is currently erroring out any time someone accepts my offer, not allowing me to complete the trade. I noticed one other user in the community forum had posted about the same issue. Here is the log file from the most recent error in my offer: Apr-22 03:15:00.488 [JavaFX Application Thread] INFO b.c.o.OpenOfferManager: OfferAvailabilityResponse arrived at peer: offerId=FTMEB-57b4d1bc-2db5-436b-99c1-261af50278b2-101; uid=921c1323-d23a-4532-9d23-a00657bcff94 Apr-22 03:15:04.125 [JavaFX Application Thread] INFO b.c.t.TradeManager: Received PayDepositRequest from wx4he67epfl34npg.onion:9999 with tradeId FTMEB-57b4d1bc-2db5-436b-99c1-261af50278b2-101 and uid 1a1299ee-92a1-4246-9c3c-25d3e3006d18 Apr-22 03:15:04.126 [JavaFX Application Thread] INFO b.c.n.a.TradeEvents: We got a new trade. id=FTMEB-57b4d1bc-2db5-436b-99c1-261af50278b2-101 Apr-22 03:15:04.126 [JavaFX Application Thread] INFO b.c.t.TaskRunner: Run task: MakerProcessPayDepositRequest Apr-22 03:15:04.126 [JavaFX Application Thread] INFO b.c.offer.Offer: Price at take-offer time: id=FTMEB, currency=BSQ, takersPrice=22350, makersPrice=22350, deviation=0.0% Apr-22 03:15:04.126 [JavaFX Application Thread] INFO b.c.t.TaskRunner: Run task: CheckIfPeerIsBanned Apr-22 03:15:04.126 [JavaFX Application Thread] INFO b.c.t.TaskRunner: Run task: MakerVerifyTakerAccount Apr-22 03:15:04.126 [JavaFX Application Thread] INFO b.c.t.TaskRunner: Run task: VerifyPeersAccountAgeWitness Apr-22 03:15:04.126 [JavaFX Application Thread] INFO b.c.t.TaskRunner: Run task: MakerVerifyTakerFeePayment Apr-22 03:15:04.127 [JavaFX Application Thread] INFO b.c.t.TaskRunner: Run task: MakerCreateAndSignContract **Apr-22 03:15:04.129 [JavaFX Application Thread] ERROR b.c.t.Task: An error occurred at task: MakerCreateAndSignContract Exception message: makerPaymentAccountPayload must not be null Apr-22 03:15:04.129 [JavaFX Application Thread] ERROR b.c.t.TaskRunner: Task failed: MakerCreateAndSignContract / errorMessage: An error occurred at task: MakerCreateAndSignContract Exception message: makerPaymentAccountPayload must not be null** **Apr-22 03:15:04.139 [JavaFX Application Thread] ERROR b.c.t.p.TradeProtocol: An error occurred at task: MakerCreateAndSignContract Exception message: makerPaymentAccountPayload must not be null** Apr-22 03:15:04.139 [JavaFX Application Thread] INFO b.c.t.p.TradeProtocol: Send AckMessage for PayDepositRequest to peer wx4he67epfl34npg.onion:9999. tradeId=FTMEB-57b4d1bc-2db5-436b-99c1-261af50278b2-101, sourceUid=1a1299ee-92a1-4246-9c3c-25d3e3006d18 Apr-22 03:15:04.140 [JavaFX Application Thread] WARN b.n.p.P2PService: We don't have the peer in our persisted peers so we don't know his capabilities. We decide to not sent the msg. peersNodeAddress=wx4he67epfl34npg.onion:9999 Apr-22 03:15:04.141 [JavaFX Application Thread] ERROR b.c.t.p.TradeProtocol: AckMessage for PayDepositRequest failed. Peer wx4he67epfl34npg.onion:9999. tradeId=FTMEB-57b4d1bc-2db5-436b-99c1-261af50278b2-101, sourceUid=1a1299ee-92a1-4246-9c3c-25d3e3006d18, errorMessage=We did not send the EncryptedMailboxMessage because the peer does not support the capability. Apr-22 03:15:04.141 [JavaFX Application Thread] WARN b.c.t.p.TradeProtocol: cleanupTradableOnFault tradeState=PREPARATION Apr-22 03:15:13.101 [JavaFX Application Thread] INFO b.c.p.p.PriceFeedService: request from provider http://xc3nh4juf2hshy7e.onion/ 61.011 sec. after last request Apr-22 03:15:14.393 [JavaFX Application Thread] INFO b.c.p.p.PriceFeedService: Market price for currency BSQ is not provided by the provider http://xc3nh4juf2hshy7e.onion/. That is expected for currencies not listed at providers. Apr-22 03:15:14.393 [JavaFX Application Thread] INFO b.c.p.p.PriceFeedService: Received new MarketPrice(currencyCode=BSQ, price=2.65E-4, timestampSec=0, isExternallyProvidedPrice=false) from provider http://xc3nh4juf2hshy7e.onion/ after 1.291 sec.
process
error at makercreateandsigncontract on not allowing trade to complete i made an offer to buy bsq with btc under the updated client and my offer is currently erroring out any time someone accepts my offer not allowing me to complete the trade i noticed one other user in the community forum had posted about the same issue here is the log file from the most recent error in my offer apr info b c o openoffermanager offeravailabilityresponse arrived at peer offerid ftmeb uid apr info b c t trademanager received paydepositrequest from onion with tradeid ftmeb and uid apr info b c n a tradeevents we got a new trade id ftmeb apr info b c t taskrunner run task makerprocesspaydepositrequest apr info b c offer offer price at take offer time id ftmeb currency bsq takersprice makersprice deviation apr info b c t taskrunner run task checkifpeerisbanned apr info b c t taskrunner run task makerverifytakeraccount apr info b c t taskrunner run task verifypeersaccountagewitness apr info b c t taskrunner run task makerverifytakerfeepayment apr info b c t taskrunner run task makercreateandsigncontract apr error b c t task an error occurred at task makercreateandsigncontract exception message makerpaymentaccountpayload must not be null apr error b c t taskrunner task failed makercreateandsigncontract errormessage an error occurred at task makercreateandsigncontract exception message makerpaymentaccountpayload must not be null apr error b c t p tradeprotocol an error occurred at task makercreateandsigncontract exception message makerpaymentaccountpayload must not be null apr info b c t p tradeprotocol send ackmessage for paydepositrequest to peer onion tradeid ftmeb sourceuid apr warn b n p we don t have the peer in our persisted peers so we don t know his capabilities we decide to not sent the msg peersnodeaddress onion apr error b c t p tradeprotocol ackmessage for paydepositrequest failed peer onion tradeid ftmeb sourceuid errormessage we did not send the encryptedmailboxmessage because the peer does not support the capability apr warn b c t p tradeprotocol cleanuptradableonfault tradestate preparation apr info b c p p pricefeedservice request from provider sec after last request apr info b c p p pricefeedservice market price for currency bsq is not provided by the provider that is expected for currencies not listed at providers apr info b c p p pricefeedservice received new marketprice currencycode bsq price timestampsec isexternallyprovidedprice false from provider after sec
1
113,657
17,150,459,572
IssuesEvent
2021-07-13 19:50:41
idonthaveafifaaddiction/serverless-webpack
https://api.github.com/repos/idonthaveafifaaddiction/serverless-webpack
opened
CVE-2020-28168 (Medium) detected in axios-0.19.2.tgz
security vulnerability
## CVE-2020-28168 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>axios-0.19.2.tgz</b></p></summary> <p>Promise based HTTP client for the browser and node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/axios/-/axios-0.19.2.tgz">https://registry.npmjs.org/axios/-/axios-0.19.2.tgz</a></p> <p>Path to dependency file: serverless-webpack/package.json</p> <p>Path to vulnerable library: serverless-webpack/node_modules/axios/package.json</p> <p> Dependency Hierarchy: - serverless-1.83.0.tgz (Root Library) - components-2.34.9.tgz - platform-client-1.1.10.tgz - :x: **axios-0.19.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/serverless-webpack/commit/8daea6a04d134bc7d69f7cca0d1a2a08967892e5">8daea6a04d134bc7d69f7cca0d1a2a08967892e5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Axios NPM package 0.21.0 contains a Server-Side Request Forgery (SSRF) vulnerability where an attacker is able to bypass a proxy by providing a URL that responds with a redirect to a restricted host or IP address. <p>Publish Date: 2020-11-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28168>CVE-2020-28168</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/axios/axios/commit/c7329fefc890050edd51e40e469a154d0117fc55">https://github.com/axios/axios/commit/c7329fefc890050edd51e40e469a154d0117fc55</a></p> <p>Release Date: 2020-11-06</p> <p>Fix Resolution: axios - 0.21.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"axios","packageVersion":"0.19.2","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"serverless:1.83.0;@serverless/components:2.34.9;@serverless/platform-client:1.1.10;axios:0.19.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"axios - 0.21.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28168","vulnerabilityDetails":"Axios NPM package 0.21.0 contains a Server-Side Request Forgery (SSRF) vulnerability where an attacker is able to bypass a proxy by providing a URL that responds with a redirect to a restricted host or IP address.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28168","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-28168 (Medium) detected in axios-0.19.2.tgz - ## CVE-2020-28168 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>axios-0.19.2.tgz</b></p></summary> <p>Promise based HTTP client for the browser and node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/axios/-/axios-0.19.2.tgz">https://registry.npmjs.org/axios/-/axios-0.19.2.tgz</a></p> <p>Path to dependency file: serverless-webpack/package.json</p> <p>Path to vulnerable library: serverless-webpack/node_modules/axios/package.json</p> <p> Dependency Hierarchy: - serverless-1.83.0.tgz (Root Library) - components-2.34.9.tgz - platform-client-1.1.10.tgz - :x: **axios-0.19.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/serverless-webpack/commit/8daea6a04d134bc7d69f7cca0d1a2a08967892e5">8daea6a04d134bc7d69f7cca0d1a2a08967892e5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Axios NPM package 0.21.0 contains a Server-Side Request Forgery (SSRF) vulnerability where an attacker is able to bypass a proxy by providing a URL that responds with a redirect to a restricted host or IP address. <p>Publish Date: 2020-11-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28168>CVE-2020-28168</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/axios/axios/commit/c7329fefc890050edd51e40e469a154d0117fc55">https://github.com/axios/axios/commit/c7329fefc890050edd51e40e469a154d0117fc55</a></p> <p>Release Date: 2020-11-06</p> <p>Fix Resolution: axios - 0.21.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"axios","packageVersion":"0.19.2","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"serverless:1.83.0;@serverless/components:2.34.9;@serverless/platform-client:1.1.10;axios:0.19.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"axios - 0.21.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28168","vulnerabilityDetails":"Axios NPM package 0.21.0 contains a Server-Side Request Forgery (SSRF) vulnerability where an attacker is able to bypass a proxy by providing a URL that responds with a redirect to a restricted host or IP address.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28168","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in axios tgz cve medium severity vulnerability vulnerable library axios tgz promise based http client for the browser and node js library home page a href path to dependency file serverless webpack package json path to vulnerable library serverless webpack node modules axios package json dependency hierarchy serverless tgz root library components tgz platform client tgz x axios tgz vulnerable library found in head commit a href found in base branch master vulnerability details axios npm package contains a server side request forgery ssrf vulnerability where an attacker is able to bypass a proxy by providing a url that responds with a redirect to a restricted host or ip address publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution axios isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree serverless serverless components serverless platform client axios isminimumfixversionavailable true minimumfixversion axios basebranches vulnerabilityidentifier cve vulnerabilitydetails axios npm package contains a server side request forgery ssrf vulnerability where an attacker is able to bypass a proxy by providing a url that responds with a redirect to a restricted host or ip address vulnerabilityurl
0
14,223
17,143,474,628
IssuesEvent
2021-07-13 12:18:00
gradle/gradle
https://api.github.com/repos/gradle/gradle
closed
Annotation processing when using kapt results in compileJava failing with "no source files"
a:bug in:annotation-processing
### Expected Behavior Compilation should succeed. ### Current Behavior ``` ./gradlew clean :sample-kotlin:assemble spits an error: * What went wrong: Execution failed for task ':sample-kotlin:compileJava'. > no source files ``` ### Context I am developing an annotation processor. ### Steps to Reproduce 1. Get https://github.com/stoyicker/test-accessors/tree/2b5d420b7b0ae5a036d69cd4f85186d221e376f0 2. Run `./gradlew clean :sample-kotlin:assemble` ### Your Environment Build scan URL: https://scans.gradle.com/s/o3642oxqlkj3k
1.0
Annotation processing when using kapt results in compileJava failing with "no source files" - ### Expected Behavior Compilation should succeed. ### Current Behavior ``` ./gradlew clean :sample-kotlin:assemble spits an error: * What went wrong: Execution failed for task ':sample-kotlin:compileJava'. > no source files ``` ### Context I am developing an annotation processor. ### Steps to Reproduce 1. Get https://github.com/stoyicker/test-accessors/tree/2b5d420b7b0ae5a036d69cd4f85186d221e376f0 2. Run `./gradlew clean :sample-kotlin:assemble` ### Your Environment Build scan URL: https://scans.gradle.com/s/o3642oxqlkj3k
process
annotation processing when using kapt results in compilejava failing with no source files expected behavior compilation should succeed current behavior gradlew clean sample kotlin assemble spits an error what went wrong execution failed for task sample kotlin compilejava no source files context i am developing an annotation processor steps to reproduce get run gradlew clean sample kotlin assemble your environment build scan url
1
70,028
9,369,081,604
IssuesEvent
2019-04-03 10:10:42
cozy/cozy-ui
https://api.github.com/repos/cozy/cozy-ui
closed
Organisation des demandes et développements
documentation
On a pris un peu de temps avec @enguerran et @GoOz pour discuter un peu plus en meta du problème de l'implémentation de l'Avatar. Les conclusions : * passer par les issues pour faire des demandes à cozy-ui permet de communiquer de manière impersonnelle ses besoins et d'avoir une discussion sur l'API * si besoin urgent, le composant peut être fait dans l'application avec possiblement l'aide de gooz concernant le css. -> tout de suite faire une issue sur cozy-ui si le composant a l'air d'être possiblement transposable sur ui. * on affinera avec gooz le backlog d'issue de manière hebdomadaire pour prioriser les demandes. @CPatchane on aura sûrement besoin de ton aide aussi :) Est ce que ca vous va ?
1.0
Organisation des demandes et développements - On a pris un peu de temps avec @enguerran et @GoOz pour discuter un peu plus en meta du problème de l'implémentation de l'Avatar. Les conclusions : * passer par les issues pour faire des demandes à cozy-ui permet de communiquer de manière impersonnelle ses besoins et d'avoir une discussion sur l'API * si besoin urgent, le composant peut être fait dans l'application avec possiblement l'aide de gooz concernant le css. -> tout de suite faire une issue sur cozy-ui si le composant a l'air d'être possiblement transposable sur ui. * on affinera avec gooz le backlog d'issue de manière hebdomadaire pour prioriser les demandes. @CPatchane on aura sûrement besoin de ton aide aussi :) Est ce que ca vous va ?
non_process
organisation des demandes et développements on a pris un peu de temps avec enguerran et gooz pour discuter un peu plus en meta du problème de l implémentation de l avatar les conclusions passer par les issues pour faire des demandes à cozy ui permet de communiquer de manière impersonnelle ses besoins et d avoir une discussion sur l api si besoin urgent le composant peut être fait dans l application avec possiblement l aide de gooz concernant le css tout de suite faire une issue sur cozy ui si le composant a l air d être possiblement transposable sur ui on affinera avec gooz le backlog d issue de manière hebdomadaire pour prioriser les demandes cpatchane on aura sûrement besoin de ton aide aussi est ce que ca vous va
0
368
2,813,483,399
IssuesEvent
2015-05-18 14:57:51
FG-Team/HCJ-Website-Builder
https://api.github.com/repos/FG-Team/HCJ-Website-Builder
closed
[DEPRECATED] xml schema for template files
Processing
Ich erstelle ein xml-Schema (xsd) für die Template-Dateien, die den Templatecode enthalten. Dieses werde ich nicht committen, sondern separat verteilen, da es keinen unmittelbaren Quellcode enthält. UPDATE 16.05.2015: Nicht notwendig, da wir eine Datenbank verwenden.
1.0
[DEPRECATED] xml schema for template files - Ich erstelle ein xml-Schema (xsd) für die Template-Dateien, die den Templatecode enthalten. Dieses werde ich nicht committen, sondern separat verteilen, da es keinen unmittelbaren Quellcode enthält. UPDATE 16.05.2015: Nicht notwendig, da wir eine Datenbank verwenden.
process
xml schema for template files ich erstelle ein xml schema xsd für die template dateien die den templatecode enthalten dieses werde ich nicht committen sondern separat verteilen da es keinen unmittelbaren quellcode enthält update nicht notwendig da wir eine datenbank verwenden
1
10,345
13,172,087,310
IssuesEvent
2020-08-11 17:48:46
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Support for RHEL 8
Pri2 automation/svc cxp process-automation/subsvc product-question triaged
[Enter feedback here] Is Linux RHEL 8 now supported for the Log Analytics Agent? I saw this conflicting on https://docs.microsoft.com/en-us/azure/azure-monitor/platform/log-analytics-agent#supported-distros Thanks --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: e38be5b8-d76d-a4f1-c014-7bf9248be2de * Version Independent ID: 976e5e90-b28c-d7ba-0495-69d92e62ea46 * Content: [Deploy a Linux Hybrid Runbook Worker in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/automation-linux-hrw-install#supported-linux-operating-systems) * Content Source: [articles/automation/automation-linux-hrw-install.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-linux-hrw-install.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
1.0
Support for RHEL 8 - [Enter feedback here] Is Linux RHEL 8 now supported for the Log Analytics Agent? I saw this conflicting on https://docs.microsoft.com/en-us/azure/azure-monitor/platform/log-analytics-agent#supported-distros Thanks --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: e38be5b8-d76d-a4f1-c014-7bf9248be2de * Version Independent ID: 976e5e90-b28c-d7ba-0495-69d92e62ea46 * Content: [Deploy a Linux Hybrid Runbook Worker in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/automation-linux-hrw-install#supported-linux-operating-systems) * Content Source: [articles/automation/automation-linux-hrw-install.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-linux-hrw-install.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
process
support for rhel is linux rhel now supported for the log analytics agent i saw this conflicting on thanks document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
1
331,560
24,312,691,019
IssuesEvent
2022-09-30 01:09:09
SHI-Labs/Neighborhood-Attention-Transformer
https://api.github.com/repos/SHI-Labs/Neighborhood-Attention-Transformer
closed
Where is the Neighborhood Attention 2D?
documentation
I wonder where is the Neighborhood Attention 2D. I can't find the code
1.0
Where is the Neighborhood Attention 2D? - I wonder where is the Neighborhood Attention 2D. I can't find the code
non_process
where is the neighborhood attention i wonder where is the neighborhood attention i can t find the code
0
16,539
21,564,857,323
IssuesEvent
2022-05-01 18:09:38
PollRobots/scheme
https://api.github.com/repos/PollRobots/scheme
closed
implement process-context procedure `command-line`
Runtime process-context
implement and test `command-line` from the process-context library
1.0
implement process-context procedure `command-line` - implement and test `command-line` from the process-context library
process
implement process context procedure command line implement and test command line from the process context library
1
11,329
14,143,515,762
IssuesEvent
2020-11-10 15:21:53
panther-labs/panther
https://api.github.com/repos/panther-labs/panther
closed
Migrate AWS.CloudTrail parser to new pantherlog
enhancement p0 story team:data processing
### Description Migrate AWS.CloudTrail parser to new pantherlog. This migration is going to provide significant performance (CPU & memory) improvements. ### Related Services panther-log-processor ### Designs Not relevant ### Acceptance Criteria - CloudTrail parser moved to use new parser factory.
1.0
Migrate AWS.CloudTrail parser to new pantherlog - ### Description Migrate AWS.CloudTrail parser to new pantherlog. This migration is going to provide significant performance (CPU & memory) improvements. ### Related Services panther-log-processor ### Designs Not relevant ### Acceptance Criteria - CloudTrail parser moved to use new parser factory.
process
migrate aws cloudtrail parser to new pantherlog description migrate aws cloudtrail parser to new pantherlog this migration is going to provide significant performance cpu memory improvements related services panther log processor designs not relevant acceptance criteria cloudtrail parser moved to use new parser factory
1
15,009
18,722,007,544
IssuesEvent
2021-11-03 12:52:27
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
opened
WASM Process Heating vs NAN
Process Heating WebAssembly
We've made a few changes to PH the last 2 sprints. Verify all of them have been changed in the ph api service. - Specifically for 618 change to excess air - need to send excessAirPerc as a fraction, instead of decimal
1.0
WASM Process Heating vs NAN - We've made a few changes to PH the last 2 sprints. Verify all of them have been changed in the ph api service. - Specifically for 618 change to excess air - need to send excessAirPerc as a fraction, instead of decimal
process
wasm process heating vs nan we ve made a few changes to ph the last sprints verify all of them have been changed in the ph api service specifically for change to excess air need to send excessairperc as a fraction instead of decimal
1
14,559
17,688,236,469
IssuesEvent
2021-08-24 06:30:36
jim-king-2000/IndustryCamera
https://api.github.com/repos/jim-king-2000/IndustryCamera
closed
[bug]: 区域管理 与UI设计不符
bug processing A
### 问题描述 - 1.区域名显示区域左右不呼应,有时只有左侧区域名称有、有时只有右侧区域名称有。 ![区域管理1](https://user-images.githubusercontent.com/89175659/130420328-43b47efc-5847-47b0-a060-a02fce365c8b.png) ![区域管理2](https://user-images.githubusercontent.com/89175659/130420335-32b0ca72-2e11-4c5d-a6f2-22102c6cc0bf.png) - 2.左侧区域名称不能选择,无选择框。 - 3.左侧区域名不能编辑。 - 4.左侧新建后,右侧没区域名显示,或者只显示右侧左侧没有显示。右侧区域名删除其中一个,右侧所有区域名消失。 ### 您预期的行为 - 参考UI设计 (编号与上述问题描述对应) - 1.左右区域名页面统一,右侧区域名部分根据左侧的选择来改变显示内容。进入页面默认左侧区域名里面没有勾选,右侧显示全部最小层级区域名。 - 2.可多层级勾选,例如在左侧勾选层级1和层级2.1则在右侧显示层级1下面的所有最小层级和层级2.1下面所有最小层级。 - 3.左侧区域名双击进行名称编辑。 - 4.新建、删除功能 能正常运行。 ### 系统表现的行为 同问题描述 ### 复现路径 每次进入 ### 辅助信息 - 浏览器版本:Chrome 92
1.0
[bug]: 区域管理 与UI设计不符 - ### 问题描述 - 1.区域名显示区域左右不呼应,有时只有左侧区域名称有、有时只有右侧区域名称有。 ![区域管理1](https://user-images.githubusercontent.com/89175659/130420328-43b47efc-5847-47b0-a060-a02fce365c8b.png) ![区域管理2](https://user-images.githubusercontent.com/89175659/130420335-32b0ca72-2e11-4c5d-a6f2-22102c6cc0bf.png) - 2.左侧区域名称不能选择,无选择框。 - 3.左侧区域名不能编辑。 - 4.左侧新建后,右侧没区域名显示,或者只显示右侧左侧没有显示。右侧区域名删除其中一个,右侧所有区域名消失。 ### 您预期的行为 - 参考UI设计 (编号与上述问题描述对应) - 1.左右区域名页面统一,右侧区域名部分根据左侧的选择来改变显示内容。进入页面默认左侧区域名里面没有勾选,右侧显示全部最小层级区域名。 - 2.可多层级勾选,例如在左侧勾选层级1和层级2.1则在右侧显示层级1下面的所有最小层级和层级2.1下面所有最小层级。 - 3.左侧区域名双击进行名称编辑。 - 4.新建、删除功能 能正常运行。 ### 系统表现的行为 同问题描述 ### 复现路径 每次进入 ### 辅助信息 - 浏览器版本:Chrome 92
process
区域管理 与ui设计不符 问题描述 区域名显示区域左右不呼应,有时只有左侧区域名称有、有时只有右侧区域名称有。 左侧区域名称不能选择,无选择框。 左侧区域名不能编辑。 左侧新建后,右侧没区域名显示,或者只显示右侧左侧没有显示。右侧区域名删除其中一个,右侧所有区域名消失。 您预期的行为 参考ui设计 (编号与上述问题描述对应) 左右区域名页面统一,右侧区域名部分根据左侧的选择来改变显示内容。进入页面默认左侧区域名里面没有勾选,右侧显示全部最小层级区域名。 可多层级勾选, 。 左侧区域名双击进行名称编辑。 新建、删除功能 能正常运行。 系统表现的行为 同问题描述 复现路径 每次进入 辅助信息 浏览器版本:chrome
1
395,568
11,688,672,034
IssuesEvent
2020-03-05 14:54:07
microsoft/PowerToys
https://api.github.com/repos/microsoft/PowerToys
closed
FancyZones gap on right edge of screen 0.15.1
Cost-Small FancyZones-Layouts Issue-Bug Priority-3 Product-FancyZones
![image](https://user-images.githubusercontent.com/57614985/75933657-6f9c0500-5e72-11ea-869a-a96797433505.png) ![image](https://user-images.githubusercontent.com/57614985/75933727-93f7e180-5e72-11ea-833a-0cb8365913d6.png) Since 0.15.0 the right most Zone on the template Zones no longer reaches the edge of the screen leaving a visible gap.
1.0
FancyZones gap on right edge of screen 0.15.1 - ![image](https://user-images.githubusercontent.com/57614985/75933657-6f9c0500-5e72-11ea-869a-a96797433505.png) ![image](https://user-images.githubusercontent.com/57614985/75933727-93f7e180-5e72-11ea-833a-0cb8365913d6.png) Since 0.15.0 the right most Zone on the template Zones no longer reaches the edge of the screen leaving a visible gap.
non_process
fancyzones gap on right edge of screen since the right most zone on the template zones no longer reaches the edge of the screen leaving a visible gap
0
197,684
14,939,411,094
IssuesEvent
2021-01-25 16:54:56
phetsims/gravity-force-lab-basics
https://api.github.com/repos/phetsims/gravity-force-lab-basics
closed
CT cannot read property copy of null
type:automated-testing
``` gravity-force-lab-basics : phet-io-state-fuzz : unbuilt https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/phet-io-wrappers/state/?sim=gravity-force-lab-basics&phetioDebug&fuzz&wrapperContinuousTest=%7B%22test%22%3A%5B%22gravity-force-lab-basics%22%2C%22phet-io-state-fuzz%22%2C%22unbuilt%22%5D%2C%22snapshotName%22%3A%22snapshot-1608697201899%22%2C%22timestamp%22%3A1608699841342%7D Uncaught Error: Cannot read property 'copy' of null TypeError: Cannot read property 'copy' of null at GFLBScreenView.globalToLocalPoint (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/nodes/Node.js:5210:31) at ShapeHitDetector.exit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/tappi/js/view/ShapeHitDetector.js:162:39) at Input.dispatchToListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1833:25) at Input.dispatchEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1794:10) at Input.exitEvents (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1761:14) at Input.removeTemporaryPointers (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:882:14) at Display.set interactive [as interactive] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/display/Display.js:773:19) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/joist/js/Sim.js:600:32 at TinyEmitter.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/axon/js/TinyEmitter.js:71:18) at BooleanProperty._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/axon/js/Property.js:280:25) id: Bayes Chrome Snapshot from 12/22/2020, 9:20:01 PM ---------------------------------- gravity-force-lab-basics : phet-io-state-fuzz : unbuilt https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/phet-io-wrappers/state/?sim=gravity-force-lab-basics&phetioDebug&fuzz&wrapperContinuousTest=%7B%22test%22%3A%5B%22gravity-force-lab-basics%22%2C%22phet-io-state-fuzz%22%2C%22unbuilt%22%5D%2C%22snapshotName%22%3A%22snapshot-1608697201899%22%2C%22timestamp%22%3A1608716083642%7D Uncaught Error: Cannot read property 'copy' of null TypeError: Cannot read property 'copy' of null at GFLBScreenView.globalToLocalPoint (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/nodes/Node.js:5210:31) at ShapeHitDetector.exit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/tappi/js/view/ShapeHitDetector.js:162:39) at Input.dispatchToListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1833:25) at Input.dispatchEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1794:10) at Input.exitEvents (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1761:14) at Input.removeTemporaryPointers (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:882:14) at Display.set interactive [as interactive] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/display/Display.js:773:19) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/joist/js/Sim.js:600:32 at TinyEmitter.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/axon/js/TinyEmitter.js:71:18) at BooleanProperty._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/axon/js/Property.js:280:25) id: Bayes Chrome Snapshot from 12/22/2020, 9:20:01 PM ---------------------------------- gravity-force-lab-basics : phet-io-wrappers-tests : assert https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/phet-io-wrappers/phet-io-wrappers-tests.html?sim=gravity-force-lab-basics&phetioDebug 8 out of 9 tests passed. 1 failed. SimTests: gravity-force-lab-basics: iframe api failed: Uncaught Error: Cannot read property 'copy' of null TypeError: Cannot read property 'copy' of null at GFLBScreenView.globalToLocalPoint (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/nodes/Node.js:5210:31) at ShapeHitDetector.exit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/tappi/js/view/ShapeHitDetector.js:162:39) at Input.dispatchToListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1833:25) at Input.dispatchEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1794:10) at Input.exitEvents (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1761:14) at Input.removeTemporaryPointers (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:882:14) at Display.set interactive [as interactive] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/display/Display.js:773:19) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/joist/js/Sim.js:600:32 at TinyEmitter.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/axon/js/TinyEmitter.js:71:18) at BooleanProperty._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/axon/js/Property.js:280:25) SimTests: gravity-force-lab-basics: iframe api failed: Uncaught Error: Uncaught Error: Assertion failed: mismatched index, possible start/end mismatch. Likely this is downstream of another error, try pausing on caught exceptions. Error: Assertion failed: mismatched index, possible start/end mismatch. Likely this is downstream of another error, try pausing on caught exceptions. at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/assert/js/assert.js:25:13) at DataStream.end (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/phet-io/js/dataStream.js:281:15) at PhetioCommandProcessor.phetioEndEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/tandem/js/PhetioObject.js:407:30) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/phet-io/js/phetioCommandProcessor.js:121:18 id: Bayes Chrome Snapshot from 12/22/2020, 9:20:01 PM ---------------------------------- gravity-force-lab-basics : phet-io-wrappers-tests : no-assert https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/phet-io-wrappers/phet-io-wrappers-tests.html?sim=gravity-force-lab-basics 8 out of 9 tests passed. 1 failed. SimTests: gravity-force-lab-basics: iframe api failed: Uncaught Error: Cannot read property 'copy' of null TypeError: Cannot read property 'copy' of null at GFLBScreenView.globalToLocalPoint (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/nodes/Node.js:5210:31) at ShapeHitDetector.exit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/tappi/js/view/ShapeHitDetector.js:162:39) at Input.dispatchToListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1833:25) at Input.dispatchEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1794:10) at Input.exitEvents (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1761:14) at Input.removeTemporaryPointers (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:882:14) at Display.set interactive [as interactive] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/display/Display.js:773:19) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/joist/js/Sim.js:600:32 at TinyEmitter.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/axon/js/TinyEmitter.js:71:18) at BooleanProperty._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/axon/js/Property.js:280:25) SimTests: gravity-force-lab-basics: iframe api failed: Uncaught Error: Uncaught Error: Assertion failed: mismatched index, possible start/end mismatch. Likely this is downstream of another error, try pausing on caught exceptions. Error: Assertion failed: mismatched index, possible start/end mismatch. Likely this is downstream of another error, try pausing on caught exceptions. at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/assert/js/assert.js:25:13) at DataStream.end (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/phet-io/js/dataStream.js:281:15) at PhetioCommandProcessor.phetioEndEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/tandem/js/PhetioObject.js:407:30) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/phet-io/js/phetioCommandProcessor.js:121:18 id: Bayes Chrome Snapshot from 12/22/2020, 9:20:01 PM ```
1.0
CT cannot read property copy of null - ``` gravity-force-lab-basics : phet-io-state-fuzz : unbuilt https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/phet-io-wrappers/state/?sim=gravity-force-lab-basics&phetioDebug&fuzz&wrapperContinuousTest=%7B%22test%22%3A%5B%22gravity-force-lab-basics%22%2C%22phet-io-state-fuzz%22%2C%22unbuilt%22%5D%2C%22snapshotName%22%3A%22snapshot-1608697201899%22%2C%22timestamp%22%3A1608699841342%7D Uncaught Error: Cannot read property 'copy' of null TypeError: Cannot read property 'copy' of null at GFLBScreenView.globalToLocalPoint (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/nodes/Node.js:5210:31) at ShapeHitDetector.exit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/tappi/js/view/ShapeHitDetector.js:162:39) at Input.dispatchToListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1833:25) at Input.dispatchEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1794:10) at Input.exitEvents (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1761:14) at Input.removeTemporaryPointers (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:882:14) at Display.set interactive [as interactive] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/display/Display.js:773:19) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/joist/js/Sim.js:600:32 at TinyEmitter.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/axon/js/TinyEmitter.js:71:18) at BooleanProperty._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/axon/js/Property.js:280:25) id: Bayes Chrome Snapshot from 12/22/2020, 9:20:01 PM ---------------------------------- gravity-force-lab-basics : phet-io-state-fuzz : unbuilt https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/phet-io-wrappers/state/?sim=gravity-force-lab-basics&phetioDebug&fuzz&wrapperContinuousTest=%7B%22test%22%3A%5B%22gravity-force-lab-basics%22%2C%22phet-io-state-fuzz%22%2C%22unbuilt%22%5D%2C%22snapshotName%22%3A%22snapshot-1608697201899%22%2C%22timestamp%22%3A1608716083642%7D Uncaught Error: Cannot read property 'copy' of null TypeError: Cannot read property 'copy' of null at GFLBScreenView.globalToLocalPoint (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/nodes/Node.js:5210:31) at ShapeHitDetector.exit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/tappi/js/view/ShapeHitDetector.js:162:39) at Input.dispatchToListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1833:25) at Input.dispatchEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1794:10) at Input.exitEvents (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1761:14) at Input.removeTemporaryPointers (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:882:14) at Display.set interactive [as interactive] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/display/Display.js:773:19) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/joist/js/Sim.js:600:32 at TinyEmitter.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/axon/js/TinyEmitter.js:71:18) at BooleanProperty._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/axon/js/Property.js:280:25) id: Bayes Chrome Snapshot from 12/22/2020, 9:20:01 PM ---------------------------------- gravity-force-lab-basics : phet-io-wrappers-tests : assert https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/phet-io-wrappers/phet-io-wrappers-tests.html?sim=gravity-force-lab-basics&phetioDebug 8 out of 9 tests passed. 1 failed. SimTests: gravity-force-lab-basics: iframe api failed: Uncaught Error: Cannot read property 'copy' of null TypeError: Cannot read property 'copy' of null at GFLBScreenView.globalToLocalPoint (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/nodes/Node.js:5210:31) at ShapeHitDetector.exit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/tappi/js/view/ShapeHitDetector.js:162:39) at Input.dispatchToListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1833:25) at Input.dispatchEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1794:10) at Input.exitEvents (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1761:14) at Input.removeTemporaryPointers (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:882:14) at Display.set interactive [as interactive] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/display/Display.js:773:19) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/joist/js/Sim.js:600:32 at TinyEmitter.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/axon/js/TinyEmitter.js:71:18) at BooleanProperty._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/axon/js/Property.js:280:25) SimTests: gravity-force-lab-basics: iframe api failed: Uncaught Error: Uncaught Error: Assertion failed: mismatched index, possible start/end mismatch. Likely this is downstream of another error, try pausing on caught exceptions. Error: Assertion failed: mismatched index, possible start/end mismatch. Likely this is downstream of another error, try pausing on caught exceptions. at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/assert/js/assert.js:25:13) at DataStream.end (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/phet-io/js/dataStream.js:281:15) at PhetioCommandProcessor.phetioEndEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/tandem/js/PhetioObject.js:407:30) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/phet-io/js/phetioCommandProcessor.js:121:18 id: Bayes Chrome Snapshot from 12/22/2020, 9:20:01 PM ---------------------------------- gravity-force-lab-basics : phet-io-wrappers-tests : no-assert https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/phet-io-wrappers/phet-io-wrappers-tests.html?sim=gravity-force-lab-basics 8 out of 9 tests passed. 1 failed. SimTests: gravity-force-lab-basics: iframe api failed: Uncaught Error: Cannot read property 'copy' of null TypeError: Cannot read property 'copy' of null at GFLBScreenView.globalToLocalPoint (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/nodes/Node.js:5210:31) at ShapeHitDetector.exit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/tappi/js/view/ShapeHitDetector.js:162:39) at Input.dispatchToListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1833:25) at Input.dispatchEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1794:10) at Input.exitEvents (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:1761:14) at Input.removeTemporaryPointers (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/input/Input.js:882:14) at Display.set interactive [as interactive] (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/scenery/js/display/Display.js:773:19) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/joist/js/Sim.js:600:32 at TinyEmitter.emit (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/axon/js/TinyEmitter.js:71:18) at BooleanProperty._notifyListeners (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/axon/js/Property.js:280:25) SimTests: gravity-force-lab-basics: iframe api failed: Uncaught Error: Uncaught Error: Assertion failed: mismatched index, possible start/end mismatch. Likely this is downstream of another error, try pausing on caught exceptions. Error: Assertion failed: mismatched index, possible start/end mismatch. Likely this is downstream of another error, try pausing on caught exceptions. at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/assert/js/assert.js:25:13) at DataStream.end (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/phet-io/js/dataStream.js:281:15) at PhetioCommandProcessor.phetioEndEvent (https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/tandem/js/PhetioObject.js:407:30) at https://bayes.colorado.edu/continuous-testing/ct-snapshots/1608697201899/phet-io/js/phetioCommandProcessor.js:121:18 id: Bayes Chrome Snapshot from 12/22/2020, 9:20:01 PM ```
non_process
ct cannot read property copy of null gravity force lab basics phet io state fuzz unbuilt uncaught error cannot read property copy of null typeerror cannot read property copy of null at gflbscreenview globaltolocalpoint at shapehitdetector exit at input dispatchtolisteners at input dispatchevent at input exitevents at input removetemporarypointers at display set interactive at at tinyemitter emit at booleanproperty notifylisteners id bayes chrome snapshot from pm gravity force lab basics phet io state fuzz unbuilt uncaught error cannot read property copy of null typeerror cannot read property copy of null at gflbscreenview globaltolocalpoint at shapehitdetector exit at input dispatchtolisteners at input dispatchevent at input exitevents at input removetemporarypointers at display set interactive at at tinyemitter emit at booleanproperty notifylisteners id bayes chrome snapshot from pm gravity force lab basics phet io wrappers tests assert out of tests passed failed simtests gravity force lab basics iframe api failed uncaught error cannot read property copy of null typeerror cannot read property copy of null at gflbscreenview globaltolocalpoint at shapehitdetector exit at input dispatchtolisteners at input dispatchevent at input exitevents at input removetemporarypointers at display set interactive at at tinyemitter emit at booleanproperty notifylisteners simtests gravity force lab basics iframe api failed uncaught error uncaught error assertion failed mismatched index possible start end mismatch likely this is downstream of another error try pausing on caught exceptions error assertion failed mismatched index possible start end mismatch likely this is downstream of another error try pausing on caught exceptions at window assertions assertfunction at datastream end at phetiocommandprocessor phetioendevent at id bayes chrome snapshot from pm gravity force lab basics phet io wrappers tests no assert out of tests passed failed simtests gravity force lab basics iframe api failed uncaught error cannot read property copy of null typeerror cannot read property copy of null at gflbscreenview globaltolocalpoint at shapehitdetector exit at input dispatchtolisteners at input dispatchevent at input exitevents at input removetemporarypointers at display set interactive at at tinyemitter emit at booleanproperty notifylisteners simtests gravity force lab basics iframe api failed uncaught error uncaught error assertion failed mismatched index possible start end mismatch likely this is downstream of another error try pausing on caught exceptions error assertion failed mismatched index possible start end mismatch likely this is downstream of another error try pausing on caught exceptions at window assertions assertfunction at datastream end at phetiocommandprocessor phetioendevent at id bayes chrome snapshot from pm
0
15,673
19,847,400,241
IssuesEvent
2022-01-21 08:25:34
ooi-data/RS01SBPS-PC01A-06-VADCPA101-streamed-vadcp_5thbeam_pd0_beam_parsed
https://api.github.com/repos/ooi-data/RS01SBPS-PC01A-06-VADCPA101-streamed-vadcp_5thbeam_pd0_beam_parsed
opened
🛑 Processing failed: KeyError
process
## Overview `KeyError` found in `processing_task` task during run ended on 2022-01-21T08:25:34.189869. ## Details Flow name: `RS01SBPS-PC01A-06-VADCPA101-streamed-vadcp_5thbeam_pd0_beam_parsed` Task name: `processing_task` Error type: `KeyError` Error message: 'vadcp_beam_error_dim_0' <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/dataset.py", line 1395, in _construct_dataarray variable = self._variables[name] KeyError: 'vadcp_beam_error_dim_0' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 344, in append_to_zarr mod_ds = _prepare_ds_to_append(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 133, in _prepare_ds_to_append existing_shape = tuple( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 134, in <genexpr> ds_to_append[dim].shape[0] for dim, size in new_var.sizes.items() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/dataset.py", line 1499, in __getitem__ return self._construct_dataarray(key) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/dataset.py", line 1397, in _construct_dataarray _, name, variable = _get_virtual_variable( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/dataset.py", line 170, in _get_virtual_variable ref_var = variables[ref_name] KeyError: 'vadcp_beam_error_dim_0' ``` </details>
1.0
🛑 Processing failed: KeyError - ## Overview `KeyError` found in `processing_task` task during run ended on 2022-01-21T08:25:34.189869. ## Details Flow name: `RS01SBPS-PC01A-06-VADCPA101-streamed-vadcp_5thbeam_pd0_beam_parsed` Task name: `processing_task` Error type: `KeyError` Error message: 'vadcp_beam_error_dim_0' <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/dataset.py", line 1395, in _construct_dataarray variable = self._variables[name] KeyError: 'vadcp_beam_error_dim_0' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 344, in append_to_zarr mod_ds = _prepare_ds_to_append(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 133, in _prepare_ds_to_append existing_shape = tuple( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 134, in <genexpr> ds_to_append[dim].shape[0] for dim, size in new_var.sizes.items() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/dataset.py", line 1499, in __getitem__ return self._construct_dataarray(key) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/dataset.py", line 1397, in _construct_dataarray _, name, variable = _get_virtual_variable( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/dataset.py", line 170, in _get_virtual_variable ref_var = variables[ref_name] KeyError: 'vadcp_beam_error_dim_0' ``` </details>
process
🛑 processing failed keyerror overview keyerror found in processing task task during run ended on details flow name streamed vadcp beam parsed task name processing task error type keyerror error message vadcp beam error dim traceback traceback most recent call last file srv conda envs notebook lib site packages xarray core dataset py line in construct dataarray variable self variables keyerror vadcp beam error dim during handling of the above exception another exception occurred traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr mod ds prepare ds to append store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in prepare ds to append existing shape tuple file srv conda envs notebook lib site packages ooi harvester processor utils py line in ds to append shape for dim size in new var sizes items file srv conda envs notebook lib site packages xarray core dataset py line in getitem return self construct dataarray key file srv conda envs notebook lib site packages xarray core dataset py line in construct dataarray name variable get virtual variable file srv conda envs notebook lib site packages xarray core dataset py line in get virtual variable ref var variables keyerror vadcp beam error dim
1
77,240
9,552,430,037
IssuesEvent
2019-05-02 16:37:14
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
opened
Use a single name for scopes & filters consistently
design feature-request
I'm opening this ticket because I think new users of Sourcegraph might get confused by the inconsistent naming of the feature called "scopes" or "filters". Example: on `/search` it's called "scopes" (`Edit search scopes`, `Add search scopes for quick filtering`, ...). In the "User settings" too. But when looking at search results, the scopes show up next to `Filters:`. I know that scopes technically are filters and the other way around, but from a new user's/admin's perspective it would be great if there was just one name for the same thing everywhere. If these two *are* distinct features, never mind, we shouldn't use a separate name for them. But maybe we should make that distinction clearer.
1.0
Use a single name for scopes & filters consistently - I'm opening this ticket because I think new users of Sourcegraph might get confused by the inconsistent naming of the feature called "scopes" or "filters". Example: on `/search` it's called "scopes" (`Edit search scopes`, `Add search scopes for quick filtering`, ...). In the "User settings" too. But when looking at search results, the scopes show up next to `Filters:`. I know that scopes technically are filters and the other way around, but from a new user's/admin's perspective it would be great if there was just one name for the same thing everywhere. If these two *are* distinct features, never mind, we shouldn't use a separate name for them. But maybe we should make that distinction clearer.
non_process
use a single name for scopes filters consistently i m opening this ticket because i think new users of sourcegraph might get confused by the inconsistent naming of the feature called scopes or filters example on search it s called scopes edit search scopes add search scopes for quick filtering in the user settings too but when looking at search results the scopes show up next to filters i know that scopes technically are filters and the other way around but from a new user s admin s perspective it would be great if there was just one name for the same thing everywhere if these two are distinct features never mind we shouldn t use a separate name for them but maybe we should make that distinction clearer
0
11,294
14,101,176,867
IssuesEvent
2020-11-06 06:15:06
pingcap/tidb
https://api.github.com/repos/pingcap/tidb
closed
coprocessor layer may encounter deadlock due to OOM
component/coprocessor severity/major type/bug
## Bug Report Please answer these questions before submitting your issue. Thanks! Currently coprocessor process may encounter deadlock in some cases. Here is one example to meet this error: We can simulate other executors consuming during the coprocessor process which cause the OOM Action. If the response channel was empty at that time, the workers would all get stucked in `waitIfNeeded` function which will cause the deadlock. ```golang // run is a worker function that get a copTask from channel, handle it and // send the result back. func (worker *copIteratorWorker) run(ctx context.Context) { defer worker.wg.Done() for task := range worker.taskCh { respCh := worker.respChan if respCh == nil { respCh = task.respChan } worker.handleTask(ctx, task, respCh) close(task.respChan) worker.maxID.setMaxIDIfLarger(task.id) worker.actionOnExceed.destroyTokenIfNeeded(func() { worker.sendRate.putToken() }) worker.memTracker.Consume(1000) // Simulating other executor memory consuming worker.actionOnExceed.waitIfNeeded() worker.memTracker.Consume(-1000) if worker.vars != nil && worker.vars.Killed != nil && atomic.LoadUint32(worker.vars.Killed) == 1 { return } select { case <-worker.finishCh: return default: } } } ``` ### 1. Minimal reproduce step (Required) <!-- a step by step guide for reproducing the bug. --> ### 2. What did you expect to see? (Required) ### 3. What did you see instead (Required) ### 4. What is your TiDB version? (Required) <!-- Paste the output of SELECT tidb_version() -->
1.0
coprocessor layer may encounter deadlock due to OOM - ## Bug Report Please answer these questions before submitting your issue. Thanks! Currently coprocessor process may encounter deadlock in some cases. Here is one example to meet this error: We can simulate other executors consuming during the coprocessor process which cause the OOM Action. If the response channel was empty at that time, the workers would all get stucked in `waitIfNeeded` function which will cause the deadlock. ```golang // run is a worker function that get a copTask from channel, handle it and // send the result back. func (worker *copIteratorWorker) run(ctx context.Context) { defer worker.wg.Done() for task := range worker.taskCh { respCh := worker.respChan if respCh == nil { respCh = task.respChan } worker.handleTask(ctx, task, respCh) close(task.respChan) worker.maxID.setMaxIDIfLarger(task.id) worker.actionOnExceed.destroyTokenIfNeeded(func() { worker.sendRate.putToken() }) worker.memTracker.Consume(1000) // Simulating other executor memory consuming worker.actionOnExceed.waitIfNeeded() worker.memTracker.Consume(-1000) if worker.vars != nil && worker.vars.Killed != nil && atomic.LoadUint32(worker.vars.Killed) == 1 { return } select { case <-worker.finishCh: return default: } } } ``` ### 1. Minimal reproduce step (Required) <!-- a step by step guide for reproducing the bug. --> ### 2. What did you expect to see? (Required) ### 3. What did you see instead (Required) ### 4. What is your TiDB version? (Required) <!-- Paste the output of SELECT tidb_version() -->
process
coprocessor layer may encounter deadlock due to oom bug report please answer these questions before submitting your issue thanks currently coprocessor process may encounter deadlock in some cases here is one example to meet this error we can simulate other executors consuming during the coprocessor process which cause the oom action if the response channel was empty at that time the workers would all get stucked in waitifneeded function which will cause the deadlock golang run is a worker function that get a coptask from channel handle it and send the result back func worker copiteratorworker run ctx context context defer worker wg done for task range worker taskch respch worker respchan if respch nil respch task respchan worker handletask ctx task respch close task respchan worker maxid setmaxidiflarger task id worker actiononexceed destroytokenifneeded func worker sendrate puttoken worker memtracker consume simulating other executor memory consuming worker actiononexceed waitifneeded worker memtracker consume if worker vars nil worker vars killed nil atomic worker vars killed return select case worker finishch return default minimal reproduce step required what did you expect to see required what did you see instead required what is your tidb version required
1
2,276
5,104,534,132
IssuesEvent
2017-01-05 01:44:52
AffiliateWP/AffiliateWP
https://api.github.com/repos/AffiliateWP/AffiliateWP
closed
Referrals: Add batch processor
batch-processing enhancement
**Issue**: Let's say there is a site that gets several thousand referrals per day. If an affiliate were to request a payout, and the referrals are being paid manually, the site admin would need to click `mark as paid` for all referrals of that specific affiliate. Of course, one can edit the "screen option" on the referrals page, and edit "Number of referrals per page" to 999, and select all and mark 999 referrals as paid. However, an error is thrown when attempting to mark all 999 referrals as paid, because the URL query becomes so long. About 500 referrals can be marked as paid at the same time before the process times out. **Solution**: Let's add batch processing for referrals of a specific affiliate and also for all referrals.
1.0
Referrals: Add batch processor - **Issue**: Let's say there is a site that gets several thousand referrals per day. If an affiliate were to request a payout, and the referrals are being paid manually, the site admin would need to click `mark as paid` for all referrals of that specific affiliate. Of course, one can edit the "screen option" on the referrals page, and edit "Number of referrals per page" to 999, and select all and mark 999 referrals as paid. However, an error is thrown when attempting to mark all 999 referrals as paid, because the URL query becomes so long. About 500 referrals can be marked as paid at the same time before the process times out. **Solution**: Let's add batch processing for referrals of a specific affiliate and also for all referrals.
process
referrals add batch processor issue let s say there is a site that gets several thousand referrals per day if an affiliate were to request a payout and the referrals are being paid manually the site admin would need to click mark as paid for all referrals of that specific affiliate of course one can edit the screen option on the referrals page and edit number of referrals per page to and select all and mark referrals as paid however an error is thrown when attempting to mark all referrals as paid because the url query becomes so long about referrals can be marked as paid at the same time before the process times out solution let s add batch processing for referrals of a specific affiliate and also for all referrals
1
8,021
11,207,454,550
IssuesEvent
2020-01-06 03:42:18
CymChad/BaseRecyclerViewAdapterHelper
https://api.github.com/repos/CymChad/BaseRecyclerViewAdapterHelper
closed
BaseViewHolder 的 setXxx 方法有的返回 BaseViewHolder 有的返回 BaseViewHolder?
processing
![image](https://user-images.githubusercontent.com/19941667/71714529-37b01b80-2e49-11ea-99f2-65a11e94fdd9.png) 既然都是返回的 this ,这样做是不是有点不太合理 版本 3.0.0-bate6
1.0
BaseViewHolder 的 setXxx 方法有的返回 BaseViewHolder 有的返回 BaseViewHolder? - ![image](https://user-images.githubusercontent.com/19941667/71714529-37b01b80-2e49-11ea-99f2-65a11e94fdd9.png) 既然都是返回的 this ,这样做是不是有点不太合理 版本 3.0.0-bate6
process
baseviewholder 的 setxxx 方法有的返回 baseviewholder 有的返回 baseviewholder 既然都是返回的 this ,这样做是不是有点不太合理 版本
1
223,142
17,104,908,022
IssuesEvent
2021-07-09 16:11:25
apache/trafficserver
https://api.github.com/repos/apache/trafficserver
opened
Add URI Signing Docs
Documentation
The URI Signing plugin has no docs. We should fix that. It's still in /experimental, but URI Signing is the new version that's being standardized by the IETF. URL Sig, which has docs, is never going to be standardized. We have new users using URL Sig because they don't know better, adding URI Signing docs will help that. We should also add a link in the URL Sig docs pointing to the URI Signing docs (when they exist), and recommending it.
1.0
Add URI Signing Docs - The URI Signing plugin has no docs. We should fix that. It's still in /experimental, but URI Signing is the new version that's being standardized by the IETF. URL Sig, which has docs, is never going to be standardized. We have new users using URL Sig because they don't know better, adding URI Signing docs will help that. We should also add a link in the URL Sig docs pointing to the URI Signing docs (when they exist), and recommending it.
non_process
add uri signing docs the uri signing plugin has no docs we should fix that it s still in experimental but uri signing is the new version that s being standardized by the ietf url sig which has docs is never going to be standardized we have new users using url sig because they don t know better adding uri signing docs will help that we should also add a link in the url sig docs pointing to the uri signing docs when they exist and recommending it
0
688
3,172,311,327
IssuesEvent
2015-09-23 07:17:01
tomchristie/django-rest-framework
https://api.github.com/repos/tomchristie/django-rest-framework
closed
Issue against Django `master`, especially Django Guardian
Process
Running against Django `master` there are a **lot** of issues. ``` 1 warnings, 265 error in 137.92 seconds ``` Most of these are caused by Django Guardian. Excluding that from the requirements leads to something more reasonable: ``` 5 failed, 239 passed, 1 warnings, 21 error in 11.27 seconds ``` Only two types of failure: ``` tests/test_utils.py:153: in setUp self.get_model = rest_framework.utils.model_meta.models.get_model E AttributeError: module 'django.db.models' has no attribute 'get_model' ... tests/browsable_api/test_browsable_api.py:36: in test_login_shown_when_logged_out self.assertContains(response, '>Log in<') .tox/py35-djangomaster/lib/python3.5/site-packages/django/test/testcases.py:398: in assertContains msg_prefix + "Couldn't find %s in response" % text_repr) E AssertionError: False is not true : Couldn't find '>Log in<' in response ``` Errors are all these: ``` .tox/py35-djangomaster/lib/python3.5/site-packages/_pytest/assertion/rewrite.py:682: ... E TypeError: Call constructor takes either 0 or 3 positional arguments ... tests/test_filters.py:11: in <module> from django.utils import unittest E ImportError: cannot import name 'unittest' ``` I don't think there's much work there but it's masked by the Guardian issues.
1.0
Issue against Django `master`, especially Django Guardian - Running against Django `master` there are a **lot** of issues. ``` 1 warnings, 265 error in 137.92 seconds ``` Most of these are caused by Django Guardian. Excluding that from the requirements leads to something more reasonable: ``` 5 failed, 239 passed, 1 warnings, 21 error in 11.27 seconds ``` Only two types of failure: ``` tests/test_utils.py:153: in setUp self.get_model = rest_framework.utils.model_meta.models.get_model E AttributeError: module 'django.db.models' has no attribute 'get_model' ... tests/browsable_api/test_browsable_api.py:36: in test_login_shown_when_logged_out self.assertContains(response, '>Log in<') .tox/py35-djangomaster/lib/python3.5/site-packages/django/test/testcases.py:398: in assertContains msg_prefix + "Couldn't find %s in response" % text_repr) E AssertionError: False is not true : Couldn't find '>Log in<' in response ``` Errors are all these: ``` .tox/py35-djangomaster/lib/python3.5/site-packages/_pytest/assertion/rewrite.py:682: ... E TypeError: Call constructor takes either 0 or 3 positional arguments ... tests/test_filters.py:11: in <module> from django.utils import unittest E ImportError: cannot import name 'unittest' ``` I don't think there's much work there but it's masked by the Guardian issues.
process
issue against django master especially django guardian running against django master there are a lot of issues warnings error in seconds most of these are caused by django guardian excluding that from the requirements leads to something more reasonable failed passed warnings error in seconds only two types of failure tests test utils py in setup self get model rest framework utils model meta models get model e attributeerror module django db models has no attribute get model tests browsable api test browsable api py in test login shown when logged out self assertcontains response log in tox djangomaster lib site packages django test testcases py in assertcontains msg prefix couldn t find s in response text repr e assertionerror false is not true couldn t find log in in response errors are all these tox djangomaster lib site packages pytest assertion rewrite py e typeerror call constructor takes either or positional arguments tests test filters py in from django utils import unittest e importerror cannot import name unittest i don t think there s much work there but it s masked by the guardian issues
1
18,834
24,738,135,579
IssuesEvent
2022-10-21 00:57:02
fertadeo/ISPC-2do-Cuat-Proyecto
https://api.github.com/repos/fertadeo/ISPC-2do-Cuat-Proyecto
closed
#TK 14.0 Diseño de footer - Administración
in process
Crear estructura HTML del footer, definir colores y tipografía
1.0
#TK 14.0 Diseño de footer - Administración - Crear estructura HTML del footer, definir colores y tipografía
process
tk diseño de footer administración crear estructura html del footer definir colores y tipografía
1
254,814
19,273,249,358
IssuesEvent
2021-12-10 08:51:15
Concordium/concordium.github.io
https://api.github.com/repos/Concordium/concordium.github.io
closed
Add documentation for enterprise identities
documentation [Prio] High [Type] Task
**Task description** Documentation for enterprise identities must be added to *mainnet* developer documentation. Note that enterprise identities are only relevant for very few enterprises (exchanges for now) but not for the general user. **Sub-tasks** - [ ] find/discuss appropriate location and create new page - [ ] review content in internal googledoc (invitation sent) - [ ] add content from internal googledoc when finalized - [ ] publish documentation Thursday (9-12) or Friday (10-12)
1.0
Add documentation for enterprise identities - **Task description** Documentation for enterprise identities must be added to *mainnet* developer documentation. Note that enterprise identities are only relevant for very few enterprises (exchanges for now) but not for the general user. **Sub-tasks** - [ ] find/discuss appropriate location and create new page - [ ] review content in internal googledoc (invitation sent) - [ ] add content from internal googledoc when finalized - [ ] publish documentation Thursday (9-12) or Friday (10-12)
non_process
add documentation for enterprise identities task description documentation for enterprise identities must be added to mainnet developer documentation note that enterprise identities are only relevant for very few enterprises exchanges for now but not for the general user sub tasks find discuss appropriate location and create new page review content in internal googledoc invitation sent add content from internal googledoc when finalized publish documentation thursday or friday
0
4,418
7,299,969,052
IssuesEvent
2018-02-26 21:54:26
vtex/formula-vtex
https://api.github.com/repos/vtex/formula-vtex
opened
Transferencia de carrinho do inStore
Frontend Processo Produto inStore
# Transferencia de carrinho do inStore ## Descrição da ideia - A ideia é múltiplos vendedores de uma loja tranferirem seus carrinhos de um pra outro de uma forma inovadora e com UX incrível
1.0
Transferencia de carrinho do inStore - # Transferencia de carrinho do inStore ## Descrição da ideia - A ideia é múltiplos vendedores de uma loja tranferirem seus carrinhos de um pra outro de uma forma inovadora e com UX incrível
process
transferencia de carrinho do instore transferencia de carrinho do instore descrição da ideia a ideia é múltiplos vendedores de uma loja tranferirem seus carrinhos de um pra outro de uma forma inovadora e com ux incrível
1