Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
45,586
12,890,010,954
IssuesEvent
2020-07-13 15:20:47
carbon-design-system/ibm-security
https://api.github.com/repos/carbon-design-system/ibm-security
closed
Add support for 'size' prop in DataTable
Defect
Add support for `size` prop in `DataTable`. The base Carbon library already supports this [option here](https://github.com/carbon-design-system/carbon/blob/fe23a6f46b001e56d31376cfebe6e523b27e77f9/packages/react/src/components/DataTable/DataTable.js#L111)
1.0
Add support for 'size' prop in DataTable - Add support for `size` prop in `DataTable`. The base Carbon library already supports this [option here](https://github.com/carbon-design-system/carbon/blob/fe23a6f46b001e56d31376cfebe6e523b27e77f9/packages/react/src/components/DataTable/DataTable.js#L111)
defect
add support for size prop in datatable add support for size prop in datatable the base carbon library already supports this
1
192,891
15,361,295,554
IssuesEvent
2021-03-01 17:57:50
dankelley/oce
https://api.github.com/repos/dankelley/oce
closed
as.ctd documentation needs updating (at least on website)
documentation website
Was just looking at the `as.ctd()` documentation on the pkgdown website at https://dankelley.github.io/oce/reference/as.ctd.html and I see as below: ![image](https://user-images.githubusercontent.com/233584/109509943-d3eaa800-7a77-11eb-9dac-2017b07594f8.png) The question is, "***CAN*** the `salinity` argument be an `rsk` object????" 😄
1.0
as.ctd documentation needs updating (at least on website) - Was just looking at the `as.ctd()` documentation on the pkgdown website at https://dankelley.github.io/oce/reference/as.ctd.html and I see as below: ![image](https://user-images.githubusercontent.com/233584/109509943-d3eaa800-7a77-11eb-9dac-2017b07594f8.png) The question is, "***CAN*** the `salinity` argument be an `rsk` object????" 😄
non_defect
as ctd documentation needs updating at least on website was just looking at the as ctd documentation on the pkgdown website at and i see as below the question is can the salinity argument be an rsk object 😄
0
18,367
3,386,974,864
IssuesEvent
2015-11-28 00:09:24
jgirald/ES2015C
https://api.github.com/repos/jgirald/ES2015C
opened
Texturas Caballero Avanzado Yamato
Character Design Medium Priority Team A Textures Yamato
### Description Aplicar texturas al modelo del personaje que representa al Caballero Yamato Avanzado. ### Acceptance Criteria Modelo acabado con las texturas incorporadas, preparado para animar. ### Estimated time effort: 2 hr
1.0
Texturas Caballero Avanzado Yamato - ### Description Aplicar texturas al modelo del personaje que representa al Caballero Yamato Avanzado. ### Acceptance Criteria Modelo acabado con las texturas incorporadas, preparado para animar. ### Estimated time effort: 2 hr
non_defect
texturas caballero avanzado yamato description aplicar texturas al modelo del personaje que representa al caballero yamato avanzado acceptance criteria modelo acabado con las texturas incorporadas preparado para animar estimated time effort hr
0
52,889
13,225,205,446
IssuesEvent
2020-08-17 20:42:11
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
Boost 1.36.0 w/ mac linker, can't find libs. (Trac #534)
Migrated from Trac defect tools/ports
Since the boost lib in I3_PORTS are not explicitly in the LD/DYLD_LIB_PATH in env-shell, the contents of the compiled I3_BUILD/lib libraries should give location of I3_PORT libs. For example, w/ boost_1.33.1: otool -L lib/libdataio.dylib gives: ... /Users/blaufuss/icework/i3tools/lib/boost-1.33.1/libboost_serialization-d.dylib (compatibility version 0.0.0, current version 0.0.0) ... Following the upgrade to 1.36.0: ... libboost_serialization-d.1.36.0.dylib (compatibility version 0.0.0, current version 0.0.0) ... Lacking any path information. And then when using these libs: ImportError: dlopen(/Users/blaufuss/icework/offline-software/trunk/build_debug/lib/icecube/icetray.so, 2): Library not loaded: libboost_python-d.1.36.0.dylib Referenced from: /Users/blaufuss/icework/offline-software/trunk/build_debug/lib/icecube/icetray.so Work around: add explicit location of $I3_PORTS/lib/boost_1.36.0 to env-shell.sh Seems something changed in the way that the port is built on mac. Linux seems fine. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/534">https://code.icecube.wisc.edu/projects/icecube/ticket/534</a>, reported by blaufussand owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2009-03-05T21:57:22", "_ts": "1236290242000000", "description": "Since the boost lib in I3_PORTS are not explicitly in the LD/DYLD_LIB_PATH in env-shell, the contents of the\ncompiled I3_BUILD/lib libraries should give location of I3_PORT libs.\n\nFor example, w/ boost_1.33.1:\n\notool -L lib/libdataio.dylib\n\ngives:\n...\n\t/Users/blaufuss/icework/i3tools/lib/boost-1.33.1/libboost_serialization-d.dylib (compatibility version 0.0.0, current version 0.0.0)\n...\n\nFollowing the upgrade to 1.36.0:\n...\n\tlibboost_serialization-d.1.36.0.dylib (compatibility version 0.0.0, current version 0.0.0)\n...\n\nLacking any path information.\n\nAnd then when using these libs:\nImportError: dlopen(/Users/blaufuss/icework/offline-software/trunk/build_debug/lib/icecube/icetray.so, 2): Library not loaded: libboost_python-d.1.36.0.dylib\n Referenced from: /Users/blaufuss/icework/offline-software/trunk/build_debug/lib/icecube/icetray.so\n\n\nWork around: add explicit location of $I3_PORTS/lib/boost_1.36.0 to env-shell.sh\n\nSeems something changed in the way that the port is built on mac.\nLinux seems fine.\n", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "time": "2009-02-09T19:24:54", "component": "tools/ports", "summary": "Boost 1.36.0 w/ mac linker, can't find libs.", "priority": "normal", "keywords": "", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
1.0
Boost 1.36.0 w/ mac linker, can't find libs. (Trac #534) - Since the boost lib in I3_PORTS are not explicitly in the LD/DYLD_LIB_PATH in env-shell, the contents of the compiled I3_BUILD/lib libraries should give location of I3_PORT libs. For example, w/ boost_1.33.1: otool -L lib/libdataio.dylib gives: ... /Users/blaufuss/icework/i3tools/lib/boost-1.33.1/libboost_serialization-d.dylib (compatibility version 0.0.0, current version 0.0.0) ... Following the upgrade to 1.36.0: ... libboost_serialization-d.1.36.0.dylib (compatibility version 0.0.0, current version 0.0.0) ... Lacking any path information. And then when using these libs: ImportError: dlopen(/Users/blaufuss/icework/offline-software/trunk/build_debug/lib/icecube/icetray.so, 2): Library not loaded: libboost_python-d.1.36.0.dylib Referenced from: /Users/blaufuss/icework/offline-software/trunk/build_debug/lib/icecube/icetray.so Work around: add explicit location of $I3_PORTS/lib/boost_1.36.0 to env-shell.sh Seems something changed in the way that the port is built on mac. Linux seems fine. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/534">https://code.icecube.wisc.edu/projects/icecube/ticket/534</a>, reported by blaufussand owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2009-03-05T21:57:22", "_ts": "1236290242000000", "description": "Since the boost lib in I3_PORTS are not explicitly in the LD/DYLD_LIB_PATH in env-shell, the contents of the\ncompiled I3_BUILD/lib libraries should give location of I3_PORT libs.\n\nFor example, w/ boost_1.33.1:\n\notool -L lib/libdataio.dylib\n\ngives:\n...\n\t/Users/blaufuss/icework/i3tools/lib/boost-1.33.1/libboost_serialization-d.dylib (compatibility version 0.0.0, current version 0.0.0)\n...\n\nFollowing the upgrade to 1.36.0:\n...\n\tlibboost_serialization-d.1.36.0.dylib (compatibility version 0.0.0, current version 0.0.0)\n...\n\nLacking any path information.\n\nAnd then when using these libs:\nImportError: dlopen(/Users/blaufuss/icework/offline-software/trunk/build_debug/lib/icecube/icetray.so, 2): Library not loaded: libboost_python-d.1.36.0.dylib\n Referenced from: /Users/blaufuss/icework/offline-software/trunk/build_debug/lib/icecube/icetray.so\n\n\nWork around: add explicit location of $I3_PORTS/lib/boost_1.36.0 to env-shell.sh\n\nSeems something changed in the way that the port is built on mac.\nLinux seems fine.\n", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "time": "2009-02-09T19:24:54", "component": "tools/ports", "summary": "Boost 1.36.0 w/ mac linker, can't find libs.", "priority": "normal", "keywords": "", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
defect
boost w mac linker can t find libs trac since the boost lib in ports are not explicitly in the ld dyld lib path in env shell the contents of the compiled build lib libraries should give location of port libs for example w boost otool l lib libdataio dylib gives users blaufuss icework lib boost libboost serialization d dylib compatibility version current version following the upgrade to libboost serialization d dylib compatibility version current version lacking any path information and then when using these libs importerror dlopen users blaufuss icework offline software trunk build debug lib icecube icetray so library not loaded libboost python d dylib referenced from users blaufuss icework offline software trunk build debug lib icecube icetray so work around add explicit location of ports lib boost to env shell sh seems something changed in the way that the port is built on mac linux seems fine migrated from json status closed changetime ts description since the boost lib in ports are not explicitly in the ld dyld lib path in env shell the contents of the ncompiled build lib libraries should give location of port libs n nfor example w boost n notool l lib libdataio dylib n ngives n n t users blaufuss icework lib boost libboost serialization d dylib compatibility version current version n n nfollowing the upgrade to n n tlibboost serialization d dylib compatibility version current version n n nlacking any path information n nand then when using these libs nimporterror dlopen users blaufuss icework offline software trunk build debug lib icecube icetray so library not loaded libboost python d dylib n referenced from users blaufuss icework offline software trunk build debug lib icecube icetray so n n nwork around add explicit location of ports lib boost to env shell sh n nseems something changed in the way that the port is built on mac nlinux seems fine n reporter blaufuss cc resolution fixed time component tools ports summary boost w mac linker can t find libs priority normal keywords milestone owner troy type defect
1
66,562
20,328,023,374
IssuesEvent
2022-02-18 08:00:46
vector-im/element-ios
https://api.github.com/repos/vector-im/element-ios
closed
Links containing markdown formatting (eg underscores) are interpreted as markdown
T-Defect A-Composer S-Major O-Occasional Z-WTF Z-Ready
### Steps to reproduce 1. Send a link like http://domain.xyz/foo/bar-_stuff-like-this_-in-it.jpg The same problem was recently fixed in Element Web, see https://github.com/vector-im/element-web/issues/4674 ### Outcome #### What did you expect? The link is send without markdown formatting. #### What happened instead? The link breaks due to being interpret as markdown. ### Your phone model iPhone 8 ### Operating system version 15.0.1 ### Application version 1.6.11 ### Homeserver _No response_ ### Will you send logs? No
1.0
Links containing markdown formatting (eg underscores) are interpreted as markdown - ### Steps to reproduce 1. Send a link like http://domain.xyz/foo/bar-_stuff-like-this_-in-it.jpg The same problem was recently fixed in Element Web, see https://github.com/vector-im/element-web/issues/4674 ### Outcome #### What did you expect? The link is send without markdown formatting. #### What happened instead? The link breaks due to being interpret as markdown. ### Your phone model iPhone 8 ### Operating system version 15.0.1 ### Application version 1.6.11 ### Homeserver _No response_ ### Will you send logs? No
defect
links containing markdown formatting eg underscores are interpreted as markdown steps to reproduce send a link like the same problem was recently fixed in element web see outcome what did you expect the link is send without markdown formatting what happened instead the link breaks due to being interpret as markdown your phone model iphone operating system version application version homeserver no response will you send logs no
1
2,698
3,005,172,924
IssuesEvent
2015-07-26 17:36:49
nightingale-media-player/nightingale-hacking
https://api.github.com/repos/nightingale-media-player/nightingale-hacking
closed
Fix mac builds
Buildsystem Mac OS X
Apparently some (or all?) mac builds are broken: http://forum.getnightingale.com/thread-429-post-5394.html#pid5394 According to @zjays it's 3267acc1883a3d33eff876c8bdca4b961c29fdeb that broke it. Their configure log looks like this: ``` `Adding configure options from /Users/Zach-Mac/GitHub/nightingale-hacking/nightingale.config: --with-macosx-sdk=/Developer/SDKs/MacOSX10.5.sdk --enable-installer --enable-official --enable-compiler-environment-checks=no rm -f /Users/Zach-Mac/GitHub/nightingale-hacking/configure /Users/Zach-Mac/GitHub/nightingale-hacking/.songbirdconfig.mk /Users/Zach-Mac/GitHub/nightingale-hacking/.songbirdconfig.out rm -f -r /Users/Zach-Mac/GitHub/nightingale-hacking/compiled Adding configure options from /Users/Zach-Mac/GitHub/nightingale-hacking/nightingale.config: --with-macosx-sdk=/Developer/SDKs/MacOSX10.5.sdk --enable-installer --enable-official --enable-compiler-environment-checks=no Nightingale Build System cd /Users/Zach-Mac/GitHub/nightingale-hacking && \ autoconf && \ rm -f -r /Users/Zach-Mac/GitHub/nightingale-hacking/autom4te.cache/ mkdir -p /Users/Zach-Mac/GitHub/nightingale-hacking/compiled /Users/Zach-Mac/GitHub/nightingale-hacking/compiled/dist cd /Users/Zach-Mac/GitHub/nightingale-hacking/compiled && \ /Users/Zach-Mac/GitHub/nightingale-hacking/configure --with-macosx-sdk=/Developer/SDKs/MacOSX10.5.sdk --enable-installer --enable-official --enable-compiler-environment-checks=no checking build system type... x86_64-apple-darwin12.6.0 checking host system type... x86_64-apple-darwin12.6.0 checking target system type... x86_64-apple-darwin12.6.0 checking for git... git configure: error: Unsupported architecture make: *** [/Users/Zach-Mac/GitHub/nightingale-hacking/compiled/config.status] Error 1 ``` Apparently setting `--disable-compiler-environment-checks` doesn't help either. From the changes in the configure update, it now detects the architecture on mac (from [line 1150 in the old version](https://github.com/nightingale-media-player/nightingale-hacking/commit/3267acc1883a3d33eff876c8bdca4b961c29fdeb#diff-93452acb07da1f0393119ce86e69e3bbL1150)) , and I'm not sure how that works for nightingale. <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/24932953-fix-mac-builds?utm_campaign=plugin&utm_content=tracker%2F230233&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F230233&utm_medium=issues&utm_source=github). </bountysource-plugin>
1.0
Fix mac builds - Apparently some (or all?) mac builds are broken: http://forum.getnightingale.com/thread-429-post-5394.html#pid5394 According to @zjays it's 3267acc1883a3d33eff876c8bdca4b961c29fdeb that broke it. Their configure log looks like this: ``` `Adding configure options from /Users/Zach-Mac/GitHub/nightingale-hacking/nightingale.config: --with-macosx-sdk=/Developer/SDKs/MacOSX10.5.sdk --enable-installer --enable-official --enable-compiler-environment-checks=no rm -f /Users/Zach-Mac/GitHub/nightingale-hacking/configure /Users/Zach-Mac/GitHub/nightingale-hacking/.songbirdconfig.mk /Users/Zach-Mac/GitHub/nightingale-hacking/.songbirdconfig.out rm -f -r /Users/Zach-Mac/GitHub/nightingale-hacking/compiled Adding configure options from /Users/Zach-Mac/GitHub/nightingale-hacking/nightingale.config: --with-macosx-sdk=/Developer/SDKs/MacOSX10.5.sdk --enable-installer --enable-official --enable-compiler-environment-checks=no Nightingale Build System cd /Users/Zach-Mac/GitHub/nightingale-hacking && \ autoconf && \ rm -f -r /Users/Zach-Mac/GitHub/nightingale-hacking/autom4te.cache/ mkdir -p /Users/Zach-Mac/GitHub/nightingale-hacking/compiled /Users/Zach-Mac/GitHub/nightingale-hacking/compiled/dist cd /Users/Zach-Mac/GitHub/nightingale-hacking/compiled && \ /Users/Zach-Mac/GitHub/nightingale-hacking/configure --with-macosx-sdk=/Developer/SDKs/MacOSX10.5.sdk --enable-installer --enable-official --enable-compiler-environment-checks=no checking build system type... x86_64-apple-darwin12.6.0 checking host system type... x86_64-apple-darwin12.6.0 checking target system type... x86_64-apple-darwin12.6.0 checking for git... git configure: error: Unsupported architecture make: *** [/Users/Zach-Mac/GitHub/nightingale-hacking/compiled/config.status] Error 1 ``` Apparently setting `--disable-compiler-environment-checks` doesn't help either. From the changes in the configure update, it now detects the architecture on mac (from [line 1150 in the old version](https://github.com/nightingale-media-player/nightingale-hacking/commit/3267acc1883a3d33eff876c8bdca4b961c29fdeb#diff-93452acb07da1f0393119ce86e69e3bbL1150)) , and I'm not sure how that works for nightingale. <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/24932953-fix-mac-builds?utm_campaign=plugin&utm_content=tracker%2F230233&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F230233&utm_medium=issues&utm_source=github). </bountysource-plugin>
non_defect
fix mac builds apparently some or all mac builds are broken according to zjays it s that broke it their configure log looks like this adding configure options from users zach mac github nightingale hacking nightingale config with macosx sdk developer sdks sdk enable installer enable official enable compiler environment checks no rm f users zach mac github nightingale hacking configure users zach mac github nightingale hacking songbirdconfig mk users zach mac github nightingale hacking songbirdconfig out rm f r users zach mac github nightingale hacking compiled adding configure options from users zach mac github nightingale hacking nightingale config with macosx sdk developer sdks sdk enable installer enable official enable compiler environment checks no nightingale build system cd users zach mac github nightingale hacking autoconf rm f r users zach mac github nightingale hacking cache mkdir p users zach mac github nightingale hacking compiled users zach mac github nightingale hacking compiled dist cd users zach mac github nightingale hacking compiled users zach mac github nightingale hacking configure with macosx sdk developer sdks sdk enable installer enable official enable compiler environment checks no checking build system type apple checking host system type apple checking target system type apple checking for git git configure error unsupported architecture make error apparently setting disable compiler environment checks doesn t help either from the changes in the configure update it now detects the architecture on mac from and i m not sure how that works for nightingale want to back this issue we accept bounties via
0
8,473
2,611,513,653
IssuesEvent
2015-02-27 05:49:37
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
patch for google issue tracker api
auto-migrated Priority-Medium Type-Defect
``` The description-blanking-out issue you described has been fixed. The email CC feature has been implemented in code, but it doesn't seem to actually do anything. I tried it, and the email I put it never received any notifications of comments on the posted issue. Because of this, I commented out three lines of code to essentially hide the "email" cc feature from the end-user. If I or someone else figures out how to make the cc work in the future, just uncomment the three lines around pagefeedback.cpp:89. - Drew Gottlieb (a.k.a. dag10) ``` Original issue reported on code.google.com by `drew1313...@gmail.com` on 8 Dec 2012 at 4:04 Attachments: * [feedback_2.diff](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-498/comment-0/feedback_2.diff)
1.0
patch for google issue tracker api - ``` The description-blanking-out issue you described has been fixed. The email CC feature has been implemented in code, but it doesn't seem to actually do anything. I tried it, and the email I put it never received any notifications of comments on the posted issue. Because of this, I commented out three lines of code to essentially hide the "email" cc feature from the end-user. If I or someone else figures out how to make the cc work in the future, just uncomment the three lines around pagefeedback.cpp:89. - Drew Gottlieb (a.k.a. dag10) ``` Original issue reported on code.google.com by `drew1313...@gmail.com` on 8 Dec 2012 at 4:04 Attachments: * [feedback_2.diff](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-498/comment-0/feedback_2.diff)
defect
patch for google issue tracker api the description blanking out issue you described has been fixed the email cc feature has been implemented in code but it doesn t seem to actually do anything i tried it and the email i put it never received any notifications of comments on the posted issue because of this i commented out three lines of code to essentially hide the email cc feature from the end user if i or someone else figures out how to make the cc work in the future just uncomment the three lines around pagefeedback cpp drew gottlieb a k a original issue reported on code google com by gmail com on dec at attachments
1
78,439
27,522,821,545
IssuesEvent
2023-03-06 16:03:56
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
closed
Formatting creates extra newlines
T-Defect A-Markdown A-Composer X-Regression S-Major O-Occasional
### Steps to reproduce Send texts containing formatting ![Screenshot_20230226-121654_1](https://user-images.githubusercontent.com/18397681/221407574-12e9b73c-dd32-453a-be73-fd6b5f8f9cfc.png) ### Outcome #### What did you expect? Same functionality as old version of Element Android and other clients. Element Desktop 1.11.23: ![image](https://user-images.githubusercontent.com/18397681/221407776-d492d40e-2f2b-4876-bee6-029bd151fae9.png) #### What happened instead? Most formatting options start a new line. This seems to only be the case when it is the first formatted text in a paragraph: ![Screenshot_20230226-125116_1](https://user-images.githubusercontent.com/18397681/221408789-0337d02f-c2ef-413e-8660-1182acfdb501.png) With the markdown: ``` one _two_ three four _five_ six ``` ### Your phone model Google Pixel 7 ### Operating system version GrapheneOS stable / Android 13 ### Application version and app store Version 1.5.25 [40105250] (F-f74a885a), Matrix SDK Version 1.5.25 (f74a885a), olm version 3.2.12, from F-Droid ### Homeserver Synapse 1.77.0 ### Will you send logs? No ### Are you willing to provide a PR? No
1.0
Formatting creates extra newlines - ### Steps to reproduce Send texts containing formatting ![Screenshot_20230226-121654_1](https://user-images.githubusercontent.com/18397681/221407574-12e9b73c-dd32-453a-be73-fd6b5f8f9cfc.png) ### Outcome #### What did you expect? Same functionality as old version of Element Android and other clients. Element Desktop 1.11.23: ![image](https://user-images.githubusercontent.com/18397681/221407776-d492d40e-2f2b-4876-bee6-029bd151fae9.png) #### What happened instead? Most formatting options start a new line. This seems to only be the case when it is the first formatted text in a paragraph: ![Screenshot_20230226-125116_1](https://user-images.githubusercontent.com/18397681/221408789-0337d02f-c2ef-413e-8660-1182acfdb501.png) With the markdown: ``` one _two_ three four _five_ six ``` ### Your phone model Google Pixel 7 ### Operating system version GrapheneOS stable / Android 13 ### Application version and app store Version 1.5.25 [40105250] (F-f74a885a), Matrix SDK Version 1.5.25 (f74a885a), olm version 3.2.12, from F-Droid ### Homeserver Synapse 1.77.0 ### Will you send logs? No ### Are you willing to provide a PR? No
defect
formatting creates extra newlines steps to reproduce send texts containing formatting outcome what did you expect same functionality as old version of element android and other clients element desktop what happened instead most formatting options start a new line this seems to only be the case when it is the first formatted text in a paragraph with the markdown one two three four five six your phone model google pixel operating system version grapheneos stable android application version and app store version f matrix sdk version olm version from f droid homeserver synapse will you send logs no are you willing to provide a pr no
1
200,738
7,011,152,550
IssuesEvent
2017-12-20 03:49:45
GoogleCloudPlatform/google-cloud-go
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-go
closed
storage: create custom domain bucket 403 error
api: storage priority: p1 type: question
# case 1: OK bkt := client.Bucket("subdomain") err := bkt.Create(c, project_id, BucketAttrs) # case 2: 403 ERROR myDomain.com is a verified bucket myDomain.com is created via https://console.cloud.google.com/storage/browser?project=***. bkt := client.Bucket("subdomain.myDomain.com") err := bkt.Create(c, project_id, BucketAttrs) googleapi: Error 403: The bucket you tried to create requires domain ownership verification., forbidden This is run on gae local SDK http://localhost:8080. https://cloud.google.com/storage/docs/json_api/v1/buckets/insert the "subdomain.myDomain.com" works fine. gsutil mb gs://subdomain.myDomain.com OK
1.0
storage: create custom domain bucket 403 error - # case 1: OK bkt := client.Bucket("subdomain") err := bkt.Create(c, project_id, BucketAttrs) # case 2: 403 ERROR myDomain.com is a verified bucket myDomain.com is created via https://console.cloud.google.com/storage/browser?project=***. bkt := client.Bucket("subdomain.myDomain.com") err := bkt.Create(c, project_id, BucketAttrs) googleapi: Error 403: The bucket you tried to create requires domain ownership verification., forbidden This is run on gae local SDK http://localhost:8080. https://cloud.google.com/storage/docs/json_api/v1/buckets/insert the "subdomain.myDomain.com" works fine. gsutil mb gs://subdomain.myDomain.com OK
non_defect
storage create custom domain bucket error case ok bkt client bucket subdomain err bkt create c project id bucketattrs case error mydomain com is a verified bucket mydomain com is created via bkt client bucket subdomain mydomain com err bkt create c project id bucketattrs googleapi error the bucket you tried to create requires domain ownership verification forbidden this is run on gae local sdk the subdomain mydomain com works fine gsutil mb gs subdomain mydomain com ok
0
34,526
7,453,062,993
IssuesEvent
2018-03-29 10:31:03
kerdokullamae/test_koik_issued
https://api.github.com/repos/kerdokullamae/test_koik_issued
closed
Ajax pageri bugid
P: highest R: fixed T: defect
**Reported by sven syld on 29 Jul 2013 15:38 UTC** KÜ otsingunimekirjas ja facetite lehitsemisel on ajaxi pageriga bugi, et peale 1. requesti läheb html struktuur paigast. Visuaalselt avaldub see nii, et lingi klikkamisel tekib vasakusse üles nurka teine, tabeli suurune kast, mis kaob tulemuste saabumisel ära.
1.0
Ajax pageri bugid - **Reported by sven syld on 29 Jul 2013 15:38 UTC** KÜ otsingunimekirjas ja facetite lehitsemisel on ajaxi pageriga bugi, et peale 1. requesti läheb html struktuur paigast. Visuaalselt avaldub see nii, et lingi klikkamisel tekib vasakusse üles nurka teine, tabeli suurune kast, mis kaob tulemuste saabumisel ära.
defect
ajax pageri bugid reported by sven syld on jul utc kü otsingunimekirjas ja facetite lehitsemisel on ajaxi pageriga bugi et peale requesti läheb html struktuur paigast visuaalselt avaldub see nii et lingi klikkamisel tekib vasakusse üles nurka teine tabeli suurune kast mis kaob tulemuste saabumisel ära
1
296,303
9,107,266,813
IssuesEvent
2019-02-21 03:31:32
thautwarm/MLStyle.jl
https://api.github.com/repos/thautwarm/MLStyle.jl
closed
can not use @match inside another macro
bug core help wanted high-priority
MWE: copying what's in here: https://github.com/thautwarm/MLStyle-Playground/blob/master/StaticCapturing.jl to a module. The scoping is not correct.
1.0
can not use @match inside another macro - MWE: copying what's in here: https://github.com/thautwarm/MLStyle-Playground/blob/master/StaticCapturing.jl to a module. The scoping is not correct.
non_defect
can not use match inside another macro mwe copying what s in here to a module the scoping is not correct
0
16,686
2,933,667,257
IssuesEvent
2015-06-30 00:56:40
Ephemerality/xray-builder
https://api.github.com/repos/Ephemerality/xray-builder
closed
[Linux] mobi_unpack.py is successful, but XRayBuilder shows error
auto-migrated Priority-Medium Type-Defect
``` I'm trying to get XRayBuilder.exe to run under Linux with Mono. Basically, it works, but it doesn't recognise the successful unpacking via mobi_unpack.py or even the latest version KindleUnpack.py. My commandline: mkdir out mono XRayBuilder.exe -u ../Mobi_Unpack_v047/lib/mobi_unpack.py -o ./out ../book.mobi Output: X-Ray Builder Version 1.40 Using ./out as an output directory. Using default Python command. Ensure Python's directory is included in your PATH environment variable. Processing blackout.mobi... Enter Shelfari URL for blackout (Enter to skip): http://www.shelfari.com/books/12345678/book Running mobi_unpack to get book data... Error unpacking mobi file: MobiUnpack 0.47 Copyright (c) 2009 Charles M. Hannum <root@ihack.net> With Additions by P. Durrant, K. Hendricks, S. Siebert, fandrieu, DiapDealer, nickredding. Unpacking Book... Palm DB type: BOOKMOBI Mobi Version: 6 Codec: utf-8 Title: MyBook Palmdoc compression firstaddl 1c3 ncxidx 1c0 exth flags 50 Processing Mobi format Ebook ... MetaData from EXTH {'Publisher': ['My Pub'], 'Language': ['en'], 'Updated Title': ['MyBook'], 'ASIN': .....} Unpack raw markup language Unpacking images, resources, fonts, etc firstaddl is 451 num_sections is 475 K8Boundary is -1 Section is 451 Warning: Section 451 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 452 Warning: Section 452 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 453 extracting image: image00003.jpeg Section is 454 Warning: Section 454 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 455 Warning: Section 455 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 456 Warning: Section 456 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 457 Warning: Section 457 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 458 Warning: Section 458 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 459 extracting image: image00009.jpeg Section is 460 extracting image: image00010.jpeg Section is 461 Warning: Section 461 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 462 Warning: Section 462 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 463 Warning: Section 463 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 464 Warning: Section 464 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 465 Warning: Section 465 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 466 Warning: Section 466 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 467 Warning: Section 467 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 468 Warning: Section 468 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 469 Warning: Section 469 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 470 Warning: Section 470 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 471 Warning: Section 471 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 472 First 4 bytes: 46 4c 49 53 Skipping FLIS section Section is 473 First 4 bytes: 46 43 49 53 Skipping FCIS section Section is 474 Skip section 474 as it contains the EOF record. Write ncx Find link anchors Insert data into html Insert hrefs into html Remove empty anchors from html Insert image references into html Error: Referenced image 17 was not recognized as a valid image Error: Referenced image 18 was not recognized as a valid image Error: Referenced image 19 was not recognized as a valid image Error: Referenced image 20 was not recognized as a valid image Error: Referenced image 1 was not recognized as a valid image Error: Referenced image 11 was not recognized as a valid image Error: Referenced image 21 was not recognized as a valid image Error: Referenced image 5 was not recognized as a valid image Error: Referenced image 13 was not recognized as a valid image Error: Referenced image 4 was not recognized as a valid image Error: Referenced image 7 was not recognized as a valid image Error: Referenced image 16 was not recognized as a valid image Error: Referenced image 15 was not recognized as a valid image Error: Referenced image 12 was not recognized as a valid image Error: Referenced image 8 was not recognized as a valid image Error: Referenced image 6 was not recognized as a valid image Error: Referenced image 14 was not recognized as a valid image Error: Referenced image 2 was not recognized as a valid image Write opf Completed Press Enter to exit. So basically, it tells me that there was an error unpacking the book file and the error log shows that everything went successfully. ``` Original issue reported on code.google.com by `mbi...@gmail.com` on 26 Nov 2014 at 8:42
1.0
[Linux] mobi_unpack.py is successful, but XRayBuilder shows error - ``` I'm trying to get XRayBuilder.exe to run under Linux with Mono. Basically, it works, but it doesn't recognise the successful unpacking via mobi_unpack.py or even the latest version KindleUnpack.py. My commandline: mkdir out mono XRayBuilder.exe -u ../Mobi_Unpack_v047/lib/mobi_unpack.py -o ./out ../book.mobi Output: X-Ray Builder Version 1.40 Using ./out as an output directory. Using default Python command. Ensure Python's directory is included in your PATH environment variable. Processing blackout.mobi... Enter Shelfari URL for blackout (Enter to skip): http://www.shelfari.com/books/12345678/book Running mobi_unpack to get book data... Error unpacking mobi file: MobiUnpack 0.47 Copyright (c) 2009 Charles M. Hannum <root@ihack.net> With Additions by P. Durrant, K. Hendricks, S. Siebert, fandrieu, DiapDealer, nickredding. Unpacking Book... Palm DB type: BOOKMOBI Mobi Version: 6 Codec: utf-8 Title: MyBook Palmdoc compression firstaddl 1c3 ncxidx 1c0 exth flags 50 Processing Mobi format Ebook ... MetaData from EXTH {'Publisher': ['My Pub'], 'Language': ['en'], 'Updated Title': ['MyBook'], 'ASIN': .....} Unpack raw markup language Unpacking images, resources, fonts, etc firstaddl is 451 num_sections is 475 K8Boundary is -1 Section is 451 Warning: Section 451 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 452 Warning: Section 452 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 453 extracting image: image00003.jpeg Section is 454 Warning: Section 454 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 455 Warning: Section 455 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 456 Warning: Section 456 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 457 Warning: Section 457 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 458 Warning: Section 458 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 459 extracting image: image00009.jpeg Section is 460 extracting image: image00010.jpeg Section is 461 Warning: Section 461 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 462 Warning: Section 462 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 463 Warning: Section 463 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 464 Warning: Section 464 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 465 Warning: Section 465 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 466 Warning: Section 466 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 467 Warning: Section 467 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 468 Warning: Section 468 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 469 Warning: Section 469 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 470 Warning: Section 470 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 471 Warning: Section 471 contains no image or an unknown image format First 4 bytes: ff d8 ff ed Section is 472 First 4 bytes: 46 4c 49 53 Skipping FLIS section Section is 473 First 4 bytes: 46 43 49 53 Skipping FCIS section Section is 474 Skip section 474 as it contains the EOF record. Write ncx Find link anchors Insert data into html Insert hrefs into html Remove empty anchors from html Insert image references into html Error: Referenced image 17 was not recognized as a valid image Error: Referenced image 18 was not recognized as a valid image Error: Referenced image 19 was not recognized as a valid image Error: Referenced image 20 was not recognized as a valid image Error: Referenced image 1 was not recognized as a valid image Error: Referenced image 11 was not recognized as a valid image Error: Referenced image 21 was not recognized as a valid image Error: Referenced image 5 was not recognized as a valid image Error: Referenced image 13 was not recognized as a valid image Error: Referenced image 4 was not recognized as a valid image Error: Referenced image 7 was not recognized as a valid image Error: Referenced image 16 was not recognized as a valid image Error: Referenced image 15 was not recognized as a valid image Error: Referenced image 12 was not recognized as a valid image Error: Referenced image 8 was not recognized as a valid image Error: Referenced image 6 was not recognized as a valid image Error: Referenced image 14 was not recognized as a valid image Error: Referenced image 2 was not recognized as a valid image Write opf Completed Press Enter to exit. So basically, it tells me that there was an error unpacking the book file and the error log shows that everything went successfully. ``` Original issue reported on code.google.com by `mbi...@gmail.com` on 26 Nov 2014 at 8:42
defect
mobi unpack py is successful but xraybuilder shows error i m trying to get xraybuilder exe to run under linux with mono basically it works but it doesn t recognise the successful unpacking via mobi unpack py or even the latest version kindleunpack py my commandline mkdir out mono xraybuilder exe u mobi unpack lib mobi unpack py o out book mobi output x ray builder version using out as an output directory using default python command ensure python s directory is included in your path environment variable processing blackout mobi enter shelfari url for blackout enter to skip running mobi unpack to get book data error unpacking mobi file mobiunpack copyright c charles m hannum with additions by p durrant k hendricks s siebert fandrieu diapdealer nickredding unpacking book palm db type bookmobi mobi version codec utf title mybook palmdoc compression firstaddl ncxidx exth flags processing mobi format ebook metadata from exth publisher language updated title asin unpack raw markup language unpacking images resources fonts etc firstaddl is num sections is is section is warning section contains no image or an unknown image format first bytes ff ff ed section is warning section contains no image or an unknown image format first bytes ff ff ed section is extracting image jpeg section is warning section contains no image or an unknown image format first bytes ff ff ed section is warning section contains no image or an unknown image format first bytes ff ff ed section is warning section contains no image or an unknown image format first bytes ff ff ed section is warning section contains no image or an unknown image format first bytes ff ff ed section is warning section contains no image or an unknown image format first bytes ff ff ed section is extracting image jpeg section is extracting image jpeg section is warning section contains no image or an unknown image format first bytes ff ff ed section is warning section contains no image or an unknown image format first bytes ff ff ed section is warning section contains no image or an unknown image format first bytes ff ff ed section is warning section contains no image or an unknown image format first bytes ff ff ed section is warning section contains no image or an unknown image format first bytes ff ff ed section is warning section contains no image or an unknown image format first bytes ff ff ed section is warning section contains no image or an unknown image format first bytes ff ff ed section is warning section contains no image or an unknown image format first bytes ff ff ed section is warning section contains no image or an unknown image format first bytes ff ff ed section is warning section contains no image or an unknown image format first bytes ff ff ed section is warning section contains no image or an unknown image format first bytes ff ff ed section is first bytes skipping flis section section is first bytes skipping fcis section section is skip section as it contains the eof record write ncx find link anchors insert data into html insert hrefs into html remove empty anchors from html insert image references into html error referenced image was not recognized as a valid image error referenced image was not recognized as a valid image error referenced image was not recognized as a valid image error referenced image was not recognized as a valid image error referenced image was not recognized as a valid image error referenced image was not recognized as a valid image error referenced image was not recognized as a valid image error referenced image was not recognized as a valid image error referenced image was not recognized as a valid image error referenced image was not recognized as a valid image error referenced image was not recognized as a valid image error referenced image was not recognized as a valid image error referenced image was not recognized as a valid image error referenced image was not recognized as a valid image error referenced image was not recognized as a valid image error referenced image was not recognized as a valid image error referenced image was not recognized as a valid image error referenced image was not recognized as a valid image write opf completed press enter to exit so basically it tells me that there was an error unpacking the book file and the error log shows that everything went successfully original issue reported on code google com by mbi gmail com on nov at
1
4,852
2,610,158,400
IssuesEvent
2015-02-26 18:50:15
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
closed
Graphics Glitch
auto-migrated Priority-Medium Type-Defect
``` Missing textures from Cad Banes Thermal Bomb ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 3:01
1.0
Graphics Glitch - ``` Missing textures from Cad Banes Thermal Bomb ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 3:01
defect
graphics glitch missing textures from cad banes thermal bomb original issue reported on code google com by gmail com on jan at
1
90,769
26,181,163,076
IssuesEvent
2023-01-02 15:46:51
CGNS/CGNS
https://api.github.com/repos/CGNS/CGNS
closed
[CGNS-78] Implement FortranCInterface_HEADER in cmake
Improvement Done Build Critical
> This issue has been imported from JIRA. Read the [original ticket here](https://cgnsorg.atlassian.net/browse/CGNS-78). - _**Created at:**_ Thu, 25 Feb 2016 09:50:28 -0600 - _**Resolved at:**_ Thu, 25 Feb 2016 10:17:26 -0600 <p>Currently, CGNS handles the C/Fortan name mangling explicitly, it uses macro and test compiles C and fortran code, i.e.</p> <p>Fortran_test/CMakeLists.txt'<br/> Fortran_test/csub.c'<br/> Fortran_test/fmain.f'<br/> Fortran_test/fortran_macros.h'</p> <p>We should instead use FortranCInterface_HEADER and let Cmake handle it. This will also help the windows build.</p>
1.0
[CGNS-78] Implement FortranCInterface_HEADER in cmake - > This issue has been imported from JIRA. Read the [original ticket here](https://cgnsorg.atlassian.net/browse/CGNS-78). - _**Created at:**_ Thu, 25 Feb 2016 09:50:28 -0600 - _**Resolved at:**_ Thu, 25 Feb 2016 10:17:26 -0600 <p>Currently, CGNS handles the C/Fortan name mangling explicitly, it uses macro and test compiles C and fortran code, i.e.</p> <p>Fortran_test/CMakeLists.txt'<br/> Fortran_test/csub.c'<br/> Fortran_test/fmain.f'<br/> Fortran_test/fortran_macros.h'</p> <p>We should instead use FortranCInterface_HEADER and let Cmake handle it. This will also help the windows build.</p>
non_defect
implement fortrancinterface header in cmake this issue has been imported from jira read the created at thu feb resolved at thu feb currently cgns handles the c fortan name mangling explicitly it uses macro and test compiles c and fortran code i e fortran test cmakelists txt fortran test csub c fortran test fmain f fortran test fortran macros h we should instead use fortrancinterface header and let cmake handle it this will also help the windows build
0
52,792
13,225,067,140
IssuesEvent
2020-08-17 20:25:21
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
radioeventbrowser (Trac #344)
Migrated from Trac RASTA defect
radioeventbrowser not build by default. Have to specify -DBUILD_TOPEVENTBROWSER=True by compiling <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/344">https://code.icecube.wisc.edu/projects/icecube/ticket/344</a>, reported by tobiasand owned by sboeser</em></summary> <p> ```json { "status": "closed", "changetime": "2012-01-17T05:10:49", "_ts": "1326777049000000", "description": "radioeventbrowser not build by default. Have to specify -DBUILD_TOPEVENTBROWSER=True by compiling", "reporter": "tobias", "cc": "", "resolution": "fixed", "time": "2012-01-16T14:11:28", "component": "RASTA", "summary": "radioeventbrowser", "priority": "normal", "keywords": "", "milestone": "", "owner": "sboeser", "type": "defect" } ``` </p> </details>
1.0
radioeventbrowser (Trac #344) - radioeventbrowser not build by default. Have to specify -DBUILD_TOPEVENTBROWSER=True by compiling <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/344">https://code.icecube.wisc.edu/projects/icecube/ticket/344</a>, reported by tobiasand owned by sboeser</em></summary> <p> ```json { "status": "closed", "changetime": "2012-01-17T05:10:49", "_ts": "1326777049000000", "description": "radioeventbrowser not build by default. Have to specify -DBUILD_TOPEVENTBROWSER=True by compiling", "reporter": "tobias", "cc": "", "resolution": "fixed", "time": "2012-01-16T14:11:28", "component": "RASTA", "summary": "radioeventbrowser", "priority": "normal", "keywords": "", "milestone": "", "owner": "sboeser", "type": "defect" } ``` </p> </details>
defect
radioeventbrowser trac radioeventbrowser not build by default have to specify dbuild topeventbrowser true by compiling migrated from json status closed changetime ts description radioeventbrowser not build by default have to specify dbuild topeventbrowser true by compiling reporter tobias cc resolution fixed time component rasta summary radioeventbrowser priority normal keywords milestone owner sboeser type defect
1
71,993
23,885,117,819
IssuesEvent
2022-09-08 06:58:48
PowerDNS/pdns
https://api.github.com/repos/PowerDNS/pdns
closed
auth: backend creation before chroot() is causing startup failures
auth defect
Commit d5385b3570b0d8c13626dcb4cd8fe5fc344db16d from issue #9464 creates backend instances before complete initialization like chroot()+chdir() and dropping privileges is done. This causes backends that open files, like the gsqlite3 backend, to fail during startup. This seems to be some unintended side-effect of the implementation.
1.0
auth: backend creation before chroot() is causing startup failures - Commit d5385b3570b0d8c13626dcb4cd8fe5fc344db16d from issue #9464 creates backend instances before complete initialization like chroot()+chdir() and dropping privileges is done. This causes backends that open files, like the gsqlite3 backend, to fail during startup. This seems to be some unintended side-effect of the implementation.
defect
auth backend creation before chroot is causing startup failures commit from issue creates backend instances before complete initialization like chroot chdir and dropping privileges is done this causes backends that open files like the backend to fail during startup this seems to be some unintended side effect of the implementation
1
11,374
3,487,879,713
IssuesEvent
2016-01-02 11:37:47
KSP-KOS/KOS
https://api.github.com/repos/KSP-KOS/KOS
closed
Q reading doesn't match kPa
bug documentation
Here is what I get when I print out `ship:Q` under the latest kOS and FAR in 1.0.5 ![untitled](https://cloud.githubusercontent.com/assets/7699577/12004375/3e72f8f6-ab22-11e5-9d80-a0b569ce37f6.png) The kOS screen refreshes once a second so the value is behind that's not the issue - the issue is what's up with the decimal placing?
1.0
Q reading doesn't match kPa - Here is what I get when I print out `ship:Q` under the latest kOS and FAR in 1.0.5 ![untitled](https://cloud.githubusercontent.com/assets/7699577/12004375/3e72f8f6-ab22-11e5-9d80-a0b569ce37f6.png) The kOS screen refreshes once a second so the value is behind that's not the issue - the issue is what's up with the decimal placing?
non_defect
q reading doesn t match kpa here is what i get when i print out ship q under the latest kos and far in the kos screen refreshes once a second so the value is behind that s not the issue the issue is what s up with the decimal placing
0
9,997
2,616,018,782
IssuesEvent
2015-03-02 01:00:21
jasonhall/bwapi
https://api.github.com/repos/jasonhall/bwapi
closed
getRemainingBuildTime() returns 750 on SiegeTanks after Siege/Unsiege
auto-migrated Component-Logic Milestone-MajorRelease Offset-hunting Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Build a Siege Tank 2. Siege the Tank (maybe need to unsiege too) 3. Use getRemainingBuildTime() on the siege tank What is the expected output? What do you see instead? getRemainingBuildTime() shall return 0, but it returns 750 What version of the product are you using? On what operating system? Windows 7 BWAPI 3.7.4 ``` Original issue reported on code.google.com by `senth.wa...@gmail.com` on 2 Aug 2012 at 12:58
1.0
getRemainingBuildTime() returns 750 on SiegeTanks after Siege/Unsiege - ``` What steps will reproduce the problem? 1. Build a Siege Tank 2. Siege the Tank (maybe need to unsiege too) 3. Use getRemainingBuildTime() on the siege tank What is the expected output? What do you see instead? getRemainingBuildTime() shall return 0, but it returns 750 What version of the product are you using? On what operating system? Windows 7 BWAPI 3.7.4 ``` Original issue reported on code.google.com by `senth.wa...@gmail.com` on 2 Aug 2012 at 12:58
defect
getremainingbuildtime returns on siegetanks after siege unsiege what steps will reproduce the problem build a siege tank siege the tank maybe need to unsiege too use getremainingbuildtime on the siege tank what is the expected output what do you see instead getremainingbuildtime shall return but it returns what version of the product are you using on what operating system windows bwapi original issue reported on code google com by senth wa gmail com on aug at
1
171,429
27,118,571,708
IssuesEvent
2023-02-15 20:41:30
mickova/Projekt_Pandik
https://api.github.com/repos/mickova/Projekt_Pandik
opened
Úprava PC wireframů
Frontend Design
# Úprava všech PC wireframů a přidání nových ### Úkoly - [x] Vytvořit wireframe na úpravu příspěvku - [x] Vytvořit wireframe na zobrazení všech příspěvků - [x] Všechny wireframy doplnit popisky o funkčnosti
1.0
Úprava PC wireframů - # Úprava všech PC wireframů a přidání nových ### Úkoly - [x] Vytvořit wireframe na úpravu příspěvku - [x] Vytvořit wireframe na zobrazení všech příspěvků - [x] Všechny wireframy doplnit popisky o funkčnosti
non_defect
úprava pc wireframů úprava všech pc wireframů a přidání nových úkoly vytvořit wireframe na úpravu příspěvku vytvořit wireframe na zobrazení všech příspěvků všechny wireframy doplnit popisky o funkčnosti
0
192,397
14,615,880,030
IssuesEvent
2020-12-22 12:19:09
mixcore/mix.core
https://api.github.com/repos/mixcore/mix.core
closed
portal > attribute > list > check some items > use delete from dropdownlist > not able to delete multible items
Fixed & Testing
![image](https://user-images.githubusercontent.com/3785721/65881868-47d5dc00-e3be-11e9-8835-86d705350635.png)
1.0
portal > attribute > list > check some items > use delete from dropdownlist > not able to delete multible items - ![image](https://user-images.githubusercontent.com/3785721/65881868-47d5dc00-e3be-11e9-8835-86d705350635.png)
non_defect
portal attribute list check some items use delete from dropdownlist not able to delete multible items
0
48,194
13,067,513,374
IssuesEvent
2020-07-31 00:42:12
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
[simulation] error in MC afterprocessing (MCPulseSeriesParticleIDMap) (Trac #1942)
Migrated from Trac combo simulation defect
I was again on the search for simulation of some MCstudies, for which I need the complete I3MCPulse-info (MCPulseMap, ParticleIDMap and I3MCTRee); It took me some while to locate an appropriate one, which I found in dataset 11843 resp. 11642 (IC86:2013 GENIE simulation), one of the latest produced datasets This dataset was produced with icesim.V04-01-11-G4C (whatever this means). All other datasets I found which are produced with an olderor similar release candidate (icesim.V04-01-11) have the necessary objects, but the after-processing for Noise-MCPulses is flawed: there are no entries in the ParticleIDMap for those pulses as originated from ParticleID(0,0) This can be identified by inspecting the I3MCPulseSeriesParticleIDMap in the frame finding no entries for the major/minor-ID(0,0) and in general a lot of entries with only a single OMKey as key and an empty vector for the value in this I3Map<OMKey, vector<...> > So whatever is the difference between these meta-releases, somebody should have a look into that the after-processing is done correctly, otherwise MC-information is permanently lost! Migrated from https://code.icecube.wisc.edu/ticket/1942 ```json { "status": "closed", "changetime": "2017-02-21T17:32:43", "description": "I was again on the search for simulation of some MCstudies, for which I need the complete I3MCPulse-info (MCPulseMap, ParticleIDMap and I3MCTRee); It took me some while to locate an appropriate one, which I found in dataset 11843 resp. 11642 (IC86:2013 GENIE simulation), one of the latest produced datasets\n\nThis dataset was produced with icesim.V04-01-11-G4C (whatever this means). All other datasets I found which are produced with an olderor similar release candidate (icesim.V04-01-11) have the necessary objects, but the after-processing for Noise-MCPulses is flawed: there are no entries in the ParticleIDMap for those pulses as originated from ParticleID(0,0)\n\nThis can be identified by inspecting the I3MCPulseSeriesParticleIDMap in the frame finding no entries for the major/minor-ID(0,0) and in general a lot of entries with only a single OMKey as key and an empty vector for the value in this I3Map<OMKey, vector<...> >\n\nSo whatever is the difference between these meta-releases, somebody should have a look into that the after-processing is done correctly, otherwise MC-information is permanently lost!", "reporter": "mzoll", "cc": "olivas, cweaver", "resolution": "fixed", "_ts": "1487698363713061", "component": "combo simulation", "summary": "[simulation] error in MC afterprocessing (MCPulseSeriesParticleIDMap)", "priority": "critical", "keywords": "MC simulation MCPulse", "time": "2017-01-26T09:02:15", "milestone": "", "owner": "juancarlos", "type": "defect" } ```
1.0
[simulation] error in MC afterprocessing (MCPulseSeriesParticleIDMap) (Trac #1942) - I was again on the search for simulation of some MCstudies, for which I need the complete I3MCPulse-info (MCPulseMap, ParticleIDMap and I3MCTRee); It took me some while to locate an appropriate one, which I found in dataset 11843 resp. 11642 (IC86:2013 GENIE simulation), one of the latest produced datasets This dataset was produced with icesim.V04-01-11-G4C (whatever this means). All other datasets I found which are produced with an olderor similar release candidate (icesim.V04-01-11) have the necessary objects, but the after-processing for Noise-MCPulses is flawed: there are no entries in the ParticleIDMap for those pulses as originated from ParticleID(0,0) This can be identified by inspecting the I3MCPulseSeriesParticleIDMap in the frame finding no entries for the major/minor-ID(0,0) and in general a lot of entries with only a single OMKey as key and an empty vector for the value in this I3Map<OMKey, vector<...> > So whatever is the difference between these meta-releases, somebody should have a look into that the after-processing is done correctly, otherwise MC-information is permanently lost! Migrated from https://code.icecube.wisc.edu/ticket/1942 ```json { "status": "closed", "changetime": "2017-02-21T17:32:43", "description": "I was again on the search for simulation of some MCstudies, for which I need the complete I3MCPulse-info (MCPulseMap, ParticleIDMap and I3MCTRee); It took me some while to locate an appropriate one, which I found in dataset 11843 resp. 11642 (IC86:2013 GENIE simulation), one of the latest produced datasets\n\nThis dataset was produced with icesim.V04-01-11-G4C (whatever this means). All other datasets I found which are produced with an olderor similar release candidate (icesim.V04-01-11) have the necessary objects, but the after-processing for Noise-MCPulses is flawed: there are no entries in the ParticleIDMap for those pulses as originated from ParticleID(0,0)\n\nThis can be identified by inspecting the I3MCPulseSeriesParticleIDMap in the frame finding no entries for the major/minor-ID(0,0) and in general a lot of entries with only a single OMKey as key and an empty vector for the value in this I3Map<OMKey, vector<...> >\n\nSo whatever is the difference between these meta-releases, somebody should have a look into that the after-processing is done correctly, otherwise MC-information is permanently lost!", "reporter": "mzoll", "cc": "olivas, cweaver", "resolution": "fixed", "_ts": "1487698363713061", "component": "combo simulation", "summary": "[simulation] error in MC afterprocessing (MCPulseSeriesParticleIDMap)", "priority": "critical", "keywords": "MC simulation MCPulse", "time": "2017-01-26T09:02:15", "milestone": "", "owner": "juancarlos", "type": "defect" } ```
defect
error in mc afterprocessing mcpulseseriesparticleidmap trac i was again on the search for simulation of some mcstudies for which i need the complete info mcpulsemap particleidmap and it took me some while to locate an appropriate one which i found in dataset resp genie simulation one of the latest produced datasets this dataset was produced with icesim whatever this means all other datasets i found which are produced with an olderor similar release candidate icesim have the necessary objects but the after processing for noise mcpulses is flawed there are no entries in the particleidmap for those pulses as originated from particleid this can be identified by inspecting the in the frame finding no entries for the major minor id and in general a lot of entries with only a single omkey as key and an empty vector for the value in this so whatever is the difference between these meta releases somebody should have a look into that the after processing is done correctly otherwise mc information is permanently lost migrated from json status closed changetime description i was again on the search for simulation of some mcstudies for which i need the complete info mcpulsemap particleidmap and it took me some while to locate an appropriate one which i found in dataset resp genie simulation one of the latest produced datasets n nthis dataset was produced with icesim whatever this means all other datasets i found which are produced with an olderor similar release candidate icesim have the necessary objects but the after processing for noise mcpulses is flawed there are no entries in the particleidmap for those pulses as originated from particleid n nthis can be identified by inspecting the in the frame finding no entries for the major minor id and in general a lot of entries with only a single omkey as key and an empty vector for the value in this n nso whatever is the difference between these meta releases somebody should have a look into that the after processing is done correctly otherwise mc information is permanently lost reporter mzoll cc olivas cweaver resolution fixed ts component combo simulation summary error in mc afterprocessing mcpulseseriesparticleidmap priority critical keywords mc simulation mcpulse time milestone owner juancarlos type defect
1
21,155
3,463,279,598
IssuesEvent
2015-12-21 09:01:44
netty/netty
https://api.github.com/repos/netty/netty
closed
HTTP/2 DefaultHttp2RemoteFlowController Stream writability notification broken
defect
`DefaultHttp2RemoteFlowController.ListenerWritabilityMonitor` no longer reliably detects when a stream's writability change occurs. `ListenerWritabilityMonitor` was implemented to avoid duplicating iteration over all streams when possible and instead was relying on the `PriorityStreamByteDistributor` to call `write` for each stream during its iteration process. However the new `StreamByteDistributor` classes do not do an iteration over all active streams and so this assumption is now invalid. The impact is isolated to stream writability change notifications and has no impact unless you explicitly add a listener, and use an allocator other than `PriorityStreamByteDistributor`.
1.0
HTTP/2 DefaultHttp2RemoteFlowController Stream writability notification broken - `DefaultHttp2RemoteFlowController.ListenerWritabilityMonitor` no longer reliably detects when a stream's writability change occurs. `ListenerWritabilityMonitor` was implemented to avoid duplicating iteration over all streams when possible and instead was relying on the `PriorityStreamByteDistributor` to call `write` for each stream during its iteration process. However the new `StreamByteDistributor` classes do not do an iteration over all active streams and so this assumption is now invalid. The impact is isolated to stream writability change notifications and has no impact unless you explicitly add a listener, and use an allocator other than `PriorityStreamByteDistributor`.
defect
http stream writability notification broken listenerwritabilitymonitor no longer reliably detects when a stream s writability change occurs listenerwritabilitymonitor was implemented to avoid duplicating iteration over all streams when possible and instead was relying on the prioritystreambytedistributor to call write for each stream during its iteration process however the new streambytedistributor classes do not do an iteration over all active streams and so this assumption is now invalid the impact is isolated to stream writability change notifications and has no impact unless you explicitly add a listener and use an allocator other than prioritystreambytedistributor
1
23,843
3,863,220,828
IssuesEvent
2016-04-08 08:26:55
FLawrence/brat-linked
https://api.github.com/repos/FLawrence/brat-linked
closed
Automated Bechdel Tester
auto-migrated Priority-High Type-Defect
``` Need page which runs the Bechdel sparql query on selected document ``` Original issue reported on code.google.com by `k.faith....@googlemail.com` on 4 Jul 2014 at 11:36
1.0
Automated Bechdel Tester - ``` Need page which runs the Bechdel sparql query on selected document ``` Original issue reported on code.google.com by `k.faith....@googlemail.com` on 4 Jul 2014 at 11:36
defect
automated bechdel tester need page which runs the bechdel sparql query on selected document original issue reported on code google com by k faith googlemail com on jul at
1
46,123
9,884,989,999
IssuesEvent
2019-06-25 00:25:51
AstrideUG/DarkBedrock
https://api.github.com/repos/AstrideUG/DarkBedrock
opened
Add AbstractCommand(Map<String, Any?>)
WithCodeExample enhancement feature_request good first issue help wanted
needs: #154 ```kotlin import net.darkdevelopers.darkbedrock.darkness.general.configs.default import net.darkdevelopers.darkbedrock.darkness.general.configs.getValue /** * Created on 25.06.2019 02:23. * @author Lars Artmann | LartyHD */ abstract class AbstractCommand(values: Map<String, Any?>) : Command { override val permission by values.default { "" } override val permissionMessage by values.default { "" } override val usage by values.default { "" } override val minLength by values.default { 0 } override val maxLength by values.default { 0 } override val aliases by values.default { setOf<String>() } } ```
1.0
Add AbstractCommand(Map<String, Any?>) - needs: #154 ```kotlin import net.darkdevelopers.darkbedrock.darkness.general.configs.default import net.darkdevelopers.darkbedrock.darkness.general.configs.getValue /** * Created on 25.06.2019 02:23. * @author Lars Artmann | LartyHD */ abstract class AbstractCommand(values: Map<String, Any?>) : Command { override val permission by values.default { "" } override val permissionMessage by values.default { "" } override val usage by values.default { "" } override val minLength by values.default { 0 } override val maxLength by values.default { 0 } override val aliases by values.default { setOf<String>() } } ```
non_defect
add abstractcommand map needs kotlin import net darkdevelopers darkbedrock darkness general configs default import net darkdevelopers darkbedrock darkness general configs getvalue created on author lars artmann lartyhd abstract class abstractcommand values map command override val permission by values default override val permissionmessage by values default override val usage by values default override val minlength by values default override val maxlength by values default override val aliases by values default setof
0
532
2,564,053,055
IssuesEvent
2015-02-06 17:11:48
hlsyounes/vhpop
https://api.github.com/repos/hlsyounes/vhpop
closed
VHPOP-2.2 does not compile with newer version of GCC
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Downloading vhpop-2.2.tar.gz_FILES and unzipping it 2. type ./configure in the terminal 3. type make What is the expected output? What do you see instead? The code should get installed without errors. But what I see is as follows: source='vhpop.cc' object='vhpop.o' libtool=no \ depfile='.deps/vhpop.Po' tmpdepfile='.deps/vhpop.TPo' \ depmode=gcc3 /bin/sh ./depcomp \ g++ -DHAVE_CONFIG_H -I. -I. -I. -g -O2 -c -o vhpop.o `test -f 'vhpop.cc' || echo './'`vhpop.cc hashing.h:55: error: ‘hash’ is not a template orderings.h:159: error: ‘hashing::hash_map’ has not been declared orderings.h:159: error: expected ‘,’ or ‘...’ before ‘<’ token orderings.h:164: error: expected template-name before ‘<’ token orderings.h:164: error: expected `{' before ‘<’ token orderings.h:164: error: expected unqualified-id before ‘<’ token orderings.h:168: error: ‘const_iterator’ in class ‘Orderings::IdMap’ does not name a type orderings.h:171: error: field ‘id_map1_’ has incomplete type orderings.h:189: error: ‘hashing::hash_map’ has not been declared orderings.h:189: error: expected ‘,’ or ‘...’ before ‘<’ token orderings.h:191: error: ‘virtual float Orderings::schedule(int) const’ and ‘float Orderings::schedule(int) const’ cannot be overloaded orderings.h:118: confused by earlier errors, bailing out make: *** [vhpop.o] Error 1 [mast@~/Desktop/RIP1/vhpop-2.2.tar.gz_FILES/vhpop-2.2]$ make clean test -z "vhpop" || rm -f vhpop rm -f *.o core *.core What version of the product are you using? On what operating system? vhpop-2.2 Please provide any additional information below. I am using it on Fedora Core 4 ``` ----- Original issue reported on code.google.com by zafarmun...@gmail.com on 19 Sep 2011 at 5:46
1.0
VHPOP-2.2 does not compile with newer version of GCC - ``` What steps will reproduce the problem? 1. Downloading vhpop-2.2.tar.gz_FILES and unzipping it 2. type ./configure in the terminal 3. type make What is the expected output? What do you see instead? The code should get installed without errors. But what I see is as follows: source='vhpop.cc' object='vhpop.o' libtool=no \ depfile='.deps/vhpop.Po' tmpdepfile='.deps/vhpop.TPo' \ depmode=gcc3 /bin/sh ./depcomp \ g++ -DHAVE_CONFIG_H -I. -I. -I. -g -O2 -c -o vhpop.o `test -f 'vhpop.cc' || echo './'`vhpop.cc hashing.h:55: error: ‘hash’ is not a template orderings.h:159: error: ‘hashing::hash_map’ has not been declared orderings.h:159: error: expected ‘,’ or ‘...’ before ‘<’ token orderings.h:164: error: expected template-name before ‘<’ token orderings.h:164: error: expected `{' before ‘<’ token orderings.h:164: error: expected unqualified-id before ‘<’ token orderings.h:168: error: ‘const_iterator’ in class ‘Orderings::IdMap’ does not name a type orderings.h:171: error: field ‘id_map1_’ has incomplete type orderings.h:189: error: ‘hashing::hash_map’ has not been declared orderings.h:189: error: expected ‘,’ or ‘...’ before ‘<’ token orderings.h:191: error: ‘virtual float Orderings::schedule(int) const’ and ‘float Orderings::schedule(int) const’ cannot be overloaded orderings.h:118: confused by earlier errors, bailing out make: *** [vhpop.o] Error 1 [mast@~/Desktop/RIP1/vhpop-2.2.tar.gz_FILES/vhpop-2.2]$ make clean test -z "vhpop" || rm -f vhpop rm -f *.o core *.core What version of the product are you using? On what operating system? vhpop-2.2 Please provide any additional information below. I am using it on Fedora Core 4 ``` ----- Original issue reported on code.google.com by zafarmun...@gmail.com on 19 Sep 2011 at 5:46
defect
vhpop does not compile with newer version of gcc what steps will reproduce the problem downloading vhpop tar gz files and unzipping it type configure in the terminal type make what is the expected output what do you see instead the code should get installed without errors but what i see is as follows source vhpop cc object vhpop o libtool no depfile deps vhpop po tmpdepfile deps vhpop tpo depmode bin sh depcomp g dhave config h i i i g c o vhpop o test f vhpop cc echo vhpop cc hashing h error ‘hash’ is not a template orderings h error ‘hashing hash map’ has not been declared orderings h error expected ‘ ’ or ‘ ’ before ‘ ’ token orderings h error expected template name before ‘ ’ token orderings h error expected before ‘ ’ token orderings h error expected unqualified id before ‘ ’ token orderings h error ‘const iterator’ in class ‘orderings idmap’ does not name a type orderings h error field ‘id ’ has incomplete type orderings h error ‘hashing hash map’ has not been declared orderings h error expected ‘ ’ or ‘ ’ before ‘ ’ token orderings h error ‘virtual float orderings schedule int const’ and ‘float orderings schedule int const’ cannot be overloaded orderings h confused by earlier errors bailing out make error make clean test z vhpop rm f vhpop rm f o core core what version of the product are you using on what operating system vhpop please provide any additional information below i am using it on fedora core original issue reported on code google com by zafarmun gmail com on sep at
1
27,315
5,331,360,084
IssuesEvent
2017-02-15 19:18:21
nexB/aboutcode-manager
https://api.github.com/repos/nexB/aboutcode-manager
opened
Add fuse 2.9.1 Scan data to /samples and document
documentation user testing
We want to add several (5-6) ScanCode data files to the /samples for ABC Manager starting with jfuse 2.9.1 as an example of a Linux userspace application written in C. This task is to: - Add a current ScanCode JSON file for fuse 2.91. to the /samples directory - Create a new section in the README for Samples which should include information such as a brief description of the Package, link to download it and documentation of the version ScanCode and options used to create the JSON file. For each new Sample Scan, we will need to follow the same process and plan for updating the Samples for new releases of ScanCode Toolkit.
1.0
Add fuse 2.9.1 Scan data to /samples and document - We want to add several (5-6) ScanCode data files to the /samples for ABC Manager starting with jfuse 2.9.1 as an example of a Linux userspace application written in C. This task is to: - Add a current ScanCode JSON file for fuse 2.91. to the /samples directory - Create a new section in the README for Samples which should include information such as a brief description of the Package, link to download it and documentation of the version ScanCode and options used to create the JSON file. For each new Sample Scan, we will need to follow the same process and plan for updating the Samples for new releases of ScanCode Toolkit.
non_defect
add fuse scan data to samples and document we want to add several scancode data files to the samples for abc manager starting with jfuse as an example of a linux userspace application written in c this task is to add a current scancode json file for fuse to the samples directory create a new section in the readme for samples which should include information such as a brief description of the package link to download it and documentation of the version scancode and options used to create the json file for each new sample scan we will need to follow the same process and plan for updating the samples for new releases of scancode toolkit
0
4,099
2,610,087,373
IssuesEvent
2015-02-26 18:26:32
chrsmith/dsdsdaadf
https://api.github.com/repos/chrsmith/dsdsdaadf
opened
深圳痤疮哪家好
auto-migrated Priority-Medium Type-Defect
``` 深圳痤疮哪家好【深圳韩方科颜全国热线400-869-1818,24小时QQ4 008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方�� �—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科� ��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康 祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治�� �粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘� �� ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:19
1.0
深圳痤疮哪家好 - ``` 深圳痤疮哪家好【深圳韩方科颜全国热线400-869-1818,24小时QQ4 008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方�� �—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科� ��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康 祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治�� �粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘� �� ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:19
defect
深圳痤疮哪家好 深圳痤疮哪家好【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方�� �—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科� ��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康 祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治�� �粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘� �� original issue reported on code google com by szft com on may at
1
63,966
18,092,323,485
IssuesEvent
2021-09-22 04:06:32
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Window Briefly Flashes White While Launching
T-Defect
### Steps to reproduce Launch Element. ### What happened? The window briefly flashes white before the loading animation is shown. I have uploaded a short screen recording of the issue to YouTube [here](https://youtu.be/tGlEVuesQjk). ### Operating system macOS 11.5.2 ### Application version 1.8.5 ### How did you install the app? from https://element.io/get-started ### Homeserver matrix.org ### Have you submitted a rageshake? No
1.0
Window Briefly Flashes White While Launching - ### Steps to reproduce Launch Element. ### What happened? The window briefly flashes white before the loading animation is shown. I have uploaded a short screen recording of the issue to YouTube [here](https://youtu.be/tGlEVuesQjk). ### Operating system macOS 11.5.2 ### Application version 1.8.5 ### How did you install the app? from https://element.io/get-started ### Homeserver matrix.org ### Have you submitted a rageshake? No
defect
window briefly flashes white while launching steps to reproduce launch element what happened the window briefly flashes white before the loading animation is shown i have uploaded a short screen recording of the issue to youtube operating system macos application version how did you install the app from homeserver matrix org have you submitted a rageshake no
1
201,382
15,802,250,468
IssuesEvent
2021-04-03 08:48:57
Leeyp/ped
https://api.github.com/repos/Leeyp/ped
opened
Incorrect example given in UG for testing (NurseSchedule)
severity.High type.DocumentationBug
The example given in the UG is incorrect. Testers who copy and paste it will not be able to proceed: ![image.png](https://raw.githubusercontent.com/Leeyp/ped/main/files/3431b173-5c69-4b84-8581-478232b7aa23.png) <!--session: 1617437426093-73ed2492-c57e-46c7-8165-64bd8167d98a-->
1.0
Incorrect example given in UG for testing (NurseSchedule) - The example given in the UG is incorrect. Testers who copy and paste it will not be able to proceed: ![image.png](https://raw.githubusercontent.com/Leeyp/ped/main/files/3431b173-5c69-4b84-8581-478232b7aa23.png) <!--session: 1617437426093-73ed2492-c57e-46c7-8165-64bd8167d98a-->
non_defect
incorrect example given in ug for testing nurseschedule the example given in the ug is incorrect testers who copy and paste it will not be able to proceed
0
789,067
27,777,270,795
IssuesEvent
2023-03-16 18:07:15
AY2223S2-CS2113-T13-4/tp
https://api.github.com/repos/AY2223S2-CS2113-T13-4/tp
closed
Create function for task count and module count
type.Story type.Enhancement priority.High
Iterate through lists and count outstanding (unmarked) tasked and modules
1.0
Create function for task count and module count - Iterate through lists and count outstanding (unmarked) tasked and modules
non_defect
create function for task count and module count iterate through lists and count outstanding unmarked tasked and modules
0
78,128
27,332,086,485
IssuesEvent
2023-02-25 19:05:57
amyjko/bookish
https://api.github.com/repos/amyjko/bookish
opened
Printing is slow
defect
In many cases, books can just be massive HTML pages with tens of thousands of DOM nodes. It's just not a scalable way to print. We should generate some other approach to this, like precomputing a PDF.
1.0
Printing is slow - In many cases, books can just be massive HTML pages with tens of thousands of DOM nodes. It's just not a scalable way to print. We should generate some other approach to this, like precomputing a PDF.
defect
printing is slow in many cases books can just be massive html pages with tens of thousands of dom nodes it s just not a scalable way to print we should generate some other approach to this like precomputing a pdf
1
206,479
7,112,703,684
IssuesEvent
2018-01-17 17:53:28
coq/coq
https://api.github.com/repos/coq/coq
closed
Coq no longer builds CoqIDE with liblablgtk2-ocaml on Ubuntu
help wanted kind: regression priority: blocker
#### Version v8.7, master #### Operating system Ubuntu (tested on Bionic, Artful, Zesty, Xenial, Trusty, Precise) #### Description of the problem The `./configure` script seems to mis-identify the version of `lablgtk2` when it's installed from the liblablgtk2-ocaml package, resulting in CoqIDE not being built: See for example [this buildlog from bionic on launchpad](https://launchpadlibrarian.net/353424761/buildlog_ubuntu-bionic-i386.coq_8.master~git~201801140203+18666-0~daily307~ubuntu18.04.1_BUILDING.txt.gz): > Selecting previously unselected package liblablgtk2-ocaml. > Preparing to unpack .../142-liblablgtk2-ocaml_2.18.5+dfsg-1build1_i386.deb ... > Unpacking liblablgtk2-ocaml (2.18.5+dfsg-1build1) ... > ... > Selecting previously unselected package liblablgtk2-ocaml-dev. > Preparing to unpack .../148-liblablgtk2-ocaml-dev_2.18.5+dfsg-1build1_i386.deb ... > Unpacking liblablgtk2-ocaml-dev (2.18.5+dfsg-1build1) ... > ... > Setting up liblablgtk2-ocaml (2.18.5+dfsg-1build1) ... > ... > Setting up liblablgtk2-ocaml-dev (2.18.5+dfsg-1build1) ... > ... > ./configure -arch Linux -prefix /usr -mandir /usr/share/man -configdir /etc/xdg/coq -emacslib /usr/share/emacs/site-lisp/coq -browser "/usr/bin/x-www-browser %s &" -with-doc no -vmbyteflags "-dllib,-lcoqrun" > You have OCaml 4.05.0. Good! > You have OCamlfind 1.7.3. Good! > You have Camlp5 7.01. Good! > You have native-code compilation. Good! > LablGtk2 found (via ocamlfind, 2.16.0), but too old (required >= 2.18.3, found 2.16.0): > => no CoqIde will be built. > > Architecture : Linux > Coq VM bytecode link flags : -dllib -lcoqrun > Other bytecode link flags : > OCaml version : 4.05.0 > OCaml binaries in : /usr/bin/ > OCaml library in : /usr/lib/ocaml > OCaml flambda flags : > Camlp5 version : 7.01 > Camlp5 binaries in : /usr/bin > Camlp5 library in : +camlp5 > Native dynamic link support : true > CoqIde : no > Documentation : None > Web browser : /usr/bin/x-www-browser %s & > Coq web site : http://coq.inria.fr/ > Bytecode VM enabled : true > Native Compiler enabled : true > > Paths for true installation: > - the Coq binaries will be copied in /usr/bin > - the Coq library will be copied in /usr/lib/coq > - the Coqide configuration files will be copied in /etc/xdg/coq > - the Coqide data files will be copied in /usr/share/coq > - the Coq man pages will be copied in /usr/share/man > - the Coq documentation will be copied in /usr/share/doc/coq > - the Coq Emacs mode will be copied in /usr/share/emacs/site-lisp/coq > - the Coqdoc LaTeX files will be copied in /usr/share/texmf/tex/latex/misc > > If anything is wrong above, please restart './configure'. > > *Warning* To compile the system for a new architecture > don't forget to do a 'make clean' before './configure'.
1.0
Coq no longer builds CoqIDE with liblablgtk2-ocaml on Ubuntu - #### Version v8.7, master #### Operating system Ubuntu (tested on Bionic, Artful, Zesty, Xenial, Trusty, Precise) #### Description of the problem The `./configure` script seems to mis-identify the version of `lablgtk2` when it's installed from the liblablgtk2-ocaml package, resulting in CoqIDE not being built: See for example [this buildlog from bionic on launchpad](https://launchpadlibrarian.net/353424761/buildlog_ubuntu-bionic-i386.coq_8.master~git~201801140203+18666-0~daily307~ubuntu18.04.1_BUILDING.txt.gz): > Selecting previously unselected package liblablgtk2-ocaml. > Preparing to unpack .../142-liblablgtk2-ocaml_2.18.5+dfsg-1build1_i386.deb ... > Unpacking liblablgtk2-ocaml (2.18.5+dfsg-1build1) ... > ... > Selecting previously unselected package liblablgtk2-ocaml-dev. > Preparing to unpack .../148-liblablgtk2-ocaml-dev_2.18.5+dfsg-1build1_i386.deb ... > Unpacking liblablgtk2-ocaml-dev (2.18.5+dfsg-1build1) ... > ... > Setting up liblablgtk2-ocaml (2.18.5+dfsg-1build1) ... > ... > Setting up liblablgtk2-ocaml-dev (2.18.5+dfsg-1build1) ... > ... > ./configure -arch Linux -prefix /usr -mandir /usr/share/man -configdir /etc/xdg/coq -emacslib /usr/share/emacs/site-lisp/coq -browser "/usr/bin/x-www-browser %s &" -with-doc no -vmbyteflags "-dllib,-lcoqrun" > You have OCaml 4.05.0. Good! > You have OCamlfind 1.7.3. Good! > You have Camlp5 7.01. Good! > You have native-code compilation. Good! > LablGtk2 found (via ocamlfind, 2.16.0), but too old (required >= 2.18.3, found 2.16.0): > => no CoqIde will be built. > > Architecture : Linux > Coq VM bytecode link flags : -dllib -lcoqrun > Other bytecode link flags : > OCaml version : 4.05.0 > OCaml binaries in : /usr/bin/ > OCaml library in : /usr/lib/ocaml > OCaml flambda flags : > Camlp5 version : 7.01 > Camlp5 binaries in : /usr/bin > Camlp5 library in : +camlp5 > Native dynamic link support : true > CoqIde : no > Documentation : None > Web browser : /usr/bin/x-www-browser %s & > Coq web site : http://coq.inria.fr/ > Bytecode VM enabled : true > Native Compiler enabled : true > > Paths for true installation: > - the Coq binaries will be copied in /usr/bin > - the Coq library will be copied in /usr/lib/coq > - the Coqide configuration files will be copied in /etc/xdg/coq > - the Coqide data files will be copied in /usr/share/coq > - the Coq man pages will be copied in /usr/share/man > - the Coq documentation will be copied in /usr/share/doc/coq > - the Coq Emacs mode will be copied in /usr/share/emacs/site-lisp/coq > - the Coqdoc LaTeX files will be copied in /usr/share/texmf/tex/latex/misc > > If anything is wrong above, please restart './configure'. > > *Warning* To compile the system for a new architecture > don't forget to do a 'make clean' before './configure'.
non_defect
coq no longer builds coqide with ocaml on ubuntu version master operating system ubuntu tested on bionic artful zesty xenial trusty precise description of the problem the configure script seems to mis identify the version of when it s installed from the ocaml package resulting in coqide not being built see for example selecting previously unselected package ocaml preparing to unpack ocaml dfsg deb unpacking ocaml dfsg selecting previously unselected package ocaml dev preparing to unpack ocaml dev dfsg deb unpacking ocaml dev dfsg setting up ocaml dfsg setting up ocaml dev dfsg configure arch linux prefix usr mandir usr share man configdir etc xdg coq emacslib usr share emacs site lisp coq browser usr bin x www browser s with doc no vmbyteflags dllib lcoqrun you have ocaml good you have ocamlfind good you have good you have native code compilation good found via ocamlfind but too old required found no coqide will be built architecture linux coq vm bytecode link flags dllib lcoqrun other bytecode link flags ocaml version ocaml binaries in usr bin ocaml library in usr lib ocaml ocaml flambda flags version binaries in usr bin library in native dynamic link support true coqide no documentation none web browser usr bin x www browser s coq web site bytecode vm enabled true native compiler enabled true paths for true installation the coq binaries will be copied in usr bin the coq library will be copied in usr lib coq the coqide configuration files will be copied in etc xdg coq the coqide data files will be copied in usr share coq the coq man pages will be copied in usr share man the coq documentation will be copied in usr share doc coq the coq emacs mode will be copied in usr share emacs site lisp coq the coqdoc latex files will be copied in usr share texmf tex latex misc if anything is wrong above please restart configure warning to compile the system for a new architecture don t forget to do a make clean before configure
0
8,722
2,611,537,296
IssuesEvent
2015-02-27 06:06:42
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
Manual placement animation is drawn twice
auto-migrated Priority-Low Type-Defect
``` Reproduction steps: 1. Make sure manual placement is activated in game scheme 2. Start match 3. Place hedgehogs Expected outcome: The teleportation animation is drawn where the hedgehog is placed, but nowhere else. Actual outcome: The teleportation animation is drawn where the hedgehog is placed but also at the top left position of the map. Tested in 0.9.20 under GNU/Linux. Here you can download a demo for 0.9.20 which demonstrates the bug: http://www.hedgewars.org/files/demos/Wuzzy/Double_Teleport_Animation_Bug.47.hwd At the beginning, 2 hedgehogs are placed by the manual placement method. After that, some ordinary teleportations happen. In my Hedgewars, the normal teleport does not cause the graphical glitch, only the manual placement does. ``` Original issue reported on code.google.com by `almikes@aol.com` on 23 Nov 2014 at 3:29
1.0
Manual placement animation is drawn twice - ``` Reproduction steps: 1. Make sure manual placement is activated in game scheme 2. Start match 3. Place hedgehogs Expected outcome: The teleportation animation is drawn where the hedgehog is placed, but nowhere else. Actual outcome: The teleportation animation is drawn where the hedgehog is placed but also at the top left position of the map. Tested in 0.9.20 under GNU/Linux. Here you can download a demo for 0.9.20 which demonstrates the bug: http://www.hedgewars.org/files/demos/Wuzzy/Double_Teleport_Animation_Bug.47.hwd At the beginning, 2 hedgehogs are placed by the manual placement method. After that, some ordinary teleportations happen. In my Hedgewars, the normal teleport does not cause the graphical glitch, only the manual placement does. ``` Original issue reported on code.google.com by `almikes@aol.com` on 23 Nov 2014 at 3:29
defect
manual placement animation is drawn twice reproduction steps make sure manual placement is activated in game scheme start match place hedgehogs expected outcome the teleportation animation is drawn where the hedgehog is placed but nowhere else actual outcome the teleportation animation is drawn where the hedgehog is placed but also at the top left position of the map tested in under gnu linux here you can download a demo for which demonstrates the bug at the beginning hedgehogs are placed by the manual placement method after that some ordinary teleportations happen in my hedgewars the normal teleport does not cause the graphical glitch only the manual placement does original issue reported on code google com by almikes aol com on nov at
1
26,352
4,682,386,796
IssuesEvent
2016-10-09 08:02:23
luigirizzo/netmap
https://api.github.com/repos/luigirizzo/netmap
closed
Something wrong with Netmap bridge
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1.[sjm@isms examples]./bridge -i netmap:eth0 -i netmap:eth1 now everythings ok. 2.pull out eth1's cable ,then insert it back to eth1. now the bridge has been broken. 3.then "ifdown eth1" and "ifup eth1",the bridge works well again. What is the expected output? What do you see instead? --On FreeBSD, it works well ,when i insert the cable back ,the bridge works fine What version of the product are you using? On what operating system? --I think it is the latest version. I download it from here at the day before yesterday (Dec. 23 2014) I use it on Centos 6.5 ,kernel version is 3.0.88. FreeBSD version is 10.1 My ethernet device is Intel Corporation 82583V Gigabit Network Connection Drivers are e1000e(linux) and em(FreeBSD). Please provide any additional information below. --When I try to add some debug print,i find a strange thing. if I add some print code like D("just test"); into the function "netmap_common_irq",the bridge works again! But it is unstable,not every packet was transmited. code like this: 2995 } else { /* TX path */ 2996 if (q >= na->num_tx_rings) 2997 return; // not a physical queue 2998 kring = na->tx_rings + q; 2999 na->nm_notify(na, q, NR_TX, 0); 3000 } 3001 // D("SJM:just test\n"); 3002 } /* netmap common irq*/ the file is $SRC/netmap/sys/dev/netmap/netmap.c ``` Original issue reported on code.google.com by `sjm....@gmail.com` on 25 Dec 2014 at 9:28
1.0
Something wrong with Netmap bridge - ``` What steps will reproduce the problem? 1.[sjm@isms examples]./bridge -i netmap:eth0 -i netmap:eth1 now everythings ok. 2.pull out eth1's cable ,then insert it back to eth1. now the bridge has been broken. 3.then "ifdown eth1" and "ifup eth1",the bridge works well again. What is the expected output? What do you see instead? --On FreeBSD, it works well ,when i insert the cable back ,the bridge works fine What version of the product are you using? On what operating system? --I think it is the latest version. I download it from here at the day before yesterday (Dec. 23 2014) I use it on Centos 6.5 ,kernel version is 3.0.88. FreeBSD version is 10.1 My ethernet device is Intel Corporation 82583V Gigabit Network Connection Drivers are e1000e(linux) and em(FreeBSD). Please provide any additional information below. --When I try to add some debug print,i find a strange thing. if I add some print code like D("just test"); into the function "netmap_common_irq",the bridge works again! But it is unstable,not every packet was transmited. code like this: 2995 } else { /* TX path */ 2996 if (q >= na->num_tx_rings) 2997 return; // not a physical queue 2998 kring = na->tx_rings + q; 2999 na->nm_notify(na, q, NR_TX, 0); 3000 } 3001 // D("SJM:just test\n"); 3002 } /* netmap common irq*/ the file is $SRC/netmap/sys/dev/netmap/netmap.c ``` Original issue reported on code.google.com by `sjm....@gmail.com` on 25 Dec 2014 at 9:28
defect
something wrong with netmap bridge what steps will reproduce the problem bridge i netmap i netmap now everythings ok pull out s cable then insert it back to now the bridge has been broken then ifdown and ifup the bridge works well again what is the expected output what do you see instead on freebsd it works well when i insert the cable back the bridge works fine what version of the product are you using on what operating system i think it is the latest version i download it from here at the day before yesterday dec i use it on centos kernel version is freebsd version is my ethernet device is intel corporation gigabit network connection drivers are linux and em freebsd please provide any additional information below when i try to add some debug print i find a strange thing if i add some print code like d just test into the function netmap common irq the bridge works again but it is unstable not every packet was transmited code like this else tx path if q na num tx rings return not a physical queue kring na tx rings q na nm notify na q nr tx d sjm just test n netmap common irq the file is src netmap sys dev netmap netmap c original issue reported on code google com by sjm gmail com on dec at
1
14,749
25,578,820,420
IssuesEvent
2022-12-01 01:27:55
CameronAuler/ENSM
https://api.github.com/repos/CameronAuler/ENSM
closed
Research and Design Project 2 Check In and Demo
course requirement
- [x] Meet with Devin to go over Research and Design Project 2. - [x] Create a research and design project 2 demo.
1.0
Research and Design Project 2 Check In and Demo - - [x] Meet with Devin to go over Research and Design Project 2. - [x] Create a research and design project 2 demo.
non_defect
research and design project check in and demo meet with devin to go over research and design project create a research and design project demo
0
115,780
17,334,578,068
IssuesEvent
2021-07-28 08:39:11
panasalap/libexpat-without-Vulnerable-Fix
https://api.github.com/repos/panasalap/libexpat-without-Vulnerable-Fix
opened
CVE-2018-20843 (High) detected in libexpatR_2_2_6
security vulnerability
## CVE-2018-20843 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>libexpatR_2_2_6</b></p></summary> <p> <p>:herb: Expat library: Fast streaming XML parser written in C99; in the process of migrating from SourceForge to GitHub</p> <p>Library home page: <a href=https://github.com/libexpat/libexpat.git>https://github.com/libexpat/libexpat.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/panasalap/libexpat-without-Vulnerable-Fix/commit/8a257a2ef5f63fe673edc667bfe62abe6f6502f3">8a257a2ef5f63fe673edc667bfe62abe6f6502f3</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>libexpat-without-Vulnerable-Fix/expat/lib/xmlparse.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In libexpat in Expat before 2.2.7, XML input including XML names that contain a large number of colons could make the XML parser consume a high amount of RAM and CPU resources while processing (enough to be usable for denial-of-service attacks). <p>Publish Date: 2019-06-24 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20843>CVE-2018-20843</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20843">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20843</a></p> <p>Release Date: 2019-06-24</p> <p>Fix Resolution: 2.2.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-20843 (High) detected in libexpatR_2_2_6 - ## CVE-2018-20843 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>libexpatR_2_2_6</b></p></summary> <p> <p>:herb: Expat library: Fast streaming XML parser written in C99; in the process of migrating from SourceForge to GitHub</p> <p>Library home page: <a href=https://github.com/libexpat/libexpat.git>https://github.com/libexpat/libexpat.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/panasalap/libexpat-without-Vulnerable-Fix/commit/8a257a2ef5f63fe673edc667bfe62abe6f6502f3">8a257a2ef5f63fe673edc667bfe62abe6f6502f3</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>libexpat-without-Vulnerable-Fix/expat/lib/xmlparse.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In libexpat in Expat before 2.2.7, XML input including XML names that contain a large number of colons could make the XML parser consume a high amount of RAM and CPU resources while processing (enough to be usable for denial-of-service attacks). <p>Publish Date: 2019-06-24 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20843>CVE-2018-20843</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20843">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20843</a></p> <p>Release Date: 2019-06-24</p> <p>Fix Resolution: 2.2.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in libexpatr cve high severity vulnerability vulnerable library libexpatr herb expat library fast streaming xml parser written in in the process of migrating from sourceforge to github library home page a href found in head commit a href found in base branch master vulnerable source files libexpat without vulnerable fix expat lib xmlparse c vulnerability details in libexpat in expat before xml input including xml names that contain a large number of colons could make the xml parser consume a high amount of ram and cpu resources while processing enough to be usable for denial of service attacks publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
164,192
20,364,357,720
IssuesEvent
2022-02-21 02:37:48
turkdevops/headless-recorder
https://api.github.com/repos/turkdevops/headless-recorder
opened
CVE-2022-0639 (Medium) detected in url-parse-1.5.3.tgz
security vulnerability
## CVE-2022-0639 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.5.3.tgz</b></p></summary> <p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p> <p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.5.3.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.5.3.tgz</a></p> <p> Dependency Hierarchy: - cli-service-4.5.0.tgz (Root Library) - webpack-dev-server-3.11.2.tgz - sockjs-client-1.5.1.tgz - :x: **url-parse-1.5.3.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.7. <p>Publish Date: 2022-02-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0639>CVE-2022-0639</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0639">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0639</a></p> <p>Release Date: 2022-02-17</p> <p>Fix Resolution: url-parse - 1.5.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-0639 (Medium) detected in url-parse-1.5.3.tgz - ## CVE-2022-0639 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.5.3.tgz</b></p></summary> <p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p> <p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.5.3.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.5.3.tgz</a></p> <p> Dependency Hierarchy: - cli-service-4.5.0.tgz (Root Library) - webpack-dev-server-3.11.2.tgz - sockjs-client-1.5.1.tgz - :x: **url-parse-1.5.3.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.7. <p>Publish Date: 2022-02-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0639>CVE-2022-0639</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0639">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0639</a></p> <p>Release Date: 2022-02-17</p> <p>Fix Resolution: url-parse - 1.5.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in url parse tgz cve medium severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href dependency hierarchy cli service tgz root library webpack dev server tgz sockjs client tgz x url parse tgz vulnerable library found in base branch master vulnerability details authorization bypass through user controlled key in npm url parse prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse step up your open source security game with whitesource
0
47,479
13,056,202,494
IssuesEvent
2020-07-30 03:58:45
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
I3Tray finish gets called due to exceptions in constructors/Configure methods (Trac #615)
IceTray Migrated from Trac defect
An exception in service or module constructors causes I3Finish to be called without there being a driving module yet, which seems to confuse I3Finish and makes it throw its own exception. The resulting second exception could be confusing to users.. Here is an example of how this looks like on the standard output: lilliput/private/minimizer/I3GSLMultiMin.cxx:105: ERROR: The "vector_bfgs2" minimizer is not available in your version of GSL. Update to version 1.14 or newer. /lilliput/private/minimizer/I3GSLMultiMin.cxx:189: FATAL: Unknown minimizer algorithm "vector_bfgs2"! Traceback (most recent call last): File "./aartfit.py", line 553, in <module> tray.Execute() File "/Users/claudio/Documents/Uni/IceTray/test/build.searecsim.release/lib/I3Tray.py", line 118, in Execute args[0].the_tray.Execute() RuntimeError: Unknown minimizer algorithm "vector_bfgs2"! I3Tray finishing... /icetray/private/icetray/I3Tray.cxx:457: FATAL: Attempt to call finish, but there is no driving module. Did you forget to call Execute()? terminate called after throwing an instance of 'std::runtime_error' what(): Attempt to call finish, but there is no driving module. Did you forget to call Execute()? Abort trap Migrated from https://code.icecube.wisc.edu/ticket/615 ```json { "status": "closed", "changetime": "2012-10-31T20:57:48", "description": "An exception in service or module constructors causes I3Finish to be called without there being a driving module yet, which seems to confuse I3Finish and makes it throw its own exception. The resulting second exception could be confusing to users..\n\n\n\nHere is an example of how this looks like on the standard output:\n\nlilliput/private/minimizer/I3GSLMultiMin.cxx:105: ERROR: The \"vector_bfgs2\" minimizer is not available in your version of GSL. Update to version 1.14 or newer.\n/lilliput/private/minimizer/I3GSLMultiMin.cxx:189: FATAL: Unknown minimizer algorithm \"vector_bfgs2\"!\nTraceback (most recent call last):\n File \"./aartfit.py\", line 553, in <module>\n tray.Execute()\n File \"/Users/claudio/Documents/Uni/IceTray/test/build.searecsim.release/lib/I3Tray.py\", line 118, in Execute\n args[0].the_tray.Execute()\nRuntimeError: Unknown minimizer algorithm \"vector_bfgs2\"!\nI3Tray finishing...\n/icetray/private/icetray/I3Tray.cxx:457: FATAL: Attempt to call finish, but there is no driving module. Did you forget to call Execute()?\nterminate called after throwing an instance of 'std::runtime_error'\n what(): Attempt to call finish, but there is no driving module. Did you forget to call Execute()?\nAbort trap\n", "reporter": "icecube", "cc": "claudio.kopper@physik.uni-erlangen.de", "resolution": "fixed", "_ts": "1351717068000000", "component": "IceTray", "summary": "I3Tray finish gets called due to exceptions in constructors/Configure methods", "priority": "normal", "keywords": "", "time": "2010-07-30T10:11:17", "milestone": "", "owner": "troy", "type": "defect" } ```
1.0
I3Tray finish gets called due to exceptions in constructors/Configure methods (Trac #615) - An exception in service or module constructors causes I3Finish to be called without there being a driving module yet, which seems to confuse I3Finish and makes it throw its own exception. The resulting second exception could be confusing to users.. Here is an example of how this looks like on the standard output: lilliput/private/minimizer/I3GSLMultiMin.cxx:105: ERROR: The "vector_bfgs2" minimizer is not available in your version of GSL. Update to version 1.14 or newer. /lilliput/private/minimizer/I3GSLMultiMin.cxx:189: FATAL: Unknown minimizer algorithm "vector_bfgs2"! Traceback (most recent call last): File "./aartfit.py", line 553, in <module> tray.Execute() File "/Users/claudio/Documents/Uni/IceTray/test/build.searecsim.release/lib/I3Tray.py", line 118, in Execute args[0].the_tray.Execute() RuntimeError: Unknown minimizer algorithm "vector_bfgs2"! I3Tray finishing... /icetray/private/icetray/I3Tray.cxx:457: FATAL: Attempt to call finish, but there is no driving module. Did you forget to call Execute()? terminate called after throwing an instance of 'std::runtime_error' what(): Attempt to call finish, but there is no driving module. Did you forget to call Execute()? Abort trap Migrated from https://code.icecube.wisc.edu/ticket/615 ```json { "status": "closed", "changetime": "2012-10-31T20:57:48", "description": "An exception in service or module constructors causes I3Finish to be called without there being a driving module yet, which seems to confuse I3Finish and makes it throw its own exception. The resulting second exception could be confusing to users..\n\n\n\nHere is an example of how this looks like on the standard output:\n\nlilliput/private/minimizer/I3GSLMultiMin.cxx:105: ERROR: The \"vector_bfgs2\" minimizer is not available in your version of GSL. Update to version 1.14 or newer.\n/lilliput/private/minimizer/I3GSLMultiMin.cxx:189: FATAL: Unknown minimizer algorithm \"vector_bfgs2\"!\nTraceback (most recent call last):\n File \"./aartfit.py\", line 553, in <module>\n tray.Execute()\n File \"/Users/claudio/Documents/Uni/IceTray/test/build.searecsim.release/lib/I3Tray.py\", line 118, in Execute\n args[0].the_tray.Execute()\nRuntimeError: Unknown minimizer algorithm \"vector_bfgs2\"!\nI3Tray finishing...\n/icetray/private/icetray/I3Tray.cxx:457: FATAL: Attempt to call finish, but there is no driving module. Did you forget to call Execute()?\nterminate called after throwing an instance of 'std::runtime_error'\n what(): Attempt to call finish, but there is no driving module. Did you forget to call Execute()?\nAbort trap\n", "reporter": "icecube", "cc": "claudio.kopper@physik.uni-erlangen.de", "resolution": "fixed", "_ts": "1351717068000000", "component": "IceTray", "summary": "I3Tray finish gets called due to exceptions in constructors/Configure methods", "priority": "normal", "keywords": "", "time": "2010-07-30T10:11:17", "milestone": "", "owner": "troy", "type": "defect" } ```
defect
finish gets called due to exceptions in constructors configure methods trac an exception in service or module constructors causes to be called without there being a driving module yet which seems to confuse and makes it throw its own exception the resulting second exception could be confusing to users here is an example of how this looks like on the standard output lilliput private minimizer cxx error the vector minimizer is not available in your version of gsl update to version or newer lilliput private minimizer cxx fatal unknown minimizer algorithm vector traceback most recent call last file aartfit py line in tray execute file users claudio documents uni icetray test build searecsim release lib py line in execute args the tray execute runtimeerror unknown minimizer algorithm vector finishing icetray private icetray cxx fatal attempt to call finish but there is no driving module did you forget to call execute terminate called after throwing an instance of std runtime error what attempt to call finish but there is no driving module did you forget to call execute abort trap migrated from json status closed changetime description an exception in service or module constructors causes to be called without there being a driving module yet which seems to confuse and makes it throw its own exception the resulting second exception could be confusing to users n n n nhere is an example of how this looks like on the standard output n nlilliput private minimizer cxx error the vector minimizer is not available in your version of gsl update to version or newer n lilliput private minimizer cxx fatal unknown minimizer algorithm vector ntraceback most recent call last n file aartfit py line in n tray execute n file users claudio documents uni icetray test build searecsim release lib py line in execute n args the tray execute nruntimeerror unknown minimizer algorithm vector finishing n icetray private icetray cxx fatal attempt to call finish but there is no driving module did you forget to call execute nterminate called after throwing an instance of std runtime error n what attempt to call finish but there is no driving module did you forget to call execute nabort trap n reporter icecube cc claudio kopper physik uni erlangen de resolution fixed ts component icetray summary finish gets called due to exceptions in constructors configure methods priority normal keywords time milestone owner troy type defect
1
38,700
8,952,334,425
IssuesEvent
2019-01-25 16:16:08
svigerske/ipopt-donotuse
https://api.github.com/repos/svigerske/ipopt-donotuse
closed
Undefined methods in Ipopt::Journal class
Ipopt defect
Issue created by migration from Trac. Original creator: ycollet Original creation time: 2010-04-30 14:25:47 Assignee: ipopt-team Version: 3.8 In the file IpJournalist.[hc]pp, the class Journal is defined. The prototype of the method SetPrintLevel is defined in the class definition but the implementation is never defined. So, is this method pure virtual ? I've got no problems under linux but when compiling under windows, visual prints some errors.
1.0
Undefined methods in Ipopt::Journal class - Issue created by migration from Trac. Original creator: ycollet Original creation time: 2010-04-30 14:25:47 Assignee: ipopt-team Version: 3.8 In the file IpJournalist.[hc]pp, the class Journal is defined. The prototype of the method SetPrintLevel is defined in the class definition but the implementation is never defined. So, is this method pure virtual ? I've got no problems under linux but when compiling under windows, visual prints some errors.
defect
undefined methods in ipopt journal class issue created by migration from trac original creator ycollet original creation time assignee ipopt team version in the file ipjournalist pp the class journal is defined the prototype of the method setprintlevel is defined in the class definition but the implementation is never defined so is this method pure virtual i ve got no problems under linux but when compiling under windows visual prints some errors
1
2,759
2,607,938,708
IssuesEvent
2015-02-26 00:29:55
chrsmithdemos/minify
https://api.github.com/repos/chrsmithdemos/minify
closed
maxAge is not set to 365 days when query-string ends with number
auto-migrated Priority-Medium Release-2.1.5 Type-Defect
``` Minify commit/version: trunk What steps will reproduce the problem? 1. append a number to the query string. The config.php says: " * Note: Despite this setting, if you include a number at the end of the * querystring, maxAge will be set to one year. E.g. /min/f=hello.css&123456" Expected output: maxAge to be one year. Actual output: maxAge was not changed. Reason: regex wrong. FIX: This is the culprit: index.php if (preg_match('/&\\d/', $_SERVER['QUERY_STRING']) || isset($_GET['v'])) my suggested fix: if (preg_match('/\\d$/', $_SERVER['QUERY_STRING']) || isset($_GET['v'])) { ``` ----- Original issue reported on code.google.com by `leo.saue...@gmail.com` on 22 Jul 2014 at 10:30
1.0
maxAge is not set to 365 days when query-string ends with number - ``` Minify commit/version: trunk What steps will reproduce the problem? 1. append a number to the query string. The config.php says: " * Note: Despite this setting, if you include a number at the end of the * querystring, maxAge will be set to one year. E.g. /min/f=hello.css&123456" Expected output: maxAge to be one year. Actual output: maxAge was not changed. Reason: regex wrong. FIX: This is the culprit: index.php if (preg_match('/&\\d/', $_SERVER['QUERY_STRING']) || isset($_GET['v'])) my suggested fix: if (preg_match('/\\d$/', $_SERVER['QUERY_STRING']) || isset($_GET['v'])) { ``` ----- Original issue reported on code.google.com by `leo.saue...@gmail.com` on 22 Jul 2014 at 10:30
defect
maxage is not set to days when query string ends with number minify commit version trunk what steps will reproduce the problem append a number to the query string the config php says note despite this setting if you include a number at the end of the querystring maxage will be set to one year e g min f hello css expected output maxage to be one year actual output maxage was not changed reason regex wrong fix this is the culprit index php if preg match d server isset get my suggested fix if preg match d server isset get original issue reported on code google com by leo saue gmail com on jul at
1
17,462
3,006,634,597
IssuesEvent
2015-07-27 11:49:33
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
closed
[Array] GetValue() throws "Invalid number of indices"
defect
```C# using Bridge; using Bridge.Html5; using System; using System.Collections.Generic; using System.Linq; namespace Demo { public class App { [Ready] public static void Main() { Console.Log(new[] { "x", "y" }.GetValue(0)); } } } ``` **Actual result**: JavaScript error `"Invalid number of indices"` **Expected result**: `"x"`
1.0
[Array] GetValue() throws "Invalid number of indices" - ```C# using Bridge; using Bridge.Html5; using System; using System.Collections.Generic; using System.Linq; namespace Demo { public class App { [Ready] public static void Main() { Console.Log(new[] { "x", "y" }.GetValue(0)); } } } ``` **Actual result**: JavaScript error `"Invalid number of indices"` **Expected result**: `"x"`
defect
getvalue throws invalid number of indices c using bridge using bridge using system using system collections generic using system linq namespace demo public class app public static void main console log new x y getvalue actual result javascript error invalid number of indices expected result x
1
193,135
6,881,900,729
IssuesEvent
2017-11-21 00:40:18
minio/minio-java
https://api.github.com/repos/minio/minio-java
closed
Provide Maven support to fetch minio jar along with all dependent jar's
priority: medium
Currently, the maven entry provided here will result in just downloading the minio-3.0.9.jar and not minio-3.0.9-all.jar. I faced quite a bit of problem trying to unit test where the call to instantiate minio object was expecting the xmlPullParser class to be present but was not there in my local repo as its not part of minio-3.0.9.jar It would be helpful if the -all.jar file is fetched as part of maven artifact entry in pom.xml rather than the trimmed one.
1.0
Provide Maven support to fetch minio jar along with all dependent jar's - Currently, the maven entry provided here will result in just downloading the minio-3.0.9.jar and not minio-3.0.9-all.jar. I faced quite a bit of problem trying to unit test where the call to instantiate minio object was expecting the xmlPullParser class to be present but was not there in my local repo as its not part of minio-3.0.9.jar It would be helpful if the -all.jar file is fetched as part of maven artifact entry in pom.xml rather than the trimmed one.
non_defect
provide maven support to fetch minio jar along with all dependent jar s currently the maven entry provided here will result in just downloading the minio jar and not minio all jar i faced quite a bit of problem trying to unit test where the call to instantiate minio object was expecting the xmlpullparser class to be present but was not there in my local repo as its not part of minio jar it would be helpful if the all jar file is fetched as part of maven artifact entry in pom xml rather than the trimmed one
0
2,988
5,346,258,174
IssuesEvent
2017-02-17 19:11:05
OSU-CS361-W17/group21_project2
https://api.github.com/repos/OSU-CS361-W17/group21_project2
closed
pop up windows
Requirement
Need a confirmation window for fire will add the scan feature to that window also adding basic interactions and confirmations so player feels the game is responsive
1.0
pop up windows - Need a confirmation window for fire will add the scan feature to that window also adding basic interactions and confirmations so player feels the game is responsive
non_defect
pop up windows need a confirmation window for fire will add the scan feature to that window also adding basic interactions and confirmations so player feels the game is responsive
0
63,464
17,672,582,722
IssuesEvent
2021-08-23 08:19:55
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
"Enable encryption in settings" should not be shown to users that are unable to modify room encryption
T-Defect S-Minor Easy A-E2EE A-Timeline X-Needs-Design O-Intermediate
### Steps to reproduce 1. Join an invite-only room where you don't have permission to modify encryption settings. 2. Be shown the following warning: ![image](https://user-images.githubusercontent.com/1342360/130156702-9819eaa8-8a7a-46d4-bf4e-456f3e23db15.png) 3. Click "Enable encryption in settings" 4. Find that the setting cannot be modified, as one does not have permission to do so. ### What happened? I became confoozed. ### What did you expect? The "Enable encryption in settings" link should not be shown if the user does not have permission to modify room settings. Additionally, the text is a bit odd, as this occurred in a room with 4 other users; so this isn't really a private message. It's just an invite-only room. ### Operating system Arch Linux ### Application version Element Desktop 1.7.34 ### How did you install the app? Arch Linux User Repository ### Have you submitted a rageshake? No
1.0
"Enable encryption in settings" should not be shown to users that are unable to modify room encryption - ### Steps to reproduce 1. Join an invite-only room where you don't have permission to modify encryption settings. 2. Be shown the following warning: ![image](https://user-images.githubusercontent.com/1342360/130156702-9819eaa8-8a7a-46d4-bf4e-456f3e23db15.png) 3. Click "Enable encryption in settings" 4. Find that the setting cannot be modified, as one does not have permission to do so. ### What happened? I became confoozed. ### What did you expect? The "Enable encryption in settings" link should not be shown if the user does not have permission to modify room settings. Additionally, the text is a bit odd, as this occurred in a room with 4 other users; so this isn't really a private message. It's just an invite-only room. ### Operating system Arch Linux ### Application version Element Desktop 1.7.34 ### How did you install the app? Arch Linux User Repository ### Have you submitted a rageshake? No
defect
enable encryption in settings should not be shown to users that are unable to modify room encryption steps to reproduce join an invite only room where you don t have permission to modify encryption settings be shown the following warning click enable encryption in settings find that the setting cannot be modified as one does not have permission to do so what happened i became confoozed what did you expect the enable encryption in settings link should not be shown if the user does not have permission to modify room settings additionally the text is a bit odd as this occurred in a room with other users so this isn t really a private message it s just an invite only room operating system arch linux application version element desktop how did you install the app arch linux user repository have you submitted a rageshake no
1
30,176
24,623,093,519
IssuesEvent
2022-10-16 06:39:09
unitaryfund/mitiq
https://api.github.com/repos/unitaryfund/mitiq
closed
CI is failing on non-linux tests
bug infrastructure priority/p0
## Issue Description CI is failing on non-linux tests. An example is given by the last commits of #1442 But it's happening in many other PRs too. I am re-running tests on #1442 to see if it was just a temporary problem. If you have any ideas on how to solve this, please comment here.
1.0
CI is failing on non-linux tests - ## Issue Description CI is failing on non-linux tests. An example is given by the last commits of #1442 But it's happening in many other PRs too. I am re-running tests on #1442 to see if it was just a temporary problem. If you have any ideas on how to solve this, please comment here.
non_defect
ci is failing on non linux tests issue description ci is failing on non linux tests an example is given by the last commits of but it s happening in many other prs too i am re running tests on to see if it was just a temporary problem if you have any ideas on how to solve this please comment here
0
385,110
26,617,988,638
IssuesEvent
2023-01-24 09:02:12
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
4.0 C# `Signal.Connect` Doesn't Work as Documented in `Object.Connect` Documentation
documentation topic:dotnet
### Godot version v4.0.beta14.mono.official [28a24639c] ### System information Windows 10, .Net SDK 6.0.405 ### Issue description The [4.0 docs for `Object.Connect`](https://docs.godotengine.org/en/latest/classes/class_object.html#class-object-method-connect) suggest that [`Signal.Connect`](https://docs.godotengine.org/en/latest/classes/class_signal.html#class-signal-method-connect) (the docs for which are lacking any C# entry) or `object.event += Signal` are the preferred ways to hook up signals in C# via code. However, trying to make use of these commands result in errors which prevent building the project. Here is the code for the class containing the statements which are erroring (you can view my Player implementation in the attached reproduction project): ``` using Godot; using System; public partial class Example : Node2D { public override void _Ready() { var button = new Button(); // C# supports passing signals as events, so we can use this idiomatic construct: button.ButtonDown += OnButtonDown; // This assumes that a `Player` class exists, which defines a `Hit` signal. var player = new Player(); // Errors with the message: // The event 'Player.Hit' can only appear on the left side of += -= player.Hit.Connect(OnPlayerHit, new Godot.Collections.Array { "sword", 100 }); // Errors with the message: // 'Example.OnPlayerHit(string, int)' is a method, which is not valid in the given context player.Hit += OnPlayerHit.Bind("sword", 100); // NOTE: this statement isn't in the docs, but I used to check whether an explicit `Callable` could use the `Bind` method // Errors with the message: // 'Callable' does not contain a definition for 'Bind' and no accessible extension method 'Bind' accepting a first argument of type 'Callable' could be found (are you missing a using directive or an assembly reference?) player.Hit += new Callable(this, nameof(OnPlayerHit)).Bind("sword", 100); } private void OnButtonDown() { GD.Print("Button down!"); } private void OnPlayerHit(string weaponType, int damage) { GD.Print(String.Format("Hit with weapon {0} for {1} damage.", weaponType, damage)); } } ``` ### Steps to reproduce To reproduce, simply open the attached reproduction project and attempt to either open the `Example.cs` script in Visual Studio, or press the `Build` button in the top right of the Editor, which should result in the errors which I documented in the Issue Description. ### Minimal reproduction project [ReproduceSignalConnectIssue.zip](https://github.com/godotengine/godot/files/10476718/ReproduceSignalConnectIssue.zip)
1.0
4.0 C# `Signal.Connect` Doesn't Work as Documented in `Object.Connect` Documentation - ### Godot version v4.0.beta14.mono.official [28a24639c] ### System information Windows 10, .Net SDK 6.0.405 ### Issue description The [4.0 docs for `Object.Connect`](https://docs.godotengine.org/en/latest/classes/class_object.html#class-object-method-connect) suggest that [`Signal.Connect`](https://docs.godotengine.org/en/latest/classes/class_signal.html#class-signal-method-connect) (the docs for which are lacking any C# entry) or `object.event += Signal` are the preferred ways to hook up signals in C# via code. However, trying to make use of these commands result in errors which prevent building the project. Here is the code for the class containing the statements which are erroring (you can view my Player implementation in the attached reproduction project): ``` using Godot; using System; public partial class Example : Node2D { public override void _Ready() { var button = new Button(); // C# supports passing signals as events, so we can use this idiomatic construct: button.ButtonDown += OnButtonDown; // This assumes that a `Player` class exists, which defines a `Hit` signal. var player = new Player(); // Errors with the message: // The event 'Player.Hit' can only appear on the left side of += -= player.Hit.Connect(OnPlayerHit, new Godot.Collections.Array { "sword", 100 }); // Errors with the message: // 'Example.OnPlayerHit(string, int)' is a method, which is not valid in the given context player.Hit += OnPlayerHit.Bind("sword", 100); // NOTE: this statement isn't in the docs, but I used to check whether an explicit `Callable` could use the `Bind` method // Errors with the message: // 'Callable' does not contain a definition for 'Bind' and no accessible extension method 'Bind' accepting a first argument of type 'Callable' could be found (are you missing a using directive or an assembly reference?) player.Hit += new Callable(this, nameof(OnPlayerHit)).Bind("sword", 100); } private void OnButtonDown() { GD.Print("Button down!"); } private void OnPlayerHit(string weaponType, int damage) { GD.Print(String.Format("Hit with weapon {0} for {1} damage.", weaponType, damage)); } } ``` ### Steps to reproduce To reproduce, simply open the attached reproduction project and attempt to either open the `Example.cs` script in Visual Studio, or press the `Build` button in the top right of the Editor, which should result in the errors which I documented in the Issue Description. ### Minimal reproduction project [ReproduceSignalConnectIssue.zip](https://github.com/godotengine/godot/files/10476718/ReproduceSignalConnectIssue.zip)
non_defect
c signal connect doesn t work as documented in object connect documentation godot version mono official system information windows net sdk issue description the suggest that the docs for which are lacking any c entry or object event signal are the preferred ways to hook up signals in c via code however trying to make use of these commands result in errors which prevent building the project here is the code for the class containing the statements which are erroring you can view my player implementation in the attached reproduction project using godot using system public partial class example public override void ready var button new button c supports passing signals as events so we can use this idiomatic construct button buttondown onbuttondown this assumes that a player class exists which defines a hit signal var player new player errors with the message the event player hit can only appear on the left side of player hit connect onplayerhit new godot collections array sword errors with the message example onplayerhit string int is a method which is not valid in the given context player hit onplayerhit bind sword note this statement isn t in the docs but i used to check whether an explicit callable could use the bind method errors with the message callable does not contain a definition for bind and no accessible extension method bind accepting a first argument of type callable could be found are you missing a using directive or an assembly reference player hit new callable this nameof onplayerhit bind sword private void onbuttondown gd print button down private void onplayerhit string weapontype int damage gd print string format hit with weapon for damage weapontype damage steps to reproduce to reproduce simply open the attached reproduction project and attempt to either open the example cs script in visual studio or press the build button in the top right of the editor which should result in the errors which i documented in the issue description minimal reproduction project
0
77,535
9,595,887,886
IssuesEvent
2019-05-09 17:10:10
publiclab/spectral-workbench
https://api.github.com/repos/publiclab/spectral-workbench
closed
Frontend for SWB authentication
design
As finalised in https://github.com/publiclab/mapknitter/issues/385, @IshaGupta18 is implementing the front end as ![image](https://user-images.githubusercontent.com/20972099/56476727-4244a380-64ba-11e9-8163-a8f1b348b4e1.png) for MK. We can use the same partial files here at SWB as the code should be exactly the same. Thanks all
1.0
Frontend for SWB authentication - As finalised in https://github.com/publiclab/mapknitter/issues/385, @IshaGupta18 is implementing the front end as ![image](https://user-images.githubusercontent.com/20972099/56476727-4244a380-64ba-11e9-8163-a8f1b348b4e1.png) for MK. We can use the same partial files here at SWB as the code should be exactly the same. Thanks all
non_defect
frontend for swb authentication as finalised in is implementing the front end as for mk we can use the same partial files here at swb as the code should be exactly the same thanks all
0
71,149
23,469,595,372
IssuesEvent
2022-08-16 20:19:09
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
opened
Invalid ARIA attributes on Vet Center Facility services in CMS
Needs refining ⭐️ Sitewide CMS 508/Accessibility 508-defect-1
## Description There are invalid aria attributes on the rows within the Facility services table - aria-vc-name & aria-required-status are not valid aria attributes. ARIA can only be used with valid values. ## Description Within the edit screens of the CMS, there are multiple components that have an `aria-describedby` attribute with a value that does not corresponds to an ID anywhere on the page. ## Screenshot ![image](https://user-images.githubusercontent.com/106678594/184977268-0e9818ab-689b-47d5-b355-77dd8b202565.png) ## Accessibility Standard WCAG version 2.0 A, [Criterion 4.2.1](https://www.w3.org/WAI/WCAG21/Understanding/name-role-value.html) ## Acceptance Criteria - [ ] Determine what the original intention of the ARIA attribute was/is. What additional context was it providing? Do we still need to provide that information in some other way? - [ ] Technical review to either remove the ARIA attribute or correct it - [ ] Change Management consulted - [ ] Implementation ticket created ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [ ] `Platform CMS Team` - [ ] `Sitewide Crew` - [ ] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [ ] `⭐️ Facilities` - [ ] `⭐️ User support`
1.0
Invalid ARIA attributes on Vet Center Facility services in CMS - ## Description There are invalid aria attributes on the rows within the Facility services table - aria-vc-name & aria-required-status are not valid aria attributes. ARIA can only be used with valid values. ## Description Within the edit screens of the CMS, there are multiple components that have an `aria-describedby` attribute with a value that does not corresponds to an ID anywhere on the page. ## Screenshot ![image](https://user-images.githubusercontent.com/106678594/184977268-0e9818ab-689b-47d5-b355-77dd8b202565.png) ## Accessibility Standard WCAG version 2.0 A, [Criterion 4.2.1](https://www.w3.org/WAI/WCAG21/Understanding/name-role-value.html) ## Acceptance Criteria - [ ] Determine what the original intention of the ARIA attribute was/is. What additional context was it providing? Do we still need to provide that information in some other way? - [ ] Technical review to either remove the ARIA attribute or correct it - [ ] Change Management consulted - [ ] Implementation ticket created ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [ ] `Platform CMS Team` - [ ] `Sitewide Crew` - [ ] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [ ] `⭐️ Facilities` - [ ] `⭐️ User support`
defect
invalid aria attributes on vet center facility services in cms description there are invalid aria attributes on the rows within the facility services table aria vc name aria required status are not valid aria attributes aria can only be used with valid values description within the edit screens of the cms there are multiple components that have an aria describedby attribute with a value that does not corresponds to an id anywhere on the page screenshot accessibility standard wcag version a acceptance criteria determine what the original intention of the aria attribute was is what additional context was it providing do we still need to provide that information in some other way technical review to either remove the aria attribute or correct it change management consulted implementation ticket created cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support
1
410,767
27,801,011,554
IssuesEvent
2023-03-17 15:50:23
distributethe6ix/70DaysOfServiceMesh
https://api.github.com/repos/distributethe6ix/70DaysOfServiceMesh
closed
Add Welcome workflow #439
documentation enhancement
<!--Describe your feature --> Add Welcome workflow that can show welcome message whenever contributors open a PR. <!-- Additional context about the feature -->
1.0
Add Welcome workflow #439 - <!--Describe your feature --> Add Welcome workflow that can show welcome message whenever contributors open a PR. <!-- Additional context about the feature -->
non_defect
add welcome workflow add welcome workflow that can show welcome message whenever contributors open a pr
0
212,862
16,501,917,422
IssuesEvent
2021-05-25 15:16:35
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
closed
Mantid Basic Course: Nice to have changes
Documentation Maintenance
From Manual testing: https://github.com/mantidproject/mantid/issues/30031 - [x] Update [Downloads image](https://docs.mantidproject.org/nightly/_images/MantidDownload_42.png) to be 5.1 or 6.0 - [x] [Place larger emphasis](https://docs.mantidproject.org/nightly/tutorials/mantid_basic_course/getting_started/getting_started.html#error-reporter) on **needing contact info** for issues to be resolved - [x] [Improve Examples for loading lots of data](https://docs.mantidproject.org/nightly/tutorials/mantid_basic_course/loading_and_displaying_data/01_loading_data.html#loading-lots-of-data-files) with Load Dialog lists, say using the 2 Mari files in TCD - [x] [Align bullet points](https://docs.mantidproject.org/nightly/tutorials/mantid_basic_course/loading_and_displaying_data/03_displaying_1D_data.html#adding-a-curve-to-an-existing-plot) - [x] Update [ISIS SANS interface instructions](https://docs.mantidproject.org/nightly/tutorials/mantid_basic_course/algorithms_workspaces_and_histories/05_interfaces.html#a-simple-walkthrough) to side step TOML inputs
1.0
Mantid Basic Course: Nice to have changes - From Manual testing: https://github.com/mantidproject/mantid/issues/30031 - [x] Update [Downloads image](https://docs.mantidproject.org/nightly/_images/MantidDownload_42.png) to be 5.1 or 6.0 - [x] [Place larger emphasis](https://docs.mantidproject.org/nightly/tutorials/mantid_basic_course/getting_started/getting_started.html#error-reporter) on **needing contact info** for issues to be resolved - [x] [Improve Examples for loading lots of data](https://docs.mantidproject.org/nightly/tutorials/mantid_basic_course/loading_and_displaying_data/01_loading_data.html#loading-lots-of-data-files) with Load Dialog lists, say using the 2 Mari files in TCD - [x] [Align bullet points](https://docs.mantidproject.org/nightly/tutorials/mantid_basic_course/loading_and_displaying_data/03_displaying_1D_data.html#adding-a-curve-to-an-existing-plot) - [x] Update [ISIS SANS interface instructions](https://docs.mantidproject.org/nightly/tutorials/mantid_basic_course/algorithms_workspaces_and_histories/05_interfaces.html#a-simple-walkthrough) to side step TOML inputs
non_defect
mantid basic course nice to have changes from manual testing update to be or on needing contact info for issues to be resolved with load dialog lists say using the mari files in tcd update to side step toml inputs
0
45,412
12,796,765,256
IssuesEvent
2020-07-02 11:03:33
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Derby emulation of all(T...) is incorrect
C: DB: Derby C: Functionality E: All Editions P: Medium T: Defect
This query here doesn't work correctly in Derby: ```java assertEquals(Arrays.asList(), create().select(TBook_ID()) .from(TBook()) .where(TBook_ID().equal(all(1, 2))) .orderBy(TBook_ID()).fetch(TBook_ID())); ``` The generated SQL is: ```sql select "TEST"."T_BOOK"."ID" from "TEST"."T_BOOK" where "TEST"."T_BOOK"."ID" = all ( select * from ( select 1 from "SYSIBM"."SYSDUMMY1" ) x union all select * from ( select 2 from "SYSIBM"."SYSDUMMY1" ) x ) order by "TEST"."T_BOOK"."ID" ``` Which is correct in most databases, but Derby raises this error: ``` 'SELECT *' only allowed in EXISTS and NOT EXISTS subqueries. ``` This probably happens in other cases too, when jOOQ automatically wraps `union` subqueries in derived tables for a variety of reasons...
1.0
Derby emulation of all(T...) is incorrect - This query here doesn't work correctly in Derby: ```java assertEquals(Arrays.asList(), create().select(TBook_ID()) .from(TBook()) .where(TBook_ID().equal(all(1, 2))) .orderBy(TBook_ID()).fetch(TBook_ID())); ``` The generated SQL is: ```sql select "TEST"."T_BOOK"."ID" from "TEST"."T_BOOK" where "TEST"."T_BOOK"."ID" = all ( select * from ( select 1 from "SYSIBM"."SYSDUMMY1" ) x union all select * from ( select 2 from "SYSIBM"."SYSDUMMY1" ) x ) order by "TEST"."T_BOOK"."ID" ``` Which is correct in most databases, but Derby raises this error: ``` 'SELECT *' only allowed in EXISTS and NOT EXISTS subqueries. ``` This probably happens in other cases too, when jOOQ automatically wraps `union` subqueries in derived tables for a variety of reasons...
defect
derby emulation of all t is incorrect this query here doesn t work correctly in derby java assertequals arrays aslist create select tbook id from tbook where tbook id equal all orderby tbook id fetch tbook id the generated sql is sql select test t book id from test t book where test t book id all select from select from sysibm x union all select from select from sysibm x order by test t book id which is correct in most databases but derby raises this error select only allowed in exists and not exists subqueries this probably happens in other cases too when jooq automatically wraps union subqueries in derived tables for a variety of reasons
1
49,813
13,187,276,107
IssuesEvent
2020-08-13 02:54:05
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
Numerical inconsistencies in SimplePropagator (Trac #2127)
Incomplete Migration Migrated from Trac combo simulation defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2127">https://code.icecube.wisc.edu/ticket/2127</a>, reported by chaack and owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2018-01-08T13:47:56", "description": "The particle returned from SimplePropagator have slightly different I3Directions compared to the input particle. This is an issue within Proposal, as the I3Particles are converted to `PROPOSALParticle` internally.\n\nConsider following code:\n\n{{{\nfrom icecube import PROPOSAL, dataclasses\nfrom icecube.icetray import I3Units\n\npropagator = PROPOSAL.SimplePropagator()\n\nmu = dataclasses.I3Particle()\nmu.type = dataclasses.I3Particle.MuMinus\nmu.pos = dataclasses.I3Position(0,0,0)\nmu.dir = dataclasses.I3Direction(n.radians(5), 0.2)\nmu.energy = 1* I3Units.PeV\nmu.time = 0 \nmu.location_type = dataclasses.I3Particle.InIce\nmu.length = NaN\n\nmu_prop = propagator.propagate(mu, 100)\nprint mu.dir == mu_prop.dir\n\n>> False\n}}}\n", "reporter": "chaack", "cc": "", "resolution": "invalid", "_ts": "1515419276810630", "component": "combo simulation", "summary": "Numerical inconsistencies in SimplePropagator", "priority": "normal", "keywords": "", "time": "2018-01-08T13:34:20", "milestone": "", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
1.0
Numerical inconsistencies in SimplePropagator (Trac #2127) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2127">https://code.icecube.wisc.edu/ticket/2127</a>, reported by chaack and owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2018-01-08T13:47:56", "description": "The particle returned from SimplePropagator have slightly different I3Directions compared to the input particle. This is an issue within Proposal, as the I3Particles are converted to `PROPOSALParticle` internally.\n\nConsider following code:\n\n{{{\nfrom icecube import PROPOSAL, dataclasses\nfrom icecube.icetray import I3Units\n\npropagator = PROPOSAL.SimplePropagator()\n\nmu = dataclasses.I3Particle()\nmu.type = dataclasses.I3Particle.MuMinus\nmu.pos = dataclasses.I3Position(0,0,0)\nmu.dir = dataclasses.I3Direction(n.radians(5), 0.2)\nmu.energy = 1* I3Units.PeV\nmu.time = 0 \nmu.location_type = dataclasses.I3Particle.InIce\nmu.length = NaN\n\nmu_prop = propagator.propagate(mu, 100)\nprint mu.dir == mu_prop.dir\n\n>> False\n}}}\n", "reporter": "chaack", "cc": "", "resolution": "invalid", "_ts": "1515419276810630", "component": "combo simulation", "summary": "Numerical inconsistencies in SimplePropagator", "priority": "normal", "keywords": "", "time": "2018-01-08T13:34:20", "milestone": "", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
defect
numerical inconsistencies in simplepropagator trac migrated from json status closed changetime description the particle returned from simplepropagator have slightly different compared to the input particle this is an issue within proposal as the are converted to proposalparticle internally n nconsider following code n n nfrom icecube import proposal dataclasses nfrom icecube icetray import n npropagator proposal simplepropagator n nmu dataclasses nmu type dataclasses muminus nmu pos dataclasses nmu dir dataclasses n radians nmu energy pev nmu time nmu location type dataclasses inice nmu length nan n nmu prop propagator propagate mu nprint mu dir mu prop dir n n false n n reporter chaack cc resolution invalid ts component combo simulation summary numerical inconsistencies in simplepropagator priority normal keywords time milestone owner jvansanten type defect
1
190,709
14,570,176,193
IssuesEvent
2020-12-17 14:04:03
zeebe-io/zeebe
https://api.github.com/repos/zeebe-io/zeebe
closed
io.netty.channel.unix.Errors$NativeIoException: bind(..) failed: Address already in use
Impact: Testing Priority: Mid Scope: broker Severity: High Status: Planned Type: Unstable Test
**Summary** Many different tests fail once in a while because of `io.netty.channel.unix.Errors$NativeIoException: bind(..) failed: Address already in use`. I know the issue is already discussed multiple times in individual flaky test issues, but I'd like to open this issue to find a solution to the bigger problem, rather than fix each individual flake. **Failures** <details><summary>Example stacktrace from `io.zeebe.broker.it.clustering.SnapshotReplicationTest.shouldReceiveNewSnapshotsOnRejoin`</summary> <pre> [ERROR] io.zeebe.broker.it.clustering.SnapshotReplicationTest.shouldReceiveNewSnapshotsOnRejoin Time elapsed: 11.895 s <<< ERROR! java.util.concurrent.CompletionException: java.util.concurrent.CompletionException: io.zeebe.util.exception.UncheckedExecutionException: Failed to start broker at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490) at java.base/java.util.concurrent.ForkJoinTask.getThrowableException(ForkJoinTask.java:600) at java.base/java.util.concurrent.ForkJoinTask.reportException(ForkJoinTask.java:678) at java.base/java.util.concurrent.ForkJoinTask.invoke(ForkJoinTask.java:737) at java.base/java.util.stream.ForEachOps$ForEachOp.evaluateParallel(ForEachOps.java:159) at java.base/java.util.stream.ForEachOps$ForEachOp$OfRef.evaluateParallel(ForEachOps.java:173) at java.base/java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:233) at java.base/java.util.stream.ReferencePipeline.forEach(ReferencePipeline.java:497) at java.base/java.util.stream.ReferencePipeline$Head.forEach(ReferencePipeline.java:661) at io.zeebe.broker.it.clustering.ClusteringRule.after(ClusteringRule.java:237) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:59) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:119) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:87) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75) at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:157) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: java.util.concurrent.CompletionException: io.zeebe.util.exception.UncheckedExecutionException: Failed to start broker at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) at java.base/java.util.concurrent.CompletableFuture.uniAcceptNow(CompletableFuture.java:743) at java.base/java.util.concurrent.CompletableFuture.uniAcceptStage(CompletableFuture.java:731) at java.base/java.util.concurrent.CompletableFuture.thenAccept(CompletableFuture.java:2108) at io.zeebe.broker.Broker.lambda$close$27(Broker.java:491) at io.zeebe.util.LogUtil.doWithMDC(LogUtil.java:21) at io.zeebe.broker.Broker.close(Broker.java:486) at java.base/java.util.stream.ForEachOps$ForEachOp$OfRef.accept(ForEachOps.java:183) at java.base/java.util.HashMap$ValueSpliterator.forEachRemaining(HashMap.java:1675) at java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:484) at java.base/java.util.stream.ForEachOps$ForEachTask.compute(ForEachOps.java:290) at java.base/java.util.concurrent.CountedCompleter.exec(CountedCompleter.java:746) at java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:290) at java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1020) at java.base/java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1656) at java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1594) at java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:177) Caused by: io.zeebe.util.exception.UncheckedExecutionException: Failed to start broker at io.zeebe.broker.Broker.internalStart(Broker.java:186) at io.zeebe.util.LogUtil.doWithMDC(LogUtil.java:21) at io.zeebe.broker.Broker.start(Broker.java:159) at io.zeebe.broker.it.clustering.ClusteringRule.startBroker(ClusteringRule.java:475) at io.zeebe.broker.it.clustering.ClusteringRule.restartBroker(ClusteringRule.java:471) at io.zeebe.broker.it.clustering.SnapshotReplicationTest.shouldReceiveNewSnapshotsOnRejoin(SnapshotReplicationTest.java:84) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:119) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:87) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75) at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:157) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: java.util.concurrent.CompletionException: io.netty.channel.unix.Errors$NativeIoException: bind(..) failed: Address already in use at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) at java.base/java.util.concurrent.CompletableFuture$UniRun.tryFire(CompletableFuture.java:777) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) at io.atomix.cluster.messaging.impl.NettyMessagingService.lambda$bind$32(NettyMessagingService.java:618) at io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:577) at io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:551) at io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:490) at io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:615) at io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:608) at io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:117) at io.netty.channel.AbstractChannel$AbstractUnsafe.safeSetFailure(AbstractChannel.java:998) at io.netty.channel.AbstractChannel$AbstractUnsafe.bind(AbstractChannel.java:552) at io.netty.channel.DefaultChannelPipeline$HeadContext.bind(DefaultChannelPipeline.java:1334) at io.netty.channel.AbstractChannelHandlerContext.invokeBind(AbstractChannelHandlerContext.java:506) at io.netty.channel.AbstractChannelHandlerContext.bind(AbstractChannelHandlerContext.java:491) at io.netty.channel.DefaultChannelPipeline.bind(DefaultChannelPipeline.java:973) at io.netty.channel.AbstractChannel.bind(AbstractChannel.java:248) at io.netty.bootstrap.AbstractBootstrap$2.run(AbstractBootstrap.java:356) at io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:164) at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:472) at io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:387) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at java.base/java.lang.Thread.run(Thread.java:834) Caused by: io.netty.channel.unix.Errors$NativeIoException: bind(..) failed: Address already in use </pre> </details> **Known tests that were flaky due to this** Issues open at time of writing: - #4940 ExporterManagerPartitionTest.shouldRunExporterForEveryPartition - #4941 BrokerTest.shouldStartAndStopBroker - #5182 WorkflowTest.shouldCompleteWorkflowInstance - #5231 RaftRolesTest.testStepDownInRoleChangedListener - #5471: SingleBrokerDataDeletionTest.shouldCompactWhenExporterHasBeenRemoved Test flakes from just today: - SnapshotReplicationTest.shouldReceiveNewSnapshotsOnRejoin - GossipClusteringTest.shouldReAddToCluster - BrokerLeaderChangeTest.shouldBecomeFollowerAfterRestartLeaderChange
2.0
io.netty.channel.unix.Errors$NativeIoException: bind(..) failed: Address already in use - **Summary** Many different tests fail once in a while because of `io.netty.channel.unix.Errors$NativeIoException: bind(..) failed: Address already in use`. I know the issue is already discussed multiple times in individual flaky test issues, but I'd like to open this issue to find a solution to the bigger problem, rather than fix each individual flake. **Failures** <details><summary>Example stacktrace from `io.zeebe.broker.it.clustering.SnapshotReplicationTest.shouldReceiveNewSnapshotsOnRejoin`</summary> <pre> [ERROR] io.zeebe.broker.it.clustering.SnapshotReplicationTest.shouldReceiveNewSnapshotsOnRejoin Time elapsed: 11.895 s <<< ERROR! java.util.concurrent.CompletionException: java.util.concurrent.CompletionException: io.zeebe.util.exception.UncheckedExecutionException: Failed to start broker at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490) at java.base/java.util.concurrent.ForkJoinTask.getThrowableException(ForkJoinTask.java:600) at java.base/java.util.concurrent.ForkJoinTask.reportException(ForkJoinTask.java:678) at java.base/java.util.concurrent.ForkJoinTask.invoke(ForkJoinTask.java:737) at java.base/java.util.stream.ForEachOps$ForEachOp.evaluateParallel(ForEachOps.java:159) at java.base/java.util.stream.ForEachOps$ForEachOp$OfRef.evaluateParallel(ForEachOps.java:173) at java.base/java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:233) at java.base/java.util.stream.ReferencePipeline.forEach(ReferencePipeline.java:497) at java.base/java.util.stream.ReferencePipeline$Head.forEach(ReferencePipeline.java:661) at io.zeebe.broker.it.clustering.ClusteringRule.after(ClusteringRule.java:237) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:59) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:119) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:87) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75) at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:157) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: java.util.concurrent.CompletionException: io.zeebe.util.exception.UncheckedExecutionException: Failed to start broker at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) at java.base/java.util.concurrent.CompletableFuture.uniAcceptNow(CompletableFuture.java:743) at java.base/java.util.concurrent.CompletableFuture.uniAcceptStage(CompletableFuture.java:731) at java.base/java.util.concurrent.CompletableFuture.thenAccept(CompletableFuture.java:2108) at io.zeebe.broker.Broker.lambda$close$27(Broker.java:491) at io.zeebe.util.LogUtil.doWithMDC(LogUtil.java:21) at io.zeebe.broker.Broker.close(Broker.java:486) at java.base/java.util.stream.ForEachOps$ForEachOp$OfRef.accept(ForEachOps.java:183) at java.base/java.util.HashMap$ValueSpliterator.forEachRemaining(HashMap.java:1675) at java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:484) at java.base/java.util.stream.ForEachOps$ForEachTask.compute(ForEachOps.java:290) at java.base/java.util.concurrent.CountedCompleter.exec(CountedCompleter.java:746) at java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:290) at java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1020) at java.base/java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1656) at java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1594) at java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:177) Caused by: io.zeebe.util.exception.UncheckedExecutionException: Failed to start broker at io.zeebe.broker.Broker.internalStart(Broker.java:186) at io.zeebe.util.LogUtil.doWithMDC(LogUtil.java:21) at io.zeebe.broker.Broker.start(Broker.java:159) at io.zeebe.broker.it.clustering.ClusteringRule.startBroker(ClusteringRule.java:475) at io.zeebe.broker.it.clustering.ClusteringRule.restartBroker(ClusteringRule.java:471) at io.zeebe.broker.it.clustering.SnapshotReplicationTest.shouldReceiveNewSnapshotsOnRejoin(SnapshotReplicationTest.java:84) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:119) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:87) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75) at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:157) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: java.util.concurrent.CompletionException: io.netty.channel.unix.Errors$NativeIoException: bind(..) failed: Address already in use at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:331) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:346) at java.base/java.util.concurrent.CompletableFuture$UniRun.tryFire(CompletableFuture.java:777) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2088) at io.atomix.cluster.messaging.impl.NettyMessagingService.lambda$bind$32(NettyMessagingService.java:618) at io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:577) at io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:551) at io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:490) at io.netty.util.concurrent.DefaultPromise.setValue0(DefaultPromise.java:615) at io.netty.util.concurrent.DefaultPromise.setFailure0(DefaultPromise.java:608) at io.netty.util.concurrent.DefaultPromise.tryFailure(DefaultPromise.java:117) at io.netty.channel.AbstractChannel$AbstractUnsafe.safeSetFailure(AbstractChannel.java:998) at io.netty.channel.AbstractChannel$AbstractUnsafe.bind(AbstractChannel.java:552) at io.netty.channel.DefaultChannelPipeline$HeadContext.bind(DefaultChannelPipeline.java:1334) at io.netty.channel.AbstractChannelHandlerContext.invokeBind(AbstractChannelHandlerContext.java:506) at io.netty.channel.AbstractChannelHandlerContext.bind(AbstractChannelHandlerContext.java:491) at io.netty.channel.DefaultChannelPipeline.bind(DefaultChannelPipeline.java:973) at io.netty.channel.AbstractChannel.bind(AbstractChannel.java:248) at io.netty.bootstrap.AbstractBootstrap$2.run(AbstractBootstrap.java:356) at io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:164) at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:472) at io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:387) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at java.base/java.lang.Thread.run(Thread.java:834) Caused by: io.netty.channel.unix.Errors$NativeIoException: bind(..) failed: Address already in use </pre> </details> **Known tests that were flaky due to this** Issues open at time of writing: - #4940 ExporterManagerPartitionTest.shouldRunExporterForEveryPartition - #4941 BrokerTest.shouldStartAndStopBroker - #5182 WorkflowTest.shouldCompleteWorkflowInstance - #5231 RaftRolesTest.testStepDownInRoleChangedListener - #5471: SingleBrokerDataDeletionTest.shouldCompactWhenExporterHasBeenRemoved Test flakes from just today: - SnapshotReplicationTest.shouldReceiveNewSnapshotsOnRejoin - GossipClusteringTest.shouldReAddToCluster - BrokerLeaderChangeTest.shouldBecomeFollowerAfterRestartLeaderChange
non_defect
io netty channel unix errors nativeioexception bind failed address already in use summary many different tests fail once in a while because of io netty channel unix errors nativeioexception bind failed address already in use i know the issue is already discussed multiple times in individual flaky test issues but i d like to open this issue to find a solution to the bigger problem rather than fix each individual flake failures example stacktrace from io zeebe broker it clustering snapshotreplicationtest shouldreceivenewsnapshotsonrejoin io zeebe broker it clustering snapshotreplicationtest shouldreceivenewsnapshotsonrejoin time elapsed s error java util concurrent completionexception java util concurrent completionexception io zeebe util exception uncheckedexecutionexception failed to start broker at java base jdk internal reflect nativeconstructoraccessorimpl native method at java base jdk internal reflect nativeconstructoraccessorimpl newinstance nativeconstructoraccessorimpl java at java base jdk internal reflect delegatingconstructoraccessorimpl newinstance delegatingconstructoraccessorimpl java at java base java lang reflect constructor newinstance constructor java at java base java util concurrent forkjointask getthrowableexception forkjointask java at java base java util concurrent forkjointask reportexception forkjointask java at java base java util concurrent forkjointask invoke forkjointask java at java base java util stream foreachops foreachop evaluateparallel foreachops java at java base java util stream foreachops foreachop ofref evaluateparallel foreachops java at java base java util stream abstractpipeline evaluate abstractpipeline java at java base java util stream referencepipeline foreach referencepipeline java at java base java util stream referencepipeline head foreach referencepipeline java at io zeebe broker it clustering clusteringrule after clusteringrule java at org junit rules externalresource evaluate externalresource java at org junit rules externalresource evaluate externalresource java at org junit rules runrules evaluate runrules java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runners suite runchild suite java at org junit runners suite runchild suite java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore junitcore run junitcore java at org apache maven surefire junitcore junitcorewrapper createrequestandrun junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper executelazy junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcoreprovider invoke junitcoreprovider java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java caused by java util concurrent completionexception io zeebe util exception uncheckedexecutionexception failed to start broker at java base java util concurrent completablefuture encodethrowable completablefuture java at java base java util concurrent completablefuture uniacceptnow completablefuture java at java base java util concurrent completablefuture uniacceptstage completablefuture java at java base java util concurrent completablefuture thenaccept completablefuture java at io zeebe broker broker lambda close broker java at io zeebe util logutil dowithmdc logutil java at io zeebe broker broker close broker java at java base java util stream foreachops foreachop ofref accept foreachops java at java base java util hashmap valuespliterator foreachremaining hashmap java at java base java util stream abstractpipeline copyinto abstractpipeline java at java base java util stream foreachops foreachtask compute foreachops java at java base java util concurrent countedcompleter exec countedcompleter java at java base java util concurrent forkjointask doexec forkjointask java at java base java util concurrent forkjoinpool workqueue toplevelexec forkjoinpool java at java base java util concurrent forkjoinpool scan forkjoinpool java at java base java util concurrent forkjoinpool runworker forkjoinpool java at java base java util concurrent forkjoinworkerthread run forkjoinworkerthread java caused by io zeebe util exception uncheckedexecutionexception failed to start broker at io zeebe broker broker internalstart broker java at io zeebe util logutil dowithmdc logutil java at io zeebe broker broker start broker java at io zeebe broker it clustering clusteringrule startbroker clusteringrule java at io zeebe broker it clustering clusteringrule restartbroker clusteringrule java at io zeebe broker it clustering snapshotreplicationtest shouldreceivenewsnapshotsonrejoin snapshotreplicationtest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit rules externalresource evaluate externalresource java at org junit rules testwatcher evaluate testwatcher java at org junit rules externalresource evaluate externalresource java at org junit rules externalresource evaluate externalresource java at org junit rules externalresource evaluate externalresource java at org junit rules runrules evaluate runrules java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runners suite runchild suite java at org junit runners suite runchild suite java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore junitcore run junitcore java at org apache maven surefire junitcore junitcorewrapper createrequestandrun junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper executelazy junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcoreprovider invoke junitcoreprovider java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java caused by java util concurrent completionexception io netty channel unix errors nativeioexception bind failed address already in use at java base java util concurrent completablefuture encodethrowable completablefuture java at java base java util concurrent completablefuture completethrowable completablefuture java at java base java util concurrent completablefuture unirun tryfire completablefuture java at java base java util concurrent completablefuture postcomplete completablefuture java at java base java util concurrent completablefuture completeexceptionally completablefuture java at io atomix cluster messaging impl nettymessagingservice lambda bind nettymessagingservice java at io netty util concurrent defaultpromise defaultpromise java at io netty util concurrent defaultpromise notifylistenersnow defaultpromise java at io netty util concurrent defaultpromise notifylisteners defaultpromise java at io netty util concurrent defaultpromise defaultpromise java at io netty util concurrent defaultpromise defaultpromise java at io netty util concurrent defaultpromise tryfailure defaultpromise java at io netty channel abstractchannel abstractunsafe safesetfailure abstractchannel java at io netty channel abstractchannel abstractunsafe bind abstractchannel java at io netty channel defaultchannelpipeline headcontext bind defaultchannelpipeline java at io netty channel abstractchannelhandlercontext invokebind abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext bind abstractchannelhandlercontext java at io netty channel defaultchannelpipeline bind defaultchannelpipeline java at io netty channel abstractchannel bind abstractchannel java at io netty bootstrap abstractbootstrap run abstractbootstrap java at io netty util concurrent abstracteventexecutor safeexecute abstracteventexecutor java at io netty util concurrent singlethreadeventexecutor runalltasks singlethreadeventexecutor java at io netty channel epoll epolleventloop run epolleventloop java at io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at io netty util internal threadexecutormap run threadexecutormap java at java base java lang thread run thread java caused by io netty channel unix errors nativeioexception bind failed address already in use known tests that were flaky due to this issues open at time of writing exportermanagerpartitiontest shouldrunexporterforeverypartition brokertest shouldstartandstopbroker workflowtest shouldcompleteworkflowinstance raftrolestest teststepdowninrolechangedlistener singlebrokerdatadeletiontest shouldcompactwhenexporterhasbeenremoved test flakes from just today snapshotreplicationtest shouldreceivenewsnapshotsonrejoin gossipclusteringtest shouldreaddtocluster brokerleaderchangetest shouldbecomefollowerafterrestartleaderchange
0
254,299
19,192,060,704
IssuesEvent
2021-12-06 02:42:49
MikeTheNose/SystemAnalysisProject
https://api.github.com/repos/MikeTheNose/SystemAnalysisProject
closed
Task 10
documentation
10.Create a decision table to capture the process logic for your system (use the reduced version on page 204, fig 7-19). The table should contain at least 2 conditions and 4 courses of action.
1.0
Task 10 - 10.Create a decision table to capture the process logic for your system (use the reduced version on page 204, fig 7-19). The table should contain at least 2 conditions and 4 courses of action.
non_defect
task create a decision table to capture the process logic for your system use the reduced version on page fig the table should contain at least conditions and courses of action
0
244,250
18,751,960,958
IssuesEvent
2021-11-05 04:02:24
AY2122S1-CS2103T-F11-4/tp
https://api.github.com/repos/AY2122S1-CS2103T-F11-4/tp
closed
[PE-D] Find Tag Command
documentation duplicate find command
![image.png](https://raw.githubusercontent.com/RonaldTanSingWei/ped/main/files/747aab19-d4b5-43f8-91f0-0bd111091c34.png) ![image.png](https://raw.githubusercontent.com/RonaldTanSingWei/ped/main/files/5fa35479-ae2c-46e6-9f1a-b807b5029d93.png) Finding tags using the find t/money returns all users with tags with money inside them, instead of only tags with money <!--session: 1635494612655-78768029-5c83-42fa-8cd5-edbbb31654ef--> <!--Version: Web v3.4.1--> ------------- Labels: `severity.Low` `type.FunctionalityBug` original: RonaldTanSingWei/ped#3
1.0
[PE-D] Find Tag Command - ![image.png](https://raw.githubusercontent.com/RonaldTanSingWei/ped/main/files/747aab19-d4b5-43f8-91f0-0bd111091c34.png) ![image.png](https://raw.githubusercontent.com/RonaldTanSingWei/ped/main/files/5fa35479-ae2c-46e6-9f1a-b807b5029d93.png) Finding tags using the find t/money returns all users with tags with money inside them, instead of only tags with money <!--session: 1635494612655-78768029-5c83-42fa-8cd5-edbbb31654ef--> <!--Version: Web v3.4.1--> ------------- Labels: `severity.Low` `type.FunctionalityBug` original: RonaldTanSingWei/ped#3
non_defect
find tag command finding tags using the find t money returns all users with tags with money inside them instead of only tags with money labels severity low type functionalitybug original ronaldtansingwei ped
0
286,277
8,785,977,023
IssuesEvent
2018-12-20 14:33:36
toobigtoignore/issf
https://api.github.com/repos/toobigtoignore/issf
closed
Certain profiles do not show the correct date range on their profile page (or in the generated report)
backend bug hackathon low priority
**Describe the bug** Profiles do not always show the correct date range. Sometimes only the first year will come up, i.e 2015 instead of 2015-2016. This also occurs when a PDF report is generated of that record. **To Reproduce** Steps to reproduce the behavior: 1. Open an SSF Profile record that you know has a date range. 2. Generate a PDF report and see that is also has the issue. **Expected behavior** Should have a year-year format like 1998-2002
1.0
Certain profiles do not show the correct date range on their profile page (or in the generated report) - **Describe the bug** Profiles do not always show the correct date range. Sometimes only the first year will come up, i.e 2015 instead of 2015-2016. This also occurs when a PDF report is generated of that record. **To Reproduce** Steps to reproduce the behavior: 1. Open an SSF Profile record that you know has a date range. 2. Generate a PDF report and see that is also has the issue. **Expected behavior** Should have a year-year format like 1998-2002
non_defect
certain profiles do not show the correct date range on their profile page or in the generated report describe the bug profiles do not always show the correct date range sometimes only the first year will come up i e instead of this also occurs when a pdf report is generated of that record to reproduce steps to reproduce the behavior open an ssf profile record that you know has a date range generate a pdf report and see that is also has the issue expected behavior should have a year year format like
0
82,084
31,935,042,328
IssuesEvent
2023-09-19 09:54:05
nats-io/nats.go
https://api.github.com/repos/nats-io/nats.go
closed
Subscribe callback can be triggered after close is called
defect
### What version were you using? We're currently using ` github.com/nats-io/nats.go v1.21 . ### What environment was the server running in? This was detected on a linux CI machine. ### Is this defect reproducible? I've copied the following snippets from our full [codebase](https://github.com/statechannels/go-nitro/blob/3dc13590424864a16e3b58a28a30ec275085335f/rpc/transport/nats/client.go#L50) to illustrate the issue. We have a function where we subscribe to messages and send those messages on an internal chan `notificationChan` ``` func (c *natsTransportClient) Subscribe() (<-chan []byte, error) { c.notificationChan = make(chan []byte) subscription, err := c.nc.Subscribe(nitroNotificationTopic, func(msg *nats.Msg) { c.notificationChan <- msg.Data }) c.natsSubscriptions = append(c.natsSubscriptions, subscription) return c.notificationChan, err } ``` In our close we first close our `natsTransport` ([which closes the underlying NATS connection](https://github.com/statechannels/go-nitro/blob/3dc13590424864a16e3b58a28a30ec275085335f/rpc/transport/nats/server.go#L46)) and then we close the `notificationChan` ``` func (c *natsTransportClient) Close() error { err := c.natsTransport.Close() if err != nil { return err } close(c.notificationChan) return nil ``` Occasionally(maybe once a week I) though on CI we see failures like this: ``` === RUN TestRpcWithNats/2_clients rpc_test.go:205: Ledger channels queried rpc_test.go:262: Payment channels queried rpc_test.go:293: Vouchers sent/received rpc_test.go:308: Ledger/virtual channels closed panic: send on closed channel goroutine 17946 [running]: github.com/statechannels/go-nitro/rpc/transport/nats.(*natsTransportClient).Subscribe.func1(0x0?) /home/runner/work/go-nitro/go-nitro/rpc/transport/nats/client.go:56 +0x25 github.com/nats-io/nats%2ego.(*Conn).waitForMsgs(0xc002343800, 0xc0045895f0) /home/runner/go/pkg/mod/github.com/nats-io/nats.go@v1.21.0/nats.go:2884 +0x3f2 created by github.com/nats-io/nats%2ego.(*Conn).subscribeLocked in goroutine 17695 /home/runner/go/pkg/mod/github.com/nats-io/nats.go@v1.21.0/nats.go:4114 +0x3c8 FAIL github.com/statechannels/go-nitro/node_test 9.077s ``` This indicates that the callback function we pass into `c.nc.Subscribe` to handle a message is being triggered **after** we've closed down the nats connection. ### Given the capability you are leveraging, describe your expectation? I would expect that once we close the underlying NATS connection we could be guaranteed that the callback function we be provided will not be called. ### Given the expectation, what is the defect you are observing? It looks like it's possible (if you get unlucky with timing) that the `Subscribe` callback can be triggered even after the connection has been closed.
1.0
Subscribe callback can be triggered after close is called - ### What version were you using? We're currently using ` github.com/nats-io/nats.go v1.21 . ### What environment was the server running in? This was detected on a linux CI machine. ### Is this defect reproducible? I've copied the following snippets from our full [codebase](https://github.com/statechannels/go-nitro/blob/3dc13590424864a16e3b58a28a30ec275085335f/rpc/transport/nats/client.go#L50) to illustrate the issue. We have a function where we subscribe to messages and send those messages on an internal chan `notificationChan` ``` func (c *natsTransportClient) Subscribe() (<-chan []byte, error) { c.notificationChan = make(chan []byte) subscription, err := c.nc.Subscribe(nitroNotificationTopic, func(msg *nats.Msg) { c.notificationChan <- msg.Data }) c.natsSubscriptions = append(c.natsSubscriptions, subscription) return c.notificationChan, err } ``` In our close we first close our `natsTransport` ([which closes the underlying NATS connection](https://github.com/statechannels/go-nitro/blob/3dc13590424864a16e3b58a28a30ec275085335f/rpc/transport/nats/server.go#L46)) and then we close the `notificationChan` ``` func (c *natsTransportClient) Close() error { err := c.natsTransport.Close() if err != nil { return err } close(c.notificationChan) return nil ``` Occasionally(maybe once a week I) though on CI we see failures like this: ``` === RUN TestRpcWithNats/2_clients rpc_test.go:205: Ledger channels queried rpc_test.go:262: Payment channels queried rpc_test.go:293: Vouchers sent/received rpc_test.go:308: Ledger/virtual channels closed panic: send on closed channel goroutine 17946 [running]: github.com/statechannels/go-nitro/rpc/transport/nats.(*natsTransportClient).Subscribe.func1(0x0?) /home/runner/work/go-nitro/go-nitro/rpc/transport/nats/client.go:56 +0x25 github.com/nats-io/nats%2ego.(*Conn).waitForMsgs(0xc002343800, 0xc0045895f0) /home/runner/go/pkg/mod/github.com/nats-io/nats.go@v1.21.0/nats.go:2884 +0x3f2 created by github.com/nats-io/nats%2ego.(*Conn).subscribeLocked in goroutine 17695 /home/runner/go/pkg/mod/github.com/nats-io/nats.go@v1.21.0/nats.go:4114 +0x3c8 FAIL github.com/statechannels/go-nitro/node_test 9.077s ``` This indicates that the callback function we pass into `c.nc.Subscribe` to handle a message is being triggered **after** we've closed down the nats connection. ### Given the capability you are leveraging, describe your expectation? I would expect that once we close the underlying NATS connection we could be guaranteed that the callback function we be provided will not be called. ### Given the expectation, what is the defect you are observing? It looks like it's possible (if you get unlucky with timing) that the `Subscribe` callback can be triggered even after the connection has been closed.
defect
subscribe callback can be triggered after close is called what version were you using we re currently using github com nats io nats go what environment was the server running in this was detected on a linux ci machine is this defect reproducible i ve copied the following snippets from our full to illustrate the issue we have a function where we subscribe to messages and send those messages on an internal chan notificationchan func c natstransportclient subscribe chan byte error c notificationchan make chan byte subscription err c nc subscribe nitronotificationtopic func msg nats msg c notificationchan msg data c natssubscriptions append c natssubscriptions subscription return c notificationchan err in our close we first close our natstransport and then we close the notificationchan func c natstransportclient close error err c natstransport close if err nil return err close c notificationchan return nil occasionally maybe once a week i though on ci we see failures like this run testrpcwithnats clients rpc test go ledger channels queried rpc test go payment channels queried rpc test go vouchers sent received rpc test go ledger virtual channels closed panic send on closed channel goroutine github com statechannels go nitro rpc transport nats natstransportclient subscribe home runner work go nitro go nitro rpc transport nats client go github com nats io nats conn waitformsgs home runner go pkg mod github com nats io nats go nats go created by github com nats io nats conn subscribelocked in goroutine home runner go pkg mod github com nats io nats go nats go fail github com statechannels go nitro node test this indicates that the callback function we pass into c nc subscribe to handle a message is being triggered after we ve closed down the nats connection given the capability you are leveraging describe your expectation i would expect that once we close the underlying nats connection we could be guaranteed that the callback function we be provided will not be called given the expectation what is the defect you are observing it looks like it s possible if you get unlucky with timing that the subscribe callback can be triggered even after the connection has been closed
1
320,557
9,782,381,276
IssuesEvent
2019-06-07 23:19:50
lightingft/appinventor-sources
https://api.github.com/repos/lightingft/appinventor-sources
reopened
Data basic properties (label and color)
Part: Component Priority: Medium Status: In Progress Type: Feature
For the Data Component, two basic properties should be implemented: - [x] Label - [x] Color
1.0
Data basic properties (label and color) - For the Data Component, two basic properties should be implemented: - [x] Label - [x] Color
non_defect
data basic properties label and color for the data component two basic properties should be implemented label color
0
37,982
8,621,502,716
IssuesEvent
2018-11-20 17:29:58
hazelcast/hazelcast-nodejs-client
https://api.github.com/repos/hazelcast/hazelcast-nodejs-client
opened
benchmark/SimpleMapBenchmark uses old .getMap api
Priority: Low Type: Defect
this benchmark does not work with current master
1.0
benchmark/SimpleMapBenchmark uses old .getMap api - this benchmark does not work with current master
defect
benchmark simplemapbenchmark uses old getmap api this benchmark does not work with current master
1
51,572
13,207,528,480
IssuesEvent
2020-08-14 23:27:46
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
USE_GOLD auto enabled on mac os x (Trac #641)
Incomplete Migration Migrated from Trac cmake defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/641">https://code.icecube.wisc.edu/projects/icecube/ticket/641</a>, reported by negaand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2014-11-23T03:37:57", "_ts": "1416713877111216", "description": "see: \nhttp://builds.icecube.wisc.edu/builders/quick_osx_10.5.8/builds/410/steps/compile/logs/stdio", "reporter": "nega", "cc": "", "resolution": "fixed", "time": "2011-06-02T16:01:41", "component": "cmake", "summary": "USE_GOLD auto enabled on mac os x", "priority": "normal", "keywords": "", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
USE_GOLD auto enabled on mac os x (Trac #641) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/641">https://code.icecube.wisc.edu/projects/icecube/ticket/641</a>, reported by negaand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2014-11-23T03:37:57", "_ts": "1416713877111216", "description": "see: \nhttp://builds.icecube.wisc.edu/builders/quick_osx_10.5.8/builds/410/steps/compile/logs/stdio", "reporter": "nega", "cc": "", "resolution": "fixed", "time": "2011-06-02T16:01:41", "component": "cmake", "summary": "USE_GOLD auto enabled on mac os x", "priority": "normal", "keywords": "", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
defect
use gold auto enabled on mac os x trac migrated from json status closed changetime ts description see n reporter nega cc resolution fixed time component cmake summary use gold auto enabled on mac os x priority normal keywords milestone owner nega type defect
1
574,784
17,023,874,201
IssuesEvent
2021-07-03 04:18:22
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Make the tags editable in the browse interface
Component: website Priority: minor Resolution: wontfix Type: enhancement
**[Submitted to the original trac issue database at 11.01am, Monday, 12th August 2013]** Yesterday I wanted to edit http://www.openstreetmap.org/browse/node/432463849 which had the wrong place value. To do that I had to open ID (or another editor). This is unneccesarily burdensome since it didn't involve any change of coordinates. It should be possible to edit, add and remove values for nodes, ways and relations with the browse interface, without having to open a map editor, kind of like Wikidata works, just as long as no coordinates are changed. I don't know if this is possible to implement or if it has been asked before, but I'm sure that I'm not the only one who has wanted to do edits like this in a more streamlined fashion?
1.0
Make the tags editable in the browse interface - **[Submitted to the original trac issue database at 11.01am, Monday, 12th August 2013]** Yesterday I wanted to edit http://www.openstreetmap.org/browse/node/432463849 which had the wrong place value. To do that I had to open ID (or another editor). This is unneccesarily burdensome since it didn't involve any change of coordinates. It should be possible to edit, add and remove values for nodes, ways and relations with the browse interface, without having to open a map editor, kind of like Wikidata works, just as long as no coordinates are changed. I don't know if this is possible to implement or if it has been asked before, but I'm sure that I'm not the only one who has wanted to do edits like this in a more streamlined fashion?
non_defect
make the tags editable in the browse interface yesterday i wanted to edit which had the wrong place value to do that i had to open id or another editor this is unneccesarily burdensome since it didn t involve any change of coordinates it should be possible to edit add and remove values for nodes ways and relations with the browse interface without having to open a map editor kind of like wikidata works just as long as no coordinates are changed i don t know if this is possible to implement or if it has been asked before but i m sure that i m not the only one who has wanted to do edits like this in a more streamlined fashion
0
25,100
4,203,768,265
IssuesEvent
2016-06-28 07:18:45
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
DataTypeException: Cannot convert from false in Jooq v.3.8.2
C: DB: MariaDB C: DB: MySQL C: Functionality P: Urgent T: Defect
I have MySQL with maven plugin generated pojos and DAOs. With the jooq v.3.7.3 I dont have this Exceptions: ``` Caused by: org.jooq.exception.MappingException: An error ocurred when mapping record to class com.stil.generated.mysql.tables.pojos.SystemProperties at org.jooq.impl.DefaultRecordMapper$MutablePOJOMapper.map(DefaultRecordMapper.java:658) ... Caused by: org.jooq.exception.DataTypeException: Cannot convert from false (class java.lang.Boolean) to class com.stil.generated.mysql.tables.pojos.SystemPropertiesHistory at org.jooq.tools.Convert$ConvertAll.fail(Convert.java:1113) ``` full stack trace http://pastebin.com/cditt1t8 I have use generated Dao function fetchOneByUniqueKey() this is my tables structure ```sql CREATE TABLE `system_properties` ( `id` int(11) NOT NULL AUTO_INCREMENT, `sys_key` varchar(32) NOT NULL, `sys_value` varchar(256) DEFAULT NULL, `description` varchar(255) NOT NULL, `order_column` int(11) DEFAULT NULL, `history` bit(1) NOT NULL DEFAULT b'0', PRIMARY KEY (`id`), UNIQUE KEY `key` (`sys_key`) ) ENGINE=InnoDB DEFAULT CHARSET=utf8; CREATE TABLE `system_properties_history` ( `id` int(11) NOT NULL AUTO_INCREMENT, `users_id` int(11) DEFAULT NULL, `properties_id` int(11) NOT NULL, `datecreated` date NOT NULL, `new_value` varchar(255) NOT NULL, PRIMARY KEY (`id`), UNIQUE KEY `properties_id_2` (`properties_id`,`datecreated`), KEY `users_id` (`users_id`), KEY `properties_id` (`properties_id`), CONSTRAINT `system_properties_history_ibfk_1` FOREIGN KEY (`properties_id`) REFERENCES `system_properties` (`id`) ON DELETE CASCADE ON UPDATE CASCADE, CONSTRAINT `system_properties_history_ibfk_2` FOREIGN KEY (`users_id`) REFERENCES `users` (`id`) ON DELETE SET NULL ON UPDATE CASCADE ) ENGINE=InnoDB DEFAULT CHARSET=utf8; ``` maybe this is problem in jooq-codegen-maven plugin with v.3.8.2 I have generated pojo: ```java public class SystemProperties implements Serializable { private static final long serialVersionUID = -442588835; private Integer id; private String sysKey; private String sysValue; private String description; private Integer orderColumn; private SystemPropertiesHistory history; //this is object instead of Boolean in MySQL table structure (there is not FK) and with jooq-codegen-maven v.3.7.3 its Boolean. } public class SystemPropertiesHistory implements Serializable { private static final long serialVersionUID = -1868690069; private Integer id; private Integer usersId; private Integer propertiesId; private Timestamp datecreated; private String newValue; } ``` jooq-codegen-maven config: ```xml <plugin> <groupId>org.jooq</groupId> <artifactId>jooq-codegen-maven</artifactId> <version>${jooq.version}</version> <executions> <execution> <id>generate-mysql</id> <phase>generate-sources</phase> <goals> <goal>generate</goal> </goals> </execution> </executions> <dependencies> <dependency> <groupId>mysql</groupId> <artifactId>mysql-connector-java</artifactId> <version>5.1.31</version> </dependency> </dependencies> <configuration> <jdbc> <driver>com.mysql.jdbc.Driver</driver> <url>${DB_URL}</url> <user>${DB_USER}</user> <password>${DB_PASS}</password> </jdbc> <generator> <database> <name>${jooq.generator.db.dialect}</name> <includes>.*</includes> <excludes></excludes> <dateAsTimestamp>true</dateAsTimestamp> <inputSchema>${DB_SCHEMA}</inputSchema> </database> <generate> <deprecated>false</deprecated> <instanceFields>true</instanceFields> <records>true</records> <relations>true</relations> <pojos>true</pojos> <daos>true</daos> </generate> <target> <packageName>com.stil.generated.mysql</packageName> <directory>target/generated-sources/jooq-mysql</directory> </target> </generator> </configuration> </plugin> ```
1.0
DataTypeException: Cannot convert from false in Jooq v.3.8.2 - I have MySQL with maven plugin generated pojos and DAOs. With the jooq v.3.7.3 I dont have this Exceptions: ``` Caused by: org.jooq.exception.MappingException: An error ocurred when mapping record to class com.stil.generated.mysql.tables.pojos.SystemProperties at org.jooq.impl.DefaultRecordMapper$MutablePOJOMapper.map(DefaultRecordMapper.java:658) ... Caused by: org.jooq.exception.DataTypeException: Cannot convert from false (class java.lang.Boolean) to class com.stil.generated.mysql.tables.pojos.SystemPropertiesHistory at org.jooq.tools.Convert$ConvertAll.fail(Convert.java:1113) ``` full stack trace http://pastebin.com/cditt1t8 I have use generated Dao function fetchOneByUniqueKey() this is my tables structure ```sql CREATE TABLE `system_properties` ( `id` int(11) NOT NULL AUTO_INCREMENT, `sys_key` varchar(32) NOT NULL, `sys_value` varchar(256) DEFAULT NULL, `description` varchar(255) NOT NULL, `order_column` int(11) DEFAULT NULL, `history` bit(1) NOT NULL DEFAULT b'0', PRIMARY KEY (`id`), UNIQUE KEY `key` (`sys_key`) ) ENGINE=InnoDB DEFAULT CHARSET=utf8; CREATE TABLE `system_properties_history` ( `id` int(11) NOT NULL AUTO_INCREMENT, `users_id` int(11) DEFAULT NULL, `properties_id` int(11) NOT NULL, `datecreated` date NOT NULL, `new_value` varchar(255) NOT NULL, PRIMARY KEY (`id`), UNIQUE KEY `properties_id_2` (`properties_id`,`datecreated`), KEY `users_id` (`users_id`), KEY `properties_id` (`properties_id`), CONSTRAINT `system_properties_history_ibfk_1` FOREIGN KEY (`properties_id`) REFERENCES `system_properties` (`id`) ON DELETE CASCADE ON UPDATE CASCADE, CONSTRAINT `system_properties_history_ibfk_2` FOREIGN KEY (`users_id`) REFERENCES `users` (`id`) ON DELETE SET NULL ON UPDATE CASCADE ) ENGINE=InnoDB DEFAULT CHARSET=utf8; ``` maybe this is problem in jooq-codegen-maven plugin with v.3.8.2 I have generated pojo: ```java public class SystemProperties implements Serializable { private static final long serialVersionUID = -442588835; private Integer id; private String sysKey; private String sysValue; private String description; private Integer orderColumn; private SystemPropertiesHistory history; //this is object instead of Boolean in MySQL table structure (there is not FK) and with jooq-codegen-maven v.3.7.3 its Boolean. } public class SystemPropertiesHistory implements Serializable { private static final long serialVersionUID = -1868690069; private Integer id; private Integer usersId; private Integer propertiesId; private Timestamp datecreated; private String newValue; } ``` jooq-codegen-maven config: ```xml <plugin> <groupId>org.jooq</groupId> <artifactId>jooq-codegen-maven</artifactId> <version>${jooq.version}</version> <executions> <execution> <id>generate-mysql</id> <phase>generate-sources</phase> <goals> <goal>generate</goal> </goals> </execution> </executions> <dependencies> <dependency> <groupId>mysql</groupId> <artifactId>mysql-connector-java</artifactId> <version>5.1.31</version> </dependency> </dependencies> <configuration> <jdbc> <driver>com.mysql.jdbc.Driver</driver> <url>${DB_URL}</url> <user>${DB_USER}</user> <password>${DB_PASS}</password> </jdbc> <generator> <database> <name>${jooq.generator.db.dialect}</name> <includes>.*</includes> <excludes></excludes> <dateAsTimestamp>true</dateAsTimestamp> <inputSchema>${DB_SCHEMA}</inputSchema> </database> <generate> <deprecated>false</deprecated> <instanceFields>true</instanceFields> <records>true</records> <relations>true</relations> <pojos>true</pojos> <daos>true</daos> </generate> <target> <packageName>com.stil.generated.mysql</packageName> <directory>target/generated-sources/jooq-mysql</directory> </target> </generator> </configuration> </plugin> ```
defect
datatypeexception cannot convert from false in jooq v i have mysql with maven plugin generated pojos and daos with the jooq v i dont have this exceptions caused by org jooq exception mappingexception an error ocurred when mapping record to class com stil generated mysql tables pojos systemproperties at org jooq impl defaultrecordmapper mutablepojomapper map defaultrecordmapper java caused by org jooq exception datatypeexception cannot convert from false class java lang boolean to class com stil generated mysql tables pojos systempropertieshistory at org jooq tools convert convertall fail convert java full stack trace i have use generated dao function fetchonebyuniquekey this is my tables structure sql create table system properties id int not null auto increment sys key varchar not null sys value varchar default null description varchar not null order column int default null history bit not null default b primary key id unique key key sys key engine innodb default charset create table system properties history id int not null auto increment users id int default null properties id int not null datecreated date not null new value varchar not null primary key id unique key properties id properties id datecreated key users id users id key properties id properties id constraint system properties history ibfk foreign key properties id references system properties id on delete cascade on update cascade constraint system properties history ibfk foreign key users id references users id on delete set null on update cascade engine innodb default charset maybe this is problem in jooq codegen maven plugin with v i have generated pojo java public class systemproperties implements serializable private static final long serialversionuid private integer id private string syskey private string sysvalue private string description private integer ordercolumn private systempropertieshistory history this is object instead of boolean in mysql table structure there is not fk and with jooq codegen maven v its boolean public class systempropertieshistory implements serializable private static final long serialversionuid private integer id private integer usersid private integer propertiesid private timestamp datecreated private string newvalue jooq codegen maven config xml org jooq jooq codegen maven jooq version generate mysql generate sources generate mysql mysql connector java com mysql jdbc driver db url db user db pass jooq generator db dialect true db schema false true true true true true com stil generated mysql target generated sources jooq mysql
1
290,645
25,082,599,869
IssuesEvent
2022-11-07 20:42:30
irods/irods
https://api.github.com/repos/irods/irods
closed
Remove `rule_texts_for_tests.py`
testing refactor
- [x] main - [x] 4-3-stable --- Consider moving these into their respsective test. Having part of the test logic in a separate file introduces more hurdles to understanding the test. _Originally posted by @korydraughn in https://github.com/irods/irods/pull/6676#discussion_r1010702458_
1.0
Remove `rule_texts_for_tests.py` - - [x] main - [x] 4-3-stable --- Consider moving these into their respsective test. Having part of the test logic in a separate file introduces more hurdles to understanding the test. _Originally posted by @korydraughn in https://github.com/irods/irods/pull/6676#discussion_r1010702458_
non_defect
remove rule texts for tests py main stable consider moving these into their respsective test having part of the test logic in a separate file introduces more hurdles to understanding the test originally posted by korydraughn in
0
77,567
27,054,726,270
IssuesEvent
2023-02-13 15:27:10
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
opened
[🐛 Bug]: Network Interception doesn't work in JavaScript properly, always giving back status 200!
I-defect needs-triaging
### What happened? I'm using Selenium with Javascript. async function mockResponseStatus(url: string){ const connection = await driver.createCDPConnection('page'); let httpResponse = new HttpResponse(url); httpResponse.addHeaders("Content-Type", "application/json") httpResponse.body = "test"; httpResponse.status = 404; httpResponse.returnStatus = 404; await driver.onIntercept(connection, httpResponse, async function () { }); } I'm trying to set up the interception but unfortunately it is always giving back status 200 - OK. I can set the body and headers but the status and method is always hardcoded. I've checked out the code and it doesn't seem right: https://github.com/SeleniumHQ/selenium/blob/trunk/javascript/node/selenium-webdriver/lib/webdriver.js ![image](https://user-images.githubusercontent.com/4748781/218498697-22c0a7e5-8e02-471a-9836-a904966a2f85.png) I can see that there were changes related to this method: ![image](https://user-images.githubusercontent.com/4748781/218499588-b06ac3c2-0396-4719-a9d2-db734fa3015d.png) https://github.com/SeleniumHQ/selenium/commit/85a785684d8ce37f99cb64c22597c44021d6fb34 Could you please fix the issue? ### How can we reproduce the issue? ```shell async function mockResponseStatus(url: string){ const connection = await driver.createCDPConnection('page'); let httpResponse = new HttpResponse(url); httpResponse.addHeaders("Content-Type", "application/json") httpResponse.body = "test"; httpResponse.status = 404; httpResponse.returnStatus = 404; await driver.onIntercept(connection, httpResponse, async function () { }); } ``` ### Relevant log output ```shell Response body in chrome browser is always: test Status is 200 ``` ### Operating System Windows 10 ### Selenium version 4.8.0 ### What are the browser(s) and version(s) where you see this issue? 109.0.5414.120 ### What are the browser driver(s) and version(s) where you see this issue? 109.0.5414.120 ### Are you using Selenium Grid? no
1.0
[🐛 Bug]: Network Interception doesn't work in JavaScript properly, always giving back status 200! - ### What happened? I'm using Selenium with Javascript. async function mockResponseStatus(url: string){ const connection = await driver.createCDPConnection('page'); let httpResponse = new HttpResponse(url); httpResponse.addHeaders("Content-Type", "application/json") httpResponse.body = "test"; httpResponse.status = 404; httpResponse.returnStatus = 404; await driver.onIntercept(connection, httpResponse, async function () { }); } I'm trying to set up the interception but unfortunately it is always giving back status 200 - OK. I can set the body and headers but the status and method is always hardcoded. I've checked out the code and it doesn't seem right: https://github.com/SeleniumHQ/selenium/blob/trunk/javascript/node/selenium-webdriver/lib/webdriver.js ![image](https://user-images.githubusercontent.com/4748781/218498697-22c0a7e5-8e02-471a-9836-a904966a2f85.png) I can see that there were changes related to this method: ![image](https://user-images.githubusercontent.com/4748781/218499588-b06ac3c2-0396-4719-a9d2-db734fa3015d.png) https://github.com/SeleniumHQ/selenium/commit/85a785684d8ce37f99cb64c22597c44021d6fb34 Could you please fix the issue? ### How can we reproduce the issue? ```shell async function mockResponseStatus(url: string){ const connection = await driver.createCDPConnection('page'); let httpResponse = new HttpResponse(url); httpResponse.addHeaders("Content-Type", "application/json") httpResponse.body = "test"; httpResponse.status = 404; httpResponse.returnStatus = 404; await driver.onIntercept(connection, httpResponse, async function () { }); } ``` ### Relevant log output ```shell Response body in chrome browser is always: test Status is 200 ``` ### Operating System Windows 10 ### Selenium version 4.8.0 ### What are the browser(s) and version(s) where you see this issue? 109.0.5414.120 ### What are the browser driver(s) and version(s) where you see this issue? 109.0.5414.120 ### Are you using Selenium Grid? no
defect
network interception doesn t work in javascript properly always giving back status what happened i m using selenium with javascript async function mockresponsestatus url string const connection await driver createcdpconnection page let httpresponse new httpresponse url httpresponse addheaders content type application json httpresponse body test httpresponse status httpresponse returnstatus await driver onintercept connection httpresponse async function i m trying to set up the interception but unfortunately it is always giving back status ok i can set the body and headers but the status and method is always hardcoded i ve checked out the code and it doesn t seem right i can see that there were changes related to this method could you please fix the issue how can we reproduce the issue shell async function mockresponsestatus url string const connection await driver createcdpconnection page let httpresponse new httpresponse url httpresponse addheaders content type application json httpresponse body test httpresponse status httpresponse returnstatus await driver onintercept connection httpresponse async function relevant log output shell response body in chrome browser is always test status is operating system windows selenium version what are the browser s and version s where you see this issue what are the browser driver s and version s where you see this issue are you using selenium grid no
1
32,106
6,715,706,308
IssuesEvent
2017-10-13 22:38:10
zotonic/zotonic
https://api.github.com/repos/zotonic/zotonic
closed
gen_smtp timeouts marked as permanent failure
defect
Right now we classify timeouts (after some retries) as permanent failure and mark the email address failed in email_status administration. This should be a temporary failure. Also, the current connect timeout in gen_smtp is hard coded to 5000 msec. This should become an option, as apparently 5000 msec is not enough for some smtp servers. /cc @michielklonhammer
1.0
gen_smtp timeouts marked as permanent failure - Right now we classify timeouts (after some retries) as permanent failure and mark the email address failed in email_status administration. This should be a temporary failure. Also, the current connect timeout in gen_smtp is hard coded to 5000 msec. This should become an option, as apparently 5000 msec is not enough for some smtp servers. /cc @michielklonhammer
defect
gen smtp timeouts marked as permanent failure right now we classify timeouts after some retries as permanent failure and mark the email address failed in email status administration this should be a temporary failure also the current connect timeout in gen smtp is hard coded to msec this should become an option as apparently msec is not enough for some smtp servers cc michielklonhammer
1
47,294
11,996,756,921
IssuesEvent
2020-04-08 17:19:06
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
opened
healthcare.api-client.dicom.dicomweb_test: test_dicomweb_delete_study failed
buildcop: issue priority: p1 type: bug
This test failed! To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop). If I'm commenting on this issue too often, add the `buildcop: quiet` label and I will stop commenting. --- commit: 442efa40a52831655aa025cd60a5c6bc064c6ca2 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/ebec6e86-7577-4ce5-bee9-a23363e7c9fd), [Sponge](http://sponge2/ebec6e86-7577-4ce5-bee9-a23363e7c9fd) status: failed
1.0
healthcare.api-client.dicom.dicomweb_test: test_dicomweb_delete_study failed - This test failed! To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop). If I'm commenting on this issue too often, add the `buildcop: quiet` label and I will stop commenting. --- commit: 442efa40a52831655aa025cd60a5c6bc064c6ca2 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/ebec6e86-7577-4ce5-bee9-a23363e7c9fd), [Sponge](http://sponge2/ebec6e86-7577-4ce5-bee9-a23363e7c9fd) status: failed
non_defect
healthcare api client dicom dicomweb test test dicomweb delete study failed this test failed to configure my behavior see if i m commenting on this issue too often add the buildcop quiet label and i will stop commenting commit buildurl status failed
0
79,936
7,733,764,236
IssuesEvent
2018-05-26 15:45:50
NMGRL/pychron
https://api.github.com/repos/NMGRL/pychron
closed
Error on deleting IA
Bug Data Specific Ready to close Tested OK Testing Required
active branch=None active analyses=66196-01,66196-02,66196-03,66196-04,66196-05,66196-06,66196-07,66196-08,66196-09 description="" Traceback ``` Traceback (most recent call last): File "/Users/mcintosh/miniconda3/envs/pychron3/lib/python3.5/site-packages/pyface/ui/qt4/action/action_item.py", line 163, in _qt4_on_triggered self.controller.perform(action, action_event) File "/Users/mcintosh/miniconda3/envs/pychron3/lib/python3.5/site-packages/traitsui/qt4/editor.py", line 269, in perform self.ui.do_undoable( self._perform, action ) File "/Users/mcintosh/miniconda3/envs/pychron3/lib/python3.5/site-packages/traitsui/ui.py", line 716, in do_undoable action( *args, **kw ) File "/Users/mcintosh/miniconda3/envs/pychron3/lib/python3.5/site-packages/traitsui/qt4/editor.py", line 291, in _perform method( info, selection ) File "/Users/mcintosh/Programming/git/pychron/pychron/envisage/browser/sample_view.py", line 392, in delete obj.delete() File "/Users/mcintosh/Programming/git/pychron/pychron/envisage/browser/interpreted_age_table.py", line 63, in delete os.remove(r.path) FileNotFoundError: [Errno 2] No such file or directory: '/Users/mcintosh/Pychron_dev/data/.dvc/repositories/Hanneman01074/661/ia/96.ia.json' ```
2.0
Error on deleting IA - active branch=None active analyses=66196-01,66196-02,66196-03,66196-04,66196-05,66196-06,66196-07,66196-08,66196-09 description="" Traceback ``` Traceback (most recent call last): File "/Users/mcintosh/miniconda3/envs/pychron3/lib/python3.5/site-packages/pyface/ui/qt4/action/action_item.py", line 163, in _qt4_on_triggered self.controller.perform(action, action_event) File "/Users/mcintosh/miniconda3/envs/pychron3/lib/python3.5/site-packages/traitsui/qt4/editor.py", line 269, in perform self.ui.do_undoable( self._perform, action ) File "/Users/mcintosh/miniconda3/envs/pychron3/lib/python3.5/site-packages/traitsui/ui.py", line 716, in do_undoable action( *args, **kw ) File "/Users/mcintosh/miniconda3/envs/pychron3/lib/python3.5/site-packages/traitsui/qt4/editor.py", line 291, in _perform method( info, selection ) File "/Users/mcintosh/Programming/git/pychron/pychron/envisage/browser/sample_view.py", line 392, in delete obj.delete() File "/Users/mcintosh/Programming/git/pychron/pychron/envisage/browser/interpreted_age_table.py", line 63, in delete os.remove(r.path) FileNotFoundError: [Errno 2] No such file or directory: '/Users/mcintosh/Pychron_dev/data/.dvc/repositories/Hanneman01074/661/ia/96.ia.json' ```
non_defect
error on deleting ia active branch none active analyses description traceback traceback most recent call last file users mcintosh envs lib site packages pyface ui action action item py line in on triggered self controller perform action action event file users mcintosh envs lib site packages traitsui editor py line in perform self ui do undoable self perform action file users mcintosh envs lib site packages traitsui ui py line in do undoable action args kw file users mcintosh envs lib site packages traitsui editor py line in perform method info selection file users mcintosh programming git pychron pychron envisage browser sample view py line in delete obj delete file users mcintosh programming git pychron pychron envisage browser interpreted age table py line in delete os remove r path filenotfounderror no such file or directory users mcintosh pychron dev data dvc repositories ia ia json
0
71,346
23,579,402,025
IssuesEvent
2022-08-23 06:07:46
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
Logger name repeated twice
T: Defect
### Expected behavior ``` 2022-08-23 08:51:15.717 [Test worker] INFO org.jooq.Constants - .... ``` The `org.jooq.Constants` correctly reflects the source of the message. ### Actual behavior ``` 2022-08-23 08:51:15.717 [Test worker] INFO org.jooq.Constants.org.jooq.Constants - ``` ``` 2022-08-23 08:51:15.743 [Test worker] DEBUG org.jooq.tools.LoggerListener.org.jooq.tools.LoggerListener - Batch size : 3 2022-08-23 08:51:15.844 [Test worker] DEBUG org.jooq.tools.LoggerListener.org.jooq.tools.LoggerListener - Executing batch query : ``` The location of the log message is duplicated: `org.jooq.Constants.org.jooq.Constants` ### Steps to reproduce the problem I'm using JOOQ with SLF4J interface (1.7.36) and [Log4J2 Implementation](https://mvnrepository.com/artifact/org.apache.logging.log4j/log4j-slf4j-impl) (2.18.0). I'm using the following lo4j2 pattern: ```xml <Console name="console" target="SYSTEM_OUT"> <PatternLayout pattern="%d{yyyy-MM-dd HH:mm:ss.SSS} %style{[%t]}{magenta} %highlight{%5level}{FATAL=red blink, ERROR=red, WARN=yellow bold, INFO=green, DEBUG=blue, TRACE=normal} %style{%logger}{cyan} - %msg%n"/> </Console> ``` I think the culprit is this code in `org.jooq.tools.JooqLogger`: ```java public static JooqLogger getLogger(Class<?> clazz) { return getLogger(clazz, clazz.getName(), -1); } // ... public static JooqLogger getLogger(Class<?> clazz, String nameSuffix, int limitMessages) { String name = clazz != null && nameSuffix != null ? (clazz.getName() + "." + nameSuffix) : clazz != null ? clazz.getName() : nameSuffix; JooqLogger result = new JooqLogger(limitMessages); // Prioritise slf4j try { result.slf4j = org.slf4j.LoggerFactory.getLogger(name); } // ... } ``` that's being called like this: ``` private static final JooqLogger log = JooqLogger.getLogger(LoggerListener.class); ``` ### Versions - jOOQ: 3.17.2 - Java: 17 LTS - Database (include vendor): Postgres 14 - OS: Linux - JDBC Driver (include name if unofficial driver): Official PG JDBC
1.0
Logger name repeated twice - ### Expected behavior ``` 2022-08-23 08:51:15.717 [Test worker] INFO org.jooq.Constants - .... ``` The `org.jooq.Constants` correctly reflects the source of the message. ### Actual behavior ``` 2022-08-23 08:51:15.717 [Test worker] INFO org.jooq.Constants.org.jooq.Constants - ``` ``` 2022-08-23 08:51:15.743 [Test worker] DEBUG org.jooq.tools.LoggerListener.org.jooq.tools.LoggerListener - Batch size : 3 2022-08-23 08:51:15.844 [Test worker] DEBUG org.jooq.tools.LoggerListener.org.jooq.tools.LoggerListener - Executing batch query : ``` The location of the log message is duplicated: `org.jooq.Constants.org.jooq.Constants` ### Steps to reproduce the problem I'm using JOOQ with SLF4J interface (1.7.36) and [Log4J2 Implementation](https://mvnrepository.com/artifact/org.apache.logging.log4j/log4j-slf4j-impl) (2.18.0). I'm using the following lo4j2 pattern: ```xml <Console name="console" target="SYSTEM_OUT"> <PatternLayout pattern="%d{yyyy-MM-dd HH:mm:ss.SSS} %style{[%t]}{magenta} %highlight{%5level}{FATAL=red blink, ERROR=red, WARN=yellow bold, INFO=green, DEBUG=blue, TRACE=normal} %style{%logger}{cyan} - %msg%n"/> </Console> ``` I think the culprit is this code in `org.jooq.tools.JooqLogger`: ```java public static JooqLogger getLogger(Class<?> clazz) { return getLogger(clazz, clazz.getName(), -1); } // ... public static JooqLogger getLogger(Class<?> clazz, String nameSuffix, int limitMessages) { String name = clazz != null && nameSuffix != null ? (clazz.getName() + "." + nameSuffix) : clazz != null ? clazz.getName() : nameSuffix; JooqLogger result = new JooqLogger(limitMessages); // Prioritise slf4j try { result.slf4j = org.slf4j.LoggerFactory.getLogger(name); } // ... } ``` that's being called like this: ``` private static final JooqLogger log = JooqLogger.getLogger(LoggerListener.class); ``` ### Versions - jOOQ: 3.17.2 - Java: 17 LTS - Database (include vendor): Postgres 14 - OS: Linux - JDBC Driver (include name if unofficial driver): Official PG JDBC
defect
logger name repeated twice expected behavior info org jooq constants the org jooq constants correctly reflects the source of the message actual behavior info org jooq constants org jooq constants debug org jooq tools loggerlistener org jooq tools loggerlistener batch size debug org jooq tools loggerlistener org jooq tools loggerlistener executing batch query the location of the log message is duplicated org jooq constants org jooq constants steps to reproduce the problem i m using jooq with interface and i m using the following pattern xml i think the culprit is this code in org jooq tools jooqlogger java public static jooqlogger getlogger class clazz return getlogger clazz clazz getname public static jooqlogger getlogger class clazz string namesuffix int limitmessages string name clazz null namesuffix null clazz getname namesuffix clazz null clazz getname namesuffix jooqlogger result new jooqlogger limitmessages prioritise try result org loggerfactory getlogger name that s being called like this private static final jooqlogger log jooqlogger getlogger loggerlistener class versions jooq java lts database include vendor postgres os linux jdbc driver include name if unofficial driver official pg jdbc
1
45,263
12,692,520,332
IssuesEvent
2020-06-21 23:04:32
cakephp/bake
https://api.github.com/repos/cakephp/bake
closed
bake --connection doesn't work for anything other than the 'default' database connection in CakePHP 4
Defect On Hold
This is a (multiple allowed): bug * CakePHP Version: 4.0.8 * Platform and Target: CentOS. Command line. ### What you did As per https://stackoverflow.com/questions/62215177/cakephp-4-bake-wont-let-me-use-a-different-database-connection-to-default I have configured `config/app_local.php` so there are connections to 2 different databases on the same server: 'Datasources' => [ 'default' => [ 'host' => 'localhost', 'username' => '***', 'password' => '***', 'database' => 'db1', 'url' => env('DATABASE_URL', null), ], 'interface_db' => [ 'host' => 'localhost', 'username' => '***', 'password' => '***', 'database' => 'db2', 'url' => env('DATABASE_URL', null), ] ] When I `bake` using the `default` database, everything works fine. When I want to connect to the database defined in the `interface_db` Datasource key, it doesn't work. php bin/cake.php bake model -cinterface_db Also tried: php bin/cake.php bake model --connection interface_db Gives an error: Error: [Cake\Datasource\Exception\MissingDatasourceConfigException] The datasource configuration "interface_db" was not found. in .../vendor/cakephp/cakephp/src/Datasource/ConnectionManager.php on line 203 ### What happened The application gives an error so nothing is baked. If I run the same command with the `default` connection: php bin/cake.php bake model or: bin/cake.php bake model --connection default It works fine. ### What you expected to happen It should work with any valid array key in Datasources when passed to `-c` or `--connection`. This worked fine in CakePHP 3.x. This is my first 4.x project so I don't know if something has broken when it's been updated.
1.0
bake --connection doesn't work for anything other than the 'default' database connection in CakePHP 4 - This is a (multiple allowed): bug * CakePHP Version: 4.0.8 * Platform and Target: CentOS. Command line. ### What you did As per https://stackoverflow.com/questions/62215177/cakephp-4-bake-wont-let-me-use-a-different-database-connection-to-default I have configured `config/app_local.php` so there are connections to 2 different databases on the same server: 'Datasources' => [ 'default' => [ 'host' => 'localhost', 'username' => '***', 'password' => '***', 'database' => 'db1', 'url' => env('DATABASE_URL', null), ], 'interface_db' => [ 'host' => 'localhost', 'username' => '***', 'password' => '***', 'database' => 'db2', 'url' => env('DATABASE_URL', null), ] ] When I `bake` using the `default` database, everything works fine. When I want to connect to the database defined in the `interface_db` Datasource key, it doesn't work. php bin/cake.php bake model -cinterface_db Also tried: php bin/cake.php bake model --connection interface_db Gives an error: Error: [Cake\Datasource\Exception\MissingDatasourceConfigException] The datasource configuration "interface_db" was not found. in .../vendor/cakephp/cakephp/src/Datasource/ConnectionManager.php on line 203 ### What happened The application gives an error so nothing is baked. If I run the same command with the `default` connection: php bin/cake.php bake model or: bin/cake.php bake model --connection default It works fine. ### What you expected to happen It should work with any valid array key in Datasources when passed to `-c` or `--connection`. This worked fine in CakePHP 3.x. This is my first 4.x project so I don't know if something has broken when it's been updated.
defect
bake connection doesn t work for anything other than the default database connection in cakephp this is a multiple allowed bug cakephp version platform and target centos command line what you did as per i have configured config app local php so there are connections to different databases on the same server datasources default host localhost username password database url env database url null interface db host localhost username password database url env database url null when i bake using the default database everything works fine when i want to connect to the database defined in the interface db datasource key it doesn t work php bin cake php bake model cinterface db also tried php bin cake php bake model connection interface db gives an error error the datasource configuration interface db was not found in vendor cakephp cakephp src datasource connectionmanager php on line what happened the application gives an error so nothing is baked if i run the same command with the default connection php bin cake php bake model or bin cake php bake model connection default it works fine what you expected to happen it should work with any valid array key in datasources when passed to c or connection this worked fine in cakephp x this is my first x project so i don t know if something has broken when it s been updated
1
77,969
27,257,957,932
IssuesEvent
2023-02-22 12:58:07
EightShapes/specs-plugin-feedback
https://api.github.com/repos/EightShapes/specs-plugin-feedback
closed
Does not work with SAP Stencils
defect wontfix
I downloaded the stencils and tried it with the plugin, but it does not work :-( It does not produce specs :-(
1.0
Does not work with SAP Stencils - I downloaded the stencils and tried it with the plugin, but it does not work :-( It does not produce specs :-(
defect
does not work with sap stencils i downloaded the stencils and tried it with the plugin but it does not work it does not produce specs
1
34,500
14,410,621,134
IssuesEvent
2020-12-04 05:22:05
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[Index Patterns] Field Format 'undefined' not found errors
Feature:FieldFormatters Feature:Index Patterns Team:AppServices blocker bug v7.11.0
**Kibana version:** master **Describe the bug:** Index patterns with scripted fields cause `Field Format 'undefined' not found errors` in Discover, breaking the data table rendering **Steps to reproduce:** 1. Add kibana_sample_data_logs data 2. Go to Discover 3. Select kibana_sample_data_logs as index pattern **Expected behavior:** Data table should be displayed, no error messages in console. **Screenshots (if relevant):** ![Bildschirmfoto 2020-11-06 um 10 04 21](https://user-images.githubusercontent.com/463851/98347404-a89b0380-2017-11eb-922a-aa50a762f767.png) **Any additional context:** It's caused by the `hour_of_day` field, which is scripted
1.0
[Index Patterns] Field Format 'undefined' not found errors - **Kibana version:** master **Describe the bug:** Index patterns with scripted fields cause `Field Format 'undefined' not found errors` in Discover, breaking the data table rendering **Steps to reproduce:** 1. Add kibana_sample_data_logs data 2. Go to Discover 3. Select kibana_sample_data_logs as index pattern **Expected behavior:** Data table should be displayed, no error messages in console. **Screenshots (if relevant):** ![Bildschirmfoto 2020-11-06 um 10 04 21](https://user-images.githubusercontent.com/463851/98347404-a89b0380-2017-11eb-922a-aa50a762f767.png) **Any additional context:** It's caused by the `hour_of_day` field, which is scripted
non_defect
field format undefined not found errors kibana version master describe the bug index patterns with scripted fields cause field format undefined not found errors in discover breaking the data table rendering steps to reproduce add kibana sample data logs data go to discover select kibana sample data logs as index pattern expected behavior data table should be displayed no error messages in console screenshots if relevant any additional context it s caused by the hour of day field which is scripted
0
15,771
2,869,064,866
IssuesEvent
2015-06-05 23:02:52
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
include additional options in pub build?
Area-Pkg Pkg-Polymer PolymerMilestone-Later Priority-Medium Triaged Type-Defect
some options we've heard requested from users: (a) allow to choose whether or not to include certain polyfills (e.g. if you intend to run only on a browser that has native support for shadow dom, for example) (b) merge files together (e.g. single deploy target) I'm not sure yet that the right solution is to make polymer transformers configurable this way. An alternative I like is to provide a separate package that optimizes your .html for deployment. For instance: &nbsp;- removes dart.js &nbsp;- it's configurable to remove stuff you don't need &nbsp;- combines .js files together if possible (addresses (b)) &nbsp;- it can also inline images and .css This might be related to issue #14537 as well
1.0
include additional options in pub build? - some options we've heard requested from users: (a) allow to choose whether or not to include certain polyfills (e.g. if you intend to run only on a browser that has native support for shadow dom, for example) (b) merge files together (e.g. single deploy target) I'm not sure yet that the right solution is to make polymer transformers configurable this way. An alternative I like is to provide a separate package that optimizes your .html for deployment. For instance: &nbsp;- removes dart.js &nbsp;- it's configurable to remove stuff you don't need &nbsp;- combines .js files together if possible (addresses (b)) &nbsp;- it can also inline images and .css This might be related to issue #14537 as well
defect
include additional options in pub build some options we ve heard requested from users a allow to choose whether or not to include certain polyfills e g if you intend to run only on a browser that has native support for shadow dom for example b merge files together e g single deploy target i m not sure yet that the right solution is to make polymer transformers configurable this way an alternative i like is to provide a separate package that optimizes your html for deployment for instance nbsp removes dart js nbsp it s configurable to remove stuff you don t need nbsp combines js files together if possible addresses b nbsp it can also inline images and css this might be related to issue as well
1
775,690
27,235,409,986
IssuesEvent
2023-02-21 15:59:12
ascheid/itsg33-pbmm-issue-gen
https://api.github.com/repos/ascheid/itsg33-pbmm-issue-gen
closed
AC-1: Access Control Policy And Procedures
Priority: P1 Control: AC-1 Class: Technical Suggested Assignment: IT Security Function ITSG-33
#Control Definition (A) The organization develops, documents, and disseminates to [Assignment: organization-defined personnel or roles]: (a) An access control policy that addresses purpose, scope, roles, responsibilities, management commitment, coordination among organizational entities, and compliance; and (b) Procedures to facilitate the implementation of the access control policy and associated access controls. (B) The organization reviews and updates the current: (a) Access control policy [Assignment: organization-defined frequency]; and (b) Access control procedures [Assignment: organization-defined frequency]. #Class Technical #Supplemental Guidance This control addresses the establishment of policy and procedures for the effective implementation of selected security controls and control enhancements in the AC family. Policy and procedures reflect applicable GC legislation and TBS policies, directives and standards. Security program policies and procedures at the organization level may make the need for system-specific policies and procedures unnecessary. The policy can be included as part of the general information security policy for organizations or conversely, can be represented by multiple policies reflecting the complex nature of certain organizations. The procedures can be established for the security program in general and for particular information systems, if needed. The organizational risk management strategy is a key factor in establishing policy and procedures #References TBS Operational Security Standard - Management of Information Technology Security [Reference 7]. #Suggested Placeholder Values X #Profile Specific Notes (A) (B) frequency [at a frequency no longer than annually] #Suggested Assignment IT Security Function #Support Teams Learning Center
1.0
AC-1: Access Control Policy And Procedures - #Control Definition (A) The organization develops, documents, and disseminates to [Assignment: organization-defined personnel or roles]: (a) An access control policy that addresses purpose, scope, roles, responsibilities, management commitment, coordination among organizational entities, and compliance; and (b) Procedures to facilitate the implementation of the access control policy and associated access controls. (B) The organization reviews and updates the current: (a) Access control policy [Assignment: organization-defined frequency]; and (b) Access control procedures [Assignment: organization-defined frequency]. #Class Technical #Supplemental Guidance This control addresses the establishment of policy and procedures for the effective implementation of selected security controls and control enhancements in the AC family. Policy and procedures reflect applicable GC legislation and TBS policies, directives and standards. Security program policies and procedures at the organization level may make the need for system-specific policies and procedures unnecessary. The policy can be included as part of the general information security policy for organizations or conversely, can be represented by multiple policies reflecting the complex nature of certain organizations. The procedures can be established for the security program in general and for particular information systems, if needed. The organizational risk management strategy is a key factor in establishing policy and procedures #References TBS Operational Security Standard - Management of Information Technology Security [Reference 7]. #Suggested Placeholder Values X #Profile Specific Notes (A) (B) frequency [at a frequency no longer than annually] #Suggested Assignment IT Security Function #Support Teams Learning Center
non_defect
ac access control policy and procedures control definition a the organization develops documents and disseminates to a an access control policy that addresses purpose scope roles responsibilities management commitment coordination among organizational entities and compliance and b procedures to facilitate the implementation of the access control policy and associated access controls b the organization reviews and updates the current a access control policy and b access control procedures class technical supplemental guidance this control addresses the establishment of policy and procedures for the effective implementation of selected security controls and control enhancements in the ac family policy and procedures reflect applicable gc legislation and tbs policies directives and standards security program policies and procedures at the organization level may make the need for system specific policies and procedures unnecessary the policy can be included as part of the general information security policy for organizations or conversely can be represented by multiple policies reflecting the complex nature of certain organizations the procedures can be established for the security program in general and for particular information systems if needed the organizational risk management strategy is a key factor in establishing policy and procedures references tbs operational security standard management of information technology security suggested placeholder values x profile specific notes a b frequency suggested assignment it security function support teams learning center
0
23,021
3,750,737,538
IssuesEvent
2016-03-11 08:50:48
contao/core
https://api.github.com/repos/contao/core
closed
E_DEPRECATED wird nicht beachtet
defect
Das errorlog von Contao wächst in wenigen Stunden um mehrere MB an, da Contao nicht meine PHP Einstellungen beachtet, das ich keine deprecated Warnungen loggen oder sehen möchte. Das Problem tritt nun extrem auf, da Deprecated Meldungen in der generateFrontendUrl eingefügt wurden.
1.0
E_DEPRECATED wird nicht beachtet - Das errorlog von Contao wächst in wenigen Stunden um mehrere MB an, da Contao nicht meine PHP Einstellungen beachtet, das ich keine deprecated Warnungen loggen oder sehen möchte. Das Problem tritt nun extrem auf, da Deprecated Meldungen in der generateFrontendUrl eingefügt wurden.
defect
e deprecated wird nicht beachtet das errorlog von contao wächst in wenigen stunden um mehrere mb an da contao nicht meine php einstellungen beachtet das ich keine deprecated warnungen loggen oder sehen möchte das problem tritt nun extrem auf da deprecated meldungen in der generatefrontendurl eingefügt wurden
1
2,083
2,603,976,007
IssuesEvent
2015-02-24 19:01:31
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳冠状沟长颗粒怎么回事
auto-migrated Priority-Medium Type-Defect
``` 沈阳冠状沟长颗粒怎么回事〓沈陽軍區政治部醫院性病〓TEL�� �024-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治� ��。位于沈陽市沈河區二緯路32號。是一所與新中國同建立共� ��煌的歷史悠久、設備精良、技術權威、專家云集,是預防、 保健、醫療、科研康復為一體的綜合性醫院。是國家首批公�� �甲等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大� ��、東南大學等知名高等院校的教學醫院。曾被中國人民解放 軍空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立�� �體二等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:19
1.0
沈阳冠状沟长颗粒怎么回事 - ``` 沈阳冠状沟长颗粒怎么回事〓沈陽軍區政治部醫院性病〓TEL�� �024-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治� ��。位于沈陽市沈河區二緯路32號。是一所與新中國同建立共� ��煌的歷史悠久、設備精良、技術權威、專家云集,是預防、 保健、醫療、科研康復為一體的綜合性醫院。是國家首批公�� �甲等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大� ��、東南大學等知名高等院校的教學醫院。曾被中國人民解放 軍空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立�� �體二等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:19
defect
沈阳冠状沟长颗粒怎么回事 沈阳冠状沟长颗粒怎么回事〓沈陽軍區政治部醫院性病〓tel�� � 〓 , � ��。 。是一所與新中國同建立共� ��煌的歷史悠久、設備精良、技術權威、專家云集,是預防、 保健、醫療、科研康復為一體的綜合性醫院。是國家首批公�� �甲等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大� ��、東南大學等知名高等院校的教學醫院。曾被中國人民解放 軍空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立�� �體二等功。 original issue reported on code google com by gmail com on jun at
1
63,698
17,863,826,665
IssuesEvent
2021-09-06 06:53:22
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
opened
Cannot send video
T-Defect
### Steps to reproduce 1. Click to + icon 2. Select MP4 video (about 113MB) 3. Send in original size 4. "Encrypting message" and few seconds later "Messages failed to send" appears 5. Retry recompress the video to 5MB and sending is successful ![obrazek](https://user-images.githubusercontent.com/15554561/132172948-78c7c449-b23e-4d53-8077-2832e6dcf218.png) It does for ### What happened? ### What did you expect? Video is sent in original size ### What happened? Video is not sent ### Your phone model Xiaomi Mi 9T ### Operating system version Android 11 ### Application version and app store Element 1.2.0 ### Homeserver bolesiny.net ### Have you submitted a rageshake? Yes
1.0
Cannot send video - ### Steps to reproduce 1. Click to + icon 2. Select MP4 video (about 113MB) 3. Send in original size 4. "Encrypting message" and few seconds later "Messages failed to send" appears 5. Retry recompress the video to 5MB and sending is successful ![obrazek](https://user-images.githubusercontent.com/15554561/132172948-78c7c449-b23e-4d53-8077-2832e6dcf218.png) It does for ### What happened? ### What did you expect? Video is sent in original size ### What happened? Video is not sent ### Your phone model Xiaomi Mi 9T ### Operating system version Android 11 ### Application version and app store Element 1.2.0 ### Homeserver bolesiny.net ### Have you submitted a rageshake? Yes
defect
cannot send video steps to reproduce click to icon select video about send in original size encrypting message and few seconds later messages failed to send appears retry recompress the video to and sending is successful it does for what happened what did you expect video is sent in original size what happened video is not sent your phone model xiaomi mi operating system version android application version and app store element homeserver bolesiny net have you submitted a rageshake yes
1
14,924
2,831,390,324
IssuesEvent
2015-05-24 15:55:10
nobodyguy/dslrdashboard
https://api.github.com/repos/nobodyguy/dslrdashboard
closed
D7100 scene mode menu is incompatible
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Select "Mode SCENE" 2. Push the scene mode button in the app and compare what you see with what the camera offers in this mode instead. What is the expected output? What do you see instead? I see a list of scene modes that seems to be derived from a different camera. I would like to see a complete list of the scene modes the D7100 supports. What version of the product are you using? On what operating system? D7100 and Android 4.1.2 on SGS3LTE. Please provide any additional information below. It looks like the current list mixes modes for which the D7100 has two different major modes. "Portrait", for instance, belongs to "Mode SCENE", while "High key" and "Low key" belong to "Mode EFFECTS", which is an extra position in the camera mode dial. With the D7100 Nikon has apparently categorized the scene and effect modes with regard to what requires JPEG (Mode EFFECTS; post-capture image processing), and what works also with RAW (Mode SCENE; pre-capture tweaks, like metering and focus mode). To support this in the app there either could be two buttons for these two major modes, or the button listing the scenes/effects would have to be context sensitive, the latter option saving some screen real estate. ``` Original issue reported on code.google.com by `uwe.doer...@gmail.com` on 9 May 2013 at 11:56
1.0
D7100 scene mode menu is incompatible - ``` What steps will reproduce the problem? 1. Select "Mode SCENE" 2. Push the scene mode button in the app and compare what you see with what the camera offers in this mode instead. What is the expected output? What do you see instead? I see a list of scene modes that seems to be derived from a different camera. I would like to see a complete list of the scene modes the D7100 supports. What version of the product are you using? On what operating system? D7100 and Android 4.1.2 on SGS3LTE. Please provide any additional information below. It looks like the current list mixes modes for which the D7100 has two different major modes. "Portrait", for instance, belongs to "Mode SCENE", while "High key" and "Low key" belong to "Mode EFFECTS", which is an extra position in the camera mode dial. With the D7100 Nikon has apparently categorized the scene and effect modes with regard to what requires JPEG (Mode EFFECTS; post-capture image processing), and what works also with RAW (Mode SCENE; pre-capture tweaks, like metering and focus mode). To support this in the app there either could be two buttons for these two major modes, or the button listing the scenes/effects would have to be context sensitive, the latter option saving some screen real estate. ``` Original issue reported on code.google.com by `uwe.doer...@gmail.com` on 9 May 2013 at 11:56
defect
scene mode menu is incompatible what steps will reproduce the problem select mode scene push the scene mode button in the app and compare what you see with what the camera offers in this mode instead what is the expected output what do you see instead i see a list of scene modes that seems to be derived from a different camera i would like to see a complete list of the scene modes the supports what version of the product are you using on what operating system and android on please provide any additional information below it looks like the current list mixes modes for which the has two different major modes portrait for instance belongs to mode scene while high key and low key belong to mode effects which is an extra position in the camera mode dial with the nikon has apparently categorized the scene and effect modes with regard to what requires jpeg mode effects post capture image processing and what works also with raw mode scene pre capture tweaks like metering and focus mode to support this in the app there either could be two buttons for these two major modes or the button listing the scenes effects would have to be context sensitive the latter option saving some screen real estate original issue reported on code google com by uwe doer gmail com on may at
1
250,717
27,111,239,095
IssuesEvent
2023-02-15 15:27:12
EliyaC/NodeGoat
https://api.github.com/repos/EliyaC/NodeGoat
closed
WS-2020-0180 (High) detected in npm-user-validate-0.1.5.tgz - autoclosed
security vulnerability
## WS-2020-0180 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>npm-user-validate-0.1.5.tgz</b></p></summary> <p>User validations for npm</p> <p>Library home page: <a href="https://registry.npmjs.org/npm-user-validate/-/npm-user-validate-0.1.5.tgz">https://registry.npmjs.org/npm-user-validate/-/npm-user-validate-0.1.5.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/npm-user-validate/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - :x: **npm-user-validate-0.1.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/EliyaC/NodeGoat/commit/2f9ac315d9e05728b7ce26ce7cf1b4e684e54fde">2f9ac315d9e05728b7ce26ce7cf1b4e684e54fde</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package npm-user-validate prior to version 1.0.1 is vulnerable to REDoS. The regex that validates a user's email took exponentially longer to process input strings that begin with the '@' character. <p>Publish Date: 2020-10-16 <p>URL: <a href=https://github.com/npm/npm-user-validate/commit/c8a87dac1a4cc6988b5418f30411a8669bef204e>WS-2020-0180</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-xgh6-85xh-479p">https://github.com/advisories/GHSA-xgh6-85xh-479p</a></p> <p>Release Date: 2020-10-16</p> <p>Fix Resolution: 1.0.1</p> </p> </details> <p></p>
True
WS-2020-0180 (High) detected in npm-user-validate-0.1.5.tgz - autoclosed - ## WS-2020-0180 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>npm-user-validate-0.1.5.tgz</b></p></summary> <p>User validations for npm</p> <p>Library home page: <a href="https://registry.npmjs.org/npm-user-validate/-/npm-user-validate-0.1.5.tgz">https://registry.npmjs.org/npm-user-validate/-/npm-user-validate-0.1.5.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npm/node_modules/npm-user-validate/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - :x: **npm-user-validate-0.1.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/EliyaC/NodeGoat/commit/2f9ac315d9e05728b7ce26ce7cf1b4e684e54fde">2f9ac315d9e05728b7ce26ce7cf1b4e684e54fde</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package npm-user-validate prior to version 1.0.1 is vulnerable to REDoS. The regex that validates a user's email took exponentially longer to process input strings that begin with the '@' character. <p>Publish Date: 2020-10-16 <p>URL: <a href=https://github.com/npm/npm-user-validate/commit/c8a87dac1a4cc6988b5418f30411a8669bef204e>WS-2020-0180</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-xgh6-85xh-479p">https://github.com/advisories/GHSA-xgh6-85xh-479p</a></p> <p>Release Date: 2020-10-16</p> <p>Fix Resolution: 1.0.1</p> </p> </details> <p></p>
non_defect
ws high detected in npm user validate tgz autoclosed ws high severity vulnerability vulnerable library npm user validate tgz user validations for npm library home page a href path to dependency file package json path to vulnerable library node modules npm node modules npm user validate package json dependency hierarchy grunt npm install tgz root library npm tgz x npm user validate tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package npm user validate prior to version is vulnerable to redos the regex that validates a user s email took exponentially longer to process input strings that begin with the character publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
57,741
16,024,609,931
IssuesEvent
2021-04-21 07:28:44
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
Missing address information on SocketTimeoutException
Module: Network I/O Source: Internal Team: Client Team: Core Type: Defect
We have stack traces as follows which makes calls to AbstractChannel.connect. The returned exception does have `address` info on it when the exception is `SocketException`, but it is missing the address info for other cases. Also this log is printed in finest level which seems much more suitable for warning level to pinpoint similar problems faster. ``` [2021-04-15 09:54:02:792 CEST] ***** ClientConnect 3 com.hazelcast.client.impl.connection.ClientConnectionManager ****:******[**********] [4.1] null java.net.SocketTimeoutException at sun.nio.ch.SocketAdaptor.connect(SocketAdaptor.java:130) at com.hazelcast.internal.networking.nio.AbstractChannel.connect(AbstractChannel.java:133) ```
1.0
Missing address information on SocketTimeoutException - We have stack traces as follows which makes calls to AbstractChannel.connect. The returned exception does have `address` info on it when the exception is `SocketException`, but it is missing the address info for other cases. Also this log is printed in finest level which seems much more suitable for warning level to pinpoint similar problems faster. ``` [2021-04-15 09:54:02:792 CEST] ***** ClientConnect 3 com.hazelcast.client.impl.connection.ClientConnectionManager ****:******[**********] [4.1] null java.net.SocketTimeoutException at sun.nio.ch.SocketAdaptor.connect(SocketAdaptor.java:130) at com.hazelcast.internal.networking.nio.AbstractChannel.connect(AbstractChannel.java:133) ```
defect
missing address information on sockettimeoutexception we have stack traces as follows which makes calls to abstractchannel connect the returned exception does have address info on it when the exception is socketexception but it is missing the address info for other cases also this log is printed in finest level which seems much more suitable for warning level to pinpoint similar problems faster clientconnect com hazelcast client impl connection clientconnectionmanager null java net sockettimeoutexception at sun nio ch socketadaptor connect socketadaptor java at com hazelcast internal networking nio abstractchannel connect abstractchannel java
1
59,044
17,015,127,742
IssuesEvent
2021-07-02 10:52:40
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
opened
Merkaartor 0.16 crashes on Shapefile opening
Component: merkaartor Priority: major Type: defect
**[Submitted to the original trac issue database at 9.55am, Thursday, 5th August 2010]** Merkaartor crashes when i try to import a shape file
1.0
Merkaartor 0.16 crashes on Shapefile opening - **[Submitted to the original trac issue database at 9.55am, Thursday, 5th August 2010]** Merkaartor crashes when i try to import a shape file
defect
merkaartor crashes on shapefile opening merkaartor crashes when i try to import a shape file
1
99,405
4,054,602,232
IssuesEvent
2016-05-24 13:01:12
dhis2/dhis2-gis
https://api.github.com/repos/dhis2/dhis2-gis
closed
Remove unused translation strings
high priority translation
There are many strings in dhis2-gis/i18n/i18n_app.properties that are never used in the source files, and they should be removed.
1.0
Remove unused translation strings - There are many strings in dhis2-gis/i18n/i18n_app.properties that are never used in the source files, and they should be removed.
non_defect
remove unused translation strings there are many strings in gis app properties that are never used in the source files and they should be removed
0
25,138
4,216,429,343
IssuesEvent
2016-06-30 09:14:38
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
NPE at InternalPartitionServiceImpl
Team: Core Type: Defect
``` 2016-06-29T08:41:52,506+0200 WARN [hz._hzInstance_1_xxx] com.hazelcast.logging.Slf4jFactory$Slf4jLogger.log(Slf4jFactory$Slf4jLogger.java:90) .20.30.224:5701 -xxx .5.5 java.lang.NullPointerException: null at com.hazelcast.partition.impl.InternalPartitionServiceImpl$RepartitioningTask.run(InternalPartitionServiceImpl.java:1647) at com.hazelcast.partition.impl.InternalPartitionServiceImpl$MigrationThread.processTask(InternalPartitionServiceImpl.java:1896) at com.hazelcast.partition.impl.InternalPartitionServiceImpl$MigrationThread.doRun(InternalPartitionServiceImpl.java:1872) at com.hazelcast.partition.impl.InternalPartitionServiceImpl$MigrationThread.run(InternalPartitionServiceImpl.java:1849) ``` version 3.5.5
1.0
NPE at InternalPartitionServiceImpl - ``` 2016-06-29T08:41:52,506+0200 WARN [hz._hzInstance_1_xxx] com.hazelcast.logging.Slf4jFactory$Slf4jLogger.log(Slf4jFactory$Slf4jLogger.java:90) .20.30.224:5701 -xxx .5.5 java.lang.NullPointerException: null at com.hazelcast.partition.impl.InternalPartitionServiceImpl$RepartitioningTask.run(InternalPartitionServiceImpl.java:1647) at com.hazelcast.partition.impl.InternalPartitionServiceImpl$MigrationThread.processTask(InternalPartitionServiceImpl.java:1896) at com.hazelcast.partition.impl.InternalPartitionServiceImpl$MigrationThread.doRun(InternalPartitionServiceImpl.java:1872) at com.hazelcast.partition.impl.InternalPartitionServiceImpl$MigrationThread.run(InternalPartitionServiceImpl.java:1849) ``` version 3.5.5
defect
npe at internalpartitionserviceimpl warn com hazelcast logging log java xxx java lang nullpointerexception null at com hazelcast partition impl internalpartitionserviceimpl repartitioningtask run internalpartitionserviceimpl java at com hazelcast partition impl internalpartitionserviceimpl migrationthread processtask internalpartitionserviceimpl java at com hazelcast partition impl internalpartitionserviceimpl migrationthread dorun internalpartitionserviceimpl java at com hazelcast partition impl internalpartitionserviceimpl migrationthread run internalpartitionserviceimpl java version
1
29,858
5,921,175,096
IssuesEvent
2017-05-22 22:14:02
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
closed
The order of partial class implementations seems to affect what properties are defined on the target object
defect in progress
In the example below I have two partial implementations of the class **TestClass**. The first definition implements the **IBase** interface, whereas the second both inherits from **BaseClass** and implements **IBase**. The order that these partial definitions are made determines what properties are on the target object (it chooses the first). As the first definition doesn't inherit from **BaseClass** it doesn't seem to have the `Y` property from the **BaseClass** properties, even though it should be inferred from the subsequent partial definition. ### Steps To Reproduce Move the first partial definition (on line 8) below the definition beneath it and you'll see that the properties that appear against the object change. https://deck.net/296544fac4622ddf1e80bb3db722a342 ```cs public class Program { public static void Main() { Console.WriteLine(new TestClass(123, 456)); } partial class TestClass : IBase { } partial class TestClass : BaseClass, IBase { public int X { get; } public TestClass(int x, int y) : base(y) { X = x; } } abstract class BaseClass { public BaseClass(int y) { Y = y; } public int Y { get; } } partial interface IBase { } } ``` ### Expected Result ```js { "X": 123, "Y": 456 } ``` ### Actual Result ```js { "X": 123 } ```
1.0
The order of partial class implementations seems to affect what properties are defined on the target object - In the example below I have two partial implementations of the class **TestClass**. The first definition implements the **IBase** interface, whereas the second both inherits from **BaseClass** and implements **IBase**. The order that these partial definitions are made determines what properties are on the target object (it chooses the first). As the first definition doesn't inherit from **BaseClass** it doesn't seem to have the `Y` property from the **BaseClass** properties, even though it should be inferred from the subsequent partial definition. ### Steps To Reproduce Move the first partial definition (on line 8) below the definition beneath it and you'll see that the properties that appear against the object change. https://deck.net/296544fac4622ddf1e80bb3db722a342 ```cs public class Program { public static void Main() { Console.WriteLine(new TestClass(123, 456)); } partial class TestClass : IBase { } partial class TestClass : BaseClass, IBase { public int X { get; } public TestClass(int x, int y) : base(y) { X = x; } } abstract class BaseClass { public BaseClass(int y) { Y = y; } public int Y { get; } } partial interface IBase { } } ``` ### Expected Result ```js { "X": 123, "Y": 456 } ``` ### Actual Result ```js { "X": 123 } ```
defect
the order of partial class implementations seems to affect what properties are defined on the target object in the example below i have two partial implementations of the class testclass the first definition implements the ibase interface whereas the second both inherits from baseclass and implements ibase the order that these partial definitions are made determines what properties are on the target object it chooses the first as the first definition doesn t inherit from baseclass it doesn t seem to have the y property from the baseclass properties even though it should be inferred from the subsequent partial definition steps to reproduce move the first partial definition on line below the definition beneath it and you ll see that the properties that appear against the object change cs public class program public static void main console writeline new testclass partial class testclass ibase partial class testclass baseclass ibase public int x get public testclass int x int y base y x x abstract class baseclass public baseclass int y y y public int y get partial interface ibase expected result js x y actual result js x
1
86,618
24,905,673,016
IssuesEvent
2022-10-29 07:52:14
aceade/tdntg
https://api.github.com/repos/aceade/tdntg
opened
Build WebGL upon commit
enhancement build
Rather than build and commit to the repo, I should have an automated build that handles this for me. - https://game.ci/docs/github/getting-started
1.0
Build WebGL upon commit - Rather than build and commit to the repo, I should have an automated build that handles this for me. - https://game.ci/docs/github/getting-started
non_defect
build webgl upon commit rather than build and commit to the repo i should have an automated build that handles this for me
0
293,454
25,293,993,903
IssuesEvent
2022-11-17 04:17:20
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: unoptimized-query-oracle/disable-rules=half failed
C-test-failure O-robot S-0-visible-logical-error O-roachtest T-sql-queries branch-release-22.2
roachtest.unoptimized-query-oracle/disable-rules=half [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7161043?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7161043?buildTab=artifacts#/unoptimized-query-oracle/disable-rules=half) on release-22.2 @ [4816df3a9d76d179ed135a2b1efb53babb5611a0](https://github.com/cockroachdb/cockroach/commits/4816df3a9d76d179ed135a2b1efb53babb5611a0): ``` Wraps: (3) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runUnoptimizedQueryOracleImpl | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/unoptimized_query_oracle.go:168 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerUnoptimizedQueryOracle.func1.1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/unoptimized_query_oracle.go:57 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runOneRoundQueryComparison | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/query_comparison_util.go:236 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runQueryComparison | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/query_comparison_util.go:66 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerUnoptimizedQueryOracle.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/unoptimized_query_oracle.go:54 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:930 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1594 Wraps: (4) expected unoptimized and optimized results to be equal |   []string( | -  { | -  "-39 years -5 mons -668 days -20:19:03.089458,31,a ,\x12,1969-09-22 21:13:20 +0000 +0000,0", | -  "-39 years -5 mons -668 days -20:19:03.089458,31,a ,\x12,1981-05-11 17:02:50.000318 +0000 +0000,0", | -  "-39 years -5 mons -668 days -20:19:03.089458,31,a ,\x12,1985-11-08 11:32:06.000741 +0000 +0000,0", | -  }, | +  nil, |   ) | sql: SELECT | '-39 years -5 mons -668 days -20:19:03.089458':::INTERVAL AS col_18662, | 31:::INT8 AS col_18663, | 'a ':::STRING AS col_18664, | e'\x12':::STRING AS col_18665, | tab_7925.col1_0 AS col_18666, | 0:::OID AS col_18667 | FROM | defaultdb.public.table1@[0] AS tab_7919 | JOIN defaultdb.public.table1@[0] AS tab_7920 | JOIN defaultdb.public.table1@[0] AS tab_7921 | JOIN defaultdb.public.table1 AS tab_7922 | JOIN defaultdb.public.table1 AS tab_7923 ON (tab_7922.col1_2) = (tab_7923.col1_2) | LEFT JOIN defaultdb.public.table1@[0] AS tab_7924 ON (tab_7923.col1_2) = (tab_7924.col1_2) | LEFT JOIN defaultdb.public.table1@[0] AS tab_7925 | JOIN defaultdb.public.table1@[0] AS tab_7926 ON (tab_7925.col1_1) = (tab_7926.col1_1) ON | (tab_7924.col1_1) = (tab_7926.col1_1) AND (tab_7924.col1_2) > (tab_7926.col1_2) ON | (tab_7921.col1_1) = (tab_7926.col1_1) AND (tab_7921.col1_1) = (tab_7924.col1_1) ON | (tab_7920.col1_0) = (tab_7926.col1_0) | AND (tab_7920.crdb_internal_mvcc_timestamp) = (tab_7924.crdb_internal_mvcc_timestamp) ON | (tab_7919.col1_0) = (tab_7920.col1_0) | AND (tab_7919.crdb_internal_mvcc_timestamp) = (tab_7921.crdb_internal_mvcc_timestamp) | ORDER BY | tab_7921.tableoid DESC, tab_7919.crdb_internal_mvcc_timestamp Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.leafError ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*unoptimized-query-oracle/disable-rules=half.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-20932
2.0
roachtest: unoptimized-query-oracle/disable-rules=half failed - roachtest.unoptimized-query-oracle/disable-rules=half [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7161043?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7161043?buildTab=artifacts#/unoptimized-query-oracle/disable-rules=half) on release-22.2 @ [4816df3a9d76d179ed135a2b1efb53babb5611a0](https://github.com/cockroachdb/cockroach/commits/4816df3a9d76d179ed135a2b1efb53babb5611a0): ``` Wraps: (3) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runUnoptimizedQueryOracleImpl | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/unoptimized_query_oracle.go:168 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerUnoptimizedQueryOracle.func1.1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/unoptimized_query_oracle.go:57 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runOneRoundQueryComparison | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/query_comparison_util.go:236 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runQueryComparison | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/query_comparison_util.go:66 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerUnoptimizedQueryOracle.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/unoptimized_query_oracle.go:54 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:930 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1594 Wraps: (4) expected unoptimized and optimized results to be equal |   []string( | -  { | -  "-39 years -5 mons -668 days -20:19:03.089458,31,a ,\x12,1969-09-22 21:13:20 +0000 +0000,0", | -  "-39 years -5 mons -668 days -20:19:03.089458,31,a ,\x12,1981-05-11 17:02:50.000318 +0000 +0000,0", | -  "-39 years -5 mons -668 days -20:19:03.089458,31,a ,\x12,1985-11-08 11:32:06.000741 +0000 +0000,0", | -  }, | +  nil, |   ) | sql: SELECT | '-39 years -5 mons -668 days -20:19:03.089458':::INTERVAL AS col_18662, | 31:::INT8 AS col_18663, | 'a ':::STRING AS col_18664, | e'\x12':::STRING AS col_18665, | tab_7925.col1_0 AS col_18666, | 0:::OID AS col_18667 | FROM | defaultdb.public.table1@[0] AS tab_7919 | JOIN defaultdb.public.table1@[0] AS tab_7920 | JOIN defaultdb.public.table1@[0] AS tab_7921 | JOIN defaultdb.public.table1 AS tab_7922 | JOIN defaultdb.public.table1 AS tab_7923 ON (tab_7922.col1_2) = (tab_7923.col1_2) | LEFT JOIN defaultdb.public.table1@[0] AS tab_7924 ON (tab_7923.col1_2) = (tab_7924.col1_2) | LEFT JOIN defaultdb.public.table1@[0] AS tab_7925 | JOIN defaultdb.public.table1@[0] AS tab_7926 ON (tab_7925.col1_1) = (tab_7926.col1_1) ON | (tab_7924.col1_1) = (tab_7926.col1_1) AND (tab_7924.col1_2) > (tab_7926.col1_2) ON | (tab_7921.col1_1) = (tab_7926.col1_1) AND (tab_7921.col1_1) = (tab_7924.col1_1) ON | (tab_7920.col1_0) = (tab_7926.col1_0) | AND (tab_7920.crdb_internal_mvcc_timestamp) = (tab_7924.crdb_internal_mvcc_timestamp) ON | (tab_7919.col1_0) = (tab_7920.col1_0) | AND (tab_7919.crdb_internal_mvcc_timestamp) = (tab_7921.crdb_internal_mvcc_timestamp) | ORDER BY | tab_7921.tableoid DESC, tab_7919.crdb_internal_mvcc_timestamp Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.leafError ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*unoptimized-query-oracle/disable-rules=half.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-20932
non_defect
roachtest unoptimized query oracle disable rules half failed roachtest unoptimized query oracle disable rules half with on release wraps attached stack trace stack trace github com cockroachdb cockroach pkg cmd roachtest tests rununoptimizedqueryoracleimpl github com cockroachdb cockroach pkg cmd roachtest tests unoptimized query oracle go github com cockroachdb cockroach pkg cmd roachtest tests registerunoptimizedqueryoracle github com cockroachdb cockroach pkg cmd roachtest tests unoptimized query oracle go github com cockroachdb cockroach pkg cmd roachtest tests runoneroundquerycomparison github com cockroachdb cockroach pkg cmd roachtest tests query comparison util go github com cockroachdb cockroach pkg cmd roachtest tests runquerycomparison github com cockroachdb cockroach pkg cmd roachtest tests query comparison util go github com cockroachdb cockroach pkg cmd roachtest tests registerunoptimizedqueryoracle github com cockroachdb cockroach pkg cmd roachtest tests unoptimized query oracle go main testrunner runtest main pkg cmd roachtest test runner go runtime goexit goroot src runtime asm s wraps expected unoptimized and optimized results to be equal    string     years mons days a   years mons days a   years mons days a     nil    sql select years mons days interval as col as col a string as col e string as col tab as col oid as col from defaultdb public as tab join defaultdb public as tab join defaultdb public as tab join defaultdb public as tab join defaultdb public as tab on tab tab left join defaultdb public as tab on tab tab left join defaultdb public as tab join defaultdb public as tab on tab tab on tab tab and tab tab on tab tab and tab tab on tab tab and tab crdb internal mvcc timestamp tab crdb internal mvcc timestamp on tab tab and tab crdb internal mvcc timestamp tab crdb internal mvcc timestamp order by tab tableoid desc tab crdb internal mvcc timestamp error types withstack withstack errutil withprefix withstack withstack errutil leaferror parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb sql queries jira issue crdb
0
315,865
23,601,273,321
IssuesEvent
2022-08-24 02:23:43
ellisvalentiner/steampipe-plugin-weatherkit
https://api.github.com/repos/ellisvalentiner/steampipe-plugin-weatherkit
closed
fix(docs): `token` missing in `docs/index.md`
bug documentation
> The configuration section in the docs/index.md file should be a replica of the .spc file. Currently, it is missing information on the token argument. mentioned in #10
1.0
fix(docs): `token` missing in `docs/index.md` - > The configuration section in the docs/index.md file should be a replica of the .spc file. Currently, it is missing information on the token argument. mentioned in #10
non_defect
fix docs token missing in docs index md the configuration section in the docs index md file should be a replica of the spc file currently it is missing information on the token argument mentioned in
0
116,447
17,369,264,384
IssuesEvent
2021-07-30 11:46:47
lukebroganws/Java-Demo
https://api.github.com/repos/lukebroganws/Java-Demo
opened
WS-2014-0034 (High) detected in commons-fileupload-1.3.1.jar
security vulnerability
## WS-2014-0034 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-fileupload-1.3.1.jar</b></p></summary> <p>The Apache Commons FileUpload component provides a simple yet flexible means of adding support for multipart file upload functionality to servlets and web applications.</p> <p>Path to dependency file: Java-Demo/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-fileupload/commons-fileupload/1.3.1/commons-fileupload-1.3.1.jar,Java-Demo/.extract/webapps/ROOT/WEB-INF/lib/commons-fileupload-1.3.1.jar,Java-Demo/target/easybuggy-1-SNAPSHOT/WEB-INF/lib/commons-fileupload-1.3.1.jar</p> <p> Dependency Hierarchy: - esapi-2.1.0.1.jar (Root Library) - :x: **commons-fileupload-1.3.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/Java-Demo/commit/9c1bc5d1780a325ef5a39962950ec2956214bf22">9c1bc5d1780a325ef5a39962950ec2956214bf22</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The class FileUploadBase in Apache Commons Fileupload before 1.4 has potential resource leak - InputStream not closed on exception. <p>Publish Date: 2014-02-17 <p>URL: <a href=https://commons.apache.org/proper/commons-fileupload/changes-report.html>WS-2014-0034</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/apache/commons-fileupload/commit/5b4881d7f75f439326f54fa554a9ca7de6d60814">https://github.com/apache/commons-fileupload/commit/5b4881d7f75f439326f54fa554a9ca7de6d60814</a></p> <p>Release Date: 2019-09-26</p> <p>Fix Resolution: 1.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-fileupload","packageName":"commons-fileupload","packageVersion":"1.3.1","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.owasp.esapi:esapi:2.1.0.1;commons-fileupload:commons-fileupload:1.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4"}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2014-0034","vulnerabilityDetails":"The class FileUploadBase in Apache Commons Fileupload before 1.4 has potential resource leak - InputStream not closed on exception.","vulnerabilityUrl":"https://commons.apache.org/proper/commons-fileupload/changes-report.html","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
WS-2014-0034 (High) detected in commons-fileupload-1.3.1.jar - ## WS-2014-0034 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-fileupload-1.3.1.jar</b></p></summary> <p>The Apache Commons FileUpload component provides a simple yet flexible means of adding support for multipart file upload functionality to servlets and web applications.</p> <p>Path to dependency file: Java-Demo/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-fileupload/commons-fileupload/1.3.1/commons-fileupload-1.3.1.jar,Java-Demo/.extract/webapps/ROOT/WEB-INF/lib/commons-fileupload-1.3.1.jar,Java-Demo/target/easybuggy-1-SNAPSHOT/WEB-INF/lib/commons-fileupload-1.3.1.jar</p> <p> Dependency Hierarchy: - esapi-2.1.0.1.jar (Root Library) - :x: **commons-fileupload-1.3.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/Java-Demo/commit/9c1bc5d1780a325ef5a39962950ec2956214bf22">9c1bc5d1780a325ef5a39962950ec2956214bf22</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The class FileUploadBase in Apache Commons Fileupload before 1.4 has potential resource leak - InputStream not closed on exception. <p>Publish Date: 2014-02-17 <p>URL: <a href=https://commons.apache.org/proper/commons-fileupload/changes-report.html>WS-2014-0034</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/apache/commons-fileupload/commit/5b4881d7f75f439326f54fa554a9ca7de6d60814">https://github.com/apache/commons-fileupload/commit/5b4881d7f75f439326f54fa554a9ca7de6d60814</a></p> <p>Release Date: 2019-09-26</p> <p>Fix Resolution: 1.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-fileupload","packageName":"commons-fileupload","packageVersion":"1.3.1","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.owasp.esapi:esapi:2.1.0.1;commons-fileupload:commons-fileupload:1.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4"}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2014-0034","vulnerabilityDetails":"The class FileUploadBase in Apache Commons Fileupload before 1.4 has potential resource leak - InputStream not closed on exception.","vulnerabilityUrl":"https://commons.apache.org/proper/commons-fileupload/changes-report.html","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_defect
ws high detected in commons fileupload jar ws high severity vulnerability vulnerable library commons fileupload jar the apache commons fileupload component provides a simple yet flexible means of adding support for multipart file upload functionality to servlets and web applications path to dependency file java demo pom xml path to vulnerable library home wss scanner repository commons fileupload commons fileupload commons fileupload jar java demo extract webapps root web inf lib commons fileupload jar java demo target easybuggy snapshot web inf lib commons fileupload jar dependency hierarchy esapi jar root library x commons fileupload jar vulnerable library found in head commit a href found in base branch main vulnerability details the class fileuploadbase in apache commons fileupload before has potential resource leak inputstream not closed on exception publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org owasp esapi esapi commons fileupload commons fileupload isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier ws vulnerabilitydetails the class fileuploadbase in apache commons fileupload before has potential resource leak inputstream not closed on exception vulnerabilityurl
0
40,076
9,819,989,775
IssuesEvent
2019-06-14 00:19:55
jccastillo0007/eFacturaT
https://api.github.com/repos/jccastillo0007/eFacturaT
opened
CxC - reporte Cartera General
bug defect
Me parece que solo trabaja bien para optibelt, pero si lo genero para la tia, no arroja la información correcta. Mi teoría es que como nosotros no manejamos número de cliente, por ello no arroja nada o muy poco. Seguramente el número de cliente, es parte de los cueris para generar el reporte. Si hay que dejarlo jalando para la tía, pero sin pegarle a optibelt.
1.0
CxC - reporte Cartera General - Me parece que solo trabaja bien para optibelt, pero si lo genero para la tia, no arroja la información correcta. Mi teoría es que como nosotros no manejamos número de cliente, por ello no arroja nada o muy poco. Seguramente el número de cliente, es parte de los cueris para generar el reporte. Si hay que dejarlo jalando para la tía, pero sin pegarle a optibelt.
defect
cxc reporte cartera general me parece que solo trabaja bien para optibelt pero si lo genero para la tia no arroja la información correcta mi teoría es que como nosotros no manejamos número de cliente por ello no arroja nada o muy poco seguramente el número de cliente es parte de los cueris para generar el reporte si hay que dejarlo jalando para la tía pero sin pegarle a optibelt
1
59,457
17,023,134,065
IssuesEvent
2021-07-03 00:31:31
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
OSM.org should move to a newer version of OpenLayers
Component: mapnik Priority: minor Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 7.59pm, Tuesday, 28th November 2006]** <[Beta]> is there a GET variable for openstreetmap to specify which base layer to show ?[[BR]] <crschmidt> [Beta]: OSM.org uses an old version of OpenLayers which doens't have that[[BR]] <[Beta]> damn them *shakes fist* ok :)[[BR]] <crschmidt> [Beta]: File a ticket for upgrading OSM.org to 2.2 into trac under slippy_map, and assign it to me[[BR]] Just to keep you on your toes : )
1.0
OSM.org should move to a newer version of OpenLayers - **[Submitted to the original trac issue database at 7.59pm, Tuesday, 28th November 2006]** <[Beta]> is there a GET variable for openstreetmap to specify which base layer to show ?[[BR]] <crschmidt> [Beta]: OSM.org uses an old version of OpenLayers which doens't have that[[BR]] <[Beta]> damn them *shakes fist* ok :)[[BR]] <crschmidt> [Beta]: File a ticket for upgrading OSM.org to 2.2 into trac under slippy_map, and assign it to me[[BR]] Just to keep you on your toes : )
defect
osm org should move to a newer version of openlayers is there a get variable for openstreetmap to specify which base layer to show osm org uses an old version of openlayers which doens t have that damn them shakes fist ok file a ticket for upgrading osm org to into trac under slippy map and assign it to me just to keep you on your toes
1
800,220
28,356,944,067
IssuesEvent
2023-04-12 08:03:05
COPRS/rs-issues
https://api.github.com/repos/COPRS/rs-issues
closed
[BUG] [SENTINEL-2] S2 SADATA products are not sent to the PRIP.
bug WERUM dev CCB priority:blocking ops Limitation
**Environment:** - Delivery tag: - Platform: OPS Orange Cloud - Configuration: Delivery tag: 1.4 Platform: OPS Orange Cloud Configuration: OPS rs-core : v1.10 **Traçability:** <!-- - Test(s): Test_name - Requirements: Req_name --> **Current Behavior:** Following product were generated without beeing sent to prip - S2_SAD **Expected Behavior:** These product should be be sent to prip **Steps To Reproduce:** Generate S2_SAD products **Test execution artefacts (i.e. logs, screenshots…)** N/A **Whenever possible, first analysis of the root cause** hereafter following compression filter: ``` app.message-filter.filter.function.expression=((payload.productFamily matches '^((S\\d.*)|(L\\d.*))(?<!(ZIP|AUX|JOB|GRANULES|REPORT|ETAD|SAD|BLANK)$)$') && (!(payload.productFamily == 'L0_SEGMENT') || ((payload.keyObjectStorage matches 'S1._(GP|HK|RF).*_RAW.*\.SAFE') && !(payload.keyObjectStorage matches 'S1._RF_RAW__0.(HH|HV|VV|VH)_.*\.SAFE')))) ``` I think S2_SAD productFamily not match with regex above. So they were not sent to compression worker and not sent to prip. hereafter sample for SADATA product S2B_OPER_AUX_SADATA_REFS_20230308T032303_V20230306T211539_20230306T225620_A031329_WF_LN This product was not sent to compression part after catalog. ![S2_SAD_CATA](https://user-images.githubusercontent.com/93319683/223708589-2c6f5cee-6771-4bd9-a9c3-7fa2954d350b.png) **Bug Generic Definition of Ready (DoR)** - [ ] The affect version in which the bug has been found is mentioned - [ ] The context and environment of the bug is detailed - [ ] The des cription of the bug is clear and unambiguous - [ ] The procedure (steps) to reproduce the bug is clearly detailed - [ ] The tested User Story / features is linked to the bug if available - [ ] Logs are attached if available - [ ] A data set attached if available **Bug Generic Definition of Done (DoD)** - [ ] the modification implemented (the solution to fix the bug) is described in the bug. - [ ] Unit tests & Continuous integration performed - Test results available - Structural Test coverage reported by SONAR - [ ] Code committed in GIT with right tag or Analysis/Trade Off documentation up-to-date in reference-system-documentation repository - [ ] Code is compliant with coding rules (SONAR Report as evidence) - [ ] Acceptance criteria of the related User story are checked and Passed
1.0
[BUG] [SENTINEL-2] S2 SADATA products are not sent to the PRIP. - **Environment:** - Delivery tag: - Platform: OPS Orange Cloud - Configuration: Delivery tag: 1.4 Platform: OPS Orange Cloud Configuration: OPS rs-core : v1.10 **Traçability:** <!-- - Test(s): Test_name - Requirements: Req_name --> **Current Behavior:** Following product were generated without beeing sent to prip - S2_SAD **Expected Behavior:** These product should be be sent to prip **Steps To Reproduce:** Generate S2_SAD products **Test execution artefacts (i.e. logs, screenshots…)** N/A **Whenever possible, first analysis of the root cause** hereafter following compression filter: ``` app.message-filter.filter.function.expression=((payload.productFamily matches '^((S\\d.*)|(L\\d.*))(?<!(ZIP|AUX|JOB|GRANULES|REPORT|ETAD|SAD|BLANK)$)$') && (!(payload.productFamily == 'L0_SEGMENT') || ((payload.keyObjectStorage matches 'S1._(GP|HK|RF).*_RAW.*\.SAFE') && !(payload.keyObjectStorage matches 'S1._RF_RAW__0.(HH|HV|VV|VH)_.*\.SAFE')))) ``` I think S2_SAD productFamily not match with regex above. So they were not sent to compression worker and not sent to prip. hereafter sample for SADATA product S2B_OPER_AUX_SADATA_REFS_20230308T032303_V20230306T211539_20230306T225620_A031329_WF_LN This product was not sent to compression part after catalog. ![S2_SAD_CATA](https://user-images.githubusercontent.com/93319683/223708589-2c6f5cee-6771-4bd9-a9c3-7fa2954d350b.png) **Bug Generic Definition of Ready (DoR)** - [ ] The affect version in which the bug has been found is mentioned - [ ] The context and environment of the bug is detailed - [ ] The des cription of the bug is clear and unambiguous - [ ] The procedure (steps) to reproduce the bug is clearly detailed - [ ] The tested User Story / features is linked to the bug if available - [ ] Logs are attached if available - [ ] A data set attached if available **Bug Generic Definition of Done (DoD)** - [ ] the modification implemented (the solution to fix the bug) is described in the bug. - [ ] Unit tests & Continuous integration performed - Test results available - Structural Test coverage reported by SONAR - [ ] Code committed in GIT with right tag or Analysis/Trade Off documentation up-to-date in reference-system-documentation repository - [ ] Code is compliant with coding rules (SONAR Report as evidence) - [ ] Acceptance criteria of the related User story are checked and Passed
non_defect
sadata products are not sent to the prip environment delivery tag platform ops orange cloud configuration delivery tag platform ops orange cloud configuration ops rs core traçability test s test name requirements req name current behavior following product were generated without beeing sent to prip sad expected behavior these product should be be sent to prip steps to reproduce generate sad products test execution artefacts i e logs screenshots… n a whenever possible first analysis of the root cause hereafter following compression filter app message filter filter function expression payload productfamily matches s d l d zip aux job granules report etad sad blank payload productfamily segment payload keyobjectstorage matches gp hk rf raw safe payload keyobjectstorage matches rf raw hh hv vv vh safe i think sad productfamily not match with regex above so they were not sent to compression worker and not sent to prip hereafter sample for sadata product oper aux sadata refs wf ln this product was not sent to compression part after catalog bug generic definition of ready dor the affect version in which the bug has been found is mentioned the context and environment of the bug is detailed the des cription of the bug is clear and unambiguous the procedure steps to reproduce the bug is clearly detailed the tested user story features is linked to the bug if available logs are attached if available a data set attached if available bug generic definition of done dod the modification implemented the solution to fix the bug is described in the bug unit tests continuous integration performed test results available structural test coverage reported by sonar code committed in git with right tag or analysis trade off documentation up to date in reference system documentation repository code is compliant with coding rules sonar report as evidence acceptance criteria of the related user story are checked and passed
0
9,532
2,615,155,539
IssuesEvent
2015-03-01 06:33:41
chrsmith/html5rocks
https://api.github.com/repos/chrsmith/html5rocks
closed
terms page
auto-migrated Milestone-4 Priority-Medium Type-Defect
``` it shows an error: Traceback (most recent call last): File "/base/python_runtime/python_lib/versions/1/google/appengine/ext/webapp/__init__.py", line 511, in __call__ handler.get(*groups) File "/base/data/home/apps/html5rocks/3.345500695094989155/main.py", line 176, in get self.render(template_path=path) File "/base/data/home/apps/html5rocks/3.345500695094989155/main.py", line 117, in render 'toc' : self.get_toc(template_path), File "/base/data/home/apps/html5rocks/3.345500695094989155/main.py", line 45, in get_toc template_text = webapp.template.render(path, {}); File "/base/python_runtime/python_lib/versions/1/google/appengine/ext/webapp/template.py", line 81, in render return t.render(Context(template_dict)) File "/base/python_runtime/python_lib/versions/1/google/appengine/ext/webapp/template.py", line 121, in wrap_render return orig_render(context) File "/base/python_runtime/python_lib/versions/third_party/django-0.96/django/template/__init__.py", line 168, in render return self.nodelist.render(context) File "/base/python_runtime/python_lib/versions/third_party/django-0.96/django/template/__init__.py", line 705, in render bits.append(self.render_node(node, context)) File "/base/python_runtime/python_lib/versions/third_party/django-0.96/django/template/__init__.py", line 718, in render_node return(node.render(context)) File "/base/python_runtime/python_lib/versions/third_party/django-0.96/django/template/loader_tags.py", line 63, in render compiled_parent = self.get_parent(context) File "/base/python_runtime/python_lib/versions/third_party/django-0.96/django/template/loader_tags.py", line 58, in get_parent raise TemplateSyntaxError, "Template %r cannot be extended, because it doesn't exist" % parent TemplateSyntaxError: Template 'base.html' cannot be extended, because it doesn't exist ``` Original issue reported on code.google.com by `antonino...@gmail.com` on 16 Nov 2010 at 9:44
1.0
terms page - ``` it shows an error: Traceback (most recent call last): File "/base/python_runtime/python_lib/versions/1/google/appengine/ext/webapp/__init__.py", line 511, in __call__ handler.get(*groups) File "/base/data/home/apps/html5rocks/3.345500695094989155/main.py", line 176, in get self.render(template_path=path) File "/base/data/home/apps/html5rocks/3.345500695094989155/main.py", line 117, in render 'toc' : self.get_toc(template_path), File "/base/data/home/apps/html5rocks/3.345500695094989155/main.py", line 45, in get_toc template_text = webapp.template.render(path, {}); File "/base/python_runtime/python_lib/versions/1/google/appengine/ext/webapp/template.py", line 81, in render return t.render(Context(template_dict)) File "/base/python_runtime/python_lib/versions/1/google/appengine/ext/webapp/template.py", line 121, in wrap_render return orig_render(context) File "/base/python_runtime/python_lib/versions/third_party/django-0.96/django/template/__init__.py", line 168, in render return self.nodelist.render(context) File "/base/python_runtime/python_lib/versions/third_party/django-0.96/django/template/__init__.py", line 705, in render bits.append(self.render_node(node, context)) File "/base/python_runtime/python_lib/versions/third_party/django-0.96/django/template/__init__.py", line 718, in render_node return(node.render(context)) File "/base/python_runtime/python_lib/versions/third_party/django-0.96/django/template/loader_tags.py", line 63, in render compiled_parent = self.get_parent(context) File "/base/python_runtime/python_lib/versions/third_party/django-0.96/django/template/loader_tags.py", line 58, in get_parent raise TemplateSyntaxError, "Template %r cannot be extended, because it doesn't exist" % parent TemplateSyntaxError: Template 'base.html' cannot be extended, because it doesn't exist ``` Original issue reported on code.google.com by `antonino...@gmail.com` on 16 Nov 2010 at 9:44
defect
terms page it shows an error traceback most recent call last file base python runtime python lib versions google appengine ext webapp init py line in call handler get groups file base data home apps main py line in get self render template path path file base data home apps main py line in render toc self get toc template path file base data home apps main py line in get toc template text webapp template render path file base python runtime python lib versions google appengine ext webapp template py line in render return t render context template dict file base python runtime python lib versions google appengine ext webapp template py line in wrap render return orig render context file base python runtime python lib versions third party django django template init py line in render return self nodelist render context file base python runtime python lib versions third party django django template init py line in render bits append self render node node context file base python runtime python lib versions third party django django template init py line in render node return node render context file base python runtime python lib versions third party django django template loader tags py line in render compiled parent self get parent context file base python runtime python lib versions third party django django template loader tags py line in get parent raise templatesyntaxerror template r cannot be extended because it doesn t exist parent templatesyntaxerror template base html cannot be extended because it doesn t exist original issue reported on code google com by antonino gmail com on nov at
1
6,639
3,870,760,192
IssuesEvent
2016-04-11 06:28:40
Samsung/iotjs
https://api.github.com/repos/Samsung/iotjs
closed
cannot build on x86_64: ./tools/build.py
bug build system
==> Build libtuv cmake /home/sihai.ysh/iotjs/deps/libtuv -DCMAKE_TOOLCHAIN_FILE=/home/sihai.ysh/iotjs/deps/libtuv/cmake/config/config_x86_64-linux.cmake -DCMAKE_BUILD_TYPE=debug -DTARGET_PLATFORM=x86_64-linux -DLIBTUV_CUSTOM_LIB_OUT=/home/sihai.ysh/iotjs/build/x86_64-linux/debug/deps/libtuv -DCMAKE_C_FLAGS=-D__LINUX__ -fno-builtin -D__x86_64__ -D__X86_64__ -g -DDEBUG -DENABLE_DEBUG_LOG -DCMAKE_CXX_FLAGS=-D__LINUX__ -fno-builtin -D__x86_64__ -D__X86_64__ -g -DDEBUG -DENABLE_DEBUG_LOG -DCMAKE_EXE_LINKER_FLAGS=-pthread -DEXTERNAL_INCLUDE_DIR= -- Build Type: [debug] CMake Error at cmake/libtuv.cmake:48 (target_include_directories): target_include_directories called with invalid arguments Call Stack (most recent call first): CMakeLists.txt:22 (include) -- Configuring incomplete, errors occurred!
1.0
cannot build on x86_64: ./tools/build.py - ==> Build libtuv cmake /home/sihai.ysh/iotjs/deps/libtuv -DCMAKE_TOOLCHAIN_FILE=/home/sihai.ysh/iotjs/deps/libtuv/cmake/config/config_x86_64-linux.cmake -DCMAKE_BUILD_TYPE=debug -DTARGET_PLATFORM=x86_64-linux -DLIBTUV_CUSTOM_LIB_OUT=/home/sihai.ysh/iotjs/build/x86_64-linux/debug/deps/libtuv -DCMAKE_C_FLAGS=-D__LINUX__ -fno-builtin -D__x86_64__ -D__X86_64__ -g -DDEBUG -DENABLE_DEBUG_LOG -DCMAKE_CXX_FLAGS=-D__LINUX__ -fno-builtin -D__x86_64__ -D__X86_64__ -g -DDEBUG -DENABLE_DEBUG_LOG -DCMAKE_EXE_LINKER_FLAGS=-pthread -DEXTERNAL_INCLUDE_DIR= -- Build Type: [debug] CMake Error at cmake/libtuv.cmake:48 (target_include_directories): target_include_directories called with invalid arguments Call Stack (most recent call first): CMakeLists.txt:22 (include) -- Configuring incomplete, errors occurred!
non_defect
cannot build on tools build py build libtuv cmake home sihai ysh iotjs deps libtuv dcmake toolchain file home sihai ysh iotjs deps libtuv cmake config config linux cmake dcmake build type debug dtarget platform linux dlibtuv custom lib out home sihai ysh iotjs build linux debug deps libtuv dcmake c flags d linux fno builtin d d g ddebug denable debug log dcmake cxx flags d linux fno builtin d d g ddebug denable debug log dcmake exe linker flags pthread dexternal include dir build type cmake error at cmake libtuv cmake target include directories target include directories called with invalid arguments call stack most recent call first cmakelists txt include configuring incomplete errors occurred
0
3,067
2,607,982,465
IssuesEvent
2015-02-26 00:50:11
chrsmithdemos/zen-coding
https://api.github.com/repos/chrsmithdemos/zen-coding
opened
Notepad++ plugin problem with many windows
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. open notepad++ open any file 2. open second copy notepad++ 3. in second copy type "head" and press ctrl+e 4. nothing happens What is the expected output? What do you see instead? When I open few windows of notepad++ and try to press ctrl+E, action happens in inactive window of notepad++ in area where was cursor later. In active windows nothing happens. If open only one copy of notepad++ there is no problem. What version of the product are you using? On what operating system? ZenCoding 0.6 + notepad++ 5.6.8 + windows 7 Please provide any additional information below. ``` ----- Original issue reported on code.google.com by `pivo...@gmail.com` on 23 Feb 2011 at 1:08
1.0
Notepad++ plugin problem with many windows - ``` What steps will reproduce the problem? 1. open notepad++ open any file 2. open second copy notepad++ 3. in second copy type "head" and press ctrl+e 4. nothing happens What is the expected output? What do you see instead? When I open few windows of notepad++ and try to press ctrl+E, action happens in inactive window of notepad++ in area where was cursor later. In active windows nothing happens. If open only one copy of notepad++ there is no problem. What version of the product are you using? On what operating system? ZenCoding 0.6 + notepad++ 5.6.8 + windows 7 Please provide any additional information below. ``` ----- Original issue reported on code.google.com by `pivo...@gmail.com` on 23 Feb 2011 at 1:08
defect
notepad plugin problem with many windows what steps will reproduce the problem open notepad open any file open second copy notepad in second copy type head and press ctrl e nothing happens what is the expected output what do you see instead when i open few windows of notepad and try to press ctrl e action happens in inactive window of notepad in area where was cursor later in active windows nothing happens if open only one copy of notepad there is no problem what version of the product are you using on what operating system zencoding notepad windows please provide any additional information below original issue reported on code google com by pivo gmail com on feb at
1
72,878
24,343,292,719
IssuesEvent
2022-10-02 01:02:52
naev/naev
https://api.github.com/repos/naev/naev
closed
Chapter 1 softlocks when landing on a planet with no shipyard
Type-Defect Priority-Critical
``` Lemuria the Illegal Waste Dumper — 今日 06:05 Start of Chapter 1 softlocks me when I land on a planet with no ability to manage the fleet ``` Probably some ships are getting somehow marked as deployed before the event is run and then the bug happens. Possible solutions: 1. Reset deployment when first adding fleet capacity 2. Make sure that nothing can be marked as deployed with no fleet capacity
1.0
Chapter 1 softlocks when landing on a planet with no shipyard - ``` Lemuria the Illegal Waste Dumper — 今日 06:05 Start of Chapter 1 softlocks me when I land on a planet with no ability to manage the fleet ``` Probably some ships are getting somehow marked as deployed before the event is run and then the bug happens. Possible solutions: 1. Reset deployment when first adding fleet capacity 2. Make sure that nothing can be marked as deployed with no fleet capacity
defect
chapter softlocks when landing on a planet with no shipyard lemuria the illegal waste dumper — 今日 start of chapter softlocks me when i land on a planet with no ability to manage the fleet probably some ships are getting somehow marked as deployed before the event is run and then the bug happens possible solutions reset deployment when first adding fleet capacity make sure that nothing can be marked as deployed with no fleet capacity
1
76,801
26,605,731,811
IssuesEvent
2023-01-23 19:10:50
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
opened
Element Crash on Video call
T-Defect
### Steps to reproduce Accept video call from remote iOS client. ### Outcome #### What did you expect? Able to have to chat a video chat #### What happened instead? Element Android crashed ### Your phone model Samsung Note 9 ### Operating system version Android 10 ### Application version and app store Element 1.5.20 ### Homeserver Synapse 1.75 ### Will you send logs? Yes ### Are you willing to provide a PR? No
1.0
Element Crash on Video call - ### Steps to reproduce Accept video call from remote iOS client. ### Outcome #### What did you expect? Able to have to chat a video chat #### What happened instead? Element Android crashed ### Your phone model Samsung Note 9 ### Operating system version Android 10 ### Application version and app store Element 1.5.20 ### Homeserver Synapse 1.75 ### Will you send logs? Yes ### Are you willing to provide a PR? No
defect
element crash on video call steps to reproduce accept video call from remote ios client outcome what did you expect able to have to chat a video chat what happened instead element android crashed your phone model samsung note operating system version android application version and app store element homeserver synapse will you send logs yes are you willing to provide a pr no
1
1,201
2,601,758,204
IssuesEvent
2015-02-24 00:33:55
chrsmith/bwapi
https://api.github.com/repos/chrsmith/bwapi
closed
BWTA doesn't cache map analysis, when using the client interface
auto-migrated Type-Defect Usability
``` What steps will reproduce the problem? 1. Run bwapi using a client (like ExampleAIClient, when ai_dll = NULL) 2. Type /analyze What is the expected output? What do you see instead? BWTA should cache its analysis, but it doesn't (no file is written to bwapi-data/BWTA) What version of the product are you using? On what operating system? BWAPI 3.3, Windows XP SP3 It still works when using ai_dll points to a proper dll module. @lowerlogic, hope it's okay I set you as owner. ``` ----- Original issue reported on code.google.com by `erez27` on 23 Nov 2010 at 1:40
1.0
BWTA doesn't cache map analysis, when using the client interface - ``` What steps will reproduce the problem? 1. Run bwapi using a client (like ExampleAIClient, when ai_dll = NULL) 2. Type /analyze What is the expected output? What do you see instead? BWTA should cache its analysis, but it doesn't (no file is written to bwapi-data/BWTA) What version of the product are you using? On what operating system? BWAPI 3.3, Windows XP SP3 It still works when using ai_dll points to a proper dll module. @lowerlogic, hope it's okay I set you as owner. ``` ----- Original issue reported on code.google.com by `erez27` on 23 Nov 2010 at 1:40
defect
bwta doesn t cache map analysis when using the client interface what steps will reproduce the problem run bwapi using a client like exampleaiclient when ai dll null type analyze what is the expected output what do you see instead bwta should cache its analysis but it doesn t no file is written to bwapi data bwta what version of the product are you using on what operating system bwapi windows xp it still works when using ai dll points to a proper dll module lowerlogic hope it s okay i set you as owner original issue reported on code google com by on nov at
1
48,761
13,184,732,093
IssuesEvent
2020-08-12 19:59:37
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
the I3::dataio::open functions are declared in two different header files (Trac #153)
Incomplete Migration Migrated from Trac dataio defect
<details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/153 , reported by boersma and owned by troy_</summary> <p> ```json { "status": "closed", "changetime": "2014-11-23T03:37:56", "description": "Namely, in:\n\nicetray/public/icetray/open.h\ndataio/public/dataio/FrameIO.h\n\nFrameIO.h does not have a corresponding FrameIO.cxx file, so I think that\nit would be good to remove that file and replace all #includes of it by\n#includes of icetray/open.h.\n\nGrepping in the current trunk of offline-software gives:\n\n* The dataio/FrameIO.h version is used in:\n./dataio/private/filter/main.cxx.aside\n./dataio/private/extract/main.cxx\n./dataio/private/strip/main.cxx\n./dataio/private/pybindings/I3LootFile.cxx\n\n* The icetray/open.h version is used in:\n./dataio/private/test/serialization-test.h\n./dataio/private/test/vectors.cxx\n./dataio/private/test/checksumming.cxx\n./dataio/private/test/speed-test.cxx\n./dataio/private/dataio/I3Reader.cxx\n./dataio/private/dataio/I3ReaderService.cxx\n./dataio/private/dataio/I3MultiWriter.cxx\n./dataio/private/dataio/I3Writer.cxx\n./icetray/private/test/iostreams.cxx\n./icetray/private/test/I3FrameTest.cxx\n./glshovel/private/i3tvtpd/spool.hpp\n\n", "reporter": "boersma", "cc": "", "resolution": "fixed", "_ts": "1416713876900096", "component": "dataio", "summary": "the I3::dataio::open functions are declared in two different header files", "priority": "normal", "keywords": "", "time": "2008-11-19T02:11:21", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
1.0
the I3::dataio::open functions are declared in two different header files (Trac #153) - <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/153 , reported by boersma and owned by troy_</summary> <p> ```json { "status": "closed", "changetime": "2014-11-23T03:37:56", "description": "Namely, in:\n\nicetray/public/icetray/open.h\ndataio/public/dataio/FrameIO.h\n\nFrameIO.h does not have a corresponding FrameIO.cxx file, so I think that\nit would be good to remove that file and replace all #includes of it by\n#includes of icetray/open.h.\n\nGrepping in the current trunk of offline-software gives:\n\n* The dataio/FrameIO.h version is used in:\n./dataio/private/filter/main.cxx.aside\n./dataio/private/extract/main.cxx\n./dataio/private/strip/main.cxx\n./dataio/private/pybindings/I3LootFile.cxx\n\n* The icetray/open.h version is used in:\n./dataio/private/test/serialization-test.h\n./dataio/private/test/vectors.cxx\n./dataio/private/test/checksumming.cxx\n./dataio/private/test/speed-test.cxx\n./dataio/private/dataio/I3Reader.cxx\n./dataio/private/dataio/I3ReaderService.cxx\n./dataio/private/dataio/I3MultiWriter.cxx\n./dataio/private/dataio/I3Writer.cxx\n./icetray/private/test/iostreams.cxx\n./icetray/private/test/I3FrameTest.cxx\n./glshovel/private/i3tvtpd/spool.hpp\n\n", "reporter": "boersma", "cc": "", "resolution": "fixed", "_ts": "1416713876900096", "component": "dataio", "summary": "the I3::dataio::open functions are declared in two different header files", "priority": "normal", "keywords": "", "time": "2008-11-19T02:11:21", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
defect
the dataio open functions are declared in two different header files trac migrated from reported by boersma and owned by troy json status closed changetime description namely in n nicetray public icetray open h ndataio public dataio frameio h n nframeio h does not have a corresponding frameio cxx file so i think that nit would be good to remove that file and replace all includes of it by n includes of icetray open h n ngrepping in the current trunk of offline software gives n n the dataio frameio h version is used in n dataio private filter main cxx aside n dataio private extract main cxx n dataio private strip main cxx n dataio private pybindings cxx n n the icetray open h version is used in n dataio private test serialization test h n dataio private test vectors cxx n dataio private test checksumming cxx n dataio private test speed test cxx n dataio private dataio cxx n dataio private dataio cxx n dataio private dataio cxx n dataio private dataio cxx n icetray private test iostreams cxx n icetray private test cxx n glshovel private spool hpp n n reporter boersma cc resolution fixed ts component dataio summary the dataio open functions are declared in two different header files priority normal keywords time milestone owner troy type defect
1
3,177
2,607,986,948
IssuesEvent
2015-02-26 00:52:13
chrsmithdemos/zen-coding
https://api.github.com/repos/chrsmithdemos/zen-coding
opened
Add support for SVG elements
auto-migrated Priority-Medium Type-Defect
``` Add support for SVG elements ``` ----- Original issue reported on code.google.com by `mcdonald...@gmail.com` on 29 Oct 2012 at 6:40
1.0
Add support for SVG elements - ``` Add support for SVG elements ``` ----- Original issue reported on code.google.com by `mcdonald...@gmail.com` on 29 Oct 2012 at 6:40
defect
add support for svg elements add support for svg elements original issue reported on code google com by mcdonald gmail com on oct at
1
49,531
13,187,227,338
IssuesEvent
2020-08-13 02:45:08
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
CVMFS - doc builds are choking on the lack of `napoleon` (Trac #1608)
Incomplete Migration Migrated from Trac defect infrastructure
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1608">https://code.icecube.wisc.edu/ticket/1608</a>, reported by nega and owned by david.schultz</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-29T19:54:11", "description": "david- could you add napoleon, or update sphinx?", "reporter": "nega", "cc": "", "resolution": "fixed", "_ts": "1459281251636355", "component": "infrastructure", "summary": "CVMFS - doc builds are choking on the lack of `napoleon`", "priority": "normal", "keywords": "cvmfs sphinx documentation", "time": "2016-03-28T21:20:33", "milestone": "", "owner": "david.schultz", "type": "defect" } ``` </p> </details>
1.0
CVMFS - doc builds are choking on the lack of `napoleon` (Trac #1608) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1608">https://code.icecube.wisc.edu/ticket/1608</a>, reported by nega and owned by david.schultz</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-29T19:54:11", "description": "david- could you add napoleon, or update sphinx?", "reporter": "nega", "cc": "", "resolution": "fixed", "_ts": "1459281251636355", "component": "infrastructure", "summary": "CVMFS - doc builds are choking on the lack of `napoleon`", "priority": "normal", "keywords": "cvmfs sphinx documentation", "time": "2016-03-28T21:20:33", "milestone": "", "owner": "david.schultz", "type": "defect" } ``` </p> </details>
defect
cvmfs doc builds are choking on the lack of napoleon trac migrated from json status closed changetime description david could you add napoleon or update sphinx reporter nega cc resolution fixed ts component infrastructure summary cvmfs doc builds are choking on the lack of napoleon priority normal keywords cvmfs sphinx documentation time milestone owner david schultz type defect
1
72,767
24,280,370,952
IssuesEvent
2022-09-28 16:51:26
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Screensharing not possible in a room
T-Defect
### Steps to reproduce 1. Using the desktop client you create a meeting in a room 2. There are several participants 3. You click on share screen 4. The screen share dialog appears, but there is nothing in it - no preview pic of your screen It works in the web client and it also works when you connect directly with another person. ### Outcome #### What did you expect? Screensharing works #### What happened instead? Not able to select the screen to share I would be very happy if this can be fiixed as it goes on my nerves in our daily meeting. ### Operating system Ubuntu ### Application version 1.10.11 ### How did you install the app? _No response_ ### Homeserver _No response_ ### Will you send logs? No
1.0
Screensharing not possible in a room - ### Steps to reproduce 1. Using the desktop client you create a meeting in a room 2. There are several participants 3. You click on share screen 4. The screen share dialog appears, but there is nothing in it - no preview pic of your screen It works in the web client and it also works when you connect directly with another person. ### Outcome #### What did you expect? Screensharing works #### What happened instead? Not able to select the screen to share I would be very happy if this can be fiixed as it goes on my nerves in our daily meeting. ### Operating system Ubuntu ### Application version 1.10.11 ### How did you install the app? _No response_ ### Homeserver _No response_ ### Will you send logs? No
defect
screensharing not possible in a room steps to reproduce using the desktop client you create a meeting in a room there are several participants you click on share screen the screen share dialog appears but there is nothing in it no preview pic of your screen it works in the web client and it also works when you connect directly with another person outcome what did you expect screensharing works what happened instead not able to select the screen to share i would be very happy if this can be fiixed as it goes on my nerves in our daily meeting operating system ubuntu application version how did you install the app no response homeserver no response will you send logs no
1