Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
464,522
| 13,326,353,043
|
IssuesEvent
|
2020-08-27 11:29:47
|
googleapis/nodejs-googleapis-common
|
https://api.github.com/repos/googleapis/nodejs-googleapis-common
|
opened
|
Synthesis failed for nodejs-googleapis-common
|
autosynth failure priority: p1 type: bug
|
Hello! Autosynth couldn't regenerate nodejs-googleapis-common. :broken_heart:
Here's the output from running `synth.py`:
```
retty=%at 019c7168faa0e56619f792693a8acdb30d6de19b
2020-08-27 04:29:35,798 autosynth [DEBUG] > Running: git checkout e96e2ccf4f10be0e99df71d31a5f3ab6c711218a
Note: checking out 'e96e2ccf4f10be0e99df71d31a5f3ab6c711218a'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at e96e2cc build: move system and samples test from Node 10 to Node 12 (#321)
2020-08-27 04:29:35,814 autosynth [DEBUG] > Running: git checkout 019c7168faa0e56619f792693a8acdb30d6de19b
Note: checking out '019c7168faa0e56619f792693a8acdb30d6de19b'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 019c716 build(java): switch to release-publish app for notifying GitHub of release status (#740)
2020-08-27 04:29:35,819 autosynth [DEBUG] > Running: git branch -f autosynth-7
2020-08-27 04:29:35,823 autosynth [DEBUG] > Running: git checkout autosynth-7
Switched to branch 'autosynth-7'
2020-08-27 04:29:35,839 autosynth [INFO] > Running synthtool
2020-08-27 04:29:35,839 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-08-27 04:29:35,839 autosynth [DEBUG] > log_file_path: /tmpfs/src/logs/nodejs-googleapis-common/7/sponge_log.log
2020-08-27 04:29:35,841 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py --
2020-08-27 04:29:36,063 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/nodejs-googleapis-common/synth.py.
On branch autosynth-7
nothing to commit, working tree clean
2020-08-27 04:29:36,206 synthtool [DEBUG] > Using precloned repo /home/kbuilder/.cache/synthtool/synthtool
DEBUG:synthtool:Using precloned repo /home/kbuilder/.cache/synthtool/synthtool
.eslintignore
.eslintrc.json
.gitattributes
.github/ISSUE_TEMPLATE/bug_report.md
.github/ISSUE_TEMPLATE/feature_request.md
.github/ISSUE_TEMPLATE/support_request.md
.github/PULL_REQUEST_TEMPLATE.md
.github/release-please.yml
.github/workflows/ci.yaml
.kokoro/.gitattributes
.kokoro/common.cfg
.kokoro/continuous/node10/common.cfg
.kokoro/continuous/node10/docs.cfg
.kokoro/continuous/node10/test.cfg
.kokoro/continuous/node12/common.cfg
.kokoro/continuous/node12/lint.cfg
.kokoro/continuous/node12/samples-test.cfg
.kokoro/continuous/node12/system-test.cfg
.kokoro/continuous/node12/test.cfg
.kokoro/docs.sh
.kokoro/lint.sh
.kokoro/populate-secrets.sh
.kokoro/presubmit/node10/common.cfg
.kokoro/presubmit/node12/common.cfg
.kokoro/presubmit/node12/samples-test.cfg
.kokoro/presubmit/node12/system-test.cfg
.kokoro/presubmit/node12/test.cfg
.kokoro/publish.sh
.kokoro/release/docs-devsite.cfg
.kokoro/release/docs-devsite.sh
.kokoro/release/docs.cfg
.kokoro/release/docs.sh
.kokoro/release/publish.cfg
.kokoro/samples-test.sh
.kokoro/system-test.sh
.kokoro/test.bat
.kokoro/test.sh
.kokoro/trampoline.sh
.mocharc.js
.nycrc
.prettierignore
.prettierrc.js
CODE_OF_CONDUCT.md
CONTRIBUTING.md
LICENSE
README.md
api-extractor.json
renovate.json
Skipping: samples/README.md
2020-08-27 04:29:36,352 synthtool [DEBUG] > Installing dependencies...
DEBUG:synthtool:Installing dependencies...
npm WARN deprecated ts-simple-ast@12.4.0: NOTICE: ts-simple-ast has been renamed to ts-morph and version reset to 1.0.0. Switch at your leisure...
npm WARN deprecated viz.js@1.8.2: no longer supported
npm WARN deprecated chokidar@2.1.8: Chokidar 2 will break on node v14+. Upgrade to chokidar 3 with 15x less dependencies.
npm WARN deprecated opn@6.0.0: The package has been renamed to `open`
npm WARN deprecated istanbul@0.4.5: This module is no longer maintained, try this instead:
npm WARN deprecated npm i nyc
npm WARN deprecated Visit https://istanbul.js.org/integrations for other alternatives.
npm WARN deprecated flat@4.1.0: Fixed a prototype pollution security issue in 4.1.0, please upgrade to ^4.1.1 or ^5.0.1.
npm WARN deprecated fsevents@1.2.13: fsevents 1 will break on node v14+ and could be using insecure binaries. Upgrade to fsevents 2.
npm WARN deprecated resolve-url@0.2.1: https://github.com/lydell/resolve-url#deprecated
npm WARN deprecated urix@0.1.0: Please see https://github.com/lydell/urix#deprecated
npm WARN deprecated core-js@2.6.11: core-js@<3 is no longer maintained and not recommended for usage due to the number of issues. Please, upgrade your dependencies to the actual version of core-js@3.
npm WARN optional SKIPPING OPTIONAL DEPENDENCY: fsevents@~2.1.2 (node_modules/chokidar/node_modules/fsevents):
npm WARN notsup SKIPPING OPTIONAL DEPENDENCY: Unsupported platform for fsevents@2.1.3: wanted {"os":"darwin","arch":"any"} (current: {"os":"linux","arch":"x64"})
npm WARN optional SKIPPING OPTIONAL DEPENDENCY: fsevents@^1.2.7 (node_modules/live-server/node_modules/chokidar/node_modules/fsevents):
npm WARN notsup SKIPPING OPTIONAL DEPENDENCY: Unsupported platform for fsevents@1.2.13: wanted {"os":"darwin","arch":"any"} (current: {"os":"linux","arch":"x64"})
npm ERR! code ETARGET
npm ERR! notarget No matching version found for @rushstack/node-core-library@3.30.0.
npm ERR! notarget In most cases you or one of your dependencies are requesting
npm ERR! notarget a package version that doesn't exist.
npm ERR! notarget
npm ERR! notarget It was specified as a dependency of '@microsoft/api-documenter'
npm ERR! notarget
npm ERR! A complete log of this run can be found in:
npm ERR! /home/kbuilder/.npm/_logs/2020-08-27T11_29_45_288Z-debug.log
2020-08-27 04:29:45,310 synthtool [ERROR] > Failed executing npm install:
None
ERROR:synthtool:Failed executing npm install:
None
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/home/kbuilder/.cache/synthtool/nodejs-googleapis-common/synth.py", line 24, in <module>
node.install()
File "/tmpfs/src/github/synthtool/synthtool/languages/node.py", line 167, in install
shell.run(["npm", "install"], hide_output=hide_output)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['npm', 'install']' returned non-zero exit status 1.
2020-08-27 04:29:45,366 autosynth [ERROR] > Synthesis failed
2020-08-27 04:29:45,367 autosynth [DEBUG] > Running: git reset --hard HEAD
HEAD is now at e96e2cc build: move system and samples test from Node 10 to Node 12 (#321)
2020-08-27 04:29:45,374 autosynth [DEBUG] > Running: git checkout autosynth
Switched to branch 'autosynth'
2020-08-27 04:29:45,379 autosynth [DEBUG] > Running: git clean -fdx
Removing __pycache__/
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 690, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 539, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 670, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 375, in synthesize_loop
has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 273, in synthesize_version_in_new_branch
synthesizer.synthesize(synth_log_path, self.environ)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/d56361e6-7265-4ad8-8ccb-dc2ff9ab5573/targets/github%2Fsynthtool;config=default/tests;query=nodejs-googleapis-common;failed=false).
|
1.0
|
Synthesis failed for nodejs-googleapis-common - Hello! Autosynth couldn't regenerate nodejs-googleapis-common. :broken_heart:
Here's the output from running `synth.py`:
```
retty=%at 019c7168faa0e56619f792693a8acdb30d6de19b
2020-08-27 04:29:35,798 autosynth [DEBUG] > Running: git checkout e96e2ccf4f10be0e99df71d31a5f3ab6c711218a
Note: checking out 'e96e2ccf4f10be0e99df71d31a5f3ab6c711218a'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at e96e2cc build: move system and samples test from Node 10 to Node 12 (#321)
2020-08-27 04:29:35,814 autosynth [DEBUG] > Running: git checkout 019c7168faa0e56619f792693a8acdb30d6de19b
Note: checking out '019c7168faa0e56619f792693a8acdb30d6de19b'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 019c716 build(java): switch to release-publish app for notifying GitHub of release status (#740)
2020-08-27 04:29:35,819 autosynth [DEBUG] > Running: git branch -f autosynth-7
2020-08-27 04:29:35,823 autosynth [DEBUG] > Running: git checkout autosynth-7
Switched to branch 'autosynth-7'
2020-08-27 04:29:35,839 autosynth [INFO] > Running synthtool
2020-08-27 04:29:35,839 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-08-27 04:29:35,839 autosynth [DEBUG] > log_file_path: /tmpfs/src/logs/nodejs-googleapis-common/7/sponge_log.log
2020-08-27 04:29:35,841 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata synth.metadata synth.py --
2020-08-27 04:29:36,063 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/nodejs-googleapis-common/synth.py.
On branch autosynth-7
nothing to commit, working tree clean
2020-08-27 04:29:36,206 synthtool [DEBUG] > Using precloned repo /home/kbuilder/.cache/synthtool/synthtool
DEBUG:synthtool:Using precloned repo /home/kbuilder/.cache/synthtool/synthtool
.eslintignore
.eslintrc.json
.gitattributes
.github/ISSUE_TEMPLATE/bug_report.md
.github/ISSUE_TEMPLATE/feature_request.md
.github/ISSUE_TEMPLATE/support_request.md
.github/PULL_REQUEST_TEMPLATE.md
.github/release-please.yml
.github/workflows/ci.yaml
.kokoro/.gitattributes
.kokoro/common.cfg
.kokoro/continuous/node10/common.cfg
.kokoro/continuous/node10/docs.cfg
.kokoro/continuous/node10/test.cfg
.kokoro/continuous/node12/common.cfg
.kokoro/continuous/node12/lint.cfg
.kokoro/continuous/node12/samples-test.cfg
.kokoro/continuous/node12/system-test.cfg
.kokoro/continuous/node12/test.cfg
.kokoro/docs.sh
.kokoro/lint.sh
.kokoro/populate-secrets.sh
.kokoro/presubmit/node10/common.cfg
.kokoro/presubmit/node12/common.cfg
.kokoro/presubmit/node12/samples-test.cfg
.kokoro/presubmit/node12/system-test.cfg
.kokoro/presubmit/node12/test.cfg
.kokoro/publish.sh
.kokoro/release/docs-devsite.cfg
.kokoro/release/docs-devsite.sh
.kokoro/release/docs.cfg
.kokoro/release/docs.sh
.kokoro/release/publish.cfg
.kokoro/samples-test.sh
.kokoro/system-test.sh
.kokoro/test.bat
.kokoro/test.sh
.kokoro/trampoline.sh
.mocharc.js
.nycrc
.prettierignore
.prettierrc.js
CODE_OF_CONDUCT.md
CONTRIBUTING.md
LICENSE
README.md
api-extractor.json
renovate.json
Skipping: samples/README.md
2020-08-27 04:29:36,352 synthtool [DEBUG] > Installing dependencies...
DEBUG:synthtool:Installing dependencies...
npm WARN deprecated ts-simple-ast@12.4.0: NOTICE: ts-simple-ast has been renamed to ts-morph and version reset to 1.0.0. Switch at your leisure...
npm WARN deprecated viz.js@1.8.2: no longer supported
npm WARN deprecated chokidar@2.1.8: Chokidar 2 will break on node v14+. Upgrade to chokidar 3 with 15x less dependencies.
npm WARN deprecated opn@6.0.0: The package has been renamed to `open`
npm WARN deprecated istanbul@0.4.5: This module is no longer maintained, try this instead:
npm WARN deprecated npm i nyc
npm WARN deprecated Visit https://istanbul.js.org/integrations for other alternatives.
npm WARN deprecated flat@4.1.0: Fixed a prototype pollution security issue in 4.1.0, please upgrade to ^4.1.1 or ^5.0.1.
npm WARN deprecated fsevents@1.2.13: fsevents 1 will break on node v14+ and could be using insecure binaries. Upgrade to fsevents 2.
npm WARN deprecated resolve-url@0.2.1: https://github.com/lydell/resolve-url#deprecated
npm WARN deprecated urix@0.1.0: Please see https://github.com/lydell/urix#deprecated
npm WARN deprecated core-js@2.6.11: core-js@<3 is no longer maintained and not recommended for usage due to the number of issues. Please, upgrade your dependencies to the actual version of core-js@3.
npm WARN optional SKIPPING OPTIONAL DEPENDENCY: fsevents@~2.1.2 (node_modules/chokidar/node_modules/fsevents):
npm WARN notsup SKIPPING OPTIONAL DEPENDENCY: Unsupported platform for fsevents@2.1.3: wanted {"os":"darwin","arch":"any"} (current: {"os":"linux","arch":"x64"})
npm WARN optional SKIPPING OPTIONAL DEPENDENCY: fsevents@^1.2.7 (node_modules/live-server/node_modules/chokidar/node_modules/fsevents):
npm WARN notsup SKIPPING OPTIONAL DEPENDENCY: Unsupported platform for fsevents@1.2.13: wanted {"os":"darwin","arch":"any"} (current: {"os":"linux","arch":"x64"})
npm ERR! code ETARGET
npm ERR! notarget No matching version found for @rushstack/node-core-library@3.30.0.
npm ERR! notarget In most cases you or one of your dependencies are requesting
npm ERR! notarget a package version that doesn't exist.
npm ERR! notarget
npm ERR! notarget It was specified as a dependency of '@microsoft/api-documenter'
npm ERR! notarget
npm ERR! A complete log of this run can be found in:
npm ERR! /home/kbuilder/.npm/_logs/2020-08-27T11_29_45_288Z-debug.log
2020-08-27 04:29:45,310 synthtool [ERROR] > Failed executing npm install:
None
ERROR:synthtool:Failed executing npm install:
None
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/home/kbuilder/.cache/synthtool/nodejs-googleapis-common/synth.py", line 24, in <module>
node.install()
File "/tmpfs/src/github/synthtool/synthtool/languages/node.py", line 167, in install
shell.run(["npm", "install"], hide_output=hide_output)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['npm', 'install']' returned non-zero exit status 1.
2020-08-27 04:29:45,366 autosynth [ERROR] > Synthesis failed
2020-08-27 04:29:45,367 autosynth [DEBUG] > Running: git reset --hard HEAD
HEAD is now at e96e2cc build: move system and samples test from Node 10 to Node 12 (#321)
2020-08-27 04:29:45,374 autosynth [DEBUG] > Running: git checkout autosynth
Switched to branch 'autosynth'
2020-08-27 04:29:45,379 autosynth [DEBUG] > Running: git clean -fdx
Removing __pycache__/
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 690, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 539, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 670, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 375, in synthesize_loop
has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 273, in synthesize_version_in_new_branch
synthesizer.synthesize(synth_log_path, self.environ)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/d56361e6-7265-4ad8-8ccb-dc2ff9ab5573/targets/github%2Fsynthtool;config=default/tests;query=nodejs-googleapis-common;failed=false).
|
non_defect
|
synthesis failed for nodejs googleapis common hello autosynth couldn t regenerate nodejs googleapis common broken heart here s the output from running synth py retty at autosynth running git checkout note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at build move system and samples test from node to node autosynth running git checkout note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at build java switch to release publish app for notifying github of release status autosynth running git branch f autosynth autosynth running git checkout autosynth switched to branch autosynth autosynth running synthtool autosynth autosynth log file path tmpfs src logs nodejs googleapis common sponge log log autosynth running tmpfs src github synthtool env bin m synthtool metadata synth metadata synth py synthtool executing home kbuilder cache synthtool nodejs googleapis common synth py on branch autosynth nothing to commit working tree clean synthtool using precloned repo home kbuilder cache synthtool synthtool debug synthtool using precloned repo home kbuilder cache synthtool synthtool eslintignore eslintrc json gitattributes github issue template bug report md github issue template feature request md github issue template support request md github pull request template md github release please yml github workflows ci yaml kokoro gitattributes kokoro common cfg kokoro continuous common cfg kokoro continuous docs cfg kokoro continuous test cfg kokoro continuous common cfg kokoro continuous lint cfg kokoro continuous samples test cfg kokoro continuous system test cfg kokoro continuous test cfg kokoro docs sh kokoro lint sh kokoro populate secrets sh kokoro presubmit common cfg kokoro presubmit common cfg kokoro presubmit samples test cfg kokoro presubmit system test cfg kokoro presubmit test cfg kokoro publish sh kokoro release docs devsite cfg kokoro release docs devsite sh kokoro release docs cfg kokoro release docs sh kokoro release publish cfg kokoro samples test sh kokoro system test sh kokoro test bat kokoro test sh kokoro trampoline sh mocharc js nycrc prettierignore prettierrc js code of conduct md contributing md license readme md api extractor json renovate json skipping samples readme md synthtool installing dependencies debug synthtool installing dependencies npm warn deprecated ts simple ast notice ts simple ast has been renamed to ts morph and version reset to switch at your leisure npm warn deprecated viz js no longer supported npm warn deprecated chokidar chokidar will break on node upgrade to chokidar with less dependencies npm warn deprecated opn the package has been renamed to open npm warn deprecated istanbul this module is no longer maintained try this instead npm warn deprecated npm i nyc npm warn deprecated visit for other alternatives npm warn deprecated flat fixed a prototype pollution security issue in please upgrade to or npm warn deprecated fsevents fsevents will break on node and could be using insecure binaries upgrade to fsevents npm warn deprecated resolve url npm warn deprecated urix please see npm warn deprecated core js core js is no longer maintained and not recommended for usage due to the number of issues please upgrade your dependencies to the actual version of core js npm warn optional skipping optional dependency fsevents node modules chokidar node modules fsevents npm warn notsup skipping optional dependency unsupported platform for fsevents wanted os darwin arch any current os linux arch npm warn optional skipping optional dependency fsevents node modules live server node modules chokidar node modules fsevents npm warn notsup skipping optional dependency unsupported platform for fsevents wanted os darwin arch any current os linux arch npm err code etarget npm err notarget no matching version found for rushstack node core library npm err notarget in most cases you or one of your dependencies are requesting npm err notarget a package version that doesn t exist npm err notarget npm err notarget it was specified as a dependency of microsoft api documenter npm err notarget npm err a complete log of this run can be found in npm err home kbuilder npm logs debug log synthtool failed executing npm install none error synthtool failed executing npm install none traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file home kbuilder cache synthtool nodejs googleapis common synth py line in node install file tmpfs src github synthtool synthtool languages node py line in install shell run hide output hide output file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status autosynth synthesis failed autosynth running git reset hard head head is now at build move system and samples test from node to node autosynth running git checkout autosynth switched to branch autosynth autosynth running git clean fdx removing pycache traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop has changes toolbox synthesize version in new branch synthesizer youngest file tmpfs src github synthtool autosynth synth py line in synthesize version in new branch synthesizer synthesize synth log path self environ file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
| 0
|
442,056
| 30,814,455,348
|
IssuesEvent
|
2023-08-01 12:39:45
|
aeon-toolkit/aeon
|
https://api.github.com/repos/aeon-toolkit/aeon
|
closed
|
[MNT] ReadTheDocs build failures due to timeout
|
documentation maintenance
|
### Describe the issue
Currently the ReadTheDocs builds for some PRs and main branch builds are timing out. I do not know the exact reason for this and it is not exactly consistent, but I suspect the API pages as removing the dependency to create them significantly reduces the build time from prior experience.
### Suggest a potential alternative/fix
Figure out what part of the docs build is taking so long and look for/discuss solutions.
### Additional context
_No response_
|
1.0
|
[MNT] ReadTheDocs build failures due to timeout - ### Describe the issue
Currently the ReadTheDocs builds for some PRs and main branch builds are timing out. I do not know the exact reason for this and it is not exactly consistent, but I suspect the API pages as removing the dependency to create them significantly reduces the build time from prior experience.
### Suggest a potential alternative/fix
Figure out what part of the docs build is taking so long and look for/discuss solutions.
### Additional context
_No response_
|
non_defect
|
readthedocs build failures due to timeout describe the issue currently the readthedocs builds for some prs and main branch builds are timing out i do not know the exact reason for this and it is not exactly consistent but i suspect the api pages as removing the dependency to create them significantly reduces the build time from prior experience suggest a potential alternative fix figure out what part of the docs build is taking so long and look for discuss solutions additional context no response
| 0
|
271,506
| 20,677,466,096
|
IssuesEvent
|
2022-03-10 10:38:42
|
tridactyl/tridactyl
|
https://api.github.com/repos/tridactyl/tridactyl
|
opened
|
investigate new tab page settings
|
P4 documentation
|
hpfr reports that you can now change the new tab page in about:preferences with Tridactyl running and Tridactyl runs on about:blank
If reproducible we should add this to the FAQ
|
1.0
|
investigate new tab page settings - hpfr reports that you can now change the new tab page in about:preferences with Tridactyl running and Tridactyl runs on about:blank
If reproducible we should add this to the FAQ
|
non_defect
|
investigate new tab page settings hpfr reports that you can now change the new tab page in about preferences with tridactyl running and tridactyl runs on about blank if reproducible we should add this to the faq
| 0
|
393,147
| 11,610,896,497
|
IssuesEvent
|
2020-02-26 04:43:05
|
earthref/MagIC
|
https://api.github.com/repos/earthref/MagIC
|
closed
|
Back button on plots should close the plots screen, not go back in the background on the browser
|
Priority - 2 Problem - 2
|
Back button on plots should close the plots screen, not go back in the background on the browser. People (I still do) will expect that back will close the plot screen and go back to the main one (In their mind the previous screen). Pressing the "x" should not be necessary (if easily done).
|
1.0
|
Back button on plots should close the plots screen, not go back in the background on the browser - Back button on plots should close the plots screen, not go back in the background on the browser. People (I still do) will expect that back will close the plot screen and go back to the main one (In their mind the previous screen). Pressing the "x" should not be necessary (if easily done).
|
non_defect
|
back button on plots should close the plots screen not go back in the background on the browser back button on plots should close the plots screen not go back in the background on the browser people i still do will expect that back will close the plot screen and go back to the main one in their mind the previous screen pressing the x should not be necessary if easily done
| 0
|
759,568
| 26,601,311,926
|
IssuesEvent
|
2023-01-23 15:57:56
|
codepandoradev/nft-marketplace-api
|
https://api.github.com/repos/codepandoradev/nft-marketplace-api
|
closed
|
Изменить в БД структуру таблицы Collections
|
task priority:hight area:app:collections not_milestone
|
## Описание
**id**
**name** - название
**description** - описание
**preview_image** - фото-превью для коллекции (маленькое). (не обязательна)
**banner_image** - большое фото, которое ставится на фон в качестве баннера (не обязательна)
**contract_creator** - контракт, через который были созданы нфт (если нфт создана через сайт - тут будет адресс нашего ск
**user_creator** - аддресс кошелька юзера, который создает коллекцию
**nft_list** - список айди нфт внутри коллекции (обязательна только для нфт, созданных нашим ск)
**isverified** - true / false - верифицирована ли коллекция
**category** - категория (art, game ..) - ссылается на другую таблицу CollectionCategory.id (не обязательна)
**network_chain_id** - chain_id сети, в которой создана коллекция. данные берутся из другой таблицы Networks.chain_id
**created_at** - дата создания коллекции
**twitter** - доп. информация (не обязательна)
**website** - доп. информация (не обязательна)
**discord** - доп. информация (не обязательна)
**telegram** -доп. информация (не обязательна)
#### Если изменений не требуется - close as not planned
|
1.0
|
Изменить в БД структуру таблицы Collections - ## Описание
**id**
**name** - название
**description** - описание
**preview_image** - фото-превью для коллекции (маленькое). (не обязательна)
**banner_image** - большое фото, которое ставится на фон в качестве баннера (не обязательна)
**contract_creator** - контракт, через который были созданы нфт (если нфт создана через сайт - тут будет адресс нашего ск
**user_creator** - аддресс кошелька юзера, который создает коллекцию
**nft_list** - список айди нфт внутри коллекции (обязательна только для нфт, созданных нашим ск)
**isverified** - true / false - верифицирована ли коллекция
**category** - категория (art, game ..) - ссылается на другую таблицу CollectionCategory.id (не обязательна)
**network_chain_id** - chain_id сети, в которой создана коллекция. данные берутся из другой таблицы Networks.chain_id
**created_at** - дата создания коллекции
**twitter** - доп. информация (не обязательна)
**website** - доп. информация (не обязательна)
**discord** - доп. информация (не обязательна)
**telegram** -доп. информация (не обязательна)
#### Если изменений не требуется - close as not planned
|
non_defect
|
изменить в бд структуру таблицы collections описание id name название description описание preview image фото превью для коллекции маленькое не обязательна banner image большое фото которое ставится на фон в качестве баннера не обязательна contract creator контракт через который были созданы нфт если нфт создана через сайт тут будет адресс нашего ск user creator аддресс кошелька юзера который создает коллекцию nft list список айди нфт внутри коллекции обязательна только для нфт созданных нашим ск isverified true false верифицирована ли коллекция category категория art game ссылается на другую таблицу collectioncategory id не обязательна network chain id chain id сети в которой создана коллекция данные берутся из другой таблицы networks chain id created at дата создания коллекции twitter доп информация не обязательна website доп информация не обязательна discord доп информация не обязательна telegram доп информация не обязательна если изменений не требуется close as not planned
| 0
|
55,064
| 14,170,641,802
|
IssuesEvent
|
2020-11-12 14:47:11
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
opened
|
Rounding errors during conversion between YearToSecond and Duration
|
T: Defect
|
### Expected behavior
When converting a Duration to a YearToSecond, and then converting back to a Duration, I expect to get the same value.
### Actual behavior
Due to rounding errors, the final value differs significantly from the original Duration.
### Steps to reproduce the problem
I found this problem using the following unit test:
```
Duration original = Duration.ofDays(730);
// After this next step, 126 hours have been added due to rounding error while converting to YearToMonth internally.
YearToSecond yts = YearToSecond.valueOf(original);
Duration converted = yts.toDuration();
// This fails:
Assert.assertEquals(original, converted);
```
Specifically, in this case the Duration of 730 days gives 17520 hours, or 63072000000 milliseconds. This is then converted (during the call to YearToSecond.valueOf(original)) to 1.99 years, which is then rounded to 1 by casting to int, and 12.15 months, rounded to 12 by cast to int. This is then passed to YearToMonth(int years, int months). The 12 month figure is there converted to a second year, yielding 2 years. The remainder (410400000 ms) is passed to DayToSecond.valueOf(double milli), which via a similar conversion process is converted to 4 days, 18 hours, for a total of 17646 hours, i.e. 126 hours more than the original value, which is substantial.
### Versions
- jOOQ: 3.12.3
- Java: 8
- OS: Linux (Kubuntu 19.10)
- JDBC Driver (include name if inofficial driver):
|
1.0
|
Rounding errors during conversion between YearToSecond and Duration - ### Expected behavior
When converting a Duration to a YearToSecond, and then converting back to a Duration, I expect to get the same value.
### Actual behavior
Due to rounding errors, the final value differs significantly from the original Duration.
### Steps to reproduce the problem
I found this problem using the following unit test:
```
Duration original = Duration.ofDays(730);
// After this next step, 126 hours have been added due to rounding error while converting to YearToMonth internally.
YearToSecond yts = YearToSecond.valueOf(original);
Duration converted = yts.toDuration();
// This fails:
Assert.assertEquals(original, converted);
```
Specifically, in this case the Duration of 730 days gives 17520 hours, or 63072000000 milliseconds. This is then converted (during the call to YearToSecond.valueOf(original)) to 1.99 years, which is then rounded to 1 by casting to int, and 12.15 months, rounded to 12 by cast to int. This is then passed to YearToMonth(int years, int months). The 12 month figure is there converted to a second year, yielding 2 years. The remainder (410400000 ms) is passed to DayToSecond.valueOf(double milli), which via a similar conversion process is converted to 4 days, 18 hours, for a total of 17646 hours, i.e. 126 hours more than the original value, which is substantial.
### Versions
- jOOQ: 3.12.3
- Java: 8
- OS: Linux (Kubuntu 19.10)
- JDBC Driver (include name if inofficial driver):
|
defect
|
rounding errors during conversion between yeartosecond and duration expected behavior when converting a duration to a yeartosecond and then converting back to a duration i expect to get the same value actual behavior due to rounding errors the final value differs significantly from the original duration steps to reproduce the problem i found this problem using the following unit test duration original duration ofdays after this next step hours have been added due to rounding error while converting to yeartomonth internally yeartosecond yts yeartosecond valueof original duration converted yts toduration this fails assert assertequals original converted specifically in this case the duration of days gives hours or milliseconds this is then converted during the call to yeartosecond valueof original to years which is then rounded to by casting to int and months rounded to by cast to int this is then passed to yeartomonth int years int months the month figure is there converted to a second year yielding years the remainder ms is passed to daytosecond valueof double milli which via a similar conversion process is converted to days hours for a total of hours i e hours more than the original value which is substantial versions jooq java os linux kubuntu jdbc driver include name if inofficial driver
| 1
|
30,362
| 24,761,018,963
|
IssuesEvent
|
2022-10-22 00:18:42
|
dotnet/roslyn
|
https://api.github.com/repos/dotnet/roslyn
|
closed
|
[Automated] PRs inserted in VS build main-33021.259
|
Area-Infrastructure untriaged vs-insertion
|
[View Complete Diff of Changes](https://github.com/dotnet/roslyn/compare/502be8b98e833f79a652f7533a4bbd4fb295bbac...9ec682f330b5a8779b4676a81f11718b5c8ef209?w=1)
- [Revert #64616 and #64576 (64873)](https://github.com/dotnet/roslyn/pull/64873)
- [Revert "Implement equality semantics for async tagger tags." (64867)](https://github.com/dotnet/roslyn/pull/64867)
- [Fix missing Analyzer property ref in build (64129)](https://github.com/dotnet/roslyn/pull/64129)
- [Fix analyzer consistency checks (64831)](https://github.com/dotnet/roslyn/pull/64831)
- [Disable Windows PDB conversion (64581)](https://github.com/dotnet/roslyn/pull/64581)
- [Disable sqlite service on full disk (64844)](https://github.com/dotnet/roslyn/pull/64844)
- [Use ParsedDocument in more commmand handlers (63579)](https://github.com/dotnet/roslyn/pull/63579)
- [Fix Color Contrast For STE Close Button (64834)](https://github.com/dotnet/roslyn/pull/64834)
- [Snippets - turn on by default in 17.5 (64829)](https://github.com/dotnet/roslyn/pull/64829)
- [Update 17.4 PreReleaseVersionLabel to 4 for preview 4 (64832)](https://github.com/dotnet/roslyn/pull/64832)
- [Additional tests for UTF-8 string literals feature (64824)](https://github.com/dotnet/roslyn/pull/64824)
- [Revert "Read required project properties from evaluation data" (64801)](https://github.com/dotnet/roslyn/pull/64801)
- [Encoding serialization (64625)](https://github.com/dotnet/roslyn/pull/64625)
- [Implemen equality semantics for async tagger tags. (64802)](https://github.com/dotnet/roslyn/pull/64802)
- [Use IAsyncEnumerable in DesignerAttributeScanning as well. (64616)](https://github.com/dotnet/roslyn/pull/64616)
- [Export correct type for VisualStudioWorkpacePartialSolutionsTestHook.cs (64799)](https://github.com/dotnet/roslyn/pull/64799)
- [Stay on the UI thread if it's blocked waiting for us (64813)](https://github.com/dotnet/roslyn/pull/64813)
- [Move parts of official build onto new vm images (64805)](https://github.com/dotnet/roslyn/pull/64805)
- [Snippets - turn feature off by default for 17.4 (64810)](https://github.com/dotnet/roslyn/pull/64810)
- [Ensure we include a projectId along with our ProjectChanged notifications (64808)](https://github.com/dotnet/roslyn/pull/64808)
- ["Generate method" codefix supports 'in' parameters (64588)](https://github.com/dotnet/roslyn/pull/64588)
- [Make `TargetFramework.NetCoreApp` use `Net70` (64791)](https://github.com/dotnet/roslyn/pull/64791)
- [VB: Align design around speculative semantic models with C#. (64751)](https://github.com/dotnet/roslyn/pull/64751)
- [Do not treat `Nullable` null value as a constant for the purpose of IL generation. (64789)](https://github.com/dotnet/roslyn/pull/64789)
- [Add more symbol publishing packages to PublishData.json (64790)](https://github.com/dotnet/roslyn/pull/64790)
- [Clarify error message when attempting to ref a property (64764)](https://github.com/dotnet/roslyn/pull/64764)
- [Don't format regions inside inactive conditionals (64551)](https://github.com/dotnet/roslyn/pull/64551)
- [Fix nullable walker cycle (64600)](https://github.com/dotnet/roslyn/pull/64600)
- [Improve SyntaxTree/Node.Dump() (62703)](https://github.com/dotnet/roslyn/pull/62703)
- [Add tests for ref fields (64707)](https://github.com/dotnet/roslyn/pull/64707)
- [Fix MSBuild tests and run them in CI again (64781)](https://github.com/dotnet/roslyn/pull/64781)
- [Read required project properties from evaluation data (64749)](https://github.com/dotnet/roslyn/pull/64749)
- [Trigger background analyzer load when document open (64786)](https://github.com/dotnet/roslyn/pull/64786)
- [Unskip CSharpSoruceGenerators.InvokeNavigateToForGeneratedFile test (64626)](https://github.com/dotnet/roslyn/pull/64626)
- [Use IAsyncEnumerable in NavigateTo OOP calls. (64576)](https://github.com/dotnet/roslyn/pull/64576)
- [Scoped implicitly-typed lambda parameters (64680)](https://github.com/dotnet/roslyn/pull/64680)
|
1.0
|
[Automated] PRs inserted in VS build main-33021.259 - [View Complete Diff of Changes](https://github.com/dotnet/roslyn/compare/502be8b98e833f79a652f7533a4bbd4fb295bbac...9ec682f330b5a8779b4676a81f11718b5c8ef209?w=1)
- [Revert #64616 and #64576 (64873)](https://github.com/dotnet/roslyn/pull/64873)
- [Revert "Implement equality semantics for async tagger tags." (64867)](https://github.com/dotnet/roslyn/pull/64867)
- [Fix missing Analyzer property ref in build (64129)](https://github.com/dotnet/roslyn/pull/64129)
- [Fix analyzer consistency checks (64831)](https://github.com/dotnet/roslyn/pull/64831)
- [Disable Windows PDB conversion (64581)](https://github.com/dotnet/roslyn/pull/64581)
- [Disable sqlite service on full disk (64844)](https://github.com/dotnet/roslyn/pull/64844)
- [Use ParsedDocument in more commmand handlers (63579)](https://github.com/dotnet/roslyn/pull/63579)
- [Fix Color Contrast For STE Close Button (64834)](https://github.com/dotnet/roslyn/pull/64834)
- [Snippets - turn on by default in 17.5 (64829)](https://github.com/dotnet/roslyn/pull/64829)
- [Update 17.4 PreReleaseVersionLabel to 4 for preview 4 (64832)](https://github.com/dotnet/roslyn/pull/64832)
- [Additional tests for UTF-8 string literals feature (64824)](https://github.com/dotnet/roslyn/pull/64824)
- [Revert "Read required project properties from evaluation data" (64801)](https://github.com/dotnet/roslyn/pull/64801)
- [Encoding serialization (64625)](https://github.com/dotnet/roslyn/pull/64625)
- [Implemen equality semantics for async tagger tags. (64802)](https://github.com/dotnet/roslyn/pull/64802)
- [Use IAsyncEnumerable in DesignerAttributeScanning as well. (64616)](https://github.com/dotnet/roslyn/pull/64616)
- [Export correct type for VisualStudioWorkpacePartialSolutionsTestHook.cs (64799)](https://github.com/dotnet/roslyn/pull/64799)
- [Stay on the UI thread if it's blocked waiting for us (64813)](https://github.com/dotnet/roslyn/pull/64813)
- [Move parts of official build onto new vm images (64805)](https://github.com/dotnet/roslyn/pull/64805)
- [Snippets - turn feature off by default for 17.4 (64810)](https://github.com/dotnet/roslyn/pull/64810)
- [Ensure we include a projectId along with our ProjectChanged notifications (64808)](https://github.com/dotnet/roslyn/pull/64808)
- ["Generate method" codefix supports 'in' parameters (64588)](https://github.com/dotnet/roslyn/pull/64588)
- [Make `TargetFramework.NetCoreApp` use `Net70` (64791)](https://github.com/dotnet/roslyn/pull/64791)
- [VB: Align design around speculative semantic models with C#. (64751)](https://github.com/dotnet/roslyn/pull/64751)
- [Do not treat `Nullable` null value as a constant for the purpose of IL generation. (64789)](https://github.com/dotnet/roslyn/pull/64789)
- [Add more symbol publishing packages to PublishData.json (64790)](https://github.com/dotnet/roslyn/pull/64790)
- [Clarify error message when attempting to ref a property (64764)](https://github.com/dotnet/roslyn/pull/64764)
- [Don't format regions inside inactive conditionals (64551)](https://github.com/dotnet/roslyn/pull/64551)
- [Fix nullable walker cycle (64600)](https://github.com/dotnet/roslyn/pull/64600)
- [Improve SyntaxTree/Node.Dump() (62703)](https://github.com/dotnet/roslyn/pull/62703)
- [Add tests for ref fields (64707)](https://github.com/dotnet/roslyn/pull/64707)
- [Fix MSBuild tests and run them in CI again (64781)](https://github.com/dotnet/roslyn/pull/64781)
- [Read required project properties from evaluation data (64749)](https://github.com/dotnet/roslyn/pull/64749)
- [Trigger background analyzer load when document open (64786)](https://github.com/dotnet/roslyn/pull/64786)
- [Unskip CSharpSoruceGenerators.InvokeNavigateToForGeneratedFile test (64626)](https://github.com/dotnet/roslyn/pull/64626)
- [Use IAsyncEnumerable in NavigateTo OOP calls. (64576)](https://github.com/dotnet/roslyn/pull/64576)
- [Scoped implicitly-typed lambda parameters (64680)](https://github.com/dotnet/roslyn/pull/64680)
|
non_defect
|
prs inserted in vs build main
| 0
|
52,222
| 13,211,408,854
|
IssuesEvent
|
2020-08-15 22:56:10
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
gcc 6.2.0 and c++11 numeric-literals (Trac #1872)
|
Incomplete Migration Migrated from Trac cmake defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1872">https://code.icecube.wisc.edu/projects/icecube/ticket/1872</a>, reported by david.schultzand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-06-12T12:46:40",
"_ts": "1560343600245810",
"description": "We need either `-std=gnu++11` or `-fext-numeric-literals`.\n\n{{{\n[ 7%] Building CXX object dataclasses/CMakeFiles/dataclasses.dir/private/dataclasses/I3DOMFunctions.cxx.o\ncd /home/dschultz/Documents/combo/trunk/build_gcc6/dataclasses && /home/dschultz/Documents/combo/trunk/build_gcc6/CMakeFiles/gfilt /usr/bin/c++ -DPROJECT=dataclasses -Ddataclasses_EXPORTS -I/home/dschultz/Documents/combo/trunk/src/cmake/tool-patches/common -I/home/dschultz/Documents/combo/trunk/src/cmake/tool-patches/boost-new -I/home/dschultz/Documents/combo/trunk/src/dataclasses/public -I/home/dschultz/Documents/combo/trunk/src/dataclasses/private -I/home/dschultz/Documents/combo/trunk/src/serialization/public -I/home/dschultz/Documents/combo/trunk/src/icetray/public -isystem /usr/local_boost_cpp14/include -isystem /usr/include/python2.7 -I/home/dschultz/Documents/combo/trunk/src/dataio/public -pipe -std=c++11 -Wall -Wno-non-virtual-dtor -Wno-deprecated -Wno-unused-variable -Wno-unused-local-typedefs -O3 -DI3_COMPILE_OUT_VERBOSE_LOGGING -fPIC -include /home/dschultz/Documents/combo/trunk/build_gcc6/CMakeFiles/I3.h -std=gnu++14 -o CMakeFiles/dataclasses.dir/private/dataclasses/I3DOMFunctions.cxx.o -c /home/dschultz/Documents/combo/trunk/src/dataclasses/private/dataclasses/I3DOMFunctions.cxx\nIn file included from /usr/local_boost_cpp14/include/boost/config.hpp:61:0,\nfrom\n /home/dschultz/Documents/combo/trunk/src/serialization/public/serialization\n /extended_type_info_typeid.hpp:24,\nfrom\n /home/dschultz/Documents/combo/trunk/src/icetray/public/icetray/i3_extended\n _type_info.h:23,\nfrom\n /home/dschultz/Documents/combo/trunk/src/icetray/public/icetray/serializati\n on.h:29,\nfrom\n /home/dschultz/Documents/combo/trunk/src/dataclasses/public/dataclasses/Uti\n lity.h:26,\nfrom\n /home/dschultz/Documents/combo/trunk/src/dataclasses/public/dataclasses/sta\n tus/I3DOMStatus.h:4,\nfrom\n /home/dschultz/Documents/combo/trunk/src/dataclasses/public/dataclasses/I3D\n OMFunctions.h:12,\nfrom /home/dschultz/Documents/combo/trunk/src/dataclasses/private/dataclasses/I3DOMFunctions.cxx:10:\n /usr/local_boost_cpp14/include/boost/math/constants/constants.hpp: In static\n member function \u2018static constexpr T boost::math::constants::detail\n ::constant_half<T>::get(const mpl_::int_<5> &)\u2019:\n/usr/local_boost_cpp14/include/boost/math/constants/constants.hpp:265:3: error:\n unable to find numeric literal operator \u2018operator\"\"Q\u2019\nBOOST_DEFINE_MATH_CONSTANT(\n half, 5.000000000000000000000000000000000000e-01, \"\n 5.000000000000000000000000000000000000000000000000000000000000000000000\n 00000000000000000000000000000000000000000e-01\")\n^\n/usr/local_boost_cpp14/include/boost/math/constants/constants.hpp:265:3: note:\n use -std=gnu++11 or -fext-numeric-literals to enable more built-in suffixes\n}}}",
"reporter": "david.schultz",
"cc": "",
"resolution": "wontfix",
"time": "2016-10-01T10:39:21",
"component": "cmake",
"summary": "gcc 6.2.0 and c++11 numeric-literals",
"priority": "major",
"keywords": "",
"milestone": "Long-Term Future",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
gcc 6.2.0 and c++11 numeric-literals (Trac #1872) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1872">https://code.icecube.wisc.edu/projects/icecube/ticket/1872</a>, reported by david.schultzand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-06-12T12:46:40",
"_ts": "1560343600245810",
"description": "We need either `-std=gnu++11` or `-fext-numeric-literals`.\n\n{{{\n[ 7%] Building CXX object dataclasses/CMakeFiles/dataclasses.dir/private/dataclasses/I3DOMFunctions.cxx.o\ncd /home/dschultz/Documents/combo/trunk/build_gcc6/dataclasses && /home/dschultz/Documents/combo/trunk/build_gcc6/CMakeFiles/gfilt /usr/bin/c++ -DPROJECT=dataclasses -Ddataclasses_EXPORTS -I/home/dschultz/Documents/combo/trunk/src/cmake/tool-patches/common -I/home/dschultz/Documents/combo/trunk/src/cmake/tool-patches/boost-new -I/home/dschultz/Documents/combo/trunk/src/dataclasses/public -I/home/dschultz/Documents/combo/trunk/src/dataclasses/private -I/home/dschultz/Documents/combo/trunk/src/serialization/public -I/home/dschultz/Documents/combo/trunk/src/icetray/public -isystem /usr/local_boost_cpp14/include -isystem /usr/include/python2.7 -I/home/dschultz/Documents/combo/trunk/src/dataio/public -pipe -std=c++11 -Wall -Wno-non-virtual-dtor -Wno-deprecated -Wno-unused-variable -Wno-unused-local-typedefs -O3 -DI3_COMPILE_OUT_VERBOSE_LOGGING -fPIC -include /home/dschultz/Documents/combo/trunk/build_gcc6/CMakeFiles/I3.h -std=gnu++14 -o CMakeFiles/dataclasses.dir/private/dataclasses/I3DOMFunctions.cxx.o -c /home/dschultz/Documents/combo/trunk/src/dataclasses/private/dataclasses/I3DOMFunctions.cxx\nIn file included from /usr/local_boost_cpp14/include/boost/config.hpp:61:0,\nfrom\n /home/dschultz/Documents/combo/trunk/src/serialization/public/serialization\n /extended_type_info_typeid.hpp:24,\nfrom\n /home/dschultz/Documents/combo/trunk/src/icetray/public/icetray/i3_extended\n _type_info.h:23,\nfrom\n /home/dschultz/Documents/combo/trunk/src/icetray/public/icetray/serializati\n on.h:29,\nfrom\n /home/dschultz/Documents/combo/trunk/src/dataclasses/public/dataclasses/Uti\n lity.h:26,\nfrom\n /home/dschultz/Documents/combo/trunk/src/dataclasses/public/dataclasses/sta\n tus/I3DOMStatus.h:4,\nfrom\n /home/dschultz/Documents/combo/trunk/src/dataclasses/public/dataclasses/I3D\n OMFunctions.h:12,\nfrom /home/dschultz/Documents/combo/trunk/src/dataclasses/private/dataclasses/I3DOMFunctions.cxx:10:\n /usr/local_boost_cpp14/include/boost/math/constants/constants.hpp: In static\n member function \u2018static constexpr T boost::math::constants::detail\n ::constant_half<T>::get(const mpl_::int_<5> &)\u2019:\n/usr/local_boost_cpp14/include/boost/math/constants/constants.hpp:265:3: error:\n unable to find numeric literal operator \u2018operator\"\"Q\u2019\nBOOST_DEFINE_MATH_CONSTANT(\n half, 5.000000000000000000000000000000000000e-01, \"\n 5.000000000000000000000000000000000000000000000000000000000000000000000\n 00000000000000000000000000000000000000000e-01\")\n^\n/usr/local_boost_cpp14/include/boost/math/constants/constants.hpp:265:3: note:\n use -std=gnu++11 or -fext-numeric-literals to enable more built-in suffixes\n}}}",
"reporter": "david.schultz",
"cc": "",
"resolution": "wontfix",
"time": "2016-10-01T10:39:21",
"component": "cmake",
"summary": "gcc 6.2.0 and c++11 numeric-literals",
"priority": "major",
"keywords": "",
"milestone": "Long-Term Future",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
defect
|
gcc and c numeric literals trac migrated from json status closed changetime ts description we need either std gnu or fext numeric literals n n n building cxx object dataclasses cmakefiles dataclasses dir private dataclasses cxx o ncd home dschultz documents combo trunk build dataclasses home dschultz documents combo trunk build cmakefiles gfilt usr bin c dproject dataclasses ddataclasses exports i home dschultz documents combo trunk src cmake tool patches common i home dschultz documents combo trunk src cmake tool patches boost new i home dschultz documents combo trunk src dataclasses public i home dschultz documents combo trunk src dataclasses private i home dschultz documents combo trunk src serialization public i home dschultz documents combo trunk src icetray public isystem usr local boost include isystem usr include i home dschultz documents combo trunk src dataio public pipe std c wall wno non virtual dtor wno deprecated wno unused variable wno unused local typedefs compile out verbose logging fpic include home dschultz documents combo trunk build cmakefiles h std gnu o cmakefiles dataclasses dir private dataclasses cxx o c home dschultz documents combo trunk src dataclasses private dataclasses cxx nin file included from usr local boost include boost config hpp nfrom n home dschultz documents combo trunk src serialization public serialization n extended type info typeid hpp nfrom n home dschultz documents combo trunk src icetray public icetray extended n type info h nfrom n home dschultz documents combo trunk src icetray public icetray serializati n on h nfrom n home dschultz documents combo trunk src dataclasses public dataclasses uti n lity h nfrom n home dschultz documents combo trunk src dataclasses public dataclasses sta n tus h nfrom n home dschultz documents combo trunk src dataclasses public dataclasses n omfunctions h nfrom home dschultz documents combo trunk src dataclasses private dataclasses cxx n usr local boost include boost math constants constants hpp in static n member function constexpr t boost math constants detail n constant half get const mpl int n usr local boost include boost math constants constants hpp error n unable to find numeric literal operator q nboost define math constant n half n n n n usr local boost include boost math constants constants hpp note n use std gnu or fext numeric literals to enable more built in suffixes n reporter david schultz cc resolution wontfix time component cmake summary gcc and c numeric literals priority major keywords milestone long term future owner nega type defect
| 1
|
25,832
| 4,469,036,489
|
IssuesEvent
|
2016-08-25 11:37:36
|
contao/core
|
https://api.github.com/repos/contao/core
|
closed
|
Renaming files on Windows machine can lead to deletion
|
defect
|
By accident I just figured out that it's not possible to rename a file with a mixed-case naming to all lower- or all uppercase due to a "fix" in [/system/modules/core/library/Contao/Files/Php.php#L121-L125](https://github.com/contao/core/blob/436f0cc99089756e7a3d2e0aa53155b1823942d6/system/modules/core/library/Contao/Files/Php.php#L121-L125).
This "fix" deletes the file and tries to rename it afterwards...
Maybe there should be a check like this to prevent the deletion if the filename is the same regardless of case-sensitivity:
```php
// Windows fix: delete the target file
if (strtoupper(substr(PHP_OS, 0, 3)) === 'WIN' && file_exists(TL_ROOT . '/' . $strNewName) && strcasecmp($strOldName, $strNewName) !== 0)
{
$this->delete($strNewName);
}
```
|
1.0
|
Renaming files on Windows machine can lead to deletion - By accident I just figured out that it's not possible to rename a file with a mixed-case naming to all lower- or all uppercase due to a "fix" in [/system/modules/core/library/Contao/Files/Php.php#L121-L125](https://github.com/contao/core/blob/436f0cc99089756e7a3d2e0aa53155b1823942d6/system/modules/core/library/Contao/Files/Php.php#L121-L125).
This "fix" deletes the file and tries to rename it afterwards...
Maybe there should be a check like this to prevent the deletion if the filename is the same regardless of case-sensitivity:
```php
// Windows fix: delete the target file
if (strtoupper(substr(PHP_OS, 0, 3)) === 'WIN' && file_exists(TL_ROOT . '/' . $strNewName) && strcasecmp($strOldName, $strNewName) !== 0)
{
$this->delete($strNewName);
}
```
|
defect
|
renaming files on windows machine can lead to deletion by accident i just figured out that it s not possible to rename a file with a mixed case naming to all lower or all uppercase due to a fix in this fix deletes the file and tries to rename it afterwards maybe there should be a check like this to prevent the deletion if the filename is the same regardless of case sensitivity php windows fix delete the target file if strtoupper substr php os win file exists tl root strnewname strcasecmp stroldname strnewname this delete strnewname
| 1
|
21,391
| 3,506,103,668
|
IssuesEvent
|
2016-01-08 03:34:45
|
CompEvol/beast2
|
https://api.github.com/repos/CompEvol/beast2
|
closed
|
EBSPAnalyser computes equal-tailed credible interval, not HPD as reported
|
defect LOW priority
|
Equal-tailed intervals can be quite different to HPDs.
|
1.0
|
EBSPAnalyser computes equal-tailed credible interval, not HPD as reported - Equal-tailed intervals can be quite different to HPDs.
|
defect
|
ebspanalyser computes equal tailed credible interval not hpd as reported equal tailed intervals can be quite different to hpds
| 1
|
65,201
| 19,268,872,697
|
IssuesEvent
|
2021-12-10 01:22:29
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
User status doesn't have “…” at the end of it if it's too long
|
T-Defect
|
### Steps to reproduce
1. Set a long status
2. Room Info > People > scroll down until your long status
### Outcome
#### What did you expect?
If the user status is long, it should have those three dots at the end of it
#### What happened instead?

### Operating system
Arch Linux
### Application version
Versi Element Nightly: 2021120801 Versi Olm: 3.2.3
### How did you install the app?
AUR
### Homeserver
matrix.org
### Will you send logs?
No
|
1.0
|
User status doesn't have “…” at the end of it if it's too long - ### Steps to reproduce
1. Set a long status
2. Room Info > People > scroll down until your long status
### Outcome
#### What did you expect?
If the user status is long, it should have those three dots at the end of it
#### What happened instead?

### Operating system
Arch Linux
### Application version
Versi Element Nightly: 2021120801 Versi Olm: 3.2.3
### How did you install the app?
AUR
### Homeserver
matrix.org
### Will you send logs?
No
|
defect
|
user status doesn t have “…” at the end of it if it s too long steps to reproduce set a long status room info people scroll down until your long status outcome what did you expect if the user status is long it should have those three dots at the end of it what happened instead operating system arch linux application version versi element nightly versi olm how did you install the app aur homeserver matrix org will you send logs no
| 1
|
29,184
| 5,579,938,460
|
IssuesEvent
|
2017-03-28 15:34:29
|
primefaces/primeng
|
https://api.github.com/repos/primefaces/primeng
|
closed
|
Lazy loaded data tables never set multiSortMeta
|
defect
|
<!--
- IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION WE MIGHT CLOSE YOUR ISSUE WITHOUT INVESTIGATING.
- IF YOU'D LIKE TO SECURE OUR RESPONSE, YOU MAY CONSIDER PRIMENG PRO SUPPORT WHERE SUPPORT IS PROVIDED WITHIN 4 hours.
-->
**I'm submitting a ...** (check one with "x")
```
[x] bug report => Search github for a similar issue or PR before submitting
[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
```
**Plunkr Case (Bug Reports)**
Don't have time to put a plunkr together at the moment, but it doesn't seem like it would be particularly useful for this bug
**Current behavior**
If a data table is set to use lazy loading, the user is expected to handle sorting based on the information emitted in the onLazyLoad and/or onSort events; however, the multiSortMeta field is never set for a lazy loaded table. As seen in the code snippet below (from datatable.ts sort() function), addSortMeta is not called if the table is lazy. Therefore, multiSortMeta will always be undefined, unless the attribute is set by the caller, which means that whichever component subscribes to the onLazyLoad event cannot know how to sort the data.
```
if(this.lazy) {
this.stopFilterPropagation = true;
this.onLazyLoad.emit(this.createLazyLoadMetadata());
}
else {
if(this.sortMode == 'multiple') {
if(!this.multiSortMeta||!metaKey) {
this.multiSortMeta = [];
}
this.addSortMeta({field: this.sortField, order: this.sortOrder});
this.sortMultiple();
}
else {
this.sortSingle();
}
}
this.onSort.emit({
field: this.sortField,
order: this.sortOrder,
multisortmeta: this.multiSortMeta
});
```
It seems that addSortMeta should be called before the onLazyLoad event is emitted, as below
```
if(this.sortMode == 'multiple') {
if(!this.multiSortMeta||!metaKey) {
this.multiSortMeta = [];
}
this.addSortMeta({field: this.sortField, order: this.sortOrder});
}
if(this.lazy) {
this.stopFilterPropagation = true;
this.onLazyLoad.emit(this.createLazyLoadMetadata());
}
else if (this.sortMode == 'multiple'){
this.sortMultiple();
}
else {
this.sortSingle();
}
this.onSort.emit({
field: this.sortField,
order: this.sortOrder,
multisortmeta: this.multiSortMeta
});
```
**Expected behavior**
multiSortMeta should be set before the onLazyLoad event is emitted so that the component listening for the event knows how the data should be sorted.
**Minimal reproduction of the problem with instructions**
Don't have time to make a plunker right now, but you should just need to create a datatable that is lazy and has sort mode set to multiple. Kind of hard to make a functional example, since the problem is that there isn't enough information emitted with the onLazyLoad event to actually sort properly in this scenario.
**What is the motivation / use case for changing the behavior?**
Be able to multisort on lazy loaded tables
**Please tell us about your environment:**
<!-- Operating system, IDE, package manager, HTTP server, ... -->
Windows 10, Visual Studio 2015 Enterprise Edition, using npm
* **Angular version:** 2.0.0
* **PrimeNG version:** 1.0.0
* **Browser:**
At least Chrome 54. Haven't tried in anything else, but doesn't seem browser-specific
* **Language:**
Typescript 2.0.8
* **Node (for AoT issues):**
5.8.0
|
1.0
|
Lazy loaded data tables never set multiSortMeta - <!--
- IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION WE MIGHT CLOSE YOUR ISSUE WITHOUT INVESTIGATING.
- IF YOU'D LIKE TO SECURE OUR RESPONSE, YOU MAY CONSIDER PRIMENG PRO SUPPORT WHERE SUPPORT IS PROVIDED WITHIN 4 hours.
-->
**I'm submitting a ...** (check one with "x")
```
[x] bug report => Search github for a similar issue or PR before submitting
[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
```
**Plunkr Case (Bug Reports)**
Don't have time to put a plunkr together at the moment, but it doesn't seem like it would be particularly useful for this bug
**Current behavior**
If a data table is set to use lazy loading, the user is expected to handle sorting based on the information emitted in the onLazyLoad and/or onSort events; however, the multiSortMeta field is never set for a lazy loaded table. As seen in the code snippet below (from datatable.ts sort() function), addSortMeta is not called if the table is lazy. Therefore, multiSortMeta will always be undefined, unless the attribute is set by the caller, which means that whichever component subscribes to the onLazyLoad event cannot know how to sort the data.
```
if(this.lazy) {
this.stopFilterPropagation = true;
this.onLazyLoad.emit(this.createLazyLoadMetadata());
}
else {
if(this.sortMode == 'multiple') {
if(!this.multiSortMeta||!metaKey) {
this.multiSortMeta = [];
}
this.addSortMeta({field: this.sortField, order: this.sortOrder});
this.sortMultiple();
}
else {
this.sortSingle();
}
}
this.onSort.emit({
field: this.sortField,
order: this.sortOrder,
multisortmeta: this.multiSortMeta
});
```
It seems that addSortMeta should be called before the onLazyLoad event is emitted, as below
```
if(this.sortMode == 'multiple') {
if(!this.multiSortMeta||!metaKey) {
this.multiSortMeta = [];
}
this.addSortMeta({field: this.sortField, order: this.sortOrder});
}
if(this.lazy) {
this.stopFilterPropagation = true;
this.onLazyLoad.emit(this.createLazyLoadMetadata());
}
else if (this.sortMode == 'multiple'){
this.sortMultiple();
}
else {
this.sortSingle();
}
this.onSort.emit({
field: this.sortField,
order: this.sortOrder,
multisortmeta: this.multiSortMeta
});
```
**Expected behavior**
multiSortMeta should be set before the onLazyLoad event is emitted so that the component listening for the event knows how the data should be sorted.
**Minimal reproduction of the problem with instructions**
Don't have time to make a plunker right now, but you should just need to create a datatable that is lazy and has sort mode set to multiple. Kind of hard to make a functional example, since the problem is that there isn't enough information emitted with the onLazyLoad event to actually sort properly in this scenario.
**What is the motivation / use case for changing the behavior?**
Be able to multisort on lazy loaded tables
**Please tell us about your environment:**
<!-- Operating system, IDE, package manager, HTTP server, ... -->
Windows 10, Visual Studio 2015 Enterprise Edition, using npm
* **Angular version:** 2.0.0
* **PrimeNG version:** 1.0.0
* **Browser:**
At least Chrome 54. Haven't tried in anything else, but doesn't seem browser-specific
* **Language:**
Typescript 2.0.8
* **Node (for AoT issues):**
5.8.0
|
defect
|
lazy loaded data tables never set multisortmeta if you don t fill out the following information we might close your issue without investigating if you d like to secure our response you may consider primeng pro support where support is provided within hours i m submitting a check one with x bug report search github for a similar issue or pr before submitting feature request please check if request is not on the roadmap already support request please do not submit support request here instead see plunkr case bug reports don t have time to put a plunkr together at the moment but it doesn t seem like it would be particularly useful for this bug current behavior if a data table is set to use lazy loading the user is expected to handle sorting based on the information emitted in the onlazyload and or onsort events however the multisortmeta field is never set for a lazy loaded table as seen in the code snippet below from datatable ts sort function addsortmeta is not called if the table is lazy therefore multisortmeta will always be undefined unless the attribute is set by the caller which means that whichever component subscribes to the onlazyload event cannot know how to sort the data if this lazy this stopfilterpropagation true this onlazyload emit this createlazyloadmetadata else if this sortmode multiple if this multisortmeta metakey this multisortmeta this addsortmeta field this sortfield order this sortorder this sortmultiple else this sortsingle this onsort emit field this sortfield order this sortorder multisortmeta this multisortmeta it seems that addsortmeta should be called before the onlazyload event is emitted as below if this sortmode multiple if this multisortmeta metakey this multisortmeta this addsortmeta field this sortfield order this sortorder if this lazy this stopfilterpropagation true this onlazyload emit this createlazyloadmetadata else if this sortmode multiple this sortmultiple else this sortsingle this onsort emit field this sortfield order this sortorder multisortmeta this multisortmeta expected behavior multisortmeta should be set before the onlazyload event is emitted so that the component listening for the event knows how the data should be sorted minimal reproduction of the problem with instructions don t have time to make a plunker right now but you should just need to create a datatable that is lazy and has sort mode set to multiple kind of hard to make a functional example since the problem is that there isn t enough information emitted with the onlazyload event to actually sort properly in this scenario what is the motivation use case for changing the behavior be able to multisort on lazy loaded tables please tell us about your environment windows visual studio enterprise edition using npm angular version primeng version browser at least chrome haven t tried in anything else but doesn t seem browser specific language typescript node for aot issues
| 1
|
25,700
| 5,192,699,345
|
IssuesEvent
|
2017-01-22 11:47:08
|
marienfressinaud/project-zero
|
https://api.github.com/repos/marienfressinaud/project-zero
|
closed
|
write CONTRIBUTING
|
P: High T: Documentation
|
CONTRIBUTING.md must include:
- how to propose an idea
- how to contribute to code
- add a `.gitmessage` file
- how to report a bug
|
1.0
|
write CONTRIBUTING - CONTRIBUTING.md must include:
- how to propose an idea
- how to contribute to code
- add a `.gitmessage` file
- how to report a bug
|
non_defect
|
write contributing contributing md must include how to propose an idea how to contribute to code add a gitmessage file how to report a bug
| 0
|
53,081
| 10,986,944,778
|
IssuesEvent
|
2019-12-02 08:05:10
|
nugu-developers/nugu-ios
|
https://api.github.com/repos/nugu-developers/nugu-ios
|
closed
|
Prepare API Stability `JadeMarble` and `KeenSense`
|
code quality enhancement in progress
|
# Overviews
## Category
- [ ] New feature
- [x] Update feature
- [ ] Update documentation
- [ ] Idea
- [ ] Architecture
## Summary
`JadeMarble` and `KeenSense` are separate from 'nugu-ios', need to prepare for API Stability.
So, Ambiguous names that using to `JadeMarble` and `KeenSense` should be more clearly defined.
## Reference
None
|
1.0
|
Prepare API Stability `JadeMarble` and `KeenSense` - # Overviews
## Category
- [ ] New feature
- [x] Update feature
- [ ] Update documentation
- [ ] Idea
- [ ] Architecture
## Summary
`JadeMarble` and `KeenSense` are separate from 'nugu-ios', need to prepare for API Stability.
So, Ambiguous names that using to `JadeMarble` and `KeenSense` should be more clearly defined.
## Reference
None
|
non_defect
|
prepare api stability jademarble and keensense overviews category new feature update feature update documentation idea architecture summary jademarble and keensense are separate from nugu ios need to prepare for api stability so ambiguous names that using to jademarble and keensense should be more clearly defined reference none
| 0
|
29,991
| 4,544,395,631
|
IssuesEvent
|
2016-09-10 17:25:08
|
FreeCodeCamp/FreeCodeCamp
|
https://api.github.com/repos/FreeCodeCamp/FreeCodeCamp
|
closed
|
Filter Arrays with .filter - can't pass
|
confirmed help wanted tests
|
Challenge [filter-arrays-with-filter](https://www.freecodecamp.com/en/challenges/object-oriented-and-functional-programming/filter-arrays-with-filter) has an issue.
User Agent is: <code>Mozilla/5.0 (Windows NT 6.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/52.0.2743.116 Safari/537.36</code>.
Please describe how to reproduce this issue, and include links to screenshots if possible.
Using the code below and it should be working but getting error message: "You should only be using filter to make the new array." I have searched and this is the correct code so I don't know what the problem is..
```js
var oldArray = [1,2,3,4,5,6,7,8,9,10];
// Only change code below this line.
var newArray = oldArray.filter(function(val) {
return val <= 5;
});
```
|
1.0
|
Filter Arrays with .filter - can't pass - Challenge [filter-arrays-with-filter](https://www.freecodecamp.com/en/challenges/object-oriented-and-functional-programming/filter-arrays-with-filter) has an issue.
User Agent is: <code>Mozilla/5.0 (Windows NT 6.1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/52.0.2743.116 Safari/537.36</code>.
Please describe how to reproduce this issue, and include links to screenshots if possible.
Using the code below and it should be working but getting error message: "You should only be using filter to make the new array." I have searched and this is the correct code so I don't know what the problem is..
```js
var oldArray = [1,2,3,4,5,6,7,8,9,10];
// Only change code below this line.
var newArray = oldArray.filter(function(val) {
return val <= 5;
});
```
|
non_defect
|
filter arrays with filter can t pass challenge has an issue user agent is mozilla windows nt applewebkit khtml like gecko chrome safari please describe how to reproduce this issue and include links to screenshots if possible using the code below and it should be working but getting error message you should only be using filter to make the new array i have searched and this is the correct code so i don t know what the problem is js var oldarray only change code below this line var newarray oldarray filter function val return val
| 0
|
12,347
| 2,693,140,148
|
IssuesEvent
|
2015-04-01 13:09:17
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
{ INSERT | UPDATE | MERGE } .. SET [ Record ] doesn't take changed flags into account
|
C: Functionality P: Medium R: Fixed T: Defect T: Incompatible change
|
The below query shouldn't update the `SOME_FIELD` value:
```java
Record record = dsl.newRecord(getTable(), entity);
record.changed(SOME_FIELD, false);
dsl.update(getTable())
.set(record)
.where(getTableId().eq(record.getValue(getTableId())))
.execute();
```
----
As this is a slightly incompatible change of behaviour, it won't be merged to jOOQ 3.5.x releases
See also:
https://groups.google.com/forum/#!topic/jooq-user/DIw1KfQaCs0
|
1.0
|
{ INSERT | UPDATE | MERGE } .. SET [ Record ] doesn't take changed flags into account - The below query shouldn't update the `SOME_FIELD` value:
```java
Record record = dsl.newRecord(getTable(), entity);
record.changed(SOME_FIELD, false);
dsl.update(getTable())
.set(record)
.where(getTableId().eq(record.getValue(getTableId())))
.execute();
```
----
As this is a slightly incompatible change of behaviour, it won't be merged to jOOQ 3.5.x releases
See also:
https://groups.google.com/forum/#!topic/jooq-user/DIw1KfQaCs0
|
defect
|
insert update merge set doesn t take changed flags into account the below query shouldn t update the some field value java record record dsl newrecord gettable entity record changed some field false dsl update gettable set record where gettableid eq record getvalue gettableid execute as this is a slightly incompatible change of behaviour it won t be merged to jooq x releases see also
| 1
|
24,774
| 11,091,983,030
|
IssuesEvent
|
2019-12-15 15:57:40
|
Ignitus/Ignitus-rest-api
|
https://api.github.com/repos/Ignitus/Ignitus-rest-api
|
closed
|
WS-2019-0064 (High) detected in handlebars-4.1.0.tgz
|
security vulnerability
|
## WS-2019-0064 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.0.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.1.0.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.1.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/Ignitus-rest-api/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/Ignitus-rest-api/node_modules/nyc/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- nyc-13.3.0.tgz (Root Library)
- istanbul-reports-2.1.1.tgz
- :x: **handlebars-4.1.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Ignitus/Ignitus-rest-api/commit/be70a6e737b2bcefe7805d2e2d591469f1f2ad7d">be70a6e737b2bcefe7805d2e2d591469f1f2ad7d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of handlebars prior to 4.0.14 are vulnerable to Prototype Pollution. Templates may alter an Objects' prototype, thus allowing an attacker to execute arbitrary code on the server.
<p>Publish Date: 2019-04-30
<p>URL: <a href=https://github.com/wycats/handlebars.js/compare/v4.1.1...v4.1.2>WS-2019-0064</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/755/versions">https://www.npmjs.com/advisories/755/versions</a></p>
<p>Release Date: 2019-04-30</p>
<p>Fix Resolution: 1.0.6-2,4.0.14,4.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2019-0064 (High) detected in handlebars-4.1.0.tgz - ## WS-2019-0064 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.0.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.1.0.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.1.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/Ignitus-rest-api/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/Ignitus-rest-api/node_modules/nyc/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- nyc-13.3.0.tgz (Root Library)
- istanbul-reports-2.1.1.tgz
- :x: **handlebars-4.1.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Ignitus/Ignitus-rest-api/commit/be70a6e737b2bcefe7805d2e2d591469f1f2ad7d">be70a6e737b2bcefe7805d2e2d591469f1f2ad7d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of handlebars prior to 4.0.14 are vulnerable to Prototype Pollution. Templates may alter an Objects' prototype, thus allowing an attacker to execute arbitrary code on the server.
<p>Publish Date: 2019-04-30
<p>URL: <a href=https://github.com/wycats/handlebars.js/compare/v4.1.1...v4.1.2>WS-2019-0064</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/755/versions">https://www.npmjs.com/advisories/755/versions</a></p>
<p>Release Date: 2019-04-30</p>
<p>Fix Resolution: 1.0.6-2,4.0.14,4.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
ws high detected in handlebars tgz ws high severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file tmp ws scm ignitus rest api package json path to vulnerable library tmp ws scm ignitus rest api node modules nyc node modules handlebars package json dependency hierarchy nyc tgz root library istanbul reports tgz x handlebars tgz vulnerable library found in head commit a href vulnerability details versions of handlebars prior to are vulnerable to prototype pollution templates may alter an objects prototype thus allowing an attacker to execute arbitrary code on the server publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
32,747
| 6,913,239,369
|
IssuesEvent
|
2017-11-28 14:43:37
|
hazelcast/hazelcast-jet
|
https://api.github.com/repos/hazelcast/hazelcast-jet
|
closed
|
WriteKafkaP might not meet at least once guarantee
|
defect kafka
|
`KafkaProducer.send` is asynchronous. The Kafka sink might theoretically lose records when snapshot is made, job fails just after it and the async call is not complete yet.
|
1.0
|
WriteKafkaP might not meet at least once guarantee - `KafkaProducer.send` is asynchronous. The Kafka sink might theoretically lose records when snapshot is made, job fails just after it and the async call is not complete yet.
|
defect
|
writekafkap might not meet at least once guarantee kafkaproducer send is asynchronous the kafka sink might theoretically lose records when snapshot is made job fails just after it and the async call is not complete yet
| 1
|
33,371
| 7,105,498,708
|
IssuesEvent
|
2018-01-16 13:53:00
|
TrevorPilley/MicroLite.Extensions.WebApi.OData
|
https://api.github.com/repos/TrevorPilley/MicroLite.Extensions.WebApi.OData
|
closed
|
Return 404 NOT FOUND if property correct but no results
|
defect
|
Applies to `GetEntityPropertyResponseAsync` and `GetEntityPropertyValueResponseAsync`
|
1.0
|
Return 404 NOT FOUND if property correct but no results - Applies to `GetEntityPropertyResponseAsync` and `GetEntityPropertyValueResponseAsync`
|
defect
|
return not found if property correct but no results applies to getentitypropertyresponseasync and getentitypropertyvalueresponseasync
| 1
|
34,265
| 7,431,749,039
|
IssuesEvent
|
2018-03-25 17:43:53
|
Yahkal/replicaisland
|
https://api.github.com/repos/Yahkal/replicaisland
|
closed
|
Blank screen when returning from screen off state
|
Priority-Medium Type-Defect auto-migrated
|
```
What steps will reproduce the problem?
1. Open a game level and put into pause (menu button)
2. Either press 'off' button to go into 'screen off' mode or allow to timeout
3. Press the 'menu' button twice to go back into 'active' mode
What is the expected output? What do you see instead?
I expect to see the same game level displayed with the 'Paused' and the level
number. I see a blank, black screen. If I press the 'back' button, no 'quit
level' dialog is displayed. If I press menu, nothing happens. I need to press
the 'home' button and restart back at the game menu.
What version of the product are you using? On what operating system?
Replica Island 1.4. myTouch 2G running (t-Mobile's) Android 2.2.1
Please provide any additional information below.
Sometimes, when I return to the game, the screen is still black and I need to
stop the app in Settings > Applications > Manage applications. I didn't see
this occurring on the previous release.
(P.S. great game, thank you!)
```
Original issue reported on code.google.com by `arcege` on 26 Jan 2011 at 2:18
|
1.0
|
Blank screen when returning from screen off state - ```
What steps will reproduce the problem?
1. Open a game level and put into pause (menu button)
2. Either press 'off' button to go into 'screen off' mode or allow to timeout
3. Press the 'menu' button twice to go back into 'active' mode
What is the expected output? What do you see instead?
I expect to see the same game level displayed with the 'Paused' and the level
number. I see a blank, black screen. If I press the 'back' button, no 'quit
level' dialog is displayed. If I press menu, nothing happens. I need to press
the 'home' button and restart back at the game menu.
What version of the product are you using? On what operating system?
Replica Island 1.4. myTouch 2G running (t-Mobile's) Android 2.2.1
Please provide any additional information below.
Sometimes, when I return to the game, the screen is still black and I need to
stop the app in Settings > Applications > Manage applications. I didn't see
this occurring on the previous release.
(P.S. great game, thank you!)
```
Original issue reported on code.google.com by `arcege` on 26 Jan 2011 at 2:18
|
defect
|
blank screen when returning from screen off state what steps will reproduce the problem open a game level and put into pause menu button either press off button to go into screen off mode or allow to timeout press the menu button twice to go back into active mode what is the expected output what do you see instead i expect to see the same game level displayed with the paused and the level number i see a blank black screen if i press the back button no quit level dialog is displayed if i press menu nothing happens i need to press the home button and restart back at the game menu what version of the product are you using on what operating system replica island mytouch running t mobile s android please provide any additional information below sometimes when i return to the game the screen is still black and i need to stop the app in settings applications manage applications i didn t see this occurring on the previous release p s great game thank you original issue reported on code google com by arcege on jan at
| 1
|
256,368
| 19,409,961,232
|
IssuesEvent
|
2021-12-20 08:26:40
|
mapeditor/tiled
|
https://api.github.com/repos/mapeditor/tiled
|
closed
|
Add a workflow that builds the manual
|
documentation
|
Seems we can use https://github.com/ammaraskar/sphinx-action for this. It would be good to trigger this on any changes to the `docs/` directory, to check for any problems in building the manual.
|
1.0
|
Add a workflow that builds the manual - Seems we can use https://github.com/ammaraskar/sphinx-action for this. It would be good to trigger this on any changes to the `docs/` directory, to check for any problems in building the manual.
|
non_defect
|
add a workflow that builds the manual seems we can use for this it would be good to trigger this on any changes to the docs directory to check for any problems in building the manual
| 0
|
246,494
| 20,868,185,579
|
IssuesEvent
|
2022-03-22 09:26:49
|
Uuvana-Studios/longvinter-windows-client
|
https://api.github.com/repos/Uuvana-Studios/longvinter-windows-client
|
closed
|
Vendor Screen Stuck/ Pix's Island
|
bug Not Tested
|
**Describe the bug**
When I arrived to the island I tried to select the vendor which didn't work at first, but after walking away the vendor screen appeared and I could not exit (pressing Esc only brought me to the main menu).
**To (Potentially) Reproduce**
Steps to reproduce the behavior:
1. Go to Pix Island
2. Click on Vendor twice
3. Walk off the dock and past the bonfire
4. Error occurs (vendor screen appears then doesn't go away)
**Expected behavior**
Clicking on the vendor should bring up the menu, and I should be allowed to exit the menu via walking away or pressing Tab
**Screenshots**
N/A
**Desktop (please complete the following information):**
- OS: Windows 11
- Game Version: Unsure
- Steam Version Unsure
|
1.0
|
Vendor Screen Stuck/ Pix's Island - **Describe the bug**
When I arrived to the island I tried to select the vendor which didn't work at first, but after walking away the vendor screen appeared and I could not exit (pressing Esc only brought me to the main menu).
**To (Potentially) Reproduce**
Steps to reproduce the behavior:
1. Go to Pix Island
2. Click on Vendor twice
3. Walk off the dock and past the bonfire
4. Error occurs (vendor screen appears then doesn't go away)
**Expected behavior**
Clicking on the vendor should bring up the menu, and I should be allowed to exit the menu via walking away or pressing Tab
**Screenshots**
N/A
**Desktop (please complete the following information):**
- OS: Windows 11
- Game Version: Unsure
- Steam Version Unsure
|
non_defect
|
vendor screen stuck pix s island describe the bug when i arrived to the island i tried to select the vendor which didn t work at first but after walking away the vendor screen appeared and i could not exit pressing esc only brought me to the main menu to potentially reproduce steps to reproduce the behavior go to pix island click on vendor twice walk off the dock and past the bonfire error occurs vendor screen appears then doesn t go away expected behavior clicking on the vendor should bring up the menu and i should be allowed to exit the menu via walking away or pressing tab screenshots n a desktop please complete the following information os windows game version unsure steam version unsure
| 0
|
532,387
| 15,555,263,679
|
IssuesEvent
|
2021-03-16 05:46:45
|
TeselaGen/openVectorEditor
|
https://api.github.com/repos/TeselaGen/openVectorEditor
|
closed
|
Editing the feature type is not changing the color in the plasmid view
|
Context: OVE Context: TeselaGen HDE Priority: 3-Normal Type: Fix or Enhance
|
Changing the feature type through right-click > Edit Feature > select a different feature type is changing the color of the feature in the sequence map but not in the plasmid view.

|
1.0
|
Editing the feature type is not changing the color in the plasmid view - Changing the feature type through right-click > Edit Feature > select a different feature type is changing the color of the feature in the sequence map but not in the plasmid view.

|
non_defect
|
editing the feature type is not changing the color in the plasmid view changing the feature type through right click edit feature select a different feature type is changing the color of the feature in the sequence map but not in the plasmid view
| 0
|
140,231
| 18,895,250,088
|
IssuesEvent
|
2021-11-15 17:09:20
|
bgoonz/searchAwesome
|
https://api.github.com/repos/bgoonz/searchAwesome
|
closed
|
CVE-2018-11698 (High) detected in lportalliferay-ce-portal-src-7.3.5-ga6-20200930172312275, node-sass-4.11.0.tgz
|
security vulnerability
|
## CVE-2018-11698 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lportalliferay-ce-portal-src-7.3.5-ga6-20200930172312275</b>, <b>node-sass-4.11.0.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.11.0.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz</a></p>
<p>Path to dependency file: searchAwesome/clones/awesome-stacks/package.json</p>
<p>Path to vulnerable library: /clones/awesome-stacks/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.11.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/bgoonz/searchAwesome/commit/8c366c860f88ff2849d4a7b7832c781154d89ece">8c366c860f88ff2849d4a7b7832c781154d89ece</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::handle_error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11698>CVE-2018-11698</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.5.5">https://github.com/sass/libsass/releases/tag/3.5.5</a></p>
<p>Release Date: 2018-06-04</p>
<p>Fix Resolution: libsass - 3.5.5;node-sass - 4.14.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-11698 (High) detected in lportalliferay-ce-portal-src-7.3.5-ga6-20200930172312275, node-sass-4.11.0.tgz - ## CVE-2018-11698 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lportalliferay-ce-portal-src-7.3.5-ga6-20200930172312275</b>, <b>node-sass-4.11.0.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.11.0.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz</a></p>
<p>Path to dependency file: searchAwesome/clones/awesome-stacks/package.json</p>
<p>Path to vulnerable library: /clones/awesome-stacks/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.11.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/bgoonz/searchAwesome/commit/8c366c860f88ff2849d4a7b7832c781154d89ece">8c366c860f88ff2849d4a7b7832c781154d89ece</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in LibSass through 3.5.4. An out-of-bounds read of a memory region was found in the function Sass::handle_error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service.
<p>Publish Date: 2018-06-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11698>CVE-2018-11698</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.5.5">https://github.com/sass/libsass/releases/tag/3.5.5</a></p>
<p>Release Date: 2018-06-04</p>
<p>Fix Resolution: libsass - 3.5.5;node-sass - 4.14.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in lportalliferay ce portal src node sass tgz cve high severity vulnerability vulnerable libraries lportalliferay ce portal src node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file searchawesome clones awesome stacks package json path to vulnerable library clones awesome stacks node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href found in base branch master vulnerability details an issue was discovered in libsass through an out of bounds read of a memory region was found in the function sass handle error which could be leveraged by an attacker to disclose information or manipulated to read from unmapped memory causing a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass node sass step up your open source security game with whitesource
| 0
|
42,153
| 10,840,735,561
|
IssuesEvent
|
2019-11-12 09:00:51
|
mozilla-lockwise/lockwise-ios
|
https://api.github.com/repos/mozilla-lockwise/lockwise-ios
|
closed
|
[iPad] crash on autofill onboarding screen
|
QA Needed blocks-release crash type: defect
|
## Steps to reproduce
1. Clean install Lockwise
2. Follow the welcome screen until you reach the autofill onboarding screen
### Actual behavior
* Lockwise will crash.
### Expected behavior
* Lockwise will not crash.
### Device & build information
* Device: iPad Pro 3rd gen
* Build version: master branch build number 3809
### Notes
* Reproducible only on iPad
* Attachments: [Firefox Lockbox 11-11-19, 3-07 PM.crash.zip](https://github.com/mozilla-lockwise/lockwise-ios/files/3831427/Firefox.Lockbox.11-11-19.3-07.PM.crash.zip)
|
1.0
|
[iPad] crash on autofill onboarding screen - ## Steps to reproduce
1. Clean install Lockwise
2. Follow the welcome screen until you reach the autofill onboarding screen
### Actual behavior
* Lockwise will crash.
### Expected behavior
* Lockwise will not crash.
### Device & build information
* Device: iPad Pro 3rd gen
* Build version: master branch build number 3809
### Notes
* Reproducible only on iPad
* Attachments: [Firefox Lockbox 11-11-19, 3-07 PM.crash.zip](https://github.com/mozilla-lockwise/lockwise-ios/files/3831427/Firefox.Lockbox.11-11-19.3-07.PM.crash.zip)
|
defect
|
crash on autofill onboarding screen steps to reproduce clean install lockwise follow the welcome screen until you reach the autofill onboarding screen actual behavior lockwise will crash expected behavior lockwise will not crash device build information device ipad pro gen build version master branch build number notes reproducible only on ipad attachments
| 1
|
193,915
| 22,261,545,892
|
IssuesEvent
|
2022-06-10 01:22:06
|
ShaikUsaf/linux-3.0.35
|
https://api.github.com/repos/ShaikUsaf/linux-3.0.35
|
opened
|
CVE-2012-4001 (Medium) detected in linuxlinux-3.0.49
|
security vulnerability
|
## CVE-2012-4001 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.0.49</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The mod_pagespeed module before 0.10.22.6 for the Apache HTTP Server does not properly verify its host name, which allows remote attackers to trigger HTTP requests to arbitrary hosts via unspecified vectors, as demonstrated by requests to intranet servers.
<p>Publish Date: 2012-09-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-4001>CVE-2012-4001</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/atniptw/PonyBuntu/commit/fe155b0056e0110eb4dd0c5935506d8370039b8c">https://github.com/atniptw/PonyBuntu/commit/fe155b0056e0110eb4dd0c5935506d8370039b8c</a></p>
<p>Release Date: 2012-09-15</p>
<p>Fix Resolution: Ubuntu-3.2.0-31.49</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2012-4001 (Medium) detected in linuxlinux-3.0.49 - ## CVE-2012-4001 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.0.49</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The mod_pagespeed module before 0.10.22.6 for the Apache HTTP Server does not properly verify its host name, which allows remote attackers to trigger HTTP requests to arbitrary hosts via unspecified vectors, as demonstrated by requests to intranet servers.
<p>Publish Date: 2012-09-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-4001>CVE-2012-4001</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/atniptw/PonyBuntu/commit/fe155b0056e0110eb4dd0c5935506d8370039b8c">https://github.com/atniptw/PonyBuntu/commit/fe155b0056e0110eb4dd0c5935506d8370039b8c</a></p>
<p>Release Date: 2012-09-15</p>
<p>Fix Resolution: Ubuntu-3.2.0-31.49</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch master vulnerable source files vulnerability details the mod pagespeed module before for the apache http server does not properly verify its host name which allows remote attackers to trigger http requests to arbitrary hosts via unspecified vectors as demonstrated by requests to intranet servers publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution ubuntu step up your open source security game with mend
| 0
|
2,415
| 2,525,846,471
|
IssuesEvent
|
2015-01-21 06:42:01
|
graybeal/ont
|
https://api.github.com/repos/graybeal/ont
|
closed
|
links to /or/sparql and /ont/sparql.html
|
1 star aquaportal enhancement imported portal Priority-Critical
|
_From [caru...@gmail.com](https://code.google.com/u/113886747689301365533/) on January 06, 2009 14:16:01_
What capability do you want added or improved? Links to http://mmisw.org/ont/sparql.html from the main ORR page Where do you want this capability to be accessible? From Tips and Tricks block in http://mmisw.org/or and as a "see examples"
link in http://mmisw.org/or/sparql
_Original issue: http://code.google.com/p/mmisw/issues/detail?id=92_
|
1.0
|
links to /or/sparql and /ont/sparql.html - _From [caru...@gmail.com](https://code.google.com/u/113886747689301365533/) on January 06, 2009 14:16:01_
What capability do you want added or improved? Links to http://mmisw.org/ont/sparql.html from the main ORR page Where do you want this capability to be accessible? From Tips and Tricks block in http://mmisw.org/or and as a "see examples"
link in http://mmisw.org/or/sparql
_Original issue: http://code.google.com/p/mmisw/issues/detail?id=92_
|
non_defect
|
links to or sparql and ont sparql html from on january what capability do you want added or improved links to from the main orr page where do you want this capability to be accessible from tips and tricks block in and as a see examples link in original issue
| 0
|
673,608
| 23,022,261,640
|
IssuesEvent
|
2022-07-22 06:13:28
|
status-im/status-desktop
|
https://api.github.com/repos/status-im/status-desktop
|
closed
|
Onboarding/Login flow improvements
|
onboarding feature general priority F2: important E:Onboarding E:PerformanceImprovements S:4
|
The idea of this issue is to improve Onboarding/Login process in a way to make it more concise, intuitive and easier for fture maintaining applying the following steps:
- remove redundant code
- merge `startup`, `login` and `onboarding` modules into a single one (since all of them are actually a single thing and share lot of same calls)
- define and organize states in a more intuitive and readable state
|
1.0
|
Onboarding/Login flow improvements - The idea of this issue is to improve Onboarding/Login process in a way to make it more concise, intuitive and easier for fture maintaining applying the following steps:
- remove redundant code
- merge `startup`, `login` and `onboarding` modules into a single one (since all of them are actually a single thing and share lot of same calls)
- define and organize states in a more intuitive and readable state
|
non_defect
|
onboarding login flow improvements the idea of this issue is to improve onboarding login process in a way to make it more concise intuitive and easier for fture maintaining applying the following steps remove redundant code merge startup login and onboarding modules into a single one since all of them are actually a single thing and share lot of same calls define and organize states in a more intuitive and readable state
| 0
|
44,084
| 11,954,085,853
|
IssuesEvent
|
2020-04-03 22:29:19
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
Datatable: stickyTopAt broken at showcase when div.layout-notification is shown
|
defect
|
## 1) Environment
- PrimeFaces version: 8.0-SNAPSHOT
- Does it work on the newest released PrimeFaces version? Not sure
- Does it work on the newest sources in GitHub? No
- Application server + version: -
- Affected browsers: all
## 2) Expected behavior
The stickyTopAt attribute points to .layout-topbar and this should be the top position.
## 3) Actual behavior
The datatable header stops at the .layout-notification element above, parially hiding the topbar.
## 4) Steps to reproduce
Go to https://www.primefaces.org/showcase/ui/data/datatable/sticky.xhtml during the Black-Friday week, in this case a div.layout-notification is rendered above the div.layout-topbar. If deleting the div.layout-notificaction during display, everything is back to normal.
## 5) Sample XHTML
See showcase
## 6) Sample bean
See showcase
If this is needed I can also try to provide an additional reproducer.
|
1.0
|
Datatable: stickyTopAt broken at showcase when div.layout-notification is shown - ## 1) Environment
- PrimeFaces version: 8.0-SNAPSHOT
- Does it work on the newest released PrimeFaces version? Not sure
- Does it work on the newest sources in GitHub? No
- Application server + version: -
- Affected browsers: all
## 2) Expected behavior
The stickyTopAt attribute points to .layout-topbar and this should be the top position.
## 3) Actual behavior
The datatable header stops at the .layout-notification element above, parially hiding the topbar.
## 4) Steps to reproduce
Go to https://www.primefaces.org/showcase/ui/data/datatable/sticky.xhtml during the Black-Friday week, in this case a div.layout-notification is rendered above the div.layout-topbar. If deleting the div.layout-notificaction during display, everything is back to normal.
## 5) Sample XHTML
See showcase
## 6) Sample bean
See showcase
If this is needed I can also try to provide an additional reproducer.
|
defect
|
datatable stickytopat broken at showcase when div layout notification is shown environment primefaces version snapshot does it work on the newest released primefaces version not sure does it work on the newest sources in github no application server version affected browsers all expected behavior the stickytopat attribute points to layout topbar and this should be the top position actual behavior the datatable header stops at the layout notification element above parially hiding the topbar steps to reproduce go to during the black friday week in this case a div layout notification is rendered above the div layout topbar if deleting the div layout notificaction during display everything is back to normal sample xhtml see showcase sample bean see showcase if this is needed i can also try to provide an additional reproducer
| 1
|
149,742
| 11,913,513,498
|
IssuesEvent
|
2020-03-31 12:08:27
|
PierrickI3/CCCRequestsForm
|
https://api.github.com/repos/PierrickI3/CCCRequestsForm
|
closed
|
Send email to requester with more details
|
enhancement ready to test
|
* Add all fields in email to requester for both CC requests and rapid response
|
1.0
|
Send email to requester with more details - * Add all fields in email to requester for both CC requests and rapid response
|
non_defect
|
send email to requester with more details add all fields in email to requester for both cc requests and rapid response
| 0
|
50,343
| 13,187,452,500
|
IssuesEvent
|
2020-08-13 03:27:39
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
geometry renderer specifies center of detector (Trac #558)
|
Migrated from Trac defect glshovel
|
for km3net... (0,0,0) isn't in the center of their detector.
<details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/558
, reported by troy and owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2009-07-20T22:14:29",
"description": "for km3net... (0,0,0) isn't in the center of their detector.",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"_ts": "1248128069000000",
"component": "glshovel",
"summary": "geometry renderer specifies center of detector",
"priority": "normal",
"keywords": "",
"time": "2009-06-12T13:52:20",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
geometry renderer specifies center of detector (Trac #558) - for km3net... (0,0,0) isn't in the center of their detector.
<details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/558
, reported by troy and owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2009-07-20T22:14:29",
"description": "for km3net... (0,0,0) isn't in the center of their detector.",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"_ts": "1248128069000000",
"component": "glshovel",
"summary": "geometry renderer specifies center of detector",
"priority": "normal",
"keywords": "",
"time": "2009-06-12T13:52:20",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
|
defect
|
geometry renderer specifies center of detector trac for isn t in the center of their detector migrated from reported by troy and owned by troy json status closed changetime description for isn t in the center of their detector reporter troy cc resolution fixed ts component glshovel summary geometry renderer specifies center of detector priority normal keywords time milestone owner troy type defect
| 1
|
153,486
| 19,706,452,024
|
IssuesEvent
|
2022-01-12 22:41:53
|
KaterinaOrg/maven-modular
|
https://api.github.com/repos/KaterinaOrg/maven-modular
|
closed
|
CVE-2020-36188 (High) detected in jackson-databind-2.9.6.jar - autoclosed
|
security vulnerability
|
## CVE-2020-36188 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /module2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- jackson-module-kotlin-2.9.6.jar (Root Library)
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/KaterinaOrg/maven-modular/commit/27eea04c852891f5a7663c22ad865308b8149942">27eea04c852891f5a7663c22ad865308b8149942</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.JNDIConnectionSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36188>CVE-2020-36188</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2996">https://github.com/FasterXML/jackson-databind/issues/2996</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/module2/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.fasterxml.jackson.module:jackson-module-kotlin:2.9.6;com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-36188","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.JNDIConnectionSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36188","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-36188 (High) detected in jackson-databind-2.9.6.jar - autoclosed - ## CVE-2020-36188 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /module2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- jackson-module-kotlin-2.9.6.jar (Root Library)
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/KaterinaOrg/maven-modular/commit/27eea04c852891f5a7663c22ad865308b8149942">27eea04c852891f5a7663c22ad865308b8149942</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.JNDIConnectionSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36188>CVE-2020-36188</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2996">https://github.com/FasterXML/jackson-databind/issues/2996</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/module2/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.fasterxml.jackson.module:jackson-module-kotlin:2.9.6;com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-36188","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.JNDIConnectionSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36188","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_defect
|
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy jackson module kotlin jar root library x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com newrelic agent deps ch qos logback core db jndiconnectionsource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree com fasterxml jackson module jackson module kotlin com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com newrelic agent deps ch qos logback core db jndiconnectionsource vulnerabilityurl
| 0
|
77,846
| 27,192,038,343
|
IssuesEvent
|
2023-02-19 22:28:10
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Pinned poll has different vote count than the one in the timeline/history
|
T-Defect S-Major A-Message-Pinning O-Occasional A-Polls Z-Labs
|
### Steps to reproduce
1. Start a poll in a reasonably high traffic room
2. Pin the poll
3. Wait N days, where N > 1
4. Open the pins to see the poll. Note the vote count increase as it fetches data from the server.
5. Click 'view message' to jump to the poll
6. Note that the poll has a different count and doesn't increase.
### Outcome
#### What did you expect?
That the poll have the same vote count regardless of the panel it's shown in.
#### What happened instead?
Presumably when I jumped to the poll in the timeline it changed which timeline segment it was looking at, so it missed most of the votes because they were in a different timeline segment.
### Operating system
Windows 10
### Application version
_No response_
### How did you install the app?
The Internet
### Homeserver
t2l.io
### Will you send logs?
No
|
1.0
|
Pinned poll has different vote count than the one in the timeline/history - ### Steps to reproduce
1. Start a poll in a reasonably high traffic room
2. Pin the poll
3. Wait N days, where N > 1
4. Open the pins to see the poll. Note the vote count increase as it fetches data from the server.
5. Click 'view message' to jump to the poll
6. Note that the poll has a different count and doesn't increase.
### Outcome
#### What did you expect?
That the poll have the same vote count regardless of the panel it's shown in.
#### What happened instead?
Presumably when I jumped to the poll in the timeline it changed which timeline segment it was looking at, so it missed most of the votes because they were in a different timeline segment.
### Operating system
Windows 10
### Application version
_No response_
### How did you install the app?
The Internet
### Homeserver
t2l.io
### Will you send logs?
No
|
defect
|
pinned poll has different vote count than the one in the timeline history steps to reproduce start a poll in a reasonably high traffic room pin the poll wait n days where n open the pins to see the poll note the vote count increase as it fetches data from the server click view message to jump to the poll note that the poll has a different count and doesn t increase outcome what did you expect that the poll have the same vote count regardless of the panel it s shown in what happened instead presumably when i jumped to the poll in the timeline it changed which timeline segment it was looking at so it missed most of the votes because they were in a different timeline segment operating system windows application version no response how did you install the app the internet homeserver io will you send logs no
| 1
|
120,950
| 25,900,095,321
|
IssuesEvent
|
2022-12-15 04:21:32
|
WordPress/openverse-api
|
https://api.github.com/repos/WordPress/openverse-api
|
closed
|
Move off of grequests
|
help wanted 🟨 priority: medium ✨ goal: improvement 💻 aspect: code 🐍 tech: python
|
## Problem
<!-- Describe a problem solved by this feature; or delete the section entirely. -->
The API currently uses [the `grequests` library](https://github.com/spyoungtech/grequests) for handling some of the `head` requests it completes when returning results (specifically in [`validate_images.py`](https://github.com/WordPress/openverse-api/blob/main/api/catalog/api/utils/validate_images.py#L37)).
## Description
<!-- Describe the feature and how it solves the problem. -->
The README for the project says:
> Note: You should probably use requests-threads or requests-futures instead.
The project has also not had a release since 2020. We might want to consider using one of those proposed alternatives.
## Alternatives
<!-- Describe any alternative solutions or features you have considered. How is this feature better? -->
Leave as is and hope for the best!
## Additional context
<!-- Add any other context about the feature here; or delete the section entirely. -->
Some more context in Sentry: https://sentry.io/share/issue/061ba99fc3df4c23bdb7643d337bbda0/
## Implementation
<!-- Replace the [ ] with [x] to check the box. -->
- [ ] 🙋 I would be interested in implementing this feature.
|
1.0
|
Move off of grequests - ## Problem
<!-- Describe a problem solved by this feature; or delete the section entirely. -->
The API currently uses [the `grequests` library](https://github.com/spyoungtech/grequests) for handling some of the `head` requests it completes when returning results (specifically in [`validate_images.py`](https://github.com/WordPress/openverse-api/blob/main/api/catalog/api/utils/validate_images.py#L37)).
## Description
<!-- Describe the feature and how it solves the problem. -->
The README for the project says:
> Note: You should probably use requests-threads or requests-futures instead.
The project has also not had a release since 2020. We might want to consider using one of those proposed alternatives.
## Alternatives
<!-- Describe any alternative solutions or features you have considered. How is this feature better? -->
Leave as is and hope for the best!
## Additional context
<!-- Add any other context about the feature here; or delete the section entirely. -->
Some more context in Sentry: https://sentry.io/share/issue/061ba99fc3df4c23bdb7643d337bbda0/
## Implementation
<!-- Replace the [ ] with [x] to check the box. -->
- [ ] 🙋 I would be interested in implementing this feature.
|
non_defect
|
move off of grequests problem the api currently uses for handling some of the head requests it completes when returning results specifically in description the readme for the project says note you should probably use requests threads or requests futures instead the project has also not had a release since we might want to consider using one of those proposed alternatives alternatives leave as is and hope for the best additional context some more context in sentry implementation 🙋 i would be interested in implementing this feature
| 0
|
53,690
| 13,262,100,818
|
IssuesEvent
|
2020-08-20 21:06:19
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
[iceprod] propagate the error code for failing IceProd tasks to pass it correctly to the batch system (Trac #1882)
|
Migrated from Trac defect iceprod
|
It seems that some IceProd tasks that fail in some of the final steps like contacting the DB, or staging out the output data, or other, are not reporting this error with a non-zero error code at the end of the process. As of now, they are happily exiting with exit code = 0 after the failure.
Returning some non-zero error code at the end of the task will be useful for us to monitor issues in production and react faster for fixing them.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1882">https://code.icecube.wisc.edu/projects/icecube/ticket/1882</a>, reported by gmerinoand owned by juancarlos</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-11-03T21:01:05",
"_ts": "1478206865363965",
"description": "It seems that some IceProd tasks that fail in some of the final steps like contacting the DB, or staging out the output data, or other, are not reporting this error with a non-zero error code at the end of the process. As of now, they are happily exiting with exit code = 0 after the failure.\n\nReturning some non-zero error code at the end of the task will be useful for us to monitor issues in production and react faster for fixing them.",
"reporter": "gmerino",
"cc": "david.schultz",
"resolution": "fixed",
"time": "2016-10-04T07:51:37",
"component": "iceprod",
"summary": "[iceprod] propagate the error code for failing IceProd tasks to pass it correctly to the batch system",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "juancarlos",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[iceprod] propagate the error code for failing IceProd tasks to pass it correctly to the batch system (Trac #1882) - It seems that some IceProd tasks that fail in some of the final steps like contacting the DB, or staging out the output data, or other, are not reporting this error with a non-zero error code at the end of the process. As of now, they are happily exiting with exit code = 0 after the failure.
Returning some non-zero error code at the end of the task will be useful for us to monitor issues in production and react faster for fixing them.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1882">https://code.icecube.wisc.edu/projects/icecube/ticket/1882</a>, reported by gmerinoand owned by juancarlos</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-11-03T21:01:05",
"_ts": "1478206865363965",
"description": "It seems that some IceProd tasks that fail in some of the final steps like contacting the DB, or staging out the output data, or other, are not reporting this error with a non-zero error code at the end of the process. As of now, they are happily exiting with exit code = 0 after the failure.\n\nReturning some non-zero error code at the end of the task will be useful for us to monitor issues in production and react faster for fixing them.",
"reporter": "gmerino",
"cc": "david.schultz",
"resolution": "fixed",
"time": "2016-10-04T07:51:37",
"component": "iceprod",
"summary": "[iceprod] propagate the error code for failing IceProd tasks to pass it correctly to the batch system",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "juancarlos",
"type": "defect"
}
```
</p>
</details>
|
defect
|
propagate the error code for failing iceprod tasks to pass it correctly to the batch system trac it seems that some iceprod tasks that fail in some of the final steps like contacting the db or staging out the output data or other are not reporting this error with a non zero error code at the end of the process as of now they are happily exiting with exit code after the failure returning some non zero error code at the end of the task will be useful for us to monitor issues in production and react faster for fixing them migrated from json status closed changetime ts description it seems that some iceprod tasks that fail in some of the final steps like contacting the db or staging out the output data or other are not reporting this error with a non zero error code at the end of the process as of now they are happily exiting with exit code after the failure n nreturning some non zero error code at the end of the task will be useful for us to monitor issues in production and react faster for fixing them reporter gmerino cc david schultz resolution fixed time component iceprod summary propagate the error code for failing iceprod tasks to pass it correctly to the batch system priority normal keywords milestone owner juancarlos type defect
| 1
|
9,724
| 3,963,992,951
|
IssuesEvent
|
2016-05-02 22:32:10
|
teotidev/guide
|
https://api.github.com/repos/teotidev/guide
|
closed
|
Export - as Sounds
|
code work
|
- checkbox list holding machines
- TextInput for prefix on file names ([TemplateName] for default)
- TextInput for postfix on file names ("" for default)
- screen footer action items
- Select All, Deselect All, Export
Figure out away to let the user edit individual machine names as well.
Plus, there is going to have to be a metadata editor that can edit the individual machine metadata to allow for pro kits exported this way
|
1.0
|
Export - as Sounds - - checkbox list holding machines
- TextInput for prefix on file names ([TemplateName] for default)
- TextInput for postfix on file names ("" for default)
- screen footer action items
- Select All, Deselect All, Export
Figure out away to let the user edit individual machine names as well.
Plus, there is going to have to be a metadata editor that can edit the individual machine metadata to allow for pro kits exported this way
|
non_defect
|
export as sounds checkbox list holding machines textinput for prefix on file names for default textinput for postfix on file names for default screen footer action items select all deselect all export figure out away to let the user edit individual machine names as well plus there is going to have to be a metadata editor that can edit the individual machine metadata to allow for pro kits exported this way
| 0
|
178,388
| 13,777,439,025
|
IssuesEvent
|
2020-10-08 10:56:45
|
wearerequired/traduttore
|
https://api.github.com/repos/wearerequired/traduttore
|
closed
|
Use PHPStan for Static Analysis
|
[Component] Tests [Type] Enhancement
|
<!--
BEFORE POSTING YOUR ISSUE:
- These comments won't show up when you submit the issue.
- Try to add as much detail as possible. Be specific!
- If you're requesting a new feature, explain why you'd like it to be added.
- Search this repository for issues and pull requests and whether it has been fixed or reported already.
- Ensure you are using the latest code before logging bugs.
-->
**Issue Overview**
Something like https://packagist.org/packages/szepeviktor/phpstan-wordpress or even just PHPStan directly with custom stubs will be very useful for finding issues in the codebase using static analysis.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Versions**
* Browser:
* PHP:
* WP-CLI:
* GlotPress:
* WordPress:
**Additional context**
<!-- List related issues or PRs against other branches -->
|
1.0
|
Use PHPStan for Static Analysis - <!--
BEFORE POSTING YOUR ISSUE:
- These comments won't show up when you submit the issue.
- Try to add as much detail as possible. Be specific!
- If you're requesting a new feature, explain why you'd like it to be added.
- Search this repository for issues and pull requests and whether it has been fixed or reported already.
- Ensure you are using the latest code before logging bugs.
-->
**Issue Overview**
Something like https://packagist.org/packages/szepeviktor/phpstan-wordpress or even just PHPStan directly with custom stubs will be very useful for finding issues in the codebase using static analysis.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Versions**
* Browser:
* PHP:
* WP-CLI:
* GlotPress:
* WordPress:
**Additional context**
<!-- List related issues or PRs against other branches -->
|
non_defect
|
use phpstan for static analysis before posting your issue these comments won t show up when you submit the issue try to add as much detail as possible be specific if you re requesting a new feature explain why you d like it to be added search this repository for issues and pull requests and whether it has been fixed or reported already ensure you are using the latest code before logging bugs issue overview something like or even just phpstan directly with custom stubs will be very useful for finding issues in the codebase using static analysis to reproduce steps to reproduce the behavior go to click on scroll down to see error expected behavior a clear and concise description of what you expected to happen screenshots if applicable add screenshots to help explain your problem versions browser php wp cli glotpress wordpress additional context
| 0
|
63,852
| 18,018,333,524
|
IssuesEvent
|
2021-09-16 16:09:52
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
opened
|
Not able to install requirements.txt file getting the below error
|
defect
|
### Describe your issue.
Building wheels for collected packages: scipy
Building wheel for scipy (setup.py) ... error
ERROR: Command errored out with exit status 1:
command: /usr/bin/python3 -u -c 'import io, os, sys, setuptools, tokenize; sys.argv[0] = '"'"'/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py'"'"'; __file__='"'"'/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py'"'"';f = getattr(tokenize, '"'"'open'"'"', open)(__file__) if os.path.exists(__file__) else io.StringIO('"'"'from setuptools import setup; setup()'"'"');code = f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' bdist_wheel -d /tmp/pip-wheel-r6jd3zfp
cwd: /tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/
Complete output (161 lines):
/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py:114: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
import imp
lapack_opt_info:
lapack_mkl_info:
customize UnixCCompiler
libraries mkl_rt not found in ['/usr/local/lib', '/usr/lib64', '/usr/lib', '/usr/lib/x86_64-linux-gnu']
NOT AVAILABLE
openblas_lapack_info:
customize UnixCCompiler
customize UnixCCompiler
libraries openblas not found in ['/usr/local/lib', '/usr/lib64', '/usr/lib', '/usr/lib/x86_64-linux-gnu']
NOT AVAILABLE
openblas_clapack_info:
customize UnixCCompiler
customize UnixCCompiler
libraries openblas,lapack not found in ['/usr/local/lib', '/usr/lib64', '/usr/lib', '/usr/lib/x86_64-linux-gnu']
NOT AVAILABLE
atlas_3_10_threads_info:
Setting PTATLAS=ATLAS
customize UnixCCompiler
libraries lapack_atlas not found in /usr/local/lib
customize UnixCCompiler
libraries tatlas,tatlas not found in /usr/local/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib64
customize UnixCCompiler
libraries tatlas,tatlas not found in /usr/lib64
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib
customize UnixCCompiler
libraries tatlas,tatlas not found in /usr/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib/x86_64-linux-gnu
customize UnixCCompiler
libraries tatlas,tatlas not found in /usr/lib/x86_64-linux-gnu
<class 'numpy.distutils.system_info.atlas_3_10_threads_info'>
NOT AVAILABLE
atlas_3_10_info:
customize UnixCCompiler
libraries lapack_atlas not found in /usr/local/lib
customize UnixCCompiler
libraries satlas,satlas not found in /usr/local/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib64
customize UnixCCompiler
libraries satlas,satlas not found in /usr/lib64
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib
customize UnixCCompiler
libraries satlas,satlas not found in /usr/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib/x86_64-linux-gnu
customize UnixCCompiler
libraries satlas,satlas not found in /usr/lib/x86_64-linux-gnu
<class 'numpy.distutils.system_info.atlas_3_10_info'>
NOT AVAILABLE
atlas_threads_info:
Setting PTATLAS=ATLAS
customize UnixCCompiler
libraries lapack_atlas not found in /usr/local/lib
customize UnixCCompiler
libraries ptf77blas,ptcblas,atlas not found in /usr/local/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib64
customize UnixCCompiler
libraries ptf77blas,ptcblas,atlas not found in /usr/lib64
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib
customize UnixCCompiler
libraries ptf77blas,ptcblas,atlas not found in /usr/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib/x86_64-linux-gnu
customize UnixCCompiler
libraries ptf77blas,ptcblas,atlas not found in /usr/lib/x86_64-linux-gnu
<class 'numpy.distutils.system_info.atlas_threads_info'>
NOT AVAILABLE
atlas_info:
customize UnixCCompiler
libraries lapack_atlas not found in /usr/local/lib
customize UnixCCompiler
libraries f77blas,cblas,atlas not found in /usr/local/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib64
customize UnixCCompiler
libraries f77blas,cblas,atlas not found in /usr/lib64
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib
customize UnixCCompiler
libraries f77blas,cblas,atlas not found in /usr/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib/x86_64-linux-gnu
customize UnixCCompiler
libraries f77blas,cblas,atlas not found in /usr/lib/x86_64-linux-gnu
<class 'numpy.distutils.system_info.atlas_info'>
NOT AVAILABLE
accelerate_info:
NOT AVAILABLE
/usr/local/lib/python3.8/dist-packages/numpy/distutils/system_info.py:639: UserWarning:
Atlas (http://math-atlas.sourceforge.net/) libraries not found.
Directories to search for the libraries can be specified in the
numpy/distutils/site.cfg file (section [atlas]) or by setting
the ATLAS environment variable.
self.calc_info()
lapack_info:
customize UnixCCompiler
libraries lapack not found in ['/usr/local/lib', '/usr/lib64', '/usr/lib', '/usr/lib/x86_64-linux-gnu']
NOT AVAILABLE
/usr/local/lib/python3.8/dist-packages/numpy/distutils/system_info.py:639: UserWarning:
Lapack (http://www.netlib.org/lapack/) libraries not found.
Directories to search for the libraries can be specified in the
numpy/distutils/site.cfg file (section [lapack]) or by setting
the LAPACK environment variable.
self.calc_info()
lapack_src_info:
NOT AVAILABLE
/usr/local/lib/python3.8/dist-packages/numpy/distutils/system_info.py:639: UserWarning:
Lapack (http://www.netlib.org/lapack/) sources not found.
Directories to search for the sources can be specified in the
numpy/distutils/site.cfg file (section [lapack_src]) or by setting
the LAPACK_SRC environment variable.
self.calc_info()
NOT AVAILABLE
Running from scipy source directory.
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py", line 474, in <module>
setup_package()
File "/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py", line 470, in setup_package
setup(**metadata)
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/core.py", line 137, in setup
config = configuration()
File "/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py", line 388, in configuration
config.add_subpackage('scipy')
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 1034, in add_subpackage
config_list = self.get_subpackage(subpackage_name, subpackage_path,
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 1000, in get_subpackage
config = self._get_configuration_from_setup_py(
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 942, in _get_configuration_from_setup_py
config = setup_module.configuration(*args)
File "scipy/setup.py", line 15, in configuration
config.add_subpackage('linalg')
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 1034, in add_subpackage
config_list = self.get_subpackage(subpackage_name, subpackage_path,
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 1000, in get_subpackage
config = self._get_configuration_from_setup_py(
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 942, in _get_configuration_from_setup_py
config = setup_module.configuration(*args)
File "scipy/linalg/setup.py", line 19, in configuration
raise NotFoundError('no lapack/blas resources found')
numpy.distutils.system_info.NotFoundError: no lapack/blas resources found
----------------------------------------
ERROR: Failed building wheel for scipy
Running setup.py clean for scipy
ERROR: Command errored out with exit status 1:
command: /usr/bin/python3 -u -c 'import io, os, sys, setuptools, tokenize; sys.argv[0] = '"'"'/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py'"'"'; __file__='"'"'/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py'"'"';f = getattr(tokenize, '"'"'open'"'"', open)(__file__) if os.path.exists(__file__) else io.StringIO('"'"'from setuptools import setup; setup()'"'"');code = f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' clean --all
cwd: /tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec
Complete output (11 lines):
/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py:114: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
import imp
`setup.py clean` is not supported, use one of the following instead:
- `git clean -xdf` (cleans all files)
- `git clean -Xdf` (cleans all versioned files, doesn't touch
files that aren't checked into the git repo)
Add `--force` to your command to use it anyway if you must (unsupported).
----------------------------------------
ERROR: Failed cleaning build dir for scipy
Failed to build scipy
Installing collected packages: scipy, python-dateutil, configparser, xgboost, pandas, minio, gensim
Attempting uninstall: scipy
Found existing installation: scipy 1.7.1
Uninstalling scipy-1.7.1:
Successfully uninstalled scipy-1.7.1
Running setup.py install for scipy ... error
ERROR: Command errored out with exit status 1:
command: /usr/bin/python3 -u -c 'import io, os, sys, setuptools, tokenize; sys.argv[0] = '"'"'/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py'"'"'; __file__='"'"'/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py'"'"';f = getattr(tokenize, '"'"'open'"'"', open)(__file__) if os.path.exists(__file__) else io.StringIO('"'"'from setuptools import setup; setup()'"'"');code = f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' install --record /tmp/pip-record-1rn3l91t/install-record.txt --single-version-externally-managed--compile --install-headers /usr/local/include/python3.8/scipy
cwd: /tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/
Complete output (170 lines):
/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py:114: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
import imp
Note: if you need reliable uninstall behavior, then install
with pip instead of using `setup.py install`:
- `pip install .` (from a git repo or downloaded source
release)
- `pip install scipy` (last SciPy release on PyPI)
lapack_opt_info:
lapack_mkl_info:
customize UnixCCompiler
libraries mkl_rt not found in ['/usr/local/lib', '/usr/lib64', '/usr/lib', '/usr/lib/x86_64-linux-gnu']
NOT AVAILABLE
openblas_lapack_info:
customize UnixCCompiler
customize UnixCCompiler
libraries openblas not found in ['/usr/local/lib', '/usr/lib64', '/usr/lib', '/usr/lib/x86_64-linux-gnu']
NOT AVAILABLE
openblas_clapack_info:
customize UnixCCompiler
customize UnixCCompiler
libraries openblas,lapack not found in ['/usr/local/lib', '/usr/lib64', '/usr/lib', '/usr/lib/x86_64-linux-gnu']
NOT AVAILABLE
atlas_3_10_threads_info:
Setting PTATLAS=ATLAS
customize UnixCCompiler
libraries lapack_atlas not found in /usr/local/lib
customize UnixCCompiler
libraries tatlas,tatlas not found in /usr/local/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib64
customize UnixCCompiler
libraries tatlas,tatlas not found in /usr/lib64
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib
customize UnixCCompiler
libraries tatlas,tatlas not found in /usr/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib/x86_64-linux-gnu
customize UnixCCompiler
libraries tatlas,tatlas not found in /usr/lib/x86_64-linux-gnu
<class 'numpy.distutils.system_info.atlas_3_10_threads_info'>
NOT AVAILABLE
atlas_3_10_info:
customize UnixCCompiler
libraries lapack_atlas not found in /usr/local/lib
customize UnixCCompiler
libraries satlas,satlas not found in /usr/local/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib64
customize UnixCCompiler
libraries satlas,satlas not found in /usr/lib64
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib
customize UnixCCompiler
libraries satlas,satlas not found in /usr/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib/x86_64-linux-gnu
customize UnixCCompiler
libraries satlas,satlas not found in /usr/lib/x86_64-linux-gnu
<class 'numpy.distutils.system_info.atlas_3_10_info'>
NOT AVAILABLE
atlas_threads_info:
Setting PTATLAS=ATLAS
customize UnixCCompiler
libraries lapack_atlas not found in /usr/local/lib
customize UnixCCompiler
libraries ptf77blas,ptcblas,atlas not found in /usr/local/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib64
customize UnixCCompiler
libraries ptf77blas,ptcblas,atlas not found in /usr/lib64
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib
customize UnixCCompiler
libraries ptf77blas,ptcblas,atlas not found in /usr/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib/x86_64-linux-gnu
customize UnixCCompiler
libraries ptf77blas,ptcblas,atlas not found in /usr/lib/x86_64-linux-gnu
<class 'numpy.distutils.system_info.atlas_threads_info'>
NOT AVAILABLE
atlas_info:
customize UnixCCompiler
libraries lapack_atlas not found in /usr/local/lib
customize UnixCCompiler
libraries f77blas,cblas,atlas not found in /usr/local/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib64
customize UnixCCompiler
libraries f77blas,cblas,atlas not found in /usr/lib64
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib
customize UnixCCompiler
libraries f77blas,cblas,atlas not found in /usr/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib/x86_64-linux-gnu
customize UnixCCompiler
libraries f77blas,cblas,atlas not found in /usr/lib/x86_64-linux-gnu
<class 'numpy.distutils.system_info.atlas_info'>
NOT AVAILABLE
accelerate_info:
NOT AVAILABLE
/usr/local/lib/python3.8/dist-packages/numpy/distutils/system_info.py:639: UserWarning:
Atlas (http://math-atlas.sourceforge.net/) libraries not found.
Directories to search for the libraries can be specified in the
numpy/distutils/site.cfg file (section [atlas]) or by setting
the ATLAS environment variable.
self.calc_info()
lapack_info:
customize UnixCCompiler
libraries lapack not found in ['/usr/local/lib', '/usr/lib64', '/usr/lib', '/usr/lib/x86_64-linux-gnu']
NOT AVAILABLE
/usr/local/lib/python3.8/dist-packages/numpy/distutils/system_info.py:639: UserWarning:
Lapack (http://www.netlib.org/lapack/) libraries not found.
Directories to search for the libraries can be specified in the
numpy/distutils/site.cfg file (section [lapack]) or by setting
the LAPACK environment variable.
self.calc_info()
lapack_src_info:
NOT AVAILABLE
/usr/local/lib/python3.8/dist-packages/numpy/distutils/system_info.py:639: UserWarning:
Lapack (http://www.netlib.org/lapack/) sources not found.
Directories to search for the sources can be specified in the
numpy/distutils/site.cfg file (section [lapack_src]) or by setting
the LAPACK_SRC environment variable.
self.calc_info()
NOT AVAILABLE
Running from scipy source directory.
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py", line 474, in <module>
setup_package()
File "/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py", line 470, in setup_package
setup(**metadata)
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/core.py", line 137, in setup
config = configuration()
File "/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py", line 388, in configuration
config.add_subpackage('scipy')
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 1034, in add_subpackage
config_list = self.get_subpackage(subpackage_name, subpackage_path,
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 1000, in get_subpackage
config = self._get_configuration_from_setup_py(
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 942, in _get_configuration_from_setup_py
config = setup_module.configuration(*args)
File "scipy/setup.py", line 15, in configuration
config.add_subpackage('linalg')
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 1034, in add_subpackage
config_list = self.get_subpackage(subpackage_name, subpackage_path,
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 1000, in get_subpackage
config = self._get_configuration_from_setup_py(
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 942, in _get_configuration_from_setup_py
config = setup_module.configuration(*args)
File "scipy/linalg/setup.py", line 19, in configuration
raise NotFoundError('no lapack/blas resources found')
numpy.distutils.system_info.NotFoundError: no lapack/blas resources found
----------------------------------------
Rolling back uninstall of scipy
Moving to /usr/local/lib/python3.8/dist-packages/scipy-1.7.1.dist-info/
from /usr/local/lib/python3.8/dist-packages/~cipy-1.7.1.dist-info
Moving to /usr/local/lib/python3.8/dist-packages/scipy.libs/
from /usr/local/lib/python3.8/dist-packages/~cipy.libs
Moving to /usr/local/lib/python3.8/dist-packages/scipy/
from /usr/local/lib/python3.8/dist-packages/~cipy
ERROR: Command errored out with exit status 1: /usr/bin/python3 -u -c 'import io, os, sys, setuptools, tokenize; sys.argv[0] = '"'"'/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py'"'"'; __file__='"'"'/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py'"'"';f = getattr(tokenize, '"'"'open'"'"', open)(__file__) if os.path.exists(__file__) else io.StringIO('"'"'from setuptools import setup; setup()'"'"');code = f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' install --record /tmp/pip-record-1rn3l91t/install-record.txt --single-version-externally-managed --compile --install-headers /usr/local/include/python3.8/scipy Check the logs for full command output.
### Reproducing Code Example
```python
Requirements.txt file
flask==1.1.1
flask-cors==3.0.8
waitress==1.4.3
uWSGI==2.0.18
pika==1.0.0
pydantic==1.1.1
elasticsearch==7.0.0
requests==2.22.0
httpretty==0.9.7
sure==1.4.11
bump2version==0.5.11
flake8==3.7.9
nltk==3.4.5
scikit-learn==0.19.2
numpy==1.16.4
scipy==1.1.0
xgboost==0.90
python-dateutil==2.8.1
minio==6.0.0
gensim==3.5.0
pandas==1.1.4
```
### Error message
```shell
sudo pip3 install -r requirements.txt
```
### SciPy/NumPy/Python version information
python 3.8
|
1.0
|
Not able to install requirements.txt file getting the below error - ### Describe your issue.
Building wheels for collected packages: scipy
Building wheel for scipy (setup.py) ... error
ERROR: Command errored out with exit status 1:
command: /usr/bin/python3 -u -c 'import io, os, sys, setuptools, tokenize; sys.argv[0] = '"'"'/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py'"'"'; __file__='"'"'/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py'"'"';f = getattr(tokenize, '"'"'open'"'"', open)(__file__) if os.path.exists(__file__) else io.StringIO('"'"'from setuptools import setup; setup()'"'"');code = f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' bdist_wheel -d /tmp/pip-wheel-r6jd3zfp
cwd: /tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/
Complete output (161 lines):
/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py:114: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
import imp
lapack_opt_info:
lapack_mkl_info:
customize UnixCCompiler
libraries mkl_rt not found in ['/usr/local/lib', '/usr/lib64', '/usr/lib', '/usr/lib/x86_64-linux-gnu']
NOT AVAILABLE
openblas_lapack_info:
customize UnixCCompiler
customize UnixCCompiler
libraries openblas not found in ['/usr/local/lib', '/usr/lib64', '/usr/lib', '/usr/lib/x86_64-linux-gnu']
NOT AVAILABLE
openblas_clapack_info:
customize UnixCCompiler
customize UnixCCompiler
libraries openblas,lapack not found in ['/usr/local/lib', '/usr/lib64', '/usr/lib', '/usr/lib/x86_64-linux-gnu']
NOT AVAILABLE
atlas_3_10_threads_info:
Setting PTATLAS=ATLAS
customize UnixCCompiler
libraries lapack_atlas not found in /usr/local/lib
customize UnixCCompiler
libraries tatlas,tatlas not found in /usr/local/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib64
customize UnixCCompiler
libraries tatlas,tatlas not found in /usr/lib64
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib
customize UnixCCompiler
libraries tatlas,tatlas not found in /usr/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib/x86_64-linux-gnu
customize UnixCCompiler
libraries tatlas,tatlas not found in /usr/lib/x86_64-linux-gnu
<class 'numpy.distutils.system_info.atlas_3_10_threads_info'>
NOT AVAILABLE
atlas_3_10_info:
customize UnixCCompiler
libraries lapack_atlas not found in /usr/local/lib
customize UnixCCompiler
libraries satlas,satlas not found in /usr/local/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib64
customize UnixCCompiler
libraries satlas,satlas not found in /usr/lib64
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib
customize UnixCCompiler
libraries satlas,satlas not found in /usr/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib/x86_64-linux-gnu
customize UnixCCompiler
libraries satlas,satlas not found in /usr/lib/x86_64-linux-gnu
<class 'numpy.distutils.system_info.atlas_3_10_info'>
NOT AVAILABLE
atlas_threads_info:
Setting PTATLAS=ATLAS
customize UnixCCompiler
libraries lapack_atlas not found in /usr/local/lib
customize UnixCCompiler
libraries ptf77blas,ptcblas,atlas not found in /usr/local/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib64
customize UnixCCompiler
libraries ptf77blas,ptcblas,atlas not found in /usr/lib64
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib
customize UnixCCompiler
libraries ptf77blas,ptcblas,atlas not found in /usr/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib/x86_64-linux-gnu
customize UnixCCompiler
libraries ptf77blas,ptcblas,atlas not found in /usr/lib/x86_64-linux-gnu
<class 'numpy.distutils.system_info.atlas_threads_info'>
NOT AVAILABLE
atlas_info:
customize UnixCCompiler
libraries lapack_atlas not found in /usr/local/lib
customize UnixCCompiler
libraries f77blas,cblas,atlas not found in /usr/local/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib64
customize UnixCCompiler
libraries f77blas,cblas,atlas not found in /usr/lib64
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib
customize UnixCCompiler
libraries f77blas,cblas,atlas not found in /usr/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib/x86_64-linux-gnu
customize UnixCCompiler
libraries f77blas,cblas,atlas not found in /usr/lib/x86_64-linux-gnu
<class 'numpy.distutils.system_info.atlas_info'>
NOT AVAILABLE
accelerate_info:
NOT AVAILABLE
/usr/local/lib/python3.8/dist-packages/numpy/distutils/system_info.py:639: UserWarning:
Atlas (http://math-atlas.sourceforge.net/) libraries not found.
Directories to search for the libraries can be specified in the
numpy/distutils/site.cfg file (section [atlas]) or by setting
the ATLAS environment variable.
self.calc_info()
lapack_info:
customize UnixCCompiler
libraries lapack not found in ['/usr/local/lib', '/usr/lib64', '/usr/lib', '/usr/lib/x86_64-linux-gnu']
NOT AVAILABLE
/usr/local/lib/python3.8/dist-packages/numpy/distutils/system_info.py:639: UserWarning:
Lapack (http://www.netlib.org/lapack/) libraries not found.
Directories to search for the libraries can be specified in the
numpy/distutils/site.cfg file (section [lapack]) or by setting
the LAPACK environment variable.
self.calc_info()
lapack_src_info:
NOT AVAILABLE
/usr/local/lib/python3.8/dist-packages/numpy/distutils/system_info.py:639: UserWarning:
Lapack (http://www.netlib.org/lapack/) sources not found.
Directories to search for the sources can be specified in the
numpy/distutils/site.cfg file (section [lapack_src]) or by setting
the LAPACK_SRC environment variable.
self.calc_info()
NOT AVAILABLE
Running from scipy source directory.
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py", line 474, in <module>
setup_package()
File "/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py", line 470, in setup_package
setup(**metadata)
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/core.py", line 137, in setup
config = configuration()
File "/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py", line 388, in configuration
config.add_subpackage('scipy')
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 1034, in add_subpackage
config_list = self.get_subpackage(subpackage_name, subpackage_path,
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 1000, in get_subpackage
config = self._get_configuration_from_setup_py(
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 942, in _get_configuration_from_setup_py
config = setup_module.configuration(*args)
File "scipy/setup.py", line 15, in configuration
config.add_subpackage('linalg')
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 1034, in add_subpackage
config_list = self.get_subpackage(subpackage_name, subpackage_path,
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 1000, in get_subpackage
config = self._get_configuration_from_setup_py(
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 942, in _get_configuration_from_setup_py
config = setup_module.configuration(*args)
File "scipy/linalg/setup.py", line 19, in configuration
raise NotFoundError('no lapack/blas resources found')
numpy.distutils.system_info.NotFoundError: no lapack/blas resources found
----------------------------------------
ERROR: Failed building wheel for scipy
Running setup.py clean for scipy
ERROR: Command errored out with exit status 1:
command: /usr/bin/python3 -u -c 'import io, os, sys, setuptools, tokenize; sys.argv[0] = '"'"'/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py'"'"'; __file__='"'"'/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py'"'"';f = getattr(tokenize, '"'"'open'"'"', open)(__file__) if os.path.exists(__file__) else io.StringIO('"'"'from setuptools import setup; setup()'"'"');code = f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' clean --all
cwd: /tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec
Complete output (11 lines):
/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py:114: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
import imp
`setup.py clean` is not supported, use one of the following instead:
- `git clean -xdf` (cleans all files)
- `git clean -Xdf` (cleans all versioned files, doesn't touch
files that aren't checked into the git repo)
Add `--force` to your command to use it anyway if you must (unsupported).
----------------------------------------
ERROR: Failed cleaning build dir for scipy
Failed to build scipy
Installing collected packages: scipy, python-dateutil, configparser, xgboost, pandas, minio, gensim
Attempting uninstall: scipy
Found existing installation: scipy 1.7.1
Uninstalling scipy-1.7.1:
Successfully uninstalled scipy-1.7.1
Running setup.py install for scipy ... error
ERROR: Command errored out with exit status 1:
command: /usr/bin/python3 -u -c 'import io, os, sys, setuptools, tokenize; sys.argv[0] = '"'"'/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py'"'"'; __file__='"'"'/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py'"'"';f = getattr(tokenize, '"'"'open'"'"', open)(__file__) if os.path.exists(__file__) else io.StringIO('"'"'from setuptools import setup; setup()'"'"');code = f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' install --record /tmp/pip-record-1rn3l91t/install-record.txt --single-version-externally-managed--compile --install-headers /usr/local/include/python3.8/scipy
cwd: /tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/
Complete output (170 lines):
/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py:114: DeprecationWarning: the imp module is deprecated in favour of importlib; see the module's documentation for alternative uses
import imp
Note: if you need reliable uninstall behavior, then install
with pip instead of using `setup.py install`:
- `pip install .` (from a git repo or downloaded source
release)
- `pip install scipy` (last SciPy release on PyPI)
lapack_opt_info:
lapack_mkl_info:
customize UnixCCompiler
libraries mkl_rt not found in ['/usr/local/lib', '/usr/lib64', '/usr/lib', '/usr/lib/x86_64-linux-gnu']
NOT AVAILABLE
openblas_lapack_info:
customize UnixCCompiler
customize UnixCCompiler
libraries openblas not found in ['/usr/local/lib', '/usr/lib64', '/usr/lib', '/usr/lib/x86_64-linux-gnu']
NOT AVAILABLE
openblas_clapack_info:
customize UnixCCompiler
customize UnixCCompiler
libraries openblas,lapack not found in ['/usr/local/lib', '/usr/lib64', '/usr/lib', '/usr/lib/x86_64-linux-gnu']
NOT AVAILABLE
atlas_3_10_threads_info:
Setting PTATLAS=ATLAS
customize UnixCCompiler
libraries lapack_atlas not found in /usr/local/lib
customize UnixCCompiler
libraries tatlas,tatlas not found in /usr/local/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib64
customize UnixCCompiler
libraries tatlas,tatlas not found in /usr/lib64
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib
customize UnixCCompiler
libraries tatlas,tatlas not found in /usr/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib/x86_64-linux-gnu
customize UnixCCompiler
libraries tatlas,tatlas not found in /usr/lib/x86_64-linux-gnu
<class 'numpy.distutils.system_info.atlas_3_10_threads_info'>
NOT AVAILABLE
atlas_3_10_info:
customize UnixCCompiler
libraries lapack_atlas not found in /usr/local/lib
customize UnixCCompiler
libraries satlas,satlas not found in /usr/local/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib64
customize UnixCCompiler
libraries satlas,satlas not found in /usr/lib64
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib
customize UnixCCompiler
libraries satlas,satlas not found in /usr/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib/x86_64-linux-gnu
customize UnixCCompiler
libraries satlas,satlas not found in /usr/lib/x86_64-linux-gnu
<class 'numpy.distutils.system_info.atlas_3_10_info'>
NOT AVAILABLE
atlas_threads_info:
Setting PTATLAS=ATLAS
customize UnixCCompiler
libraries lapack_atlas not found in /usr/local/lib
customize UnixCCompiler
libraries ptf77blas,ptcblas,atlas not found in /usr/local/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib64
customize UnixCCompiler
libraries ptf77blas,ptcblas,atlas not found in /usr/lib64
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib
customize UnixCCompiler
libraries ptf77blas,ptcblas,atlas not found in /usr/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib/x86_64-linux-gnu
customize UnixCCompiler
libraries ptf77blas,ptcblas,atlas not found in /usr/lib/x86_64-linux-gnu
<class 'numpy.distutils.system_info.atlas_threads_info'>
NOT AVAILABLE
atlas_info:
customize UnixCCompiler
libraries lapack_atlas not found in /usr/local/lib
customize UnixCCompiler
libraries f77blas,cblas,atlas not found in /usr/local/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib64
customize UnixCCompiler
libraries f77blas,cblas,atlas not found in /usr/lib64
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib
customize UnixCCompiler
libraries f77blas,cblas,atlas not found in /usr/lib
customize UnixCCompiler
libraries lapack_atlas not found in /usr/lib/x86_64-linux-gnu
customize UnixCCompiler
libraries f77blas,cblas,atlas not found in /usr/lib/x86_64-linux-gnu
<class 'numpy.distutils.system_info.atlas_info'>
NOT AVAILABLE
accelerate_info:
NOT AVAILABLE
/usr/local/lib/python3.8/dist-packages/numpy/distutils/system_info.py:639: UserWarning:
Atlas (http://math-atlas.sourceforge.net/) libraries not found.
Directories to search for the libraries can be specified in the
numpy/distutils/site.cfg file (section [atlas]) or by setting
the ATLAS environment variable.
self.calc_info()
lapack_info:
customize UnixCCompiler
libraries lapack not found in ['/usr/local/lib', '/usr/lib64', '/usr/lib', '/usr/lib/x86_64-linux-gnu']
NOT AVAILABLE
/usr/local/lib/python3.8/dist-packages/numpy/distutils/system_info.py:639: UserWarning:
Lapack (http://www.netlib.org/lapack/) libraries not found.
Directories to search for the libraries can be specified in the
numpy/distutils/site.cfg file (section [lapack]) or by setting
the LAPACK environment variable.
self.calc_info()
lapack_src_info:
NOT AVAILABLE
/usr/local/lib/python3.8/dist-packages/numpy/distutils/system_info.py:639: UserWarning:
Lapack (http://www.netlib.org/lapack/) sources not found.
Directories to search for the sources can be specified in the
numpy/distutils/site.cfg file (section [lapack_src]) or by setting
the LAPACK_SRC environment variable.
self.calc_info()
NOT AVAILABLE
Running from scipy source directory.
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py", line 474, in <module>
setup_package()
File "/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py", line 470, in setup_package
setup(**metadata)
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/core.py", line 137, in setup
config = configuration()
File "/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py", line 388, in configuration
config.add_subpackage('scipy')
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 1034, in add_subpackage
config_list = self.get_subpackage(subpackage_name, subpackage_path,
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 1000, in get_subpackage
config = self._get_configuration_from_setup_py(
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 942, in _get_configuration_from_setup_py
config = setup_module.configuration(*args)
File "scipy/setup.py", line 15, in configuration
config.add_subpackage('linalg')
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 1034, in add_subpackage
config_list = self.get_subpackage(subpackage_name, subpackage_path,
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 1000, in get_subpackage
config = self._get_configuration_from_setup_py(
File "/usr/local/lib/python3.8/dist-packages/numpy/distutils/misc_util.py", line 942, in _get_configuration_from_setup_py
config = setup_module.configuration(*args)
File "scipy/linalg/setup.py", line 19, in configuration
raise NotFoundError('no lapack/blas resources found')
numpy.distutils.system_info.NotFoundError: no lapack/blas resources found
----------------------------------------
Rolling back uninstall of scipy
Moving to /usr/local/lib/python3.8/dist-packages/scipy-1.7.1.dist-info/
from /usr/local/lib/python3.8/dist-packages/~cipy-1.7.1.dist-info
Moving to /usr/local/lib/python3.8/dist-packages/scipy.libs/
from /usr/local/lib/python3.8/dist-packages/~cipy.libs
Moving to /usr/local/lib/python3.8/dist-packages/scipy/
from /usr/local/lib/python3.8/dist-packages/~cipy
ERROR: Command errored out with exit status 1: /usr/bin/python3 -u -c 'import io, os, sys, setuptools, tokenize; sys.argv[0] = '"'"'/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py'"'"'; __file__='"'"'/tmp/pip-install-b_ad1fui/scipy_bebacd68c6e14294901683aca68151ec/setup.py'"'"';f = getattr(tokenize, '"'"'open'"'"', open)(__file__) if os.path.exists(__file__) else io.StringIO('"'"'from setuptools import setup; setup()'"'"');code = f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' install --record /tmp/pip-record-1rn3l91t/install-record.txt --single-version-externally-managed --compile --install-headers /usr/local/include/python3.8/scipy Check the logs for full command output.
### Reproducing Code Example
```python
Requirements.txt file
flask==1.1.1
flask-cors==3.0.8
waitress==1.4.3
uWSGI==2.0.18
pika==1.0.0
pydantic==1.1.1
elasticsearch==7.0.0
requests==2.22.0
httpretty==0.9.7
sure==1.4.11
bump2version==0.5.11
flake8==3.7.9
nltk==3.4.5
scikit-learn==0.19.2
numpy==1.16.4
scipy==1.1.0
xgboost==0.90
python-dateutil==2.8.1
minio==6.0.0
gensim==3.5.0
pandas==1.1.4
```
### Error message
```shell
sudo pip3 install -r requirements.txt
```
### SciPy/NumPy/Python version information
python 3.8
|
defect
|
not able to install requirements txt file getting the below error describe your issue building wheels for collected packages scipy building wheel for scipy setup py error error command errored out with exit status command usr bin u c import io os sys setuptools tokenize sys argv tmp pip install b scipy setup py file tmp pip install b scipy setup py f getattr tokenize open open file if os path exists file else io stringio from setuptools import setup setup code f read replace r n n f close exec compile code file exec bdist wheel d tmp pip wheel cwd tmp pip install b scipy complete output lines tmp pip install b scipy setup py deprecationwarning the imp module is deprecated in favour of importlib see the module s documentation for alternative uses import imp lapack opt info lapack mkl info customize unixccompiler libraries mkl rt not found in not available openblas lapack info customize unixccompiler customize unixccompiler libraries openblas not found in not available openblas clapack info customize unixccompiler customize unixccompiler libraries openblas lapack not found in not available atlas threads info setting ptatlas atlas customize unixccompiler libraries lapack atlas not found in usr local lib customize unixccompiler libraries tatlas tatlas not found in usr local lib customize unixccompiler libraries lapack atlas not found in usr customize unixccompiler libraries tatlas tatlas not found in usr customize unixccompiler libraries lapack atlas not found in usr lib customize unixccompiler libraries tatlas tatlas not found in usr lib customize unixccompiler libraries lapack atlas not found in usr lib linux gnu customize unixccompiler libraries tatlas tatlas not found in usr lib linux gnu not available atlas info customize unixccompiler libraries lapack atlas not found in usr local lib customize unixccompiler libraries satlas satlas not found in usr local lib customize unixccompiler libraries lapack atlas not found in usr customize unixccompiler libraries satlas satlas not found in usr customize unixccompiler libraries lapack atlas not found in usr lib customize unixccompiler libraries satlas satlas not found in usr lib customize unixccompiler libraries lapack atlas not found in usr lib linux gnu customize unixccompiler libraries satlas satlas not found in usr lib linux gnu not available atlas threads info setting ptatlas atlas customize unixccompiler libraries lapack atlas not found in usr local lib customize unixccompiler libraries ptcblas atlas not found in usr local lib customize unixccompiler libraries lapack atlas not found in usr customize unixccompiler libraries ptcblas atlas not found in usr customize unixccompiler libraries lapack atlas not found in usr lib customize unixccompiler libraries ptcblas atlas not found in usr lib customize unixccompiler libraries lapack atlas not found in usr lib linux gnu customize unixccompiler libraries ptcblas atlas not found in usr lib linux gnu not available atlas info customize unixccompiler libraries lapack atlas not found in usr local lib customize unixccompiler libraries cblas atlas not found in usr local lib customize unixccompiler libraries lapack atlas not found in usr customize unixccompiler libraries cblas atlas not found in usr customize unixccompiler libraries lapack atlas not found in usr lib customize unixccompiler libraries cblas atlas not found in usr lib customize unixccompiler libraries lapack atlas not found in usr lib linux gnu customize unixccompiler libraries cblas atlas not found in usr lib linux gnu not available accelerate info not available usr local lib dist packages numpy distutils system info py userwarning atlas libraries not found directories to search for the libraries can be specified in the numpy distutils site cfg file section or by setting the atlas environment variable self calc info lapack info customize unixccompiler libraries lapack not found in not available usr local lib dist packages numpy distutils system info py userwarning lapack libraries not found directories to search for the libraries can be specified in the numpy distutils site cfg file section or by setting the lapack environment variable self calc info lapack src info not available usr local lib dist packages numpy distutils system info py userwarning lapack sources not found directories to search for the sources can be specified in the numpy distutils site cfg file section or by setting the lapack src environment variable self calc info not available running from scipy source directory traceback most recent call last file line in file tmp pip install b scipy setup py line in setup package file tmp pip install b scipy setup py line in setup package setup metadata file usr local lib dist packages numpy distutils core py line in setup config configuration file tmp pip install b scipy setup py line in configuration config add subpackage scipy file usr local lib dist packages numpy distutils misc util py line in add subpackage config list self get subpackage subpackage name subpackage path file usr local lib dist packages numpy distutils misc util py line in get subpackage config self get configuration from setup py file usr local lib dist packages numpy distutils misc util py line in get configuration from setup py config setup module configuration args file scipy setup py line in configuration config add subpackage linalg file usr local lib dist packages numpy distutils misc util py line in add subpackage config list self get subpackage subpackage name subpackage path file usr local lib dist packages numpy distutils misc util py line in get subpackage config self get configuration from setup py file usr local lib dist packages numpy distutils misc util py line in get configuration from setup py config setup module configuration args file scipy linalg setup py line in configuration raise notfounderror no lapack blas resources found numpy distutils system info notfounderror no lapack blas resources found error failed building wheel for scipy running setup py clean for scipy error command errored out with exit status command usr bin u c import io os sys setuptools tokenize sys argv tmp pip install b scipy setup py file tmp pip install b scipy setup py f getattr tokenize open open file if os path exists file else io stringio from setuptools import setup setup code f read replace r n n f close exec compile code file exec clean all cwd tmp pip install b scipy complete output lines tmp pip install b scipy setup py deprecationwarning the imp module is deprecated in favour of importlib see the module s documentation for alternative uses import imp setup py clean is not supported use one of the following instead git clean xdf cleans all files git clean xdf cleans all versioned files doesn t touch files that aren t checked into the git repo add force to your command to use it anyway if you must unsupported error failed cleaning build dir for scipy failed to build scipy installing collected packages scipy python dateutil configparser xgboost pandas minio gensim attempting uninstall scipy found existing installation scipy uninstalling scipy successfully uninstalled scipy running setup py install for scipy error error command errored out with exit status command usr bin u c import io os sys setuptools tokenize sys argv tmp pip install b scipy setup py file tmp pip install b scipy setup py f getattr tokenize open open file if os path exists file else io stringio from setuptools import setup setup code f read replace r n n f close exec compile code file exec install record tmp pip record install record txt single version externally managed compile install headers usr local include scipy cwd tmp pip install b scipy complete output lines tmp pip install b scipy setup py deprecationwarning the imp module is deprecated in favour of importlib see the module s documentation for alternative uses import imp note if you need reliable uninstall behavior then install with pip instead of using setup py install pip install from a git repo or downloaded source release pip install scipy last scipy release on pypi lapack opt info lapack mkl info customize unixccompiler libraries mkl rt not found in not available openblas lapack info customize unixccompiler customize unixccompiler libraries openblas not found in not available openblas clapack info customize unixccompiler customize unixccompiler libraries openblas lapack not found in not available atlas threads info setting ptatlas atlas customize unixccompiler libraries lapack atlas not found in usr local lib customize unixccompiler libraries tatlas tatlas not found in usr local lib customize unixccompiler libraries lapack atlas not found in usr customize unixccompiler libraries tatlas tatlas not found in usr customize unixccompiler libraries lapack atlas not found in usr lib customize unixccompiler libraries tatlas tatlas not found in usr lib customize unixccompiler libraries lapack atlas not found in usr lib linux gnu customize unixccompiler libraries tatlas tatlas not found in usr lib linux gnu not available atlas info customize unixccompiler libraries lapack atlas not found in usr local lib customize unixccompiler libraries satlas satlas not found in usr local lib customize unixccompiler libraries lapack atlas not found in usr customize unixccompiler libraries satlas satlas not found in usr customize unixccompiler libraries lapack atlas not found in usr lib customize unixccompiler libraries satlas satlas not found in usr lib customize unixccompiler libraries lapack atlas not found in usr lib linux gnu customize unixccompiler libraries satlas satlas not found in usr lib linux gnu not available atlas threads info setting ptatlas atlas customize unixccompiler libraries lapack atlas not found in usr local lib customize unixccompiler libraries ptcblas atlas not found in usr local lib customize unixccompiler libraries lapack atlas not found in usr customize unixccompiler libraries ptcblas atlas not found in usr customize unixccompiler libraries lapack atlas not found in usr lib customize unixccompiler libraries ptcblas atlas not found in usr lib customize unixccompiler libraries lapack atlas not found in usr lib linux gnu customize unixccompiler libraries ptcblas atlas not found in usr lib linux gnu not available atlas info customize unixccompiler libraries lapack atlas not found in usr local lib customize unixccompiler libraries cblas atlas not found in usr local lib customize unixccompiler libraries lapack atlas not found in usr customize unixccompiler libraries cblas atlas not found in usr customize unixccompiler libraries lapack atlas not found in usr lib customize unixccompiler libraries cblas atlas not found in usr lib customize unixccompiler libraries lapack atlas not found in usr lib linux gnu customize unixccompiler libraries cblas atlas not found in usr lib linux gnu not available accelerate info not available usr local lib dist packages numpy distutils system info py userwarning atlas libraries not found directories to search for the libraries can be specified in the numpy distutils site cfg file section or by setting the atlas environment variable self calc info lapack info customize unixccompiler libraries lapack not found in not available usr local lib dist packages numpy distutils system info py userwarning lapack libraries not found directories to search for the libraries can be specified in the numpy distutils site cfg file section or by setting the lapack environment variable self calc info lapack src info not available usr local lib dist packages numpy distutils system info py userwarning lapack sources not found directories to search for the sources can be specified in the numpy distutils site cfg file section or by setting the lapack src environment variable self calc info not available running from scipy source directory traceback most recent call last file line in file tmp pip install b scipy setup py line in setup package file tmp pip install b scipy setup py line in setup package setup metadata file usr local lib dist packages numpy distutils core py line in setup config configuration file tmp pip install b scipy setup py line in configuration config add subpackage scipy file usr local lib dist packages numpy distutils misc util py line in add subpackage config list self get subpackage subpackage name subpackage path file usr local lib dist packages numpy distutils misc util py line in get subpackage config self get configuration from setup py file usr local lib dist packages numpy distutils misc util py line in get configuration from setup py config setup module configuration args file scipy setup py line in configuration config add subpackage linalg file usr local lib dist packages numpy distutils misc util py line in add subpackage config list self get subpackage subpackage name subpackage path file usr local lib dist packages numpy distutils misc util py line in get subpackage config self get configuration from setup py file usr local lib dist packages numpy distutils misc util py line in get configuration from setup py config setup module configuration args file scipy linalg setup py line in configuration raise notfounderror no lapack blas resources found numpy distutils system info notfounderror no lapack blas resources found rolling back uninstall of scipy moving to usr local lib dist packages scipy dist info from usr local lib dist packages cipy dist info moving to usr local lib dist packages scipy libs from usr local lib dist packages cipy libs moving to usr local lib dist packages scipy from usr local lib dist packages cipy error command errored out with exit status usr bin u c import io os sys setuptools tokenize sys argv tmp pip install b scipy setup py file tmp pip install b scipy setup py f getattr tokenize open open file if os path exists file else io stringio from setuptools import setup setup code f read replace r n n f close exec compile code file exec install record tmp pip record install record txt single version externally managed compile install headers usr local include scipy check the logs for full command output reproducing code example python requirements txt file flask flask cors waitress uwsgi pika pydantic elasticsearch requests httpretty sure nltk scikit learn numpy scipy xgboost python dateutil minio gensim pandas error message shell sudo install r requirements txt scipy numpy python version information python
| 1
|
87,938
| 25,257,368,616
|
IssuesEvent
|
2022-11-15 19:19:22
|
RamenDR/ramen
|
https://api.github.com/repos/RamenDR/ramen
|
closed
|
openshift-preflight: HasRequiredLabel failed
|
bug build
|
# Background
The [openshift-preflight](https://github.com/redhat-openshift-ecosystem/openshift-preflight/) tool checks to see if containers meet requirements for Openshift Certification. When running the preflight scan, the `HasRequiredLabel` check fails. The `HasRequiredLabel` output follows:
```bash
{
"name": "HasRequiredLabel",
"elapsed_time": 0,
"description": "Checking if the required labels (name, vendor, version, release, summary, description) are present in the container metadata.",
"help": "Check Check HasRequiredLabel encountered an error. Please review the preflight.log file for more information.",
"suggestion": "Add the following labels to your Dockerfile or Containerfile: name, vendor, version, release, summary, description",
"knowledgebase_url": "https://access.redhat.com/documentation/en-us/red_hat_software_certification/8.45/html/red_hat_openshift_software_certification_policy_guide/assembly-requirements-for-container-images_openshift-sw-cert-policy-introduction",
"check_url": "https://access.redhat.com/documentation/en-us/red_hat_software_certification/8.45/html/red_hat_openshift_software_certification_policy_guide/assembly-requirements-for-container-images_openshift-sw-cert-policy-introduction"
},
```
# Proposed Fix and Commentary
A proposed fix for this issue is in [PR83 of ShioRamen](https://github.com/RamenDR/ShioRamen/pull/83). The labels mentioned above must be present, but the text itself may be debated, so I'm starting the discussion here. The following is added to `Dockerfile`:
```bash
# Add labels to image
LABEL name="Ramen DR operator" \
vendor="github.com/RamenDR/ramen" \
version="1.0" \
summary="Provides disaster recovery and relocations services for workloads and their persistent data" \
description="Deploy Ramen DR operator"
```
The purpose of opening this issue is to get feedback on the implementation details to assist in a later PR for bringing the remainder of ShioRamen updates into Ramen.
|
1.0
|
openshift-preflight: HasRequiredLabel failed - # Background
The [openshift-preflight](https://github.com/redhat-openshift-ecosystem/openshift-preflight/) tool checks to see if containers meet requirements for Openshift Certification. When running the preflight scan, the `HasRequiredLabel` check fails. The `HasRequiredLabel` output follows:
```bash
{
"name": "HasRequiredLabel",
"elapsed_time": 0,
"description": "Checking if the required labels (name, vendor, version, release, summary, description) are present in the container metadata.",
"help": "Check Check HasRequiredLabel encountered an error. Please review the preflight.log file for more information.",
"suggestion": "Add the following labels to your Dockerfile or Containerfile: name, vendor, version, release, summary, description",
"knowledgebase_url": "https://access.redhat.com/documentation/en-us/red_hat_software_certification/8.45/html/red_hat_openshift_software_certification_policy_guide/assembly-requirements-for-container-images_openshift-sw-cert-policy-introduction",
"check_url": "https://access.redhat.com/documentation/en-us/red_hat_software_certification/8.45/html/red_hat_openshift_software_certification_policy_guide/assembly-requirements-for-container-images_openshift-sw-cert-policy-introduction"
},
```
# Proposed Fix and Commentary
A proposed fix for this issue is in [PR83 of ShioRamen](https://github.com/RamenDR/ShioRamen/pull/83). The labels mentioned above must be present, but the text itself may be debated, so I'm starting the discussion here. The following is added to `Dockerfile`:
```bash
# Add labels to image
LABEL name="Ramen DR operator" \
vendor="github.com/RamenDR/ramen" \
version="1.0" \
summary="Provides disaster recovery and relocations services for workloads and their persistent data" \
description="Deploy Ramen DR operator"
```
The purpose of opening this issue is to get feedback on the implementation details to assist in a later PR for bringing the remainder of ShioRamen updates into Ramen.
|
non_defect
|
openshift preflight hasrequiredlabel failed background the tool checks to see if containers meet requirements for openshift certification when running the preflight scan the hasrequiredlabel check fails the hasrequiredlabel output follows bash name hasrequiredlabel elapsed time description checking if the required labels name vendor version release summary description are present in the container metadata help check check hasrequiredlabel encountered an error please review the preflight log file for more information suggestion add the following labels to your dockerfile or containerfile name vendor version release summary description knowledgebase url check url proposed fix and commentary a proposed fix for this issue is in the labels mentioned above must be present but the text itself may be debated so i m starting the discussion here the following is added to dockerfile bash add labels to image label name ramen dr operator vendor github com ramendr ramen version summary provides disaster recovery and relocations services for workloads and their persistent data description deploy ramen dr operator the purpose of opening this issue is to get feedback on the implementation details to assist in a later pr for bringing the remainder of shioramen updates into ramen
| 0
|
48,843
| 13,184,755,442
|
IssuesEvent
|
2020-08-12 20:02:06
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
from icecube import superdst is broken (Trac #306)
|
Incomplete Migration Migrated from Trac combo reconstruction defect
|
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/306
, reported by blaufuss and owned by jacobi_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2011-09-09T14:52:28",
"description": "In icerec V04-00-01:\n\ntrying to use the line:\n\nfrom icecube import superdst\nyield:\n\n\nTraceback (most recent call last):\n File \"./dst_test.py\", line 13, in <module>\n from icecube import icetray, dataclasses, dataio, superdst\n File \"/data/i3home/blaufuss/icework/icerec/V04-00-01/build_d/lib/icecube/superdst/__init__.py\", line 4, in <module>\n load_pybindings(__name__,__path__)\n File \"/data/i3home/blaufuss/icework/icerec/V04-00-01/build_d/lib/icecube/load_pybindings.py\", line 53, in load_pybindings\n m = imp.load_dynamic(name, path[0] + \".so\")\nImportError: /data/i3home/blaufuss/icework/icerec/V04-00-01/build_d/lib/icecube/superdst.so: cannot open shared object file: No such file or directory\n\n\nA work around of changing to \n\nload(\"superdst\")\n\nexists",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1315579948000000",
"component": "combo reconstruction",
"summary": "from icecube import superdst is broken",
"priority": "normal",
"keywords": "",
"time": "2011-09-09T14:09:13",
"milestone": "",
"owner": "jacobi",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
from icecube import superdst is broken (Trac #306) - <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/306
, reported by blaufuss and owned by jacobi_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2011-09-09T14:52:28",
"description": "In icerec V04-00-01:\n\ntrying to use the line:\n\nfrom icecube import superdst\nyield:\n\n\nTraceback (most recent call last):\n File \"./dst_test.py\", line 13, in <module>\n from icecube import icetray, dataclasses, dataio, superdst\n File \"/data/i3home/blaufuss/icework/icerec/V04-00-01/build_d/lib/icecube/superdst/__init__.py\", line 4, in <module>\n load_pybindings(__name__,__path__)\n File \"/data/i3home/blaufuss/icework/icerec/V04-00-01/build_d/lib/icecube/load_pybindings.py\", line 53, in load_pybindings\n m = imp.load_dynamic(name, path[0] + \".so\")\nImportError: /data/i3home/blaufuss/icework/icerec/V04-00-01/build_d/lib/icecube/superdst.so: cannot open shared object file: No such file or directory\n\n\nA work around of changing to \n\nload(\"superdst\")\n\nexists",
"reporter": "blaufuss",
"cc": "",
"resolution": "fixed",
"_ts": "1315579948000000",
"component": "combo reconstruction",
"summary": "from icecube import superdst is broken",
"priority": "normal",
"keywords": "",
"time": "2011-09-09T14:09:13",
"milestone": "",
"owner": "jacobi",
"type": "defect"
}
```
</p>
</details>
|
defect
|
from icecube import superdst is broken trac migrated from reported by blaufuss and owned by jacobi json status closed changetime description in icerec n ntrying to use the line n nfrom icecube import superdst nyield n n ntraceback most recent call last n file dst test py line in n from icecube import icetray dataclasses dataio superdst n file data blaufuss icework icerec build d lib icecube superdst init py line in n load pybindings name path n file data blaufuss icework icerec build d lib icecube load pybindings py line in load pybindings n m imp load dynamic name path so nimporterror data blaufuss icework icerec build d lib icecube superdst so cannot open shared object file no such file or directory n n na work around of changing to n nload superdst n nexists reporter blaufuss cc resolution fixed ts component combo reconstruction summary from icecube import superdst is broken priority normal keywords time milestone owner jacobi type defect
| 1
|
50,925
| 13,187,979,245
|
IssuesEvent
|
2020-08-13 05:12:10
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
IceRec.IC2011-L2_V12-08-00_IceSim4compat_V5 compatible with version 10 of I3DOMCalibration (Trac #1691)
|
Migrated from Trac combo reconstruction defect
|
The latest IceRec for 2011 (IceSim4 compat) is not compatible with version I3DOMCalibration. We need a new release with a newer version of dataclasses that will support the official GCD.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1691">https://code.icecube.wisc.edu/ticket/1691</a>, reported by juancarlos and owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:58",
"description": "The latest IceRec for 2011 (IceSim4 compat) is not compatible with version I3DOMCalibration. We need a new release with a newer version of dataclasses that will support the official GCD.",
"reporter": "juancarlos",
"cc": "javierg@udel.edu",
"resolution": "wontfix",
"_ts": "1550067178841456",
"component": "combo reconstruction",
"summary": "IceRec.IC2011-L2_V12-08-00_IceSim4compat_V5 compatible with version 10 of I3DOMCalibration",
"priority": "critical",
"keywords": "IceRec",
"time": "2016-05-04T19:04:57",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
IceRec.IC2011-L2_V12-08-00_IceSim4compat_V5 compatible with version 10 of I3DOMCalibration (Trac #1691) - The latest IceRec for 2011 (IceSim4 compat) is not compatible with version I3DOMCalibration. We need a new release with a newer version of dataclasses that will support the official GCD.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1691">https://code.icecube.wisc.edu/ticket/1691</a>, reported by juancarlos and owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:58",
"description": "The latest IceRec for 2011 (IceSim4 compat) is not compatible with version I3DOMCalibration. We need a new release with a newer version of dataclasses that will support the official GCD.",
"reporter": "juancarlos",
"cc": "javierg@udel.edu",
"resolution": "wontfix",
"_ts": "1550067178841456",
"component": "combo reconstruction",
"summary": "IceRec.IC2011-L2_V12-08-00_IceSim4compat_V5 compatible with version 10 of I3DOMCalibration",
"priority": "critical",
"keywords": "IceRec",
"time": "2016-05-04T19:04:57",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
defect
|
icerec compatible with version of trac the latest icerec for compat is not compatible with version we need a new release with a newer version of dataclasses that will support the official gcd migrated from json status closed changetime description the latest icerec for compat is not compatible with version we need a new release with a newer version of dataclasses that will support the official gcd reporter juancarlos cc javierg udel edu resolution wontfix ts component combo reconstruction summary icerec compatible with version of priority critical keywords icerec time milestone owner olivas type defect
| 1
|
41,748
| 10,587,368,076
|
IssuesEvent
|
2019-10-08 21:56:34
|
techo/voluntariado-eventual
|
https://api.github.com/repos/techo/voluntariado-eventual
|
closed
|
BE: Errores al crea una actividad estan desacomodados
|
Defecto Mejora
|
**Describí el error**
Errores al crea una actividad estan desacomodados
**Para reproducirlo**
Pasos para reproducir el comportamiento:
-. Ir a TESTING
-. Menu - ADMIN - Crear una Actividad
-. Presiono el boton "Guardar"
-. Aparecen los "Errores" pero estan desacomodados. Por ejemplo:
"El campo localidad debe tener un valor valido" y despues viene, por ejemplo:
"El campo Oficina es requerido"
**Comportamiento esperando**
-. Ir a TESTING
-. Menu - ADMIN - Crear una Actividad
-. Presiono el boton "Guardar"
-. Aparecen los errores acomodados por solapa: Informacion General, Ubicacion de la Actividad, Puntos de Encuentro, Inscripciones, Construccion
**Capturas de pantalla**

**Si estás en una computadora (por favor completá la siguiente información):**
- Navegador [por ejemplo: chrome, explorar, safari]
**Smartphone (completá la siguiente informaicón):**
- Dispositivo: [por ejemplo: Huawei GW, iPhone6, Samsung J2]
- Sistema operativo: [por ejemplo: Android4, iOS8.1]
- Navegador [por ejemplo: navegador del celu, Chrome, Safari]
**Contexto adicional**
Toda otra cosa que ayude a explicar lo que pasó.
|
1.0
|
BE: Errores al crea una actividad estan desacomodados - **Describí el error**
Errores al crea una actividad estan desacomodados
**Para reproducirlo**
Pasos para reproducir el comportamiento:
-. Ir a TESTING
-. Menu - ADMIN - Crear una Actividad
-. Presiono el boton "Guardar"
-. Aparecen los "Errores" pero estan desacomodados. Por ejemplo:
"El campo localidad debe tener un valor valido" y despues viene, por ejemplo:
"El campo Oficina es requerido"
**Comportamiento esperando**
-. Ir a TESTING
-. Menu - ADMIN - Crear una Actividad
-. Presiono el boton "Guardar"
-. Aparecen los errores acomodados por solapa: Informacion General, Ubicacion de la Actividad, Puntos de Encuentro, Inscripciones, Construccion
**Capturas de pantalla**

**Si estás en una computadora (por favor completá la siguiente información):**
- Navegador [por ejemplo: chrome, explorar, safari]
**Smartphone (completá la siguiente informaicón):**
- Dispositivo: [por ejemplo: Huawei GW, iPhone6, Samsung J2]
- Sistema operativo: [por ejemplo: Android4, iOS8.1]
- Navegador [por ejemplo: navegador del celu, Chrome, Safari]
**Contexto adicional**
Toda otra cosa que ayude a explicar lo que pasó.
|
defect
|
be errores al crea una actividad estan desacomodados describí el error errores al crea una actividad estan desacomodados para reproducirlo pasos para reproducir el comportamiento ir a testing menu admin crear una actividad presiono el boton guardar aparecen los errores pero estan desacomodados por ejemplo el campo localidad debe tener un valor valido y despues viene por ejemplo el campo oficina es requerido comportamiento esperando ir a testing menu admin crear una actividad presiono el boton guardar aparecen los errores acomodados por solapa informacion general ubicacion de la actividad puntos de encuentro inscripciones construccion capturas de pantalla si estás en una computadora por favor completá la siguiente información navegador smartphone completá la siguiente informaicón dispositivo sistema operativo navegador contexto adicional toda otra cosa que ayude a explicar lo que pasó
| 1
|
5,786
| 2,610,215,458
|
IssuesEvent
|
2015-02-26 19:08:43
|
chrsmith/somefinders
|
https://api.github.com/repos/chrsmith/somefinders
|
opened
|
решебник чертов воробьёв.pdf
|
auto-migrated Priority-Medium Type-Defect
|
```
'''Вилиор Беспалов'''
День добрый никак не могу найти .решебник
чертов воробьёв.pdf. где то видел уже
'''Альфред Петров'''
Вот хороший сайт где можно скачать
http://bit.ly/17vyxgV
'''Болеслав Воробьёв'''
Спасибо вроде то но просит телефон вводить
'''Анвар Шестаков'''
Неа все ок у меня ничего не списало
'''Бертольд Константинов'''
Не это не влияет на баланс
Информация о файле: решебник чертов
воробьёв.pdf
Загружен: В этом месяце
Скачан раз: 705
Рейтинг: 1379
Средняя скорость скачивания: 123
Похожих файлов: 19
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 4:49
|
1.0
|
решебник чертов воробьёв.pdf - ```
'''Вилиор Беспалов'''
День добрый никак не могу найти .решебник
чертов воробьёв.pdf. где то видел уже
'''Альфред Петров'''
Вот хороший сайт где можно скачать
http://bit.ly/17vyxgV
'''Болеслав Воробьёв'''
Спасибо вроде то но просит телефон вводить
'''Анвар Шестаков'''
Неа все ок у меня ничего не списало
'''Бертольд Константинов'''
Не это не влияет на баланс
Информация о файле: решебник чертов
воробьёв.pdf
Загружен: В этом месяце
Скачан раз: 705
Рейтинг: 1379
Средняя скорость скачивания: 123
Похожих файлов: 19
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 4:49
|
defect
|
решебник чертов воробьёв pdf вилиор беспалов день добрый никак не могу найти решебник чертов воробьёв pdf где то видел уже альфред петров вот хороший сайт где можно скачать болеслав воробьёв спасибо вроде то но просит телефон вводить анвар шестаков неа все ок у меня ничего не списало бертольд константинов не это не влияет на баланс информация о файле решебник чертов воробьёв pdf загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at
| 1
|
594,342
| 18,043,520,029
|
IssuesEvent
|
2021-09-18 13:21:05
|
lardemua/atom
|
https://api.github.com/repos/lardemua/atom
|
closed
|
Improvement of atom_rviz_plugin readme
|
enhancement low priority
|
Hi @MiguelDRPina ,
I suggest that you improve the readme with some photos or even videos of the calibration panel.
Use the atom readme as a model.
|
1.0
|
Improvement of atom_rviz_plugin readme - Hi @MiguelDRPina ,
I suggest that you improve the readme with some photos or even videos of the calibration panel.
Use the atom readme as a model.
|
non_defect
|
improvement of atom rviz plugin readme hi migueldrpina i suggest that you improve the readme with some photos or even videos of the calibration panel use the atom readme as a model
| 0
|
100,300
| 4,082,563,306
|
IssuesEvent
|
2016-05-31 13:20:37
|
minj/foxtrick
|
https://api.github.com/repos/minj/foxtrick
|
closed
|
Disallow removing ExtraShortcuts preferences link on Android
|
accepted enhancement Platform-Android Priority-Medium usability
|
It's a real foot-gun.
Mozilla is not really enthusiastic with fixing [HTML options](https://bugzilla.mozilla.org/show_bug.cgi?id=1192607) it seems.
|
1.0
|
Disallow removing ExtraShortcuts preferences link on Android - It's a real foot-gun.
Mozilla is not really enthusiastic with fixing [HTML options](https://bugzilla.mozilla.org/show_bug.cgi?id=1192607) it seems.
|
non_defect
|
disallow removing extrashortcuts preferences link on android it s a real foot gun mozilla is not really enthusiastic with fixing it seems
| 0
|
137,894
| 11,166,542,770
|
IssuesEvent
|
2019-12-27 13:42:22
|
one-lightning/TestingUp
|
https://api.github.com/repos/one-lightning/TestingUp
|
closed
|
[TC] [CTRL+Z] hotkey to undo the text changes
|
testcase testing
|
**Preconditions**
**1.** Notepad installed on device
**2.** Existing a notepad file with name " My first document" on the Desktop
**Steps**
**1.** Find the notepad document on the desktop
**2.** Open document
**3.** Click on text area
**4.** Edit document (add text or delete)
**5.** Choose "undo" button from "Edit Menu"
**Expected Result**
"Undo" button in Edit Menushould works
|
2.0
|
[TC] [CTRL+Z] hotkey to undo the text changes - **Preconditions**
**1.** Notepad installed on device
**2.** Existing a notepad file with name " My first document" on the Desktop
**Steps**
**1.** Find the notepad document on the desktop
**2.** Open document
**3.** Click on text area
**4.** Edit document (add text or delete)
**5.** Choose "undo" button from "Edit Menu"
**Expected Result**
"Undo" button in Edit Menushould works
|
non_defect
|
hotkey to undo the text changes preconditions notepad installed on device existing a notepad file with name my first document on the desktop steps find the notepad document on the desktop open document click on text area edit document add text or delete choose undo button from edit menu expected result undo button in edit menushould works
| 0
|
42,878
| 11,349,483,947
|
IssuesEvent
|
2020-01-24 05:09:06
|
idaholab/raven
|
https://api.github.com/repos/idaholab/raven
|
closed
|
If illegal XML named variables are present, the code crashes
|
defect devel priority_critical
|
--------
Issue Description
--------
##### What did you expect to see happen?
Ability to load CSV files with invalid XML variables (e.g. containing "/", etc)
##### What did you see instead?
If a variable has an invalid XML character (e.g. "/", "$", etc.), the XML utilis methods (used to write out the metadata) is going to replace it with a "." when constructing the XML metadata (node <dim>).
Since the new variables do not match any of the output variables, those variables are not written in the auxiliary metadata XML file. This causes the code to crash when trying to reload the CSVs. When the variables are read back, the code does not find the Index of the "illegal" variables (not written in the XML) defined in the <Output> block.
For example, When we dump the following variables we get this warning:
```
XML UTILS: Replacing illegal tag characters in "u-234_atoms/barn-cm": u-234_atoms.barn-cm
XML UTILS: Replacing illegal tag characters in "ag-107_atoms/barn-cm": ag-107_atoms.barn-cm
XML UTILS: Replacing illegal tag characters in "u-235_atoms/barn-cm": u-235_atoms.barn-cm
```
When they are read back, the illegal named variables (e.g. ag-107_atoms/barn-cm) are not found in the XML and consequentially are not stored in the dict ```self._pivotParams```. This leaves variables in the output space without any index linked to them and consequentially, when the method ```asDataset``` gets called, an unexpected crash happens:
```
( 43.79 sec) HistorySet : DEBUG -> Reading data from "/Users/alfoa/projects/raven_github/raven/dh_chenj/samples/originalModelCSV/history_time_evolution_2017.csv.csv"
( 43.81 sec) HistorySet : DEBUG -> Reading data from "/Users/alfoa/projects/raven_github/raven/dh_chenj/samples/originalModelCSV/history_time_evolution_2018.csv.csv"
( 43.83 sec) HistorySet : DEBUG -> Reading data from "/Users/alfoa/projects/raven_github/raven/dh_chenj/samples/originalModelCSV/history_time_evolution_2019.csv.csv"
Traceback (most recent call last):
File "/Users/alfoa/projects/raven_github/raven/framework/Driver.py", line 281, in <module>
raven()
File "/Users/alfoa/projects/raven_github/raven/framework/Driver.py", line 234, in raven
simulation.run()
File "/Users/alfoa/projects/raven_github/raven/framework/Simulation.py", line 798, in run
stepInstance.takeAstep(stepInputDict)
File "/Users/alfoa/projects/raven_github/raven/framework/Steps.py", line 312, in takeAstep
self._localTakeAstepRun(inDictionary)
File "/Users/alfoa/projects/raven_github/raven/framework/Steps.py", line 947, in _localTakeAstepRun
outputs[i].load(inDictionary['Input'][i].getPath(),'csv',**options)
File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 433, in load
self._fromCSV(fileName,**kwargs)
File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/HistorySet.py", line 154, in _fromCSV
self.load(data,style='dict',dims=self.getDimensions())
File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 435, in load
self._fromDict(fileName,**kwargs)
File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 1257, in _fromDict
self.asDataset()
File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 260, in asDataset
return self._convertToXrDataset()
File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 1079, in _convertToXrDataset
arrays[var][self.sampleTag] += firstSample
KeyError: u'u-234_atoms/barn-cm'
```
##### Do you have a suggested fix for the development team?
Either
the illegal XML character should be replaced with a "token" recognizable in the loading stage => we will able to identify it and connect them back with the original variables
or
we move from :
```
<dims>
<totals_watts>time</totals_watts>
</dims>
```
to
```
<dims>
<var name="totals_watts">time</var>
<var name="ag-107_atoms/barn-cm">time</var>
</dims>
```
or
change the DataObject.py method ```setPivotParameter``` to:
```
def setPivotParams(self,params):
"""
Sets the pivot parameters for variables.
@ In, params, dict, var:[params] as str:list(str)
@ Out, None
"""
# TODO typechecking, assertions
coords = set().union(*params.values())
for coord in coords:
if coord not in self._pivotParams:
self._pivotParams[coord] = list(var for var in params.keys() if coord in params[var])
else:
self._pivotParams[coord] = list(set(list(var for var in params.keys() if coord in params[var]) + self._pivotParams[coord]))
```
##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [x] 1. Is it tagged with a type: defect or improvement?
- [x] 2. Is it tagged with a priority: critical, normal or minor?
- [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)?
- [x] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
|
1.0
|
If illegal XML named variables are present, the code crashes - --------
Issue Description
--------
##### What did you expect to see happen?
Ability to load CSV files with invalid XML variables (e.g. containing "/", etc)
##### What did you see instead?
If a variable has an invalid XML character (e.g. "/", "$", etc.), the XML utilis methods (used to write out the metadata) is going to replace it with a "." when constructing the XML metadata (node <dim>).
Since the new variables do not match any of the output variables, those variables are not written in the auxiliary metadata XML file. This causes the code to crash when trying to reload the CSVs. When the variables are read back, the code does not find the Index of the "illegal" variables (not written in the XML) defined in the <Output> block.
For example, When we dump the following variables we get this warning:
```
XML UTILS: Replacing illegal tag characters in "u-234_atoms/barn-cm": u-234_atoms.barn-cm
XML UTILS: Replacing illegal tag characters in "ag-107_atoms/barn-cm": ag-107_atoms.barn-cm
XML UTILS: Replacing illegal tag characters in "u-235_atoms/barn-cm": u-235_atoms.barn-cm
```
When they are read back, the illegal named variables (e.g. ag-107_atoms/barn-cm) are not found in the XML and consequentially are not stored in the dict ```self._pivotParams```. This leaves variables in the output space without any index linked to them and consequentially, when the method ```asDataset``` gets called, an unexpected crash happens:
```
( 43.79 sec) HistorySet : DEBUG -> Reading data from "/Users/alfoa/projects/raven_github/raven/dh_chenj/samples/originalModelCSV/history_time_evolution_2017.csv.csv"
( 43.81 sec) HistorySet : DEBUG -> Reading data from "/Users/alfoa/projects/raven_github/raven/dh_chenj/samples/originalModelCSV/history_time_evolution_2018.csv.csv"
( 43.83 sec) HistorySet : DEBUG -> Reading data from "/Users/alfoa/projects/raven_github/raven/dh_chenj/samples/originalModelCSV/history_time_evolution_2019.csv.csv"
Traceback (most recent call last):
File "/Users/alfoa/projects/raven_github/raven/framework/Driver.py", line 281, in <module>
raven()
File "/Users/alfoa/projects/raven_github/raven/framework/Driver.py", line 234, in raven
simulation.run()
File "/Users/alfoa/projects/raven_github/raven/framework/Simulation.py", line 798, in run
stepInstance.takeAstep(stepInputDict)
File "/Users/alfoa/projects/raven_github/raven/framework/Steps.py", line 312, in takeAstep
self._localTakeAstepRun(inDictionary)
File "/Users/alfoa/projects/raven_github/raven/framework/Steps.py", line 947, in _localTakeAstepRun
outputs[i].load(inDictionary['Input'][i].getPath(),'csv',**options)
File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 433, in load
self._fromCSV(fileName,**kwargs)
File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/HistorySet.py", line 154, in _fromCSV
self.load(data,style='dict',dims=self.getDimensions())
File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 435, in load
self._fromDict(fileName,**kwargs)
File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 1257, in _fromDict
self.asDataset()
File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 260, in asDataset
return self._convertToXrDataset()
File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 1079, in _convertToXrDataset
arrays[var][self.sampleTag] += firstSample
KeyError: u'u-234_atoms/barn-cm'
```
##### Do you have a suggested fix for the development team?
Either
the illegal XML character should be replaced with a "token" recognizable in the loading stage => we will able to identify it and connect them back with the original variables
or
we move from :
```
<dims>
<totals_watts>time</totals_watts>
</dims>
```
to
```
<dims>
<var name="totals_watts">time</var>
<var name="ag-107_atoms/barn-cm">time</var>
</dims>
```
or
change the DataObject.py method ```setPivotParameter``` to:
```
def setPivotParams(self,params):
"""
Sets the pivot parameters for variables.
@ In, params, dict, var:[params] as str:list(str)
@ Out, None
"""
# TODO typechecking, assertions
coords = set().union(*params.values())
for coord in coords:
if coord not in self._pivotParams:
self._pivotParams[coord] = list(var for var in params.keys() if coord in params[var])
else:
self._pivotParams[coord] = list(set(list(var for var in params.keys() if coord in params[var]) + self._pivotParams[coord]))
```
##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [x] 1. Is it tagged with a type: defect or improvement?
- [x] 2. Is it tagged with a priority: critical, normal or minor?
- [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)?
- [x] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
|
defect
|
if illegal xml named variables are present the code crashes issue description what did you expect to see happen ability to load csv files with invalid xml variables e g containing etc what did you see instead if a variable has an invalid xml character e g etc the xml utilis methods used to write out the metadata is going to replace it with a when constructing the xml metadata node since the new variables do not match any of the output variables those variables are not written in the auxiliary metadata xml file this causes the code to crash when trying to reload the csvs when the variables are read back the code does not find the index of the illegal variables not written in the xml defined in the block for example when we dump the following variables we get this warning xml utils replacing illegal tag characters in u atoms barn cm u atoms barn cm xml utils replacing illegal tag characters in ag atoms barn cm ag atoms barn cm xml utils replacing illegal tag characters in u atoms barn cm u atoms barn cm when they are read back the illegal named variables e g ag atoms barn cm are not found in the xml and consequentially are not stored in the dict self pivotparams this leaves variables in the output space without any index linked to them and consequentially when the method asdataset gets called an unexpected crash happens sec historyset debug reading data from users alfoa projects raven github raven dh chenj samples originalmodelcsv history time evolution csv csv sec historyset debug reading data from users alfoa projects raven github raven dh chenj samples originalmodelcsv history time evolution csv csv sec historyset debug reading data from users alfoa projects raven github raven dh chenj samples originalmodelcsv history time evolution csv csv traceback most recent call last file users alfoa projects raven github raven framework driver py line in raven file users alfoa projects raven github raven framework driver py line in raven simulation run file users alfoa projects raven github raven framework simulation py line in run stepinstance takeastep stepinputdict file users alfoa projects raven github raven framework steps py line in takeastep self localtakeasteprun indictionary file users alfoa projects raven github raven framework steps py line in localtakeasteprun outputs load indictionary getpath csv options file users alfoa projects raven github raven framework dataobjects dataset py line in load self fromcsv filename kwargs file users alfoa projects raven github raven framework dataobjects historyset py line in fromcsv self load data style dict dims self getdimensions file users alfoa projects raven github raven framework dataobjects dataset py line in load self fromdict filename kwargs file users alfoa projects raven github raven framework dataobjects dataset py line in fromdict self asdataset file users alfoa projects raven github raven framework dataobjects dataset py line in asdataset return self converttoxrdataset file users alfoa projects raven github raven framework dataobjects dataset py line in converttoxrdataset arrays firstsample keyerror u u atoms barn cm do you have a suggested fix for the development team either the illegal xml character should be replaced with a token recognizable in the loading stage we will able to identify it and connect them back with the original variables or we move from time to time time or change the dataobject py method setpivotparameter to def setpivotparams self params sets the pivot parameters for variables in params dict var as str list str out none todo typechecking assertions coords set union params values for coord in coords if coord not in self pivotparams self pivotparams list var for var in params keys if coord in params else self pivotparams list set list var for var in params keys if coord in params self pivotparams please attach the input file s that generate this error the simpler the input the faster we can find the issue for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or improvement is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest stable branch if yes is there any issue tagged with stable create if needed if the issue is being closed without a merge request has an explanation of why it is being closed been provided
| 1
|
53,757
| 13,262,242,758
|
IssuesEvent
|
2020-08-20 21:22:47
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
[cvmfs] compatibility between py2-v3 and icerec version V05-07 (few variables cause problems) (Trac #2018)
|
Migrated from Trac cvmfs defect
|
I checked out icerec version V05-007 and built it with py2-v3. I then added a few projects to the metaproject. It seemed to be working properly except that I get errors when I try to write some variables to files, like SplineMPE_MillipedeHighEnergyMIEFitParams or MPEFitParaboloidFitParams. I attached the script.
The error is: ERROR (icetray): frame caught exception "unregistered class" while loading class type "I3ParaboloidFitParams" at key "MPEFitParaboloidFitParams"
and
ERROR (I3TableWriter): Frame object 'MPEFitParaboloidFitParams' could not be deserialized and will not be booked.
The exact same script works perfectly when using the same icerec version built on py2-v2.
I put the line to run this script at the top of the main.py.
It is the first time I'm writing a ticket, apologies if I didn't follow the good practices. Thank you!
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2018">https://code.icecube.wisc.edu/projects/icecube/ticket/2018</a>, reported by stephanie.bronand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:15:13",
"_ts": "1550067313248429",
"description": "I checked out icerec version V05-007 and built it with py2-v3. I then added a few projects to the metaproject. It seemed to be working properly except that I get errors when I try to write some variables to files, like SplineMPE_MillipedeHighEnergyMIEFitParams or MPEFitParaboloidFitParams. I attached the script.\n\nThe error is: ERROR (icetray): frame caught exception \"unregistered class\" while loading class type \"I3ParaboloidFitParams\" at key \"MPEFitParaboloidFitParams\"\n\nand \n\nERROR (I3TableWriter): Frame object 'MPEFitParaboloidFitParams' could not be deserialized and will not be booked.\n\nThe exact same script works perfectly when using the same icerec version built on py2-v2. \n\nI put the line to run this script at the top of the main.py. \n\nIt is the first time I'm writing a ticket, apologies if I didn't follow the good practices. Thank you!\n",
"reporter": "stephanie.bron",
"cc": "david.schultz",
"resolution": "worksforme",
"time": "2017-05-11T13:36:21",
"component": "cvmfs",
"summary": "[cvmfs] compatibility between py2-v3 and icerec version V05-07 (few variables cause problems)",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[cvmfs] compatibility between py2-v3 and icerec version V05-07 (few variables cause problems) (Trac #2018) - I checked out icerec version V05-007 and built it with py2-v3. I then added a few projects to the metaproject. It seemed to be working properly except that I get errors when I try to write some variables to files, like SplineMPE_MillipedeHighEnergyMIEFitParams or MPEFitParaboloidFitParams. I attached the script.
The error is: ERROR (icetray): frame caught exception "unregistered class" while loading class type "I3ParaboloidFitParams" at key "MPEFitParaboloidFitParams"
and
ERROR (I3TableWriter): Frame object 'MPEFitParaboloidFitParams' could not be deserialized and will not be booked.
The exact same script works perfectly when using the same icerec version built on py2-v2.
I put the line to run this script at the top of the main.py.
It is the first time I'm writing a ticket, apologies if I didn't follow the good practices. Thank you!
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2018">https://code.icecube.wisc.edu/projects/icecube/ticket/2018</a>, reported by stephanie.bronand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:15:13",
"_ts": "1550067313248429",
"description": "I checked out icerec version V05-007 and built it with py2-v3. I then added a few projects to the metaproject. It seemed to be working properly except that I get errors when I try to write some variables to files, like SplineMPE_MillipedeHighEnergyMIEFitParams or MPEFitParaboloidFitParams. I attached the script.\n\nThe error is: ERROR (icetray): frame caught exception \"unregistered class\" while loading class type \"I3ParaboloidFitParams\" at key \"MPEFitParaboloidFitParams\"\n\nand \n\nERROR (I3TableWriter): Frame object 'MPEFitParaboloidFitParams' could not be deserialized and will not be booked.\n\nThe exact same script works perfectly when using the same icerec version built on py2-v2. \n\nI put the line to run this script at the top of the main.py. \n\nIt is the first time I'm writing a ticket, apologies if I didn't follow the good practices. Thank you!\n",
"reporter": "stephanie.bron",
"cc": "david.schultz",
"resolution": "worksforme",
"time": "2017-05-11T13:36:21",
"component": "cvmfs",
"summary": "[cvmfs] compatibility between py2-v3 and icerec version V05-07 (few variables cause problems)",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
|
defect
|
compatibility between and icerec version few variables cause problems trac i checked out icerec version and built it with i then added a few projects to the metaproject it seemed to be working properly except that i get errors when i try to write some variables to files like splinempe millipedehighenergymiefitparams or mpefitparaboloidfitparams i attached the script the error is error icetray frame caught exception unregistered class while loading class type at key mpefitparaboloidfitparams and error frame object mpefitparaboloidfitparams could not be deserialized and will not be booked the exact same script works perfectly when using the same icerec version built on i put the line to run this script at the top of the main py it is the first time i m writing a ticket apologies if i didn t follow the good practices thank you migrated from json status closed changetime ts description i checked out icerec version and built it with i then added a few projects to the metaproject it seemed to be working properly except that i get errors when i try to write some variables to files like splinempe millipedehighenergymiefitparams or mpefitparaboloidfitparams i attached the script n nthe error is error icetray frame caught exception unregistered class while loading class type at key mpefitparaboloidfitparams n nand n nerror frame object mpefitparaboloidfitparams could not be deserialized and will not be booked n nthe exact same script works perfectly when using the same icerec version built on n ni put the line to run this script at the top of the main py n nit is the first time i m writing a ticket apologies if i didn t follow the good practices thank you n reporter stephanie bron cc david schultz resolution worksforme time component cvmfs summary compatibility between and icerec version few variables cause problems priority normal keywords milestone owner nega type defect
| 1
|
62,247
| 8,583,411,484
|
IssuesEvent
|
2018-11-13 19:39:41
|
mapbox/mapbox-gl-native
|
https://api.github.com/repos/mapbox/mapbox-gl-native
|
closed
|
[iOS][documentation] Link relevant API classes to `/ios-sdk/maps/examples`
|
documentation iOS
|
### Quick overview
Mapbox hosts example code for popular implementations of the iOS Maps SDK at www.mapbox.com/ios-sdk/maps/examples. To help interested users understand correct/recommended usage, these examples should be linked within the API docs, similar to the GL JS API's "Related" sections:

### Examples to be linked:
There are currently 39 live examples on https://www.mapbox.com/ios-sdk/maps/examples. I think it should be possible to link all of them to relevant parts of the documentation.
|
1.0
|
[iOS][documentation] Link relevant API classes to `/ios-sdk/maps/examples` - ### Quick overview
Mapbox hosts example code for popular implementations of the iOS Maps SDK at www.mapbox.com/ios-sdk/maps/examples. To help interested users understand correct/recommended usage, these examples should be linked within the API docs, similar to the GL JS API's "Related" sections:

### Examples to be linked:
There are currently 39 live examples on https://www.mapbox.com/ios-sdk/maps/examples. I think it should be possible to link all of them to relevant parts of the documentation.
|
non_defect
|
link relevant api classes to ios sdk maps examples quick overview mapbox hosts example code for popular implementations of the ios maps sdk at to help interested users understand correct recommended usage these examples should be linked within the api docs similar to the gl js api s related sections examples to be linked there are currently live examples on i think it should be possible to link all of them to relevant parts of the documentation
| 0
|
40,829
| 10,171,384,178
|
IssuesEvent
|
2019-08-08 08:16:13
|
frescobaldi/frescobaldi
|
https://api.github.com/repos/frescobaldi/frescobaldi
|
opened
|
Icons for dark themes
|
defect
|
I've started using Frescobaldi with a dark theme to better match my other desktop settings, to do so I've installed the `adwaita-qt` package for Linux Mint and then chose the "Adwaita Dark" them in Frescobaldi's preferences.
However, this doesn't show proper icons, rendering the UI something close to unusable:


If I'm not mistaken Frescobaldi *includes* icons from the Tango library and loads them - and there is only one set of icons that simply doesn't match a dark theme.
I'm not really sure what the best way would be to proceed. I think there should be *at least* the option of choosing between two different sets of icons - and we would somehow have to provide them. Or there should be an option to use icons from the given theme - but I don't have any idea how that works yet. And of course there's the fact that we have a number of custom icons provided, which are *not* part of any external theme.
|
1.0
|
Icons for dark themes - I've started using Frescobaldi with a dark theme to better match my other desktop settings, to do so I've installed the `adwaita-qt` package for Linux Mint and then chose the "Adwaita Dark" them in Frescobaldi's preferences.
However, this doesn't show proper icons, rendering the UI something close to unusable:


If I'm not mistaken Frescobaldi *includes* icons from the Tango library and loads them - and there is only one set of icons that simply doesn't match a dark theme.
I'm not really sure what the best way would be to proceed. I think there should be *at least* the option of choosing between two different sets of icons - and we would somehow have to provide them. Or there should be an option to use icons from the given theme - but I don't have any idea how that works yet. And of course there's the fact that we have a number of custom icons provided, which are *not* part of any external theme.
|
defect
|
icons for dark themes i ve started using frescobaldi with a dark theme to better match my other desktop settings to do so i ve installed the adwaita qt package for linux mint and then chose the adwaita dark them in frescobaldi s preferences however this doesn t show proper icons rendering the ui something close to unusable if i m not mistaken frescobaldi includes icons from the tango library and loads them and there is only one set of icons that simply doesn t match a dark theme i m not really sure what the best way would be to proceed i think there should be at least the option of choosing between two different sets of icons and we would somehow have to provide them or there should be an option to use icons from the given theme but i don t have any idea how that works yet and of course there s the fact that we have a number of custom icons provided which are not part of any external theme
| 1
|
249,110
| 26,884,771,551
|
IssuesEvent
|
2023-02-06 01:28:37
|
nidhi7598/linux-4.19.72
|
https://api.github.com/repos/nidhi7598/linux-4.19.72
|
closed
|
WS-2021-0213 (High) detected in linuxlinux-4.19.254 - autoclosed
|
security vulnerability
|
## WS-2021-0213 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.254</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72/commit/10a8c99e4f60044163c159867bc6f5452c1c36e5">10a8c99e4f60044163c159867bc6f5452c1c36e5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Linux/kernel in versions v3.5-rc1--v5.13 is vulnerable to possible buffer overflow in wl1251_cmd_scan
<p>Publish Date: 2021-05-31
<p>URL: <a href=https://github.com/torvalds/linux/commit/57ad99ae3c6738ba87bad259bb57c641ca68ebf6>WS-2021-0213</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1001488">https://osv.dev/vulnerability/UVI-2021-1001488</a></p>
<p>Release Date: 2021-05-31</p>
<p>Fix Resolution: v5.14-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2021-0213 (High) detected in linuxlinux-4.19.254 - autoclosed - ## WS-2021-0213 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.254</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72/commit/10a8c99e4f60044163c159867bc6f5452c1c36e5">10a8c99e4f60044163c159867bc6f5452c1c36e5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Linux/kernel in versions v3.5-rc1--v5.13 is vulnerable to possible buffer overflow in wl1251_cmd_scan
<p>Publish Date: 2021-05-31
<p>URL: <a href=https://github.com/torvalds/linux/commit/57ad99ae3c6738ba87bad259bb57c641ca68ebf6>WS-2021-0213</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/UVI-2021-1001488">https://osv.dev/vulnerability/UVI-2021-1001488</a></p>
<p>Release Date: 2021-05-31</p>
<p>Fix Resolution: v5.14-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
ws high detected in linuxlinux autoclosed ws high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details linux kernel in versions is vulnerable to possible buffer overflow in cmd scan publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
717,998
| 24,699,819,268
|
IssuesEvent
|
2022-10-19 14:34:11
|
SimplyVC/panic
|
https://api.github.com/repos/SimplyVC/panic
|
opened
|
Alert Router must latest & updated mappings
|
bug 5 SP Priority: High
|
<!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺
v ✰ Thanks for opening an issue! ✰
v Before smashing the submit button please review the template.
☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -->
### Summary of bug
<!-- Concisely describe the bug in question and what effect this has on the end user or system. -->
### 'As is' behaviour
<!-- Describe the current behaviour of the feature/component. -->
PANIC Alerter currently keeps the same mapping when no configs are recieved.
### 'To be' behaviour
<!-- Describe the intended behaviour of the feature/component. -->
When a channel is either deleted/removed from a configuration, the router must not include this mapping in the configuration.
### Steps to reproduce
<!-- What actions (in order) should the user go through to reproduce the issue? -->
1) Created a valid Substrate Polkadot chain on PANIC with a single valid Slack configuration.
2) Started getting alerts, so works as intended.
3) Removed the channel from this config.
4) `alert_router.log` still shows the following **after recieving a new alert**:
```
19/10/2022 01:46:28 PM - AlertRouter - INFO - ['634fff30a7d1b84a6b792d88']
19/10/2022 01:47:18 PM - AlertRouter - INFO - Obtaining list of channels to alert
19/10/2022 01:47:18 PM - AlertRouter - INFO - ['634fff30a7d1b84a6b792d88']
19/10/2022 01:49:38 PM - AlertRouter - INFO - Obtaining list of channels to alert
19/10/2022 01:49:38 PM - AlertRouter - INFO - ['634fff30a7d1b84a6b792d88']
```
### Acceptance criteria
<!-- This is a scenario/s which the bug fix must satisfy for the solution to be deemed successful. -->
**Given**: PANIC running with a configuration including a channel
**When**: Channel removed from the config
**Then**: Alert router receives the correct updated mapping
|
1.0
|
Alert Router must latest & updated mappings - <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺
v ✰ Thanks for opening an issue! ✰
v Before smashing the submit button please review the template.
☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -->
### Summary of bug
<!-- Concisely describe the bug in question and what effect this has on the end user or system. -->
### 'As is' behaviour
<!-- Describe the current behaviour of the feature/component. -->
PANIC Alerter currently keeps the same mapping when no configs are recieved.
### 'To be' behaviour
<!-- Describe the intended behaviour of the feature/component. -->
When a channel is either deleted/removed from a configuration, the router must not include this mapping in the configuration.
### Steps to reproduce
<!-- What actions (in order) should the user go through to reproduce the issue? -->
1) Created a valid Substrate Polkadot chain on PANIC with a single valid Slack configuration.
2) Started getting alerts, so works as intended.
3) Removed the channel from this config.
4) `alert_router.log` still shows the following **after recieving a new alert**:
```
19/10/2022 01:46:28 PM - AlertRouter - INFO - ['634fff30a7d1b84a6b792d88']
19/10/2022 01:47:18 PM - AlertRouter - INFO - Obtaining list of channels to alert
19/10/2022 01:47:18 PM - AlertRouter - INFO - ['634fff30a7d1b84a6b792d88']
19/10/2022 01:49:38 PM - AlertRouter - INFO - Obtaining list of channels to alert
19/10/2022 01:49:38 PM - AlertRouter - INFO - ['634fff30a7d1b84a6b792d88']
```
### Acceptance criteria
<!-- This is a scenario/s which the bug fix must satisfy for the solution to be deemed successful. -->
**Given**: PANIC running with a configuration including a channel
**When**: Channel removed from the config
**Then**: Alert router receives the correct updated mapping
|
non_defect
|
alert router must latest updated mappings ☺ v ✰ thanks for opening an issue ✰ v before smashing the submit button please review the template ☺ summary of bug as is behaviour panic alerter currently keeps the same mapping when no configs are recieved to be behaviour when a channel is either deleted removed from a configuration the router must not include this mapping in the configuration steps to reproduce created a valid substrate polkadot chain on panic with a single valid slack configuration started getting alerts so works as intended removed the channel from this config alert router log still shows the following after recieving a new alert pm alertrouter info pm alertrouter info obtaining list of channels to alert pm alertrouter info pm alertrouter info obtaining list of channels to alert pm alertrouter info acceptance criteria given panic running with a configuration including a channel when channel removed from the config then alert router receives the correct updated mapping
| 0
|
81,199
| 30,753,278,328
|
IssuesEvent
|
2023-07-28 21:42:07
|
zed-industries/community
|
https://api.github.com/repos/zed-industries/community
|
opened
|
Svelte "Zero-effort type safety" not working in `.ts` files
|
defect triage admin read
|
### Check for existing issues
- [X] Completed
### Describe the bug / provide steps to reproduce it
In SvelteKit projects, types are automatically inferred for many `export`ed values ([Zero-effort type safety](https://svelte.dev/blog/zero-config-type-safety)).
In `+page.svelte`, it works:
<img width="614" alt="image" src="https://github.com/zed-industries/community/assets/11315492/cb34705f-9d76-4bcb-be5c-df037edcd4d5">
But in `+page.ts` it does not:
<img width="780" alt="image" src="https://github.com/zed-industries/community/assets/11315492/d3cdda3f-1943-49d9-b7bc-88de82893c11">
### Environment
Zed: v0.96.3 (stable)
OS: macOS 13.5.0
Memory: 32 GiB
Architecture: x86_64
### If applicable, add mockups / screenshots to help explain present your vision of the feature
_No response_
### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue.
If you only need the most recent lines, you can run the `zed: open log` command palette action to see the last 1000.
_No response_
|
1.0
|
Svelte "Zero-effort type safety" not working in `.ts` files - ### Check for existing issues
- [X] Completed
### Describe the bug / provide steps to reproduce it
In SvelteKit projects, types are automatically inferred for many `export`ed values ([Zero-effort type safety](https://svelte.dev/blog/zero-config-type-safety)).
In `+page.svelte`, it works:
<img width="614" alt="image" src="https://github.com/zed-industries/community/assets/11315492/cb34705f-9d76-4bcb-be5c-df037edcd4d5">
But in `+page.ts` it does not:
<img width="780" alt="image" src="https://github.com/zed-industries/community/assets/11315492/d3cdda3f-1943-49d9-b7bc-88de82893c11">
### Environment
Zed: v0.96.3 (stable)
OS: macOS 13.5.0
Memory: 32 GiB
Architecture: x86_64
### If applicable, add mockups / screenshots to help explain present your vision of the feature
_No response_
### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue.
If you only need the most recent lines, you can run the `zed: open log` command palette action to see the last 1000.
_No response_
|
defect
|
svelte zero effort type safety not working in ts files check for existing issues completed describe the bug provide steps to reproduce it in sveltekit projects types are automatically inferred for many export ed values in page svelte it works img width alt image src but in page ts it does not img width alt image src environment zed stable os macos memory gib architecture if applicable add mockups screenshots to help explain present your vision of the feature no response if applicable attach your library logs zed zed log file to this issue if you only need the most recent lines you can run the zed open log command palette action to see the last no response
| 1
|
20,457
| 2,622,849,476
|
IssuesEvent
|
2015-03-04 08:04:23
|
max99x/pagemon-chrome-ext
|
https://api.github.com/repos/max99x/pagemon-chrome-ext
|
closed
|
Feature request: Ignore filter (regexp)
|
auto-migrated Priority-Medium Type-Enhancement
|
```
Great extension (oops forgot to rate it, will do immediately)
There are already some similar requests, i've noticed.
maybe possible to add in advanced mode a multiline field where one could enter
regexps to ignore changes (this would solve views, visitors, counters and etc.)
```
Original issue reported on code.google.com by `sh8an...@gmail.com` on 26 Jun 2011 at 1:18
|
1.0
|
Feature request: Ignore filter (regexp) - ```
Great extension (oops forgot to rate it, will do immediately)
There are already some similar requests, i've noticed.
maybe possible to add in advanced mode a multiline field where one could enter
regexps to ignore changes (this would solve views, visitors, counters and etc.)
```
Original issue reported on code.google.com by `sh8an...@gmail.com` on 26 Jun 2011 at 1:18
|
non_defect
|
feature request ignore filter regexp great extension oops forgot to rate it will do immediately there are already some similar requests i ve noticed maybe possible to add in advanced mode a multiline field where one could enter regexps to ignore changes this would solve views visitors counters and etc original issue reported on code google com by gmail com on jun at
| 0
|
116,494
| 14,970,149,213
|
IssuesEvent
|
2021-01-27 19:11:08
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
closed
|
Chapter 31 Orientation - Research and Implement Accessibility standards for video components
|
508/Accessibility CH31 design frontend vsa vsa-ebenefits
|
## Description
As an engineer, I want the components I build to be as accessible as possible, so the broadest spectrum of users possible can utilize my applications. Components that leverage video assets should be presented in an accessible way that includes the use of captions, the provision of a transcript, and the use of progressive enhancement.
## Considerations
- Use of the `<video>` tag to embed video
- Using source sets to provide thumbnails at varying screen sizes
- Find a way to include transcripts
## Tasks
- [x] Research the use of `<video>` tags and replace the use of iframes with them
- [x] Research and implement source sets for screen sizes
- [x] Explore whether iFrames should be removed in favor of `<video>` tags
## Acceptance criteria
- [ ] ~~Transcript has been provided for the user~~
- [x] A list of next steps for iteration has been determined
**Next steps**
- [ ] Varying screen sizes have been addressed
- [ ] Pull request has been made into `vets-website`
|
1.0
|
Chapter 31 Orientation - Research and Implement Accessibility standards for video components - ## Description
As an engineer, I want the components I build to be as accessible as possible, so the broadest spectrum of users possible can utilize my applications. Components that leverage video assets should be presented in an accessible way that includes the use of captions, the provision of a transcript, and the use of progressive enhancement.
## Considerations
- Use of the `<video>` tag to embed video
- Using source sets to provide thumbnails at varying screen sizes
- Find a way to include transcripts
## Tasks
- [x] Research the use of `<video>` tags and replace the use of iframes with them
- [x] Research and implement source sets for screen sizes
- [x] Explore whether iFrames should be removed in favor of `<video>` tags
## Acceptance criteria
- [ ] ~~Transcript has been provided for the user~~
- [x] A list of next steps for iteration has been determined
**Next steps**
- [ ] Varying screen sizes have been addressed
- [ ] Pull request has been made into `vets-website`
|
non_defect
|
chapter orientation research and implement accessibility standards for video components description as an engineer i want the components i build to be as accessible as possible so the broadest spectrum of users possible can utilize my applications components that leverage video assets should be presented in an accessible way that includes the use of captions the provision of a transcript and the use of progressive enhancement considerations use of the tag to embed video using source sets to provide thumbnails at varying screen sizes find a way to include transcripts tasks research the use of tags and replace the use of iframes with them research and implement source sets for screen sizes explore whether iframes should be removed in favor of tags acceptance criteria transcript has been provided for the user a list of next steps for iteration has been determined next steps varying screen sizes have been addressed pull request has been made into vets website
| 0
|
113,014
| 14,355,667,210
|
IssuesEvent
|
2020-11-30 10:25:20
|
opengovsg/postmangovsg
|
https://api.github.com/repos/opengovsg/postmangovsg
|
closed
|
Add instructions for deduplicating in excel
|
P1 chore design
|
We have decided to remove the deduplication feature in the frontend. We need to optimise the UX by adding a sentence to redirect users to our guide if they need help figuring out how to do deduplication in excel.
**Link to guide**: https://guide.postman.gov.sg/quick-start#remove-duplicates-in-excel
<img width="839" alt="Screenshot 2020-09-23 at 3 07 14 PM" src="https://user-images.githubusercontent.com/39106042/93978651-dd783180-fdae-11ea-97f3-d76d0da49c2b.png">
|
1.0
|
Add instructions for deduplicating in excel - We have decided to remove the deduplication feature in the frontend. We need to optimise the UX by adding a sentence to redirect users to our guide if they need help figuring out how to do deduplication in excel.
**Link to guide**: https://guide.postman.gov.sg/quick-start#remove-duplicates-in-excel
<img width="839" alt="Screenshot 2020-09-23 at 3 07 14 PM" src="https://user-images.githubusercontent.com/39106042/93978651-dd783180-fdae-11ea-97f3-d76d0da49c2b.png">
|
non_defect
|
add instructions for deduplicating in excel we have decided to remove the deduplication feature in the frontend we need to optimise the ux by adding a sentence to redirect users to our guide if they need help figuring out how to do deduplication in excel link to guide img width alt screenshot at pm src
| 0
|
622,065
| 19,605,992,567
|
IssuesEvent
|
2022-01-06 09:32:41
|
vignetteapp/vignette
|
https://api.github.com/repos/vignetteapp/vignette
|
closed
|
Investigate Project Reunion Integration
|
good first issue priority:low
|
In the next week we will be testing Windows 11 so we can test out our integration to Windows 11 early. We should investigate how we can publish to the Store. Right now, as part of the new Windows Apps SDK, we can use the MSIX format for Windows and distribute to MS Store, then Steam for others. This should allow us to be more flexible with our deployment pattern.
|
1.0
|
Investigate Project Reunion Integration - In the next week we will be testing Windows 11 so we can test out our integration to Windows 11 early. We should investigate how we can publish to the Store. Right now, as part of the new Windows Apps SDK, we can use the MSIX format for Windows and distribute to MS Store, then Steam for others. This should allow us to be more flexible with our deployment pattern.
|
non_defect
|
investigate project reunion integration in the next week we will be testing windows so we can test out our integration to windows early we should investigate how we can publish to the store right now as part of the new windows apps sdk we can use the msix format for windows and distribute to ms store then steam for others this should allow us to be more flexible with our deployment pattern
| 0
|
157,128
| 19,914,244,597
|
IssuesEvent
|
2022-01-25 20:34:53
|
dmyers87/boomstrap-react
|
https://api.github.com/repos/dmyers87/boomstrap-react
|
opened
|
WS-2018-0589 (Low) detected in nwmatcher-1.3.9.tgz
|
security vulnerability
|
## WS-2018-0589 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nwmatcher-1.3.9.tgz</b></p></summary>
<p>A CSS3-compliant JavaScript selector engine.</p>
<p>Library home page: <a href="https://registry.npmjs.org/nwmatcher/-/nwmatcher-1.3.9.tgz">https://registry.npmjs.org/nwmatcher/-/nwmatcher-1.3.9.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/nwmatcher/package.json</p>
<p>
Dependency Hierarchy:
- jest-cli-0.2.2.tgz (Root Library)
- jsdom-0.10.6.tgz
- :x: **nwmatcher-1.3.9.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/dmyers87/boomstrap-react/commit/56ff85f974b05cab00c2299011cfbdf611dd773d">56ff85f974b05cab00c2299011cfbdf611dd773d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Regular Expression vulnerability was found in nwmatcher before 1.4.4. The fix replacing multiple repeated instances of the "\s*" pattern.
<p>Publish Date: 2018-03-05
<p>URL: <a href=https://github.com/dperini/nwmatcher/commit/9dcc2b039beeabd18327a5ebaa537625872e16f0>WS-2018-0589</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/dperini/nwmatcher/commit/9dcc2b039beeabd18327a5ebaa537625872e16f0">https://github.com/dperini/nwmatcher/commit/9dcc2b039beeabd18327a5ebaa537625872e16f0</a></p>
<p>Release Date: 2018-03-05</p>
<p>Fix Resolution: 1.4.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"nwmatcher","packageVersion":"1.3.9","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"jest-cli:0.2.2;jsdom:0.10.6;nwmatcher:1.3.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.4","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2018-0589","vulnerabilityDetails":"A Regular Expression vulnerability was found in nwmatcher before 1.4.4. The fix replacing multiple repeated instances of the \"\\s*\" pattern.","vulnerabilityUrl":"https://github.com/dperini/nwmatcher/commit/9dcc2b039beeabd18327a5ebaa537625872e16f0","cvss3Severity":"low","cvss3Score":"3.7","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
WS-2018-0589 (Low) detected in nwmatcher-1.3.9.tgz - ## WS-2018-0589 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nwmatcher-1.3.9.tgz</b></p></summary>
<p>A CSS3-compliant JavaScript selector engine.</p>
<p>Library home page: <a href="https://registry.npmjs.org/nwmatcher/-/nwmatcher-1.3.9.tgz">https://registry.npmjs.org/nwmatcher/-/nwmatcher-1.3.9.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/nwmatcher/package.json</p>
<p>
Dependency Hierarchy:
- jest-cli-0.2.2.tgz (Root Library)
- jsdom-0.10.6.tgz
- :x: **nwmatcher-1.3.9.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/dmyers87/boomstrap-react/commit/56ff85f974b05cab00c2299011cfbdf611dd773d">56ff85f974b05cab00c2299011cfbdf611dd773d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Regular Expression vulnerability was found in nwmatcher before 1.4.4. The fix replacing multiple repeated instances of the "\s*" pattern.
<p>Publish Date: 2018-03-05
<p>URL: <a href=https://github.com/dperini/nwmatcher/commit/9dcc2b039beeabd18327a5ebaa537625872e16f0>WS-2018-0589</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/dperini/nwmatcher/commit/9dcc2b039beeabd18327a5ebaa537625872e16f0">https://github.com/dperini/nwmatcher/commit/9dcc2b039beeabd18327a5ebaa537625872e16f0</a></p>
<p>Release Date: 2018-03-05</p>
<p>Fix Resolution: 1.4.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"nwmatcher","packageVersion":"1.3.9","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"jest-cli:0.2.2;jsdom:0.10.6;nwmatcher:1.3.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.4","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2018-0589","vulnerabilityDetails":"A Regular Expression vulnerability was found in nwmatcher before 1.4.4. The fix replacing multiple repeated instances of the \"\\s*\" pattern.","vulnerabilityUrl":"https://github.com/dperini/nwmatcher/commit/9dcc2b039beeabd18327a5ebaa537625872e16f0","cvss3Severity":"low","cvss3Score":"3.7","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_defect
|
ws low detected in nwmatcher tgz ws low severity vulnerability vulnerable library nwmatcher tgz a compliant javascript selector engine library home page a href path to dependency file package json path to vulnerable library node modules nwmatcher package json dependency hierarchy jest cli tgz root library jsdom tgz x nwmatcher tgz vulnerable library found in head commit a href found in base branch master vulnerability details a regular expression vulnerability was found in nwmatcher before the fix replacing multiple repeated instances of the s pattern publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree jest cli jsdom nwmatcher isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier ws vulnerabilitydetails a regular expression vulnerability was found in nwmatcher before the fix replacing multiple repeated instances of the s pattern vulnerabilityurl
| 0
|
69,928
| 22,751,011,314
|
IssuesEvent
|
2022-07-07 13:08:34
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Event tile info line lost left padding
|
T-Defect
|
### Steps to reproduce
1. Open https://matrix.to/#/!bijaLdadorKgNGtHdA:matrix.org/$ymlC3BbdUs7EdOeT6IxRy2p9Kn4I0PjsJvUwRQ5H1s8?via=matrix.org&via=tchncs.de&via=envs.net
2. Select a near message
### Outcome
#### What did you expect?
The info message should have the left padding.

#### What happened instead?
The padding is gone due to `:not()` pseudo class which increases specificity.

### Operating system
Debian
### Browser information
Firefox ESR 91
### URL for webapp
localhost
### Application version
develop branch
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
Event tile info line lost left padding - ### Steps to reproduce
1. Open https://matrix.to/#/!bijaLdadorKgNGtHdA:matrix.org/$ymlC3BbdUs7EdOeT6IxRy2p9Kn4I0PjsJvUwRQ5H1s8?via=matrix.org&via=tchncs.de&via=envs.net
2. Select a near message
### Outcome
#### What did you expect?
The info message should have the left padding.

#### What happened instead?
The padding is gone due to `:not()` pseudo class which increases specificity.

### Operating system
Debian
### Browser information
Firefox ESR 91
### URL for webapp
localhost
### Application version
develop branch
### Homeserver
_No response_
### Will you send logs?
No
|
defect
|
event tile info line lost left padding steps to reproduce open select a near message outcome what did you expect the info message should have the left padding what happened instead the padding is gone due to not pseudo class which increases specificity operating system debian browser information firefox esr url for webapp localhost application version develop branch homeserver no response will you send logs no
| 1
|
26,397
| 4,218,417,454
|
IssuesEvent
|
2016-06-30 15:53:29
|
xcat2/xcat-core
|
https://api.github.com/repos/xcat2/xcat-core
|
closed
|
bmcdiscover using range is not working as expected
|
priority:high status:pending test:testcase_requested type:bug xCAT 2.12.1 Sprint 2
|
For the bmcdiscover command, we allow a `--range` to be used:
```
--range Specify one or more IP ranges acceptable to nmap. \
IP rance can be hostnames, IP addresses, networks, etc. \
A single IP address (10.1.2.3) or an IP range
```
However, running this to specify a range and IP address doesn't work as expected:
1 Single IPs work
```
[root@fs1 discovery]# bmcdiscover --range 50.5.31.1
50.5.31.1,8335-GTB,100469A,,
[root@fs1 discovery]# bmcdiscover --range 192.168.5.165
192.168.5.165,8335-GTB,100464A,,
```
2 Using commas does not work
```
[root@fs1 discovery]# bmcdiscover --range 192.168.5.165,50.5.31.1
Error: No bmc found.
```
3 Using space separated IP address does not return both BMCs
```
[root@fs1 discovery]# bmcdiscover --range 192.168.5.165 50.5.31.1
192.168.5.165,8335-GTB,100464A,,
[root@fs1 discovery]#
```
|
2.0
|
bmcdiscover using range is not working as expected - For the bmcdiscover command, we allow a `--range` to be used:
```
--range Specify one or more IP ranges acceptable to nmap. \
IP rance can be hostnames, IP addresses, networks, etc. \
A single IP address (10.1.2.3) or an IP range
```
However, running this to specify a range and IP address doesn't work as expected:
1 Single IPs work
```
[root@fs1 discovery]# bmcdiscover --range 50.5.31.1
50.5.31.1,8335-GTB,100469A,,
[root@fs1 discovery]# bmcdiscover --range 192.168.5.165
192.168.5.165,8335-GTB,100464A,,
```
2 Using commas does not work
```
[root@fs1 discovery]# bmcdiscover --range 192.168.5.165,50.5.31.1
Error: No bmc found.
```
3 Using space separated IP address does not return both BMCs
```
[root@fs1 discovery]# bmcdiscover --range 192.168.5.165 50.5.31.1
192.168.5.165,8335-GTB,100464A,,
[root@fs1 discovery]#
```
|
non_defect
|
bmcdiscover using range is not working as expected for the bmcdiscover command we allow a range to be used range specify one or more ip ranges acceptable to nmap ip rance can be hostnames ip addresses networks etc a single ip address or an ip range however running this to specify a range and ip address doesn t work as expected single ips work bmcdiscover range gtb bmcdiscover range gtb using commas does not work bmcdiscover range error no bmc found using space separated ip address does not return both bmcs bmcdiscover range gtb
| 0
|
137,751
| 12,793,102,224
|
IssuesEvent
|
2020-07-02 03:06:11
|
IllumiDesk/illumidesk
|
https://api.github.com/repos/IllumiDesk/illumidesk
|
closed
|
Add docstrings and types to lms_grades.py
|
documentation
|
Add docstrings and types to all classes and functions defined in `src/illumidesk/handlers/lms_grades.py`.
|
1.0
|
Add docstrings and types to lms_grades.py - Add docstrings and types to all classes and functions defined in `src/illumidesk/handlers/lms_grades.py`.
|
non_defect
|
add docstrings and types to lms grades py add docstrings and types to all classes and functions defined in src illumidesk handlers lms grades py
| 0
|
36,695
| 8,068,258,099
|
IssuesEvent
|
2018-08-05 18:12:37
|
spockframework/spock
|
https://api.github.com/repos/spockframework/spock
|
closed
|
Mock invocation order is based on method return order, not invocation
|
Module-Core Type-Defect
|
Originally reported on Google Code with ID 353
```
This fails:
def "spock bug"() {
given:
def r1 = Mock(Runnable)
def r2 = Mock(Runnable)
def r3 = { r1.run() }
when:
r3()
then:
1 * r1.run() >> { r2.run() }
then:
1 * r2.run()
}
While this passes:
def "spock bug"() {
given:
def r1 = Mock(Runnable)
def r2 = Mock(Runnable)
def r3 = { r1.run() }
when:
r3()
then:
1 * r2.run()
then:
1 * r1.run() >> { r2.run() }
}
Spock thinks r2.run() was called first because it returned before r1.run().
```
Reported by `luke@gradleware.com` on 2014-03-12 21:53:34
|
1.0
|
Mock invocation order is based on method return order, not invocation - Originally reported on Google Code with ID 353
```
This fails:
def "spock bug"() {
given:
def r1 = Mock(Runnable)
def r2 = Mock(Runnable)
def r3 = { r1.run() }
when:
r3()
then:
1 * r1.run() >> { r2.run() }
then:
1 * r2.run()
}
While this passes:
def "spock bug"() {
given:
def r1 = Mock(Runnable)
def r2 = Mock(Runnable)
def r3 = { r1.run() }
when:
r3()
then:
1 * r2.run()
then:
1 * r1.run() >> { r2.run() }
}
Spock thinks r2.run() was called first because it returned before r1.run().
```
Reported by `luke@gradleware.com` on 2014-03-12 21:53:34
|
defect
|
mock invocation order is based on method return order not invocation originally reported on google code with id this fails def spock bug given def mock runnable def mock runnable def run when then run run then run while this passes def spock bug given def mock runnable def mock runnable def run when then run then run run spock thinks run was called first because it returned before run reported by luke gradleware com on
| 1
|
37,818
| 8,519,589,672
|
IssuesEvent
|
2018-11-01 15:05:04
|
GoldenSoftwareLtd/gedemin
|
https://api.github.com/repos/GoldenSoftwareLtd/gedemin
|
closed
|
01. Договор на услуги
|
Accountancy Priority-Medium Type-Defect
|
Originally reported on Google Code with ID 1896
```
- в договоре автоматически не заполняется, а в счет-фактуре необходимо
вставить поле \"ОКПО\",
- в счет-фактуре втавить поле \"адрес плательщики\",
- в договоре, счет-фактуре вставить поле \"почтовый индекс Исполнителя и
Заказчика,
- в счет-фактуре и протоколе вместо Поставщик писать Исполнитель (по
аналогии с договором и актом),
- в протоколе в шапке писать: согласования свободных отпускных цен на
товары, тарифов на работы (услуги) между ООО \"Рога и копыта\" и УП \"
Копыта и рога\" на поставку товара, выполнение (оказание) работы (услуги):,
- в договоре вставить поле \"в том числе НДС(20%)\" - прописью,
- в договоре (в таблице) не округляется \"сумма с НДС\"
- в счет фактуре (при оказании нескольких услуг) не суммируются \"Общая
сумма НДС\" и \"Общая сумма\".
```
Reported by `gs1994` on 2010-02-25 10:10:26
|
1.0
|
01. Договор на услуги - Originally reported on Google Code with ID 1896
```
- в договоре автоматически не заполняется, а в счет-фактуре необходимо
вставить поле \"ОКПО\",
- в счет-фактуре втавить поле \"адрес плательщики\",
- в договоре, счет-фактуре вставить поле \"почтовый индекс Исполнителя и
Заказчика,
- в счет-фактуре и протоколе вместо Поставщик писать Исполнитель (по
аналогии с договором и актом),
- в протоколе в шапке писать: согласования свободных отпускных цен на
товары, тарифов на работы (услуги) между ООО \"Рога и копыта\" и УП \"
Копыта и рога\" на поставку товара, выполнение (оказание) работы (услуги):,
- в договоре вставить поле \"в том числе НДС(20%)\" - прописью,
- в договоре (в таблице) не округляется \"сумма с НДС\"
- в счет фактуре (при оказании нескольких услуг) не суммируются \"Общая
сумма НДС\" и \"Общая сумма\".
```
Reported by `gs1994` on 2010-02-25 10:10:26
|
defect
|
договор на услуги originally reported on google code with id в договоре автоматически не заполняется а в счет фактуре необходимо вставить поле окпо в счет фактуре втавить поле адрес плательщики в договоре счет фактуре вставить поле почтовый индекс исполнителя и заказчика в счет фактуре и протоколе вместо поставщик писать исполнитель по аналогии с договором и актом в протоколе в шапке писать согласования свободных отпускных цен на товары тарифов на работы услуги между ооо рога и копыта и уп копыта и рога на поставку товара выполнение оказание работы услуги в договоре вставить поле в том числе ндс прописью в договоре в таблице не округляется сумма с ндс в счет фактуре при оказании нескольких услуг не суммируются общая сумма ндс и общая сумма reported by on
| 1
|
81,916
| 31,810,319,251
|
IssuesEvent
|
2023-09-13 16:22:02
|
SeleniumHQ/selenium
|
https://api.github.com/repos/SeleniumHQ/selenium
|
opened
|
[🐛 Bug]: Testing WebView2 with automatic download of the Edge WebDriver no longer works
|
I-defect needs-triaging
|
### What happened?
We are using Selenium for testing code inside a WebView2 instance from C# code. Originally we used the below code to create the `EdgeDriver`, as per [Microsoft documentation](https://learn.microsoft.com/en-us/microsoft-edge/webview2/how-to/webdriver#attaching-microsoft-edge-webdriver-to-the-launched-webview2-app):
```csharp
using OpenQA.Selenium.Edge;
var options = new EdgeOptions { DebuggerAddress = "localhost:12345", UseWebView = true };
using var driver = new EdgeDriver(options);
```
This broke with (I believe) version 4.9, producing the following error:
```
Unhandled exception. OpenQA.Selenium.WebDriverException: Unable to locate driver with path: msedgedriver.exe, for more information on how to install drivers see https://www.selenium.dev/documentation/webdriver/getting_started/instal
l_drivers/
---> OpenQA.Selenium.WebDriverException: Invalid response from process (code 65): selenium-manager/windows/selenium-manager.exe --browser "webview2" --output json
Invalid browser name: webview2
at OpenQA.Selenium.SeleniumManager.RunCommand(String fileName, String arguments)
at OpenQA.Selenium.SeleniumManager.DriverPath(DriverOptions options)
at OpenQA.Selenium.DriverFinder.VerifyDriverServicePath(DriverService service, DriverOptions options)
--- End of inner exception stack trace ---
at OpenQA.Selenium.DriverFinder.VerifyDriverServicePath(DriverService service, DriverOptions options)
at OpenQA.Selenium.Edge.EdgeDriverService.CreateDefaultService(EdgeOptions options)
at OpenQA.Selenium.Edge.EdgeDriver..ctor(EdgeOptions options)
at Program.<Main>$(String[] args)
```
Assuming at the time that this was an intentional change, we removed the `UseWebView` property and instead added `BrowserVersion`. The latter was necessary in order to avoid Selenium picking up the version of the actual Edge browser (which might differ from the used WebView2 version for various reasons) as the basis for the Edge WebDriver version it downloads.
```csharp
using OpenQA.Selenium.Edge;
var options = new EdgeOptions { DebuggerAddress = "localhost:12345", BrowserVersion = "116" };
using var driver = new EdgeDriver(options);
```
This however now also broke with Selenium 4.12.* and produces the following error:
```
Unhandled exception. OpenQA.Selenium.NoSuchDriverException: Unable to obtain MicrosoftEdge using Selenium Manager; For documentation on this error, please visit: https://www.selenium.dev/documentation/webdriver/troubleshooting/error
s/driver_location
---> OpenQA.Selenium.WebDriverException: Error starting process: D:\projects\Eng21\infra\workitems\6372-deps-update\ConsoleApp\bin\Debug\net6.0\selenium-manager/windows/selenium-manager.exe --browser "MicrosoftEdge" --output json
--browser-version 116
---> OpenQA.Selenium.WebDriverException: Selenium Manager process exited abnormally with 65 code: D:\projects\Eng21\infra\workitems\6372-deps-update\ConsoleApp\bin\Debug\net6.0\selenium-manager/windows/selenium-manager.exe --brows
er "MicrosoftEdge" --output json --browser-version 116
Error Output >>
Standard Output >>
{
"logs": [
{
"level": "ERROR",
"timestamp": 1694621742,
"message": "edge 116 cannot be downloaded"
}
],
"result": {
"code": 65,
"message": "edge 116 cannot be downloaded",
"driver_path": "",
"browser_path": ""
}
}
at OpenQA.Selenium.SeleniumManager.RunCommand(String fileName, String arguments)
--- End of inner exception stack trace ---
at OpenQA.Selenium.SeleniumManager.RunCommand(String fileName, String arguments)
at OpenQA.Selenium.SeleniumManager.DriverPath(DriverOptions options)
at OpenQA.Selenium.DriverFinder.FullPath(DriverOptions options)
--- End of inner exception stack trace ---
at OpenQA.Selenium.DriverFinder.FullPath(DriverOptions options)
at OpenQA.Selenium.Chromium.ChromiumDriver.GenerateDriverServiceCommandExecutor(DriverService service, DriverOptions options, TimeSpan commandTimeout)
at OpenQA.Selenium.Chromium.ChromiumDriver..ctor(ChromiumDriverService service, ChromiumOptions options, TimeSpan commandTimeout)
at OpenQA.Selenium.Edge.EdgeDriver..ctor(EdgeDriverService service, EdgeOptions options, TimeSpan commandTimeout)
at OpenQA.Selenium.Edge.EdgeDriver..ctor(EdgeDriverService service, EdgeOptions options)
at OpenQA.Selenium.Edge.EdgeDriver..ctor(EdgeOptions options)
at Program.<Main>$(String[] args)
```
### How can we reproduce the issue?
```shell
See code examples in the above description.
```
### Relevant log output
```shell
According to the documentation there is no logger for .NET, so I am not sure where to get logs, aside from the exception stack traces above.
```
### Operating System
Windows 10, Windows 11
### Selenium version
4.12.4
### What are the browser(s) and version(s) where you see this issue?
Edge WebView2 116
### What are the browser driver(s) and version(s) where you see this issue?
Whichever one Selenium downloads (or doesn't in this case)
### Are you using Selenium Grid?
No
|
1.0
|
[🐛 Bug]: Testing WebView2 with automatic download of the Edge WebDriver no longer works - ### What happened?
We are using Selenium for testing code inside a WebView2 instance from C# code. Originally we used the below code to create the `EdgeDriver`, as per [Microsoft documentation](https://learn.microsoft.com/en-us/microsoft-edge/webview2/how-to/webdriver#attaching-microsoft-edge-webdriver-to-the-launched-webview2-app):
```csharp
using OpenQA.Selenium.Edge;
var options = new EdgeOptions { DebuggerAddress = "localhost:12345", UseWebView = true };
using var driver = new EdgeDriver(options);
```
This broke with (I believe) version 4.9, producing the following error:
```
Unhandled exception. OpenQA.Selenium.WebDriverException: Unable to locate driver with path: msedgedriver.exe, for more information on how to install drivers see https://www.selenium.dev/documentation/webdriver/getting_started/instal
l_drivers/
---> OpenQA.Selenium.WebDriverException: Invalid response from process (code 65): selenium-manager/windows/selenium-manager.exe --browser "webview2" --output json
Invalid browser name: webview2
at OpenQA.Selenium.SeleniumManager.RunCommand(String fileName, String arguments)
at OpenQA.Selenium.SeleniumManager.DriverPath(DriverOptions options)
at OpenQA.Selenium.DriverFinder.VerifyDriverServicePath(DriverService service, DriverOptions options)
--- End of inner exception stack trace ---
at OpenQA.Selenium.DriverFinder.VerifyDriverServicePath(DriverService service, DriverOptions options)
at OpenQA.Selenium.Edge.EdgeDriverService.CreateDefaultService(EdgeOptions options)
at OpenQA.Selenium.Edge.EdgeDriver..ctor(EdgeOptions options)
at Program.<Main>$(String[] args)
```
Assuming at the time that this was an intentional change, we removed the `UseWebView` property and instead added `BrowserVersion`. The latter was necessary in order to avoid Selenium picking up the version of the actual Edge browser (which might differ from the used WebView2 version for various reasons) as the basis for the Edge WebDriver version it downloads.
```csharp
using OpenQA.Selenium.Edge;
var options = new EdgeOptions { DebuggerAddress = "localhost:12345", BrowserVersion = "116" };
using var driver = new EdgeDriver(options);
```
This however now also broke with Selenium 4.12.* and produces the following error:
```
Unhandled exception. OpenQA.Selenium.NoSuchDriverException: Unable to obtain MicrosoftEdge using Selenium Manager; For documentation on this error, please visit: https://www.selenium.dev/documentation/webdriver/troubleshooting/error
s/driver_location
---> OpenQA.Selenium.WebDriverException: Error starting process: D:\projects\Eng21\infra\workitems\6372-deps-update\ConsoleApp\bin\Debug\net6.0\selenium-manager/windows/selenium-manager.exe --browser "MicrosoftEdge" --output json
--browser-version 116
---> OpenQA.Selenium.WebDriverException: Selenium Manager process exited abnormally with 65 code: D:\projects\Eng21\infra\workitems\6372-deps-update\ConsoleApp\bin\Debug\net6.0\selenium-manager/windows/selenium-manager.exe --brows
er "MicrosoftEdge" --output json --browser-version 116
Error Output >>
Standard Output >>
{
"logs": [
{
"level": "ERROR",
"timestamp": 1694621742,
"message": "edge 116 cannot be downloaded"
}
],
"result": {
"code": 65,
"message": "edge 116 cannot be downloaded",
"driver_path": "",
"browser_path": ""
}
}
at OpenQA.Selenium.SeleniumManager.RunCommand(String fileName, String arguments)
--- End of inner exception stack trace ---
at OpenQA.Selenium.SeleniumManager.RunCommand(String fileName, String arguments)
at OpenQA.Selenium.SeleniumManager.DriverPath(DriverOptions options)
at OpenQA.Selenium.DriverFinder.FullPath(DriverOptions options)
--- End of inner exception stack trace ---
at OpenQA.Selenium.DriverFinder.FullPath(DriverOptions options)
at OpenQA.Selenium.Chromium.ChromiumDriver.GenerateDriverServiceCommandExecutor(DriverService service, DriverOptions options, TimeSpan commandTimeout)
at OpenQA.Selenium.Chromium.ChromiumDriver..ctor(ChromiumDriverService service, ChromiumOptions options, TimeSpan commandTimeout)
at OpenQA.Selenium.Edge.EdgeDriver..ctor(EdgeDriverService service, EdgeOptions options, TimeSpan commandTimeout)
at OpenQA.Selenium.Edge.EdgeDriver..ctor(EdgeDriverService service, EdgeOptions options)
at OpenQA.Selenium.Edge.EdgeDriver..ctor(EdgeOptions options)
at Program.<Main>$(String[] args)
```
### How can we reproduce the issue?
```shell
See code examples in the above description.
```
### Relevant log output
```shell
According to the documentation there is no logger for .NET, so I am not sure where to get logs, aside from the exception stack traces above.
```
### Operating System
Windows 10, Windows 11
### Selenium version
4.12.4
### What are the browser(s) and version(s) where you see this issue?
Edge WebView2 116
### What are the browser driver(s) and version(s) where you see this issue?
Whichever one Selenium downloads (or doesn't in this case)
### Are you using Selenium Grid?
No
|
defect
|
testing with automatic download of the edge webdriver no longer works what happened we are using selenium for testing code inside a instance from c code originally we used the below code to create the edgedriver as per csharp using openqa selenium edge var options new edgeoptions debuggeraddress localhost usewebview true using var driver new edgedriver options this broke with i believe version producing the following error unhandled exception openqa selenium webdriverexception unable to locate driver with path msedgedriver exe for more information on how to install drivers see l drivers openqa selenium webdriverexception invalid response from process code selenium manager windows selenium manager exe browser output json invalid browser name at openqa selenium seleniummanager runcommand string filename string arguments at openqa selenium seleniummanager driverpath driveroptions options at openqa selenium driverfinder verifydriverservicepath driverservice service driveroptions options end of inner exception stack trace at openqa selenium driverfinder verifydriverservicepath driverservice service driveroptions options at openqa selenium edge edgedriverservice createdefaultservice edgeoptions options at openqa selenium edge edgedriver ctor edgeoptions options at program string args assuming at the time that this was an intentional change we removed the usewebview property and instead added browserversion the latter was necessary in order to avoid selenium picking up the version of the actual edge browser which might differ from the used version for various reasons as the basis for the edge webdriver version it downloads csharp using openqa selenium edge var options new edgeoptions debuggeraddress localhost browserversion using var driver new edgedriver options this however now also broke with selenium and produces the following error unhandled exception openqa selenium nosuchdriverexception unable to obtain microsoftedge using selenium manager for documentation on this error please visit s driver location openqa selenium webdriverexception error starting process d projects infra workitems deps update consoleapp bin debug selenium manager windows selenium manager exe browser microsoftedge output json browser version openqa selenium webdriverexception selenium manager process exited abnormally with code d projects infra workitems deps update consoleapp bin debug selenium manager windows selenium manager exe brows er microsoftedge output json browser version error output standard output logs level error timestamp message edge cannot be downloaded result code message edge cannot be downloaded driver path browser path at openqa selenium seleniummanager runcommand string filename string arguments end of inner exception stack trace at openqa selenium seleniummanager runcommand string filename string arguments at openqa selenium seleniummanager driverpath driveroptions options at openqa selenium driverfinder fullpath driveroptions options end of inner exception stack trace at openqa selenium driverfinder fullpath driveroptions options at openqa selenium chromium chromiumdriver generatedriverservicecommandexecutor driverservice service driveroptions options timespan commandtimeout at openqa selenium chromium chromiumdriver ctor chromiumdriverservice service chromiumoptions options timespan commandtimeout at openqa selenium edge edgedriver ctor edgedriverservice service edgeoptions options timespan commandtimeout at openqa selenium edge edgedriver ctor edgedriverservice service edgeoptions options at openqa selenium edge edgedriver ctor edgeoptions options at program string args how can we reproduce the issue shell see code examples in the above description relevant log output shell according to the documentation there is no logger for net so i am not sure where to get logs aside from the exception stack traces above operating system windows windows selenium version what are the browser s and version s where you see this issue edge what are the browser driver s and version s where you see this issue whichever one selenium downloads or doesn t in this case are you using selenium grid no
| 1
|
49,784
| 13,187,269,753
|
IssuesEvent
|
2020-08-13 02:52:53
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
steamshovel times relative to particle (Trac #2061)
|
Incomplete Migration Migrated from Trac combo core defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2061">https://code.icecube.wisc.edu/ticket/2061</a>, reported by jgonzalez and owned by sander.vanheule</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2017-08-18T07:22:34",
"description": "When I set the event times according to a particle (lower-left drop down menu), it seems that steamshovel now assumes that the time interval starts at a depth of about 1000 m. This is convenient for in-ice, but not so for IceTop since the interval starts 3 microseconds late. Is this hardcoded somewhere? Each particle has a vertex properly set, perhaps one can use that vertex somehow?\n\nI'm using offline-software V17-05-00.",
"reporter": "jgonzalez",
"cc": "",
"resolution": "fixed",
"_ts": "1503040954763004",
"component": "combo core",
"summary": "steamshovel times relative to particle",
"priority": "normal",
"keywords": "",
"time": "2017-08-01T15:52:29",
"milestone": "",
"owner": "sander.vanheule",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
steamshovel times relative to particle (Trac #2061) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2061">https://code.icecube.wisc.edu/ticket/2061</a>, reported by jgonzalez and owned by sander.vanheule</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2017-08-18T07:22:34",
"description": "When I set the event times according to a particle (lower-left drop down menu), it seems that steamshovel now assumes that the time interval starts at a depth of about 1000 m. This is convenient for in-ice, but not so for IceTop since the interval starts 3 microseconds late. Is this hardcoded somewhere? Each particle has a vertex properly set, perhaps one can use that vertex somehow?\n\nI'm using offline-software V17-05-00.",
"reporter": "jgonzalez",
"cc": "",
"resolution": "fixed",
"_ts": "1503040954763004",
"component": "combo core",
"summary": "steamshovel times relative to particle",
"priority": "normal",
"keywords": "",
"time": "2017-08-01T15:52:29",
"milestone": "",
"owner": "sander.vanheule",
"type": "defect"
}
```
</p>
</details>
|
defect
|
steamshovel times relative to particle trac migrated from json status closed changetime description when i set the event times according to a particle lower left drop down menu it seems that steamshovel now assumes that the time interval starts at a depth of about m this is convenient for in ice but not so for icetop since the interval starts microseconds late is this hardcoded somewhere each particle has a vertex properly set perhaps one can use that vertex somehow n ni m using offline software reporter jgonzalez cc resolution fixed ts component combo core summary steamshovel times relative to particle priority normal keywords time milestone owner sander vanheule type defect
| 1
|
49,313
| 13,186,606,247
|
IssuesEvent
|
2020-08-13 00:43:28
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
[cscd-llh] needs more tests (Trac #1168)
|
Incomplete Migration Migrated from Trac combo reconstruction defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1168">https://code.icecube.wisc.edu/ticket/1168</a>, reported by hdembinski and owned by tpalczewski</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-08-21T13:27:59",
"description": "the project has some cpp-tests, but the pybindings are not tested. the script CscdLlhTest.py is practically a test should be converted into a test, by moving it to resources/test/test_cscd_llh.py\n\ncurrently, the test coverage is rather low, see\nhttp://software.icecube.wisc.edu/coverage/00_LATEST/\n\nespecially for\nprivate/converter\ncscd-llh/private/pdf\ncscd-llh/public/cscd-llh/pdf\n",
"reporter": "hdembinski",
"cc": "",
"resolution": "fixed",
"_ts": "1440163679305068",
"component": "combo reconstruction",
"summary": "[cscd-llh] needs more tests",
"priority": "normal",
"keywords": "",
"time": "2015-08-18T20:31:03",
"milestone": "",
"owner": "tpalczewski",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[cscd-llh] needs more tests (Trac #1168) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1168">https://code.icecube.wisc.edu/ticket/1168</a>, reported by hdembinski and owned by tpalczewski</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-08-21T13:27:59",
"description": "the project has some cpp-tests, but the pybindings are not tested. the script CscdLlhTest.py is practically a test should be converted into a test, by moving it to resources/test/test_cscd_llh.py\n\ncurrently, the test coverage is rather low, see\nhttp://software.icecube.wisc.edu/coverage/00_LATEST/\n\nespecially for\nprivate/converter\ncscd-llh/private/pdf\ncscd-llh/public/cscd-llh/pdf\n",
"reporter": "hdembinski",
"cc": "",
"resolution": "fixed",
"_ts": "1440163679305068",
"component": "combo reconstruction",
"summary": "[cscd-llh] needs more tests",
"priority": "normal",
"keywords": "",
"time": "2015-08-18T20:31:03",
"milestone": "",
"owner": "tpalczewski",
"type": "defect"
}
```
</p>
</details>
|
defect
|
needs more tests trac migrated from json status closed changetime description the project has some cpp tests but the pybindings are not tested the script cscdllhtest py is practically a test should be converted into a test by moving it to resources test test cscd llh py n ncurrently the test coverage is rather low see n for nprivate converter ncscd llh private pdf ncscd llh public cscd llh pdf n reporter hdembinski cc resolution fixed ts component combo reconstruction summary needs more tests priority normal keywords time milestone owner tpalczewski type defect
| 1
|
37,763
| 8,515,042,808
|
IssuesEvent
|
2018-10-31 20:22:02
|
Gubaer/josm-scripting-plugin
|
https://api.github.com/repos/Gubaer/josm-scripting-plugin
|
closed
|
Cannot find method HelpAwareOptionPane::showOptionDialog with latest JOSM
|
defect
|
It works with 14178, but it fails with latest JOSM version 14382:

I do not see any changes in HelpAwareOptionPane.java file since May, though.
|
1.0
|
Cannot find method HelpAwareOptionPane::showOptionDialog with latest JOSM - It works with 14178, but it fails with latest JOSM version 14382:

I do not see any changes in HelpAwareOptionPane.java file since May, though.
|
defect
|
cannot find method helpawareoptionpane showoptiondialog with latest josm it works with but it fails with latest josm version i do not see any changes in helpawareoptionpane java file since may though
| 1
|
750,438
| 26,202,054,428
|
IssuesEvent
|
2023-01-03 18:27:07
|
jernejvivod/mimic-iii-explorer
|
https://api.github.com/repos/jernejvivod/mimic-iii-explorer
|
closed
|
Implement edge-case tests for MimicEntityManager
|
Priority: LOW
|
Write tests for functionality with non-existent IDs, empty foreign-key paths, and invalid entity names.
|
1.0
|
Implement edge-case tests for MimicEntityManager - Write tests for functionality with non-existent IDs, empty foreign-key paths, and invalid entity names.
|
non_defect
|
implement edge case tests for mimicentitymanager write tests for functionality with non existent ids empty foreign key paths and invalid entity names
| 0
|
440,431
| 30,745,356,604
|
IssuesEvent
|
2023-07-28 14:39:13
|
kartoza/SAEOSS-Portal
|
https://api.github.com/repos/kartoza/SAEOSS-Portal
|
opened
|
Documentation: Developer mkdocstring email notification
|
documentation
|
Update mkdocstrings for email notification
|
1.0
|
Documentation: Developer mkdocstring email notification - Update mkdocstrings for email notification
|
non_defect
|
documentation developer mkdocstring email notification update mkdocstrings for email notification
| 0
|
18,087
| 10,880,816,139
|
IssuesEvent
|
2019-11-17 13:49:47
|
terraform-providers/terraform-provider-azurerm
|
https://api.github.com/repos/terraform-providers/terraform-provider-azurerm
|
closed
|
private_ip_address not filled in azurerm_network_interface resource and datasource
|
bug good first issue service/network-interfaces
|
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
### Terraform (and AzureRM Provider) Version
Terraform v0.12.12
terraform-provider-azurerm_v1.34.0_x4
### Affected Resource(s)
Resource azurerm_network_interface
Datasource azurerm_network_interface
### Terraform Configuration Files
```hcl
resource "azurerm_network_interface" "nic_0" {
name = "${var.hostname}-nic_0"
location = data.azurerm_resource_group.rg_vm.location
resource_group_name = data.azurerm_resource_group.rg_vm.name
ip_configuration {
name = "ip-configuration"
subnet_id = data.azurerm_subnet.subnet.id
private_ip_address_allocation = "dynamic"
primary = "true"
}
ip_configuration {
name = "ip-configuration2"
subnet_id = data.azurerm_subnet.subnet.id
private_ip_address_allocation = "dynamic"
}
}
data "azurerm_network_interface" "nic_0" {
name = "${var.hostname}-nic_0"
resource_group_name = data.azurerm_resource_group.rg_vm.name
}
output "resource_ip_configuration" {
value = azurerm_network_interface.nic_0.ip_configuration
}
output "data_ip_configuration" {
value = data.azurerm_network_interface.nic_0.ip_configuration
}
```
### Expected Behavior
All Attributes of ìp_configuration` should be filled and shown in outputs.
### Actual Behavior
`private_ip_address` is not filled:
```
data_ip_configuration = [
{
"application_gateway_backend_address_pools_ids" = []
"application_security_group_ids" = []
"load_balancer_backend_address_pools_ids" = []
"load_balancer_inbound_nat_rules_ids" = []
"name" = "ip-configuration"
"primary" = true
"private_ip_address" = ""
"private_ip_address_allocation" = "dynamic"
"private_ip_address_version" = "IPv4"
"public_ip_address_id" = ""
"subnet_id" = "/subscriptions/[...]"
},
{
"application_gateway_backend_address_pools_ids" = []
"application_security_group_ids" = []
"load_balancer_backend_address_pools_ids" = []
"load_balancer_inbound_nat_rules_ids" = []
"name" = "ip-configuration2"
"primary" = false
"private_ip_address" = ""
"private_ip_address_allocation" = "dynamic"
"private_ip_address_version" = "IPv4"
"public_ip_address_id" = ""
"subnet_id" = "/subscriptions/[...]"
},
]
resource_ip_configuration = [
{
"application_gateway_backend_address_pools_ids" = []
"application_security_group_ids" = []
"load_balancer_backend_address_pools_ids" = []
"load_balancer_inbound_nat_rules_ids" = []
"name" = "ip-configuration"
"primary" = true
"private_ip_address" = ""
"private_ip_address_allocation" = "dynamic"
"private_ip_address_version" = "IPv4"
"public_ip_address_id" = ""
"subnet_id" = "/subscriptions/[...]"
},
{
"application_gateway_backend_address_pools_ids" = []
"application_security_group_ids" = []
"load_balancer_backend_address_pools_ids" = []
"load_balancer_inbound_nat_rules_ids" = []
"name" = "ip-configuration2"
"primary" = false
"private_ip_address" = ""
"private_ip_address_allocation" = "dynamic"
"private_ip_address_version" = "IPv4"
"public_ip_address_id" = ""
"subnet_id" = "/subscriptions/[...]"
},
]
```
### Steps to Reproduce
`terraform apply`
|
1.0
|
private_ip_address not filled in azurerm_network_interface resource and datasource - ### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
### Terraform (and AzureRM Provider) Version
Terraform v0.12.12
terraform-provider-azurerm_v1.34.0_x4
### Affected Resource(s)
Resource azurerm_network_interface
Datasource azurerm_network_interface
### Terraform Configuration Files
```hcl
resource "azurerm_network_interface" "nic_0" {
name = "${var.hostname}-nic_0"
location = data.azurerm_resource_group.rg_vm.location
resource_group_name = data.azurerm_resource_group.rg_vm.name
ip_configuration {
name = "ip-configuration"
subnet_id = data.azurerm_subnet.subnet.id
private_ip_address_allocation = "dynamic"
primary = "true"
}
ip_configuration {
name = "ip-configuration2"
subnet_id = data.azurerm_subnet.subnet.id
private_ip_address_allocation = "dynamic"
}
}
data "azurerm_network_interface" "nic_0" {
name = "${var.hostname}-nic_0"
resource_group_name = data.azurerm_resource_group.rg_vm.name
}
output "resource_ip_configuration" {
value = azurerm_network_interface.nic_0.ip_configuration
}
output "data_ip_configuration" {
value = data.azurerm_network_interface.nic_0.ip_configuration
}
```
### Expected Behavior
All Attributes of ìp_configuration` should be filled and shown in outputs.
### Actual Behavior
`private_ip_address` is not filled:
```
data_ip_configuration = [
{
"application_gateway_backend_address_pools_ids" = []
"application_security_group_ids" = []
"load_balancer_backend_address_pools_ids" = []
"load_balancer_inbound_nat_rules_ids" = []
"name" = "ip-configuration"
"primary" = true
"private_ip_address" = ""
"private_ip_address_allocation" = "dynamic"
"private_ip_address_version" = "IPv4"
"public_ip_address_id" = ""
"subnet_id" = "/subscriptions/[...]"
},
{
"application_gateway_backend_address_pools_ids" = []
"application_security_group_ids" = []
"load_balancer_backend_address_pools_ids" = []
"load_balancer_inbound_nat_rules_ids" = []
"name" = "ip-configuration2"
"primary" = false
"private_ip_address" = ""
"private_ip_address_allocation" = "dynamic"
"private_ip_address_version" = "IPv4"
"public_ip_address_id" = ""
"subnet_id" = "/subscriptions/[...]"
},
]
resource_ip_configuration = [
{
"application_gateway_backend_address_pools_ids" = []
"application_security_group_ids" = []
"load_balancer_backend_address_pools_ids" = []
"load_balancer_inbound_nat_rules_ids" = []
"name" = "ip-configuration"
"primary" = true
"private_ip_address" = ""
"private_ip_address_allocation" = "dynamic"
"private_ip_address_version" = "IPv4"
"public_ip_address_id" = ""
"subnet_id" = "/subscriptions/[...]"
},
{
"application_gateway_backend_address_pools_ids" = []
"application_security_group_ids" = []
"load_balancer_backend_address_pools_ids" = []
"load_balancer_inbound_nat_rules_ids" = []
"name" = "ip-configuration2"
"primary" = false
"private_ip_address" = ""
"private_ip_address_allocation" = "dynamic"
"private_ip_address_version" = "IPv4"
"public_ip_address_id" = ""
"subnet_id" = "/subscriptions/[...]"
},
]
```
### Steps to Reproduce
`terraform apply`
|
non_defect
|
private ip address not filled in azurerm network interface resource and datasource community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform and azurerm provider version terraform terraform provider azurerm affected resource s resource azurerm network interface datasource azurerm network interface terraform configuration files hcl resource azurerm network interface nic name var hostname nic location data azurerm resource group rg vm location resource group name data azurerm resource group rg vm name ip configuration name ip configuration subnet id data azurerm subnet subnet id private ip address allocation dynamic primary true ip configuration name ip subnet id data azurerm subnet subnet id private ip address allocation dynamic data azurerm network interface nic name var hostname nic resource group name data azurerm resource group rg vm name output resource ip configuration value azurerm network interface nic ip configuration output data ip configuration value data azurerm network interface nic ip configuration expected behavior all attributes of ìp configuration should be filled and shown in outputs actual behavior private ip address is not filled data ip configuration application gateway backend address pools ids application security group ids load balancer backend address pools ids load balancer inbound nat rules ids name ip configuration primary true private ip address private ip address allocation dynamic private ip address version public ip address id subnet id subscriptions application gateway backend address pools ids application security group ids load balancer backend address pools ids load balancer inbound nat rules ids name ip primary false private ip address private ip address allocation dynamic private ip address version public ip address id subnet id subscriptions resource ip configuration application gateway backend address pools ids application security group ids load balancer backend address pools ids load balancer inbound nat rules ids name ip configuration primary true private ip address private ip address allocation dynamic private ip address version public ip address id subnet id subscriptions application gateway backend address pools ids application security group ids load balancer backend address pools ids load balancer inbound nat rules ids name ip primary false private ip address private ip address allocation dynamic private ip address version public ip address id subnet id subscriptions steps to reproduce terraform apply
| 0
|
53,243
| 13,261,242,980
|
IssuesEvent
|
2020-08-20 19:32:16
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
dataclasses.get_most_energetic_cascade returns cascade with lower energy than dataclasses.get_most_energetic_inice_cascade (Trac #1030)
|
Migrated from Trac combo core defect
|
When using dataclasses.get_most_energetic_cascade, it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses.get_most_energetic_inice_cascade.
An example file where this happens, e.g. for the first event, can be found at /data/user/hbretz/cascades_vs_inice_cascades.i3
```text
In [43]: dataclasses.get_most_energetic_inice_cascade(tree).energy
Out[43]: 9226.722570765558
In [44]: dataclasses.get_most_energetic_cascade(tree).energy
Out[44]: 16.27508624107153
```
This issue occured in icerec trunk, r133025.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1030">https://code.icecube.wisc.edu/projects/icecube/ticket/1030</a>, reported by hbretzand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-08-10T20:31:17",
"_ts": "1439238677802127",
"description": "When using dataclasses.get_most_energetic_cascade, it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses.get_most_energetic_inice_cascade.\n\nAn example file where this happens, e.g. for the first event, can be found at /data/user/hbretz/cascades_vs_inice_cascades.i3\n\n\n{{{\nIn [43]: dataclasses.get_most_energetic_inice_cascade(tree).energy\nOut[43]: 9226.722570765558\n\nIn [44]: dataclasses.get_most_energetic_cascade(tree).energy\nOut[44]: 16.27508624107153\n}}}\n\nThis issue occured in icerec trunk, r133025.\n",
"reporter": "hbretz",
"cc": "david.schultz@icecube.wisc.edu",
"resolution": "fixed",
"time": "2015-06-24T15:41:58",
"component": "combo core",
"summary": "dataclasses.get_most_energetic_cascade returns cascade with lower energy than dataclasses.get_most_energetic_inice_cascade",
"priority": "blocker",
"keywords": "",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
dataclasses.get_most_energetic_cascade returns cascade with lower energy than dataclasses.get_most_energetic_inice_cascade (Trac #1030) - When using dataclasses.get_most_energetic_cascade, it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses.get_most_energetic_inice_cascade.
An example file where this happens, e.g. for the first event, can be found at /data/user/hbretz/cascades_vs_inice_cascades.i3
```text
In [43]: dataclasses.get_most_energetic_inice_cascade(tree).energy
Out[43]: 9226.722570765558
In [44]: dataclasses.get_most_energetic_cascade(tree).energy
Out[44]: 16.27508624107153
```
This issue occured in icerec trunk, r133025.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1030">https://code.icecube.wisc.edu/projects/icecube/ticket/1030</a>, reported by hbretzand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-08-10T20:31:17",
"_ts": "1439238677802127",
"description": "When using dataclasses.get_most_energetic_cascade, it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses.get_most_energetic_inice_cascade.\n\nAn example file where this happens, e.g. for the first event, can be found at /data/user/hbretz/cascades_vs_inice_cascades.i3\n\n\n{{{\nIn [43]: dataclasses.get_most_energetic_inice_cascade(tree).energy\nOut[43]: 9226.722570765558\n\nIn [44]: dataclasses.get_most_energetic_cascade(tree).energy\nOut[44]: 16.27508624107153\n}}}\n\nThis issue occured in icerec trunk, r133025.\n",
"reporter": "hbretz",
"cc": "david.schultz@icecube.wisc.edu",
"resolution": "fixed",
"time": "2015-06-24T15:41:58",
"component": "combo core",
"summary": "dataclasses.get_most_energetic_cascade returns cascade with lower energy than dataclasses.get_most_energetic_inice_cascade",
"priority": "blocker",
"keywords": "",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
defect
|
dataclasses get most energetic cascade returns cascade with lower energy than dataclasses get most energetic inice cascade trac when using dataclasses get most energetic cascade it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses get most energetic inice cascade an example file where this happens e g for the first event can be found at data user hbretz cascades vs inice cascades text in dataclasses get most energetic inice cascade tree energy out in dataclasses get most energetic cascade tree energy out this issue occured in icerec trunk migrated from json status closed changetime ts description when using dataclasses get most energetic cascade it can happen that the cascade returned by the method has a lower energy than the one returned by dataclasses get most energetic inice cascade n nan example file where this happens e g for the first event can be found at data user hbretz cascades vs inice cascades n n n nin dataclasses get most energetic inice cascade tree energy nout n nin dataclasses get most energetic cascade tree energy nout n n nthis issue occured in icerec trunk n reporter hbretz cc david schultz icecube wisc edu resolution fixed time component combo core summary dataclasses get most energetic cascade returns cascade with lower energy than dataclasses get most energetic inice cascade priority blocker keywords milestone owner olivas type defect
| 1
|
649,601
| 21,316,298,383
|
IssuesEvent
|
2022-04-16 10:36:29
|
Dadangdut33/simple-prayertime-reminder
|
https://api.github.com/repos/Dadangdut33/simple-prayertime-reminder
|
closed
|
Prayertime handler and its integration
|
enhancement High priority
|
**Is your feature request related to a problem? Please describe.**
We need a handler to get each prayertime to display and store.
**Describe the solution you'd like**
Create the methods in the handler folder under the praytime.ts file name
- [x] the praytime handler
- [x] integrated successfully
|
1.0
|
Prayertime handler and its integration - **Is your feature request related to a problem? Please describe.**
We need a handler to get each prayertime to display and store.
**Describe the solution you'd like**
Create the methods in the handler folder under the praytime.ts file name
- [x] the praytime handler
- [x] integrated successfully
|
non_defect
|
prayertime handler and its integration is your feature request related to a problem please describe we need a handler to get each prayertime to display and store describe the solution you d like create the methods in the handler folder under the praytime ts file name the praytime handler integrated successfully
| 0
|
116,246
| 14,938,824,418
|
IssuesEvent
|
2021-01-25 16:13:09
|
BarryCap/BarryCap.github.io
|
https://api.github.com/repos/BarryCap/BarryCap.github.io
|
closed
|
Remove outline selection of footer
|
bug design
|
The **footer has a border which appears when hovering it**.
|
1.0
|
Remove outline selection of footer - The **footer has a border which appears when hovering it**.
|
non_defect
|
remove outline selection of footer the footer has a border which appears when hovering it
| 0
|
235,360
| 19,340,634,719
|
IssuesEvent
|
2021-12-15 03:47:01
|
1Copenut/api-learning
|
https://api.github.com/repos/1Copenut/api-learning
|
closed
|
Set up a `docker-compose.dev.yaml` for local development
|
testing docker
|
* https://www.docker.com/blog/how-to-setup-your-local-node-js-development-environment-using-docker-part-2/ has a good blueprint for using Docker for local development too.
|
1.0
|
Set up a `docker-compose.dev.yaml` for local development - * https://www.docker.com/blog/how-to-setup-your-local-node-js-development-environment-using-docker-part-2/ has a good blueprint for using Docker for local development too.
|
non_defect
|
set up a docker compose dev yaml for local development has a good blueprint for using docker for local development too
| 0
|
66,556
| 20,323,853,562
|
IssuesEvent
|
2022-02-18 02:46:09
|
idaholab/moose
|
https://api.github.com/repos/idaholab/moose
|
closed
|
Sporatic test failure in stochastic_tools
|
T: defect P: normal C: Modules/Stochastic Tools
|
## Bug Description
https://civet.inl.gov/job/958376/
## Steps to Reproduce
```
rt --re polynomial_regression/errors -p2
```
## Impact
Difficulty merging unrelated PRs.
|
1.0
|
Sporatic test failure in stochastic_tools - ## Bug Description
https://civet.inl.gov/job/958376/
## Steps to Reproduce
```
rt --re polynomial_regression/errors -p2
```
## Impact
Difficulty merging unrelated PRs.
|
defect
|
sporatic test failure in stochastic tools bug description steps to reproduce rt re polynomial regression errors impact difficulty merging unrelated prs
| 1
|
5,196
| 7,974,012,858
|
IssuesEvent
|
2018-07-17 02:43:30
|
pelias/pelias
|
https://api.github.com/repos/pelias/pelias
|
closed
|
Harden dependency on libpostal
|
processed
|
Currently, if pelias-config specifies `parser: addressit` instead of the expected `parser: libpostal`, exceptions are thrown upon incoming requests. Since we have no intention of supporting Pelias without `libpostal` we should make that clear to users by removing the config parameter and documenting the hard dependency more explicitly.
Startup of API should also fail if libpostal is not properly installed or could not be loaded for some reason.
|
1.0
|
Harden dependency on libpostal - Currently, if pelias-config specifies `parser: addressit` instead of the expected `parser: libpostal`, exceptions are thrown upon incoming requests. Since we have no intention of supporting Pelias without `libpostal` we should make that clear to users by removing the config parameter and documenting the hard dependency more explicitly.
Startup of API should also fail if libpostal is not properly installed or could not be loaded for some reason.
|
non_defect
|
harden dependency on libpostal currently if pelias config specifies parser addressit instead of the expected parser libpostal exceptions are thrown upon incoming requests since we have no intention of supporting pelias without libpostal we should make that clear to users by removing the config parameter and documenting the hard dependency more explicitly startup of api should also fail if libpostal is not properly installed or could not be loaded for some reason
| 0
|
248,333
| 26,785,120,232
|
IssuesEvent
|
2023-02-01 01:41:31
|
ConnectionMaster/create-probot-app
|
https://api.github.com/repos/ConnectionMaster/create-probot-app
|
opened
|
CVE-2022-25881 (Medium) detected in http-cache-semantics-3.8.1.tgz
|
security vulnerability
|
## CVE-2022-25881 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>http-cache-semantics-3.8.1.tgz</b></p></summary>
<p>Parses Cache-Control and other headers. Helps building correct HTTP caches and proxies</p>
<p>Library home page: <a href="https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-3.8.1.tgz">https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-3.8.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/npm/node_modules/http-cache-semantics/package.json</p>
<p>
Dependency Hierarchy:
- npm-6.14.11.tgz (Root Library)
- npm-registry-fetch-4.0.7.tgz
- make-fetch-happen-5.0.2.tgz
- :x: **http-cache-semantics-3.8.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ConnectionMaster/create-probot-app/commit/885809abccc4313bd892be901d1adc0141fa9f71">885809abccc4313bd892be901d1adc0141fa9f71</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects versions of the package http-cache-semantics before 4.1.1. The issue can be exploited via malicious request header values sent to a server, when that server reads the cache policy from the request using this library.
<p>Publish Date: 2023-01-31
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25881>CVE-2022-25881</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-25881">https://www.cve.org/CVERecord?id=CVE-2022-25881</a></p>
<p>Release Date: 2023-01-31</p>
<p>Fix Resolution: http-cache-semantics - 4.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-25881 (Medium) detected in http-cache-semantics-3.8.1.tgz - ## CVE-2022-25881 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>http-cache-semantics-3.8.1.tgz</b></p></summary>
<p>Parses Cache-Control and other headers. Helps building correct HTTP caches and proxies</p>
<p>Library home page: <a href="https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-3.8.1.tgz">https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-3.8.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/npm/node_modules/http-cache-semantics/package.json</p>
<p>
Dependency Hierarchy:
- npm-6.14.11.tgz (Root Library)
- npm-registry-fetch-4.0.7.tgz
- make-fetch-happen-5.0.2.tgz
- :x: **http-cache-semantics-3.8.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ConnectionMaster/create-probot-app/commit/885809abccc4313bd892be901d1adc0141fa9f71">885809abccc4313bd892be901d1adc0141fa9f71</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects versions of the package http-cache-semantics before 4.1.1. The issue can be exploited via malicious request header values sent to a server, when that server reads the cache policy from the request using this library.
<p>Publish Date: 2023-01-31
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25881>CVE-2022-25881</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-25881">https://www.cve.org/CVERecord?id=CVE-2022-25881</a></p>
<p>Release Date: 2023-01-31</p>
<p>Fix Resolution: http-cache-semantics - 4.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in http cache semantics tgz cve medium severity vulnerability vulnerable library http cache semantics tgz parses cache control and other headers helps building correct http caches and proxies library home page a href path to dependency file package json path to vulnerable library node modules npm node modules http cache semantics package json dependency hierarchy npm tgz root library npm registry fetch tgz make fetch happen tgz x http cache semantics tgz vulnerable library found in head commit a href found in base branch master vulnerability details this affects versions of the package http cache semantics before the issue can be exploited via malicious request header values sent to a server when that server reads the cache policy from the request using this library publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution http cache semantics step up your open source security game with mend
| 0
|
59,953
| 17,023,298,121
|
IssuesEvent
|
2021-07-03 01:17:57
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
rendering of bridges meeting at same level
|
Component: mapnik Priority: minor Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 10.02am, Saturday, 20th September 2008]**
Mapnik has a problem rendering an intersection on a bridge. It draws two bridges, where one arbitrary path seemingly spans the other.
See for example http://www.openstreetmap.org/?lat=49.542075&lon=8.661282&zoom=18 which tries to render three ramps meeting at an overhead cycle path.
|
1.0
|
rendering of bridges meeting at same level - **[Submitted to the original trac issue database at 10.02am, Saturday, 20th September 2008]**
Mapnik has a problem rendering an intersection on a bridge. It draws two bridges, where one arbitrary path seemingly spans the other.
See for example http://www.openstreetmap.org/?lat=49.542075&lon=8.661282&zoom=18 which tries to render three ramps meeting at an overhead cycle path.
|
defect
|
rendering of bridges meeting at same level mapnik has a problem rendering an intersection on a bridge it draws two bridges where one arbitrary path seemingly spans the other see for example which tries to render three ramps meeting at an overhead cycle path
| 1
|
13,899
| 2,789,663,998
|
IssuesEvent
|
2015-05-08 20:42:56
|
google/google-visualization-api-issues
|
https://api.github.com/repos/google/google-visualization-api-issues
|
closed
|
Visualization tooltip border gets cutoff for chart endpoints.
|
Priority-Medium Type-Defect
|
Original [issue 17](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=17) created by google-admin on 2009-08-04T21:01:31.000Z:
Visualization tooltip border gets cutoff for chart endpoints.
<b>What steps will reproduce the problem? Please provide a link to a</b>
<b>demonstration page if at all possible, or attach code.</b>
1. Create a line chart that is w=600 and h=180.
2. Give the line chart 12 data points with values ranging from 800,000 to
1,000,000.
3. When the tooltip for the two endpoints is displayed, sometimes the left
or right border is cutoff.
NOTE: I am using the gwt visualization api.
<b>What component is this issue related to (PieChart, LineChart, DataTable,</b>
<b>Query, etc)?</b>
I've seen this with the line chart but it's possible that other charts
encounter this issue as well.
<b>Are you using the test environment (version 1.1)?</b>
<b>(If you are not sure, answer NO)</b>
No.
<b>What operating system and browser are you using?</b>
Vista SP1.
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
|
1.0
|
Visualization tooltip border gets cutoff for chart endpoints. - Original [issue 17](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=17) created by google-admin on 2009-08-04T21:01:31.000Z:
Visualization tooltip border gets cutoff for chart endpoints.
<b>What steps will reproduce the problem? Please provide a link to a</b>
<b>demonstration page if at all possible, or attach code.</b>
1. Create a line chart that is w=600 and h=180.
2. Give the line chart 12 data points with values ranging from 800,000 to
1,000,000.
3. When the tooltip for the two endpoints is displayed, sometimes the left
or right border is cutoff.
NOTE: I am using the gwt visualization api.
<b>What component is this issue related to (PieChart, LineChart, DataTable,</b>
<b>Query, etc)?</b>
I've seen this with the line chart but it's possible that other charts
encounter this issue as well.
<b>Are you using the test environment (version 1.1)?</b>
<b>(If you are not sure, answer NO)</b>
No.
<b>What operating system and browser are you using?</b>
Vista SP1.
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
|
defect
|
visualization tooltip border gets cutoff for chart endpoints original created by google admin on visualization tooltip border gets cutoff for chart endpoints what steps will reproduce the problem please provide a link to a demonstration page if at all possible or attach code create a line chart that is w and h give the line chart data points with values ranging from to when the tooltip for the two endpoints is displayed sometimes the left or right border is cutoff note i am using the gwt visualization api what component is this issue related to piechart linechart datatable query etc i ve seen this with the line chart but it s possible that other charts encounter this issue as well are you using the test environment version if you are not sure answer no no what operating system and browser are you using vista for developers viewing this issue please click the star icon to be notified of future changes and to let us know how many of you are interested in seeing it resolved
| 1
|
57,982
| 16,240,076,209
|
IssuesEvent
|
2021-05-07 08:26:45
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
opened
|
TABLE.fields() not compatible with dsl.select(fields) in Kotlin
|
T: Defect
|
In Kotlin, the `TABLE.fields()` method returns an `Array` of fields and the `DSLContext.select()` expects a `Collection` of fields. They are therefor not compatible with each other.
### Expected behavior
This (code that works in Java) I believe should also work in Kotlin:
```
dsl.select(MY_TABLE.fields())
```
### Actual behavior
The code will not compile, with the following error:
```
Kotlin: None of the following functions can be called with the arguments supplied:
public abstract fun select(p0: (MutableCollection<out SelectFieldOrAsterisk!>..Collection<SelectFieldOrAsterisk!>?)): SelectSelectStep<Record!> defined in org.jooq.DSLContext
public abstract fun select(vararg p0: SelectFieldOrAsterisk!): SelectSelectStep<Record!> defined in org.jooq.DSLContext
public abstract fun <T1 : Any!> select(p0: SelectField<TypeVariable(T1)!>!): SelectSelectStep<Record1<TypeVariable(T1)!>!> defined in org.jooq.DSLContext
```
Unless I manually use the `toList()` extension function:
```
dsl.select(MY_TABLE.fields().toList())
```
### Steps to reproduce the problem
Generate code for a table and try the following code, using Kotlin:
```
dsl.select(MY_TABLE.fields())
```
### Versions
- jOOQ: 3.14.8
- Java: 11
- Kotlin 1.4.32
- Database (include vendor): PostgreSQL 12
- OS: Fedora 32
- JDBC Driver (include name if inofficial driver):
|
1.0
|
TABLE.fields() not compatible with dsl.select(fields) in Kotlin - In Kotlin, the `TABLE.fields()` method returns an `Array` of fields and the `DSLContext.select()` expects a `Collection` of fields. They are therefor not compatible with each other.
### Expected behavior
This (code that works in Java) I believe should also work in Kotlin:
```
dsl.select(MY_TABLE.fields())
```
### Actual behavior
The code will not compile, with the following error:
```
Kotlin: None of the following functions can be called with the arguments supplied:
public abstract fun select(p0: (MutableCollection<out SelectFieldOrAsterisk!>..Collection<SelectFieldOrAsterisk!>?)): SelectSelectStep<Record!> defined in org.jooq.DSLContext
public abstract fun select(vararg p0: SelectFieldOrAsterisk!): SelectSelectStep<Record!> defined in org.jooq.DSLContext
public abstract fun <T1 : Any!> select(p0: SelectField<TypeVariable(T1)!>!): SelectSelectStep<Record1<TypeVariable(T1)!>!> defined in org.jooq.DSLContext
```
Unless I manually use the `toList()` extension function:
```
dsl.select(MY_TABLE.fields().toList())
```
### Steps to reproduce the problem
Generate code for a table and try the following code, using Kotlin:
```
dsl.select(MY_TABLE.fields())
```
### Versions
- jOOQ: 3.14.8
- Java: 11
- Kotlin 1.4.32
- Database (include vendor): PostgreSQL 12
- OS: Fedora 32
- JDBC Driver (include name if inofficial driver):
|
defect
|
table fields not compatible with dsl select fields in kotlin in kotlin the table fields method returns an array of fields and the dslcontext select expects a collection of fields they are therefor not compatible with each other expected behavior this code that works in java i believe should also work in kotlin dsl select my table fields actual behavior the code will not compile with the following error kotlin none of the following functions can be called with the arguments supplied public abstract fun select mutablecollection collection selectselectstep defined in org jooq dslcontext public abstract fun select vararg selectfieldorasterisk selectselectstep defined in org jooq dslcontext public abstract fun select selectfield selectselectstep defined in org jooq dslcontext unless i manually use the tolist extension function dsl select my table fields tolist steps to reproduce the problem generate code for a table and try the following code using kotlin dsl select my table fields versions jooq java kotlin database include vendor postgresql os fedora jdbc driver include name if inofficial driver
| 1
|
51,693
| 21,779,373,462
|
IssuesEvent
|
2022-05-13 17:00:20
|
cityofaustin/atd-data-tech
|
https://api.github.com/repos/cityofaustin/atd-data-tech
|
closed
|
[Meeting] Shared Mobility Digital License Application Feedback
|
Type: Meeting Service: Apps Type: Enhancement Product: Shared Mobility Reporting Tracker Workgroup: SMS
|
### Objective
Update Mobility Services Division Manager and Permitting Supervisor on how the Digital License Application for Smart Mobility works and get buy-in for a digital permit approval process
### Participants
Jacob C., Courtney H., Joseph A., Mary V. Diana M., Patrick M., Karo E.
### Agenda
None
------
- [x] Schedule meeting
- [ ] Optional: Schedule debrief
- [x] Meet and take notes
- [ ] Optional: Debrief with DTS team members
- [ ] Create resulting issues
|
1.0
|
[Meeting] Shared Mobility Digital License Application Feedback - ### Objective
Update Mobility Services Division Manager and Permitting Supervisor on how the Digital License Application for Smart Mobility works and get buy-in for a digital permit approval process
### Participants
Jacob C., Courtney H., Joseph A., Mary V. Diana M., Patrick M., Karo E.
### Agenda
None
------
- [x] Schedule meeting
- [ ] Optional: Schedule debrief
- [x] Meet and take notes
- [ ] Optional: Debrief with DTS team members
- [ ] Create resulting issues
|
non_defect
|
shared mobility digital license application feedback objective update mobility services division manager and permitting supervisor on how the digital license application for smart mobility works and get buy in for a digital permit approval process participants jacob c courtney h joseph a mary v diana m patrick m karo e agenda none schedule meeting optional schedule debrief meet and take notes optional debrief with dts team members create resulting issues
| 0
|
92,728
| 8,376,979,042
|
IssuesEvent
|
2018-10-05 21:57:34
|
apache/incubator-mxnet
|
https://api.github.com/repos/apache/incubator-mxnet
|
closed
|
test_gluon.test_conv has fixed seed that can mask flakiness
|
Flaky Test
|
The unit test in title have been using fixed seed to mask flakiness. Suggested action:
1. Evaluate whether the test is flaky without fixed seed. If not, remove seed. Else move to 2
2. If test is flaky, determine whether it's an actual uncaught edge case. If so, fix the operator. Else move to 3
3. If numerical instability is inevitable, adjust tolerance level appropriately.
|
1.0
|
test_gluon.test_conv has fixed seed that can mask flakiness - The unit test in title have been using fixed seed to mask flakiness. Suggested action:
1. Evaluate whether the test is flaky without fixed seed. If not, remove seed. Else move to 2
2. If test is flaky, determine whether it's an actual uncaught edge case. If so, fix the operator. Else move to 3
3. If numerical instability is inevitable, adjust tolerance level appropriately.
|
non_defect
|
test gluon test conv has fixed seed that can mask flakiness the unit test in title have been using fixed seed to mask flakiness suggested action evaluate whether the test is flaky without fixed seed if not remove seed else move to if test is flaky determine whether it s an actual uncaught edge case if so fix the operator else move to if numerical instability is inevitable adjust tolerance level appropriately
| 0
|
20,932
| 3,437,552,369
|
IssuesEvent
|
2015-12-13 09:37:47
|
Cockatrice/Cockatrice
|
https://api.github.com/repos/Cockatrice/Cockatrice
|
closed
|
Morph card issue
|
App - Cockatrice Defect - Game Rules Compliance
|
From #1651:
When you have a morph creature in play and you double-click a creature card from your hand that is the same as the face-down creature, it enters the battlefield on top of the morph, telling your opponent, that the morph is the same creature as the one you played from your hand.
|
1.0
|
Morph card issue - From #1651:
When you have a morph creature in play and you double-click a creature card from your hand that is the same as the face-down creature, it enters the battlefield on top of the morph, telling your opponent, that the morph is the same creature as the one you played from your hand.
|
defect
|
morph card issue from when you have a morph creature in play and you double click a creature card from your hand that is the same as the face down creature it enters the battlefield on top of the morph telling your opponent that the morph is the same creature as the one you played from your hand
| 1
|
185,829
| 21,843,817,049
|
IssuesEvent
|
2022-05-18 01:13:20
|
Whizkevina/uchi-sidebar-clone
|
https://api.github.com/repos/Whizkevina/uchi-sidebar-clone
|
opened
|
CVE-2022-29353 (Medium) detected in graphql-upload-8.1.0.tgz
|
security vulnerability
|
## CVE-2022-29353 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>graphql-upload-8.1.0.tgz</b></p></summary>
<p>Middleware and an Upload scalar to add support for GraphQL multipart requests (file uploads via queries and mutations) to various Node.js GraphQL servers.</p>
<p>Library home page: <a href="https://registry.npmjs.org/graphql-upload/-/graphql-upload-8.1.0.tgz">https://registry.npmjs.org/graphql-upload/-/graphql-upload-8.1.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/graphql-upload/package.json</p>
<p>
Dependency Hierarchy:
- apollo-server-2.15.1.tgz (Root Library)
- apollo-server-core-2.18.2.tgz
- :x: **graphql-upload-8.1.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Whizkevina/uchi-sidebar-clone/commit/5405eeecb088ab7acf45ef51e052988d72c3fe7f">5405eeecb088ab7acf45ef51e052988d72c3fe7f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An arbitrary file upload vulnerability in the file upload module of Graphql-upload v13.0.0 allows attackers to execute arbitrary code via a crafted filename.
<p>Publish Date: 2022-05-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-29353>CVE-2022-29353</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-29353">https://nvd.nist.gov/vuln/detail/CVE-2022-29353</a></p>
<p>Release Date: 2022-05-16</p>
<p>Fix Resolution: no_fix</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-29353 (Medium) detected in graphql-upload-8.1.0.tgz - ## CVE-2022-29353 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>graphql-upload-8.1.0.tgz</b></p></summary>
<p>Middleware and an Upload scalar to add support for GraphQL multipart requests (file uploads via queries and mutations) to various Node.js GraphQL servers.</p>
<p>Library home page: <a href="https://registry.npmjs.org/graphql-upload/-/graphql-upload-8.1.0.tgz">https://registry.npmjs.org/graphql-upload/-/graphql-upload-8.1.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/graphql-upload/package.json</p>
<p>
Dependency Hierarchy:
- apollo-server-2.15.1.tgz (Root Library)
- apollo-server-core-2.18.2.tgz
- :x: **graphql-upload-8.1.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Whizkevina/uchi-sidebar-clone/commit/5405eeecb088ab7acf45ef51e052988d72c3fe7f">5405eeecb088ab7acf45ef51e052988d72c3fe7f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An arbitrary file upload vulnerability in the file upload module of Graphql-upload v13.0.0 allows attackers to execute arbitrary code via a crafted filename.
<p>Publish Date: 2022-05-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-29353>CVE-2022-29353</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-29353">https://nvd.nist.gov/vuln/detail/CVE-2022-29353</a></p>
<p>Release Date: 2022-05-16</p>
<p>Fix Resolution: no_fix</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in graphql upload tgz cve medium severity vulnerability vulnerable library graphql upload tgz middleware and an upload scalar to add support for graphql multipart requests file uploads via queries and mutations to various node js graphql servers library home page a href path to dependency file package json path to vulnerable library node modules graphql upload package json dependency hierarchy apollo server tgz root library apollo server core tgz x graphql upload tgz vulnerable library found in head commit a href found in base branch main vulnerability details an arbitrary file upload vulnerability in the file upload module of graphql upload allows attackers to execute arbitrary code via a crafted filename publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution no fix step up your open source security game with whitesource
| 0
|
266,708
| 23,255,450,309
|
IssuesEvent
|
2022-08-04 08:52:11
|
tutao/tutanota
|
https://api.github.com/repos/tutao/tutanota
|
closed
|
Doesn't show message body (text) anymore on PaleMoon
|
bug not-testable
|
- [v] This is not a feature request (existing functionality does not work, **not** missing functionality).
I will request features on [forum](https://www.reddit.com/r/tutanota/) or via support.
- [v] I've searched and did not find a similar issue.
**Bug in web app**
**Describe the bug**
As the topic says. I open any message - no text is shown.
**To Reproduce**
Just open the mail box in Pale Moon
**Expected behavior**
Was working fine a month ago that's for sure.
**Screenshots**
This is how every message looks like: https://i.gyazo.com/53bb363d57569056b3d4eff8cfb94a12.png
Error console after loading: https://i.gyazo.com/f18d732b82618c2a95ab5dbd2c8b3a5d.png
<details>
<summary>What the error on the page says (it was always there, never bothered too much)</summary>
```
Feedback v3.98.12 - DbError - UNKNOWN - Browser
Client: Browser
Type: UNKNOWN
Tutanota version: 3.98.12
Timestamp (UTC): Mon, 25 Jul 2022 20:44:46 GMT
User agent:
Mozilla/5.0 (Windows NT 10.0; WOW64; rv:4.8) Goanna/20170101 PaleMoon/29.4.4
DbError
Error message: IndexedDbTransaction get().catch
OSes: {"0":"MetaData"}
event:{"code":11,"message":"An attempt was made to use an object that is not, or is no longer, usable","result":2152923147,"name":"InvalidStateError","filename":"https://mail.tutanota.com/worker.js","lineNumber":12,"columnNumber":0,"stack":"get/<@https://mail.tutanota.com/worker.js:12:22116\nget@https://mail.tutanota.com/worker.js:12:22083\ninit@https://mail.tutanota.com/worker.js:12:65193\n","INDEX_SIZE_ERR":1,"DOMSTRING_SIZE_ERR":2,"HIERARCHY_REQUEST_ERR":3,"WRONG_DOCUMENT_ERR":4,"INVALID_CHARACTER_ERR":5,"NO_DATA_ALLOWED_ERR":6,"NO_MODIFICATION_ALLOWED_ERR":7,"NOT_FOUND_ERR":8,"NOT_SUPPORTED_ERR":9,"INUSE_ATTRIBUTE_ERR":10,"INVALID_STATE_ERR":11,"SYNTAX_ERR":12,"INVALID_MODIFICATION_ERR":13,"NAMESPACE_ERR":14,"INVALID_ACCESS_ERR":15,"VALIDATION_ERR":16,"TYPE_MISMATCH_ERR":17,"SECURITY_ERR":18,"NETWORK_ERR":19,"ABORT_ERR":20,"URL_MISMATCH_ERR":21,"QUOTA_EXCEEDED_ERR":22,"TIMEOUT_ERR":23,"INVALID_NODE_TYPE_ERR":24,"DATA_CLONE_ERR":25}
transaction.error: <null>
event.target: <null>
event.target.error: <null>
custom.target: <null>
custom.target.error: <null>
Stacktrace:
DbError. IndexedDbTransaction get().catch
OSes: {"0":"MetaData"}
event:{"code":11,"message":"An attempt was made to use an object that is not, or is no longer, usable","result":2152923147,"name":"InvalidStateError","filename":"https://mail.tutanota.com/worker.js","lineNumber":12,"columnNumber":0,"stack":"get/<@https://mail.tutanota.com/worker.js:12:22116\nget@https://mail.tutanota.com/worker.js:12:22083\ninit@https://mail.tutanota.com/worker.js:12:65193\n","INDEX_SIZE_ERR":1,"DOMSTRING_SIZE_ERR":2,"HIERARCHY_REQUEST_ERR":3,"WRONG_DOCUMENT_ERR":4,"INVALID_CHARACTER_ERR":5,"NO_DATA_ALLOWED_ERR":6,"NO_MODIFICATION_ALLOWED_ERR":7,"NOT_FOUND_ERR":8,"NOT_SUPPORTED_ERR":9,"INUSE_ATTRIBUTE_ERR":10,"INVALID_STATE_ERR":11,"SYNTAX_ERR":12,"INVALID_MODIFICATION_ERR":13,"NAMESPACE_ERR":14,"INVALID_ACCESS_ERR":15,"VALIDATION_ERR":16,"TYPE_MISMATCH_ERR":17,"SECURITY_ERR":18,"NETWORK_ERR":19,"ABORT_ERR":20,"URL_MISMATCH_ERR":21,"QUOTA_EXCEEDED_ERR":22,"TIMEOUT_ERR":23,"INVALID_NODE_TYPE_ERR":24,"DATA_CLONE_ERR":25}
transaction.error: <null>
event.target: <null>
event.target.error: <null>
custom.target: <null>
custom.target.error: <null>
_handleDbError@https://mail.tutanota.com/worker.js:12:23927
get/<@https://mail.tutanota.com/worker.js:12:22292
get@https://mail.tutanota.com/worker.js:12:22083
init@https://mail.tutanota.com/worker.js:12:65193
```
</details>
**Desktop (please complete the following information):**
- OS: W10
- Browser Pale Moon
- Version 31.1.1 32-bit
|
1.0
|
Doesn't show message body (text) anymore on PaleMoon - - [v] This is not a feature request (existing functionality does not work, **not** missing functionality).
I will request features on [forum](https://www.reddit.com/r/tutanota/) or via support.
- [v] I've searched and did not find a similar issue.
**Bug in web app**
**Describe the bug**
As the topic says. I open any message - no text is shown.
**To Reproduce**
Just open the mail box in Pale Moon
**Expected behavior**
Was working fine a month ago that's for sure.
**Screenshots**
This is how every message looks like: https://i.gyazo.com/53bb363d57569056b3d4eff8cfb94a12.png
Error console after loading: https://i.gyazo.com/f18d732b82618c2a95ab5dbd2c8b3a5d.png
<details>
<summary>What the error on the page says (it was always there, never bothered too much)</summary>
```
Feedback v3.98.12 - DbError - UNKNOWN - Browser
Client: Browser
Type: UNKNOWN
Tutanota version: 3.98.12
Timestamp (UTC): Mon, 25 Jul 2022 20:44:46 GMT
User agent:
Mozilla/5.0 (Windows NT 10.0; WOW64; rv:4.8) Goanna/20170101 PaleMoon/29.4.4
DbError
Error message: IndexedDbTransaction get().catch
OSes: {"0":"MetaData"}
event:{"code":11,"message":"An attempt was made to use an object that is not, or is no longer, usable","result":2152923147,"name":"InvalidStateError","filename":"https://mail.tutanota.com/worker.js","lineNumber":12,"columnNumber":0,"stack":"get/<@https://mail.tutanota.com/worker.js:12:22116\nget@https://mail.tutanota.com/worker.js:12:22083\ninit@https://mail.tutanota.com/worker.js:12:65193\n","INDEX_SIZE_ERR":1,"DOMSTRING_SIZE_ERR":2,"HIERARCHY_REQUEST_ERR":3,"WRONG_DOCUMENT_ERR":4,"INVALID_CHARACTER_ERR":5,"NO_DATA_ALLOWED_ERR":6,"NO_MODIFICATION_ALLOWED_ERR":7,"NOT_FOUND_ERR":8,"NOT_SUPPORTED_ERR":9,"INUSE_ATTRIBUTE_ERR":10,"INVALID_STATE_ERR":11,"SYNTAX_ERR":12,"INVALID_MODIFICATION_ERR":13,"NAMESPACE_ERR":14,"INVALID_ACCESS_ERR":15,"VALIDATION_ERR":16,"TYPE_MISMATCH_ERR":17,"SECURITY_ERR":18,"NETWORK_ERR":19,"ABORT_ERR":20,"URL_MISMATCH_ERR":21,"QUOTA_EXCEEDED_ERR":22,"TIMEOUT_ERR":23,"INVALID_NODE_TYPE_ERR":24,"DATA_CLONE_ERR":25}
transaction.error: <null>
event.target: <null>
event.target.error: <null>
custom.target: <null>
custom.target.error: <null>
Stacktrace:
DbError. IndexedDbTransaction get().catch
OSes: {"0":"MetaData"}
event:{"code":11,"message":"An attempt was made to use an object that is not, or is no longer, usable","result":2152923147,"name":"InvalidStateError","filename":"https://mail.tutanota.com/worker.js","lineNumber":12,"columnNumber":0,"stack":"get/<@https://mail.tutanota.com/worker.js:12:22116\nget@https://mail.tutanota.com/worker.js:12:22083\ninit@https://mail.tutanota.com/worker.js:12:65193\n","INDEX_SIZE_ERR":1,"DOMSTRING_SIZE_ERR":2,"HIERARCHY_REQUEST_ERR":3,"WRONG_DOCUMENT_ERR":4,"INVALID_CHARACTER_ERR":5,"NO_DATA_ALLOWED_ERR":6,"NO_MODIFICATION_ALLOWED_ERR":7,"NOT_FOUND_ERR":8,"NOT_SUPPORTED_ERR":9,"INUSE_ATTRIBUTE_ERR":10,"INVALID_STATE_ERR":11,"SYNTAX_ERR":12,"INVALID_MODIFICATION_ERR":13,"NAMESPACE_ERR":14,"INVALID_ACCESS_ERR":15,"VALIDATION_ERR":16,"TYPE_MISMATCH_ERR":17,"SECURITY_ERR":18,"NETWORK_ERR":19,"ABORT_ERR":20,"URL_MISMATCH_ERR":21,"QUOTA_EXCEEDED_ERR":22,"TIMEOUT_ERR":23,"INVALID_NODE_TYPE_ERR":24,"DATA_CLONE_ERR":25}
transaction.error: <null>
event.target: <null>
event.target.error: <null>
custom.target: <null>
custom.target.error: <null>
_handleDbError@https://mail.tutanota.com/worker.js:12:23927
get/<@https://mail.tutanota.com/worker.js:12:22292
get@https://mail.tutanota.com/worker.js:12:22083
init@https://mail.tutanota.com/worker.js:12:65193
```
</details>
**Desktop (please complete the following information):**
- OS: W10
- Browser Pale Moon
- Version 31.1.1 32-bit
|
non_defect
|
doesn t show message body text anymore on palemoon this is not a feature request existing functionality does not work not missing functionality i will request features on or via support i ve searched and did not find a similar issue bug in web app describe the bug as the topic says i open any message no text is shown to reproduce just open the mail box in pale moon expected behavior was working fine a month ago that s for sure screenshots this is how every message looks like error console after loading what the error on the page says it was always there never bothered too much feedback dberror unknown browser client browser type unknown tutanota version timestamp utc mon jul gmt user agent mozilla windows nt rv goanna palemoon dberror error message indexeddbtransaction get catch oses metadata event code message an attempt was made to use an object that is not or is no longer usable result name invalidstateerror filename transaction error event target event target error custom target custom target error stacktrace dberror indexeddbtransaction get catch oses metadata event code message an attempt was made to use an object that is not or is no longer usable result name invalidstateerror filename transaction error event target event target error custom target custom target error handledberror get get init desktop please complete the following information os browser pale moon version bit
| 0
|
46,428
| 13,055,911,061
|
IssuesEvent
|
2020-07-30 03:05:45
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
tpx doc improvement (Trac #1187)
|
Incomplete Migration Migrated from Trac combo reconstruction defect
|
Migrated from https://code.icecube.wisc.edu/ticket/1187
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "rst doc does not list maintainer at top of page\nalso add link to doxygen documentaion",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "tpx doc improvement",
"priority": "normal",
"keywords": "",
"time": "2015-08-19T13:36:23",
"milestone": "",
"owner": "karg",
"type": "defect"
}
```
|
1.0
|
tpx doc improvement (Trac #1187) - Migrated from https://code.icecube.wisc.edu/ticket/1187
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "rst doc does not list maintainer at top of page\nalso add link to doxygen documentaion",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "tpx doc improvement",
"priority": "normal",
"keywords": "",
"time": "2015-08-19T13:36:23",
"milestone": "",
"owner": "karg",
"type": "defect"
}
```
|
defect
|
tpx doc improvement trac migrated from json status closed changetime description rst doc does not list maintainer at top of page nalso add link to doxygen documentaion reporter kjmeagher cc resolution fixed ts component combo reconstruction summary tpx doc improvement priority normal keywords time milestone owner karg type defect
| 1
|
52,859
| 13,225,173,499
|
IssuesEvent
|
2020-08-17 20:38:14
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
Disentangle Geant4 from clsim (Trac #481)
|
Migrated from Trac combo simulation defect
|
In testing simprod scripts I came to realize that as long as Geant4 is found in the ports clsim will use it resulting in a HUGE slowdown for photon propagation. This means that we need to have entirely different builds for G4 and sans G4. Even then, there is no way to disable this via CMake flags unless you hack the CMakeLists.txt in clsim.
In my opinion, G4 propagation should not be done in clsim but should either be a separate module (in a different project) or at least be a service used by clsim if enabled.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/481">https://code.icecube.wisc.edu/projects/icecube/ticket/481</a>, reported by juancarlosand owned by claudio.kopper</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-03-11T14:43:34",
"_ts": "1426085014817363",
"description": "In testing simprod scripts I came to realize that as long as Geant4 is found in the ports clsim will use it resulting in a HUGE slowdown for photon propagation. This means that we need to have entirely different builds for G4 and sans G4. Even then, there is no way to disable this via CMake flags unless you hack the CMakeLists.txt in clsim.\nIn my opinion, G4 propagation should not be done in clsim but should either be a separate module (in a different project) or at least be a service used by clsim if enabled.",
"reporter": "juancarlos",
"cc": "",
"resolution": "worksforme",
"time": "2014-01-28T18:11:39",
"component": "combo simulation",
"summary": "Disentangle Geant4 from clsim",
"priority": "major",
"keywords": "",
"milestone": "",
"owner": "claudio.kopper",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
Disentangle Geant4 from clsim (Trac #481) - In testing simprod scripts I came to realize that as long as Geant4 is found in the ports clsim will use it resulting in a HUGE slowdown for photon propagation. This means that we need to have entirely different builds for G4 and sans G4. Even then, there is no way to disable this via CMake flags unless you hack the CMakeLists.txt in clsim.
In my opinion, G4 propagation should not be done in clsim but should either be a separate module (in a different project) or at least be a service used by clsim if enabled.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/481">https://code.icecube.wisc.edu/projects/icecube/ticket/481</a>, reported by juancarlosand owned by claudio.kopper</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-03-11T14:43:34",
"_ts": "1426085014817363",
"description": "In testing simprod scripts I came to realize that as long as Geant4 is found in the ports clsim will use it resulting in a HUGE slowdown for photon propagation. This means that we need to have entirely different builds for G4 and sans G4. Even then, there is no way to disable this via CMake flags unless you hack the CMakeLists.txt in clsim.\nIn my opinion, G4 propagation should not be done in clsim but should either be a separate module (in a different project) or at least be a service used by clsim if enabled.",
"reporter": "juancarlos",
"cc": "",
"resolution": "worksforme",
"time": "2014-01-28T18:11:39",
"component": "combo simulation",
"summary": "Disentangle Geant4 from clsim",
"priority": "major",
"keywords": "",
"milestone": "",
"owner": "claudio.kopper",
"type": "defect"
}
```
</p>
</details>
|
defect
|
disentangle from clsim trac in testing simprod scripts i came to realize that as long as is found in the ports clsim will use it resulting in a huge slowdown for photon propagation this means that we need to have entirely different builds for and sans even then there is no way to disable this via cmake flags unless you hack the cmakelists txt in clsim in my opinion propagation should not be done in clsim but should either be a separate module in a different project or at least be a service used by clsim if enabled migrated from json status closed changetime ts description in testing simprod scripts i came to realize that as long as is found in the ports clsim will use it resulting in a huge slowdown for photon propagation this means that we need to have entirely different builds for and sans even then there is no way to disable this via cmake flags unless you hack the cmakelists txt in clsim nin my opinion propagation should not be done in clsim but should either be a separate module in a different project or at least be a service used by clsim if enabled reporter juancarlos cc resolution worksforme time component combo simulation summary disentangle from clsim priority major keywords milestone owner claudio kopper type defect
| 1
|
63,232
| 17,468,364,999
|
IssuesEvent
|
2021-08-06 20:40:30
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Sticky headings jockey for position causing scroll jitter in room list
|
T-Defect A-Room-List
|
### Steps to reproduce
This bug depends on your viewport height compared to one of your room list sub-sections.
1. Start with a big viewport, and have a room subsection expanded where all of it fits on screen
1. Resize to shrink your viewport so that it's just big enough to show the whole room subsection. Then shrink a bit more so that the headings overlap when you scroll down
1. Scroll to the bottom of the room list
1. Notice the jitter
I have reproduced this with Chrome on Windows and the Element desktop app on macOS but seems like it would happen in any environment.
Related issues:
- https://github.com/vector-im/element-web/issues/18366
- https://github.com/vector-im/element-web/issues/14085
### What happened?
The whole scrollable room list area jitters up and down as the different sticky headings jockey for position (jumpy).
https://user-images.githubusercontent.com/558581/128567603-cc95887d-05e2-45e5-b8c6-8e9c17b205cd.mp4
### What did you expect?
The headings would have a steady state no matter how big the viewport is.
### Operating system
Windows, macOS
### Browser information
Chrome, Element desktop app
### URL for webapp
develop.element.io
|
1.0
|
Sticky headings jockey for position causing scroll jitter in room list - ### Steps to reproduce
This bug depends on your viewport height compared to one of your room list sub-sections.
1. Start with a big viewport, and have a room subsection expanded where all of it fits on screen
1. Resize to shrink your viewport so that it's just big enough to show the whole room subsection. Then shrink a bit more so that the headings overlap when you scroll down
1. Scroll to the bottom of the room list
1. Notice the jitter
I have reproduced this with Chrome on Windows and the Element desktop app on macOS but seems like it would happen in any environment.
Related issues:
- https://github.com/vector-im/element-web/issues/18366
- https://github.com/vector-im/element-web/issues/14085
### What happened?
The whole scrollable room list area jitters up and down as the different sticky headings jockey for position (jumpy).
https://user-images.githubusercontent.com/558581/128567603-cc95887d-05e2-45e5-b8c6-8e9c17b205cd.mp4
### What did you expect?
The headings would have a steady state no matter how big the viewport is.
### Operating system
Windows, macOS
### Browser information
Chrome, Element desktop app
### URL for webapp
develop.element.io
|
defect
|
sticky headings jockey for position causing scroll jitter in room list steps to reproduce this bug depends on your viewport height compared to one of your room list sub sections start with a big viewport and have a room subsection expanded where all of it fits on screen resize to shrink your viewport so that it s just big enough to show the whole room subsection then shrink a bit more so that the headings overlap when you scroll down scroll to the bottom of the room list notice the jitter i have reproduced this with chrome on windows and the element desktop app on macos but seems like it would happen in any environment related issues what happened the whole scrollable room list area jitters up and down as the different sticky headings jockey for position jumpy what did you expect the headings would have a steady state no matter how big the viewport is operating system windows macos browser information chrome element desktop app url for webapp develop element io
| 1
|
44,869
| 12,417,436,841
|
IssuesEvent
|
2020-05-22 20:42:43
|
CocoaPods/CocoaPods
|
https://api.github.com/repos/CocoaPods/CocoaPods
|
closed
|
Unknown object version -- 53
|
t2:defect
|
I am using **CocoaPods 1.9.2** (and Flutter 1.17.1).
* [X] I've read and understood the [*CONTRIBUTING* guidelines and have done my best effort to follow](https://github.com/CocoaPods/CocoaPods/blob/master/CONTRIBUTING.md).
# Report
## What did you do?
I have a `ios/Runner.xcodeproj/project.pbxproj` file with an `objectVersion = 53;` -- this was written by XCode.
`pod update` reports **Unknown object version.**
```
RuntimeError - [Xcodeproj] Unknown object version.
/Library/Ruby/Gems/2.6.0/gems/xcodeproj-1.15.0/lib/xcodeproj/project.rb:227:in `initialize_from_file'
/Library/Ruby/Gems/2.6.0/gems/xcodeproj-1.15.0/lib/xcodeproj/project.rb:112:in `open'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer/analyzer.rb:1177:in `block (2 levels) in inspect_targets_to_integrate'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer/analyzer.rb:1176:in `each'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer/analyzer.rb:1176:in `block in inspect_targets_to_integrate'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/user_interface.rb:64:in `section'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer/analyzer.rb:1171:in `inspect_targets_to_integrate'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer/analyzer.rb:106:in `analyze'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer.rb:410:in `analyze'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer.rb:235:in `block in resolve_dependencies'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/user_interface.rb:64:in `section'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer.rb:234:in `resolve_dependencies'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer.rb:156:in `install!'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/command/update.rb:63:in `run'
/Library/Ruby/Gems/2.6.0/gems/claide-1.0.3/lib/claide/command.rb:334:in `run'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/command.rb:52:in `run'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/bin/pod:55:in `<top (required)>'
/usr/local/bin/pod:23:in `load'
/usr/local/bin/pod:23:in `<main>'
```
## What did you expect to happen?
I did not expect a _Runtime Error_.
## What happened instead?
`pod` is broken.
A workaround is to hack the `objectVersion` -- changing from 53 to 52 does the trick. However, then `pod` makes a few other changes to the file that make me uncomfortable. Here is the `diff`:
```
diff --git a/ios/Runner.xcodeproj/project.pbxproj b/ios/Runner.xcodeproj/project.pbxproj
index 0f4a643..5c5ced2 100644
--- a/ios/Runner.xcodeproj/project.pbxproj
+++ b/ios/Runner.xcodeproj/project.pbxproj
@@ -3,7 +3,7 @@
archiveVersion = 1;
classes = {
};
- objectVersion = 53;
+ objectVersion = 52;
objects = {
/* Begin PBXBuildFile section */
@@ -231,10 +231,10 @@
buildActionMask = 2147483647;
files = (
);
- inputPaths = (
+ inputFileListPaths = (
);
name = "[CP] Embed Pods Frameworks";
- outputPaths = (
+ outputFileListPaths = (
);
runOnlyForDeploymentPostprocessing = 0;
shellPath = /bin/sh;
@@ -314,10 +314,10 @@
buildActionMask = 2147483647;
files = (
);
- inputPaths = (
+ inputFileListPaths = (
);
name = "[CP] Copy Pods Resources";
- outputPaths = (
+ outputFileListPaths = (
);
runOnlyForDeploymentPostprocessing = 0;
shellPath = /bin/sh;
```
(_Is it normal for `pod` to rewrite this file?_ I would have guessed that it was read-only, and only XCode should modify it.)
## CocoaPods Environment
### Stack
```
CocoaPods : 1.9.2
Ruby : ruby 2.6.3p62 (2019-04-16 revision 67580) [universal.x86_64-darwin19]
RubyGems : 3.0.3
Host : Mac OS X 10.15.4 (19E287)
Xcode : 11.5 (11E608c)
Git : git version 2.24.3 (Apple Git-128)
Ruby lib dir : /System/Library/Frameworks/Ruby.framework/Versions/2.6/usr/lib
Repositories : master - git - https://github.com/CocoaPods/Specs.git @ 7a2a87b70aac3e71c0002b6ae4d4b6dc6090ae81
trunk - CDN - https://cdn.cocoapods.org/
```
### Installation Source
```
Executable Path: /usr/local/bin/pod
```
### Plugins
```
cocoapods-deintegrate : 1.0.4
cocoapods-plugins : 1.0.0
cocoapods-search : 1.0.0
cocoapods-stats : 1.1.0
cocoapods-trunk : 1.4.1
cocoapods-try : 1.1.0
```
### Podfile
```ruby
# Uncomment this line to define a global platform for your project
platform :ios, '10.1'
# CocoaPods analytics sends network stats synchronously affecting flutter build latency.
ENV['COCOAPODS_DISABLE_STATS'] = 'true'
project 'Runner', {
'Debug' => :debug,
'Profile' => :release,
'Release' => :release,
}
def parse_KV_file(file, separator='=')
file_abs_path = File.expand_path(file)
if !File.exists? file_abs_path
return [];
end
generated_key_values = {}
skip_line_start_symbols = ["#", "/"]
File.foreach(file_abs_path) do |line|
next if skip_line_start_symbols.any? { |symbol| line =~ /^\s*#{symbol}/ }
plugin = line.split(pattern=separator)
if plugin.length == 2
podname = plugin[0].strip()
path = plugin[1].strip()
podpath = File.expand_path("#{path}", file_abs_path)
generated_key_values[podname] = podpath
else
puts "Invalid plugin specification: #{line}"
end
end
generated_key_values
end
target 'Runner' do
use_frameworks!
use_modular_headers!
# Flutter Pod
copied_flutter_dir = File.join(__dir__, 'Flutter')
copied_framework_path = File.join(copied_flutter_dir, 'Flutter.framework')
copied_podspec_path = File.join(copied_flutter_dir, 'Flutter.podspec')
unless File.exist?(copied_framework_path) && File.exist?(copied_podspec_path)
# Copy Flutter.framework and Flutter.podspec to Flutter/ to have something to link against if the xcode backend script has not run yet.
# That script will copy the correct debug/profile/release version of the framework based on the currently selected Xcode configuration.
# CocoaPods will not embed the framework on pod install (before any build phases can generate) if the dylib does not exist.
generated_xcode_build_settings_path = File.join(copied_flutter_dir, 'Generated.xcconfig')
unless File.exist?(generated_xcode_build_settings_path)
raise "Generated.xcconfig must exist. If you're running pod install manually, make sure flutter pub get is executed first"
end
generated_xcode_build_settings = parse_KV_file(generated_xcode_build_settings_path)
cached_framework_dir = generated_xcode_build_settings['FLUTTER_FRAMEWORK_DIR'];
unless File.exist?(copied_framework_path)
FileUtils.cp_r(File.join(cached_framework_dir, 'Flutter.framework'), copied_flutter_dir)
end
unless File.exist?(copied_podspec_path)
FileUtils.cp(File.join(cached_framework_dir, 'Flutter.podspec'), copied_flutter_dir)
end
end
# Keep pod path relative so it can be checked into Podfile.lock.
pod 'Flutter', :path => 'Flutter'
# Plugin Pods
# Prepare symlinks folder. We use symlinks to avoid having Podfile.lock
# referring to absolute paths on developers' machines.
system('rm -rf .symlinks')
system('mkdir -p .symlinks/plugins')
plugin_pods = parse_KV_file('../.flutter-plugins')
plugin_pods.each do |name, path|
symlink = File.join('.symlinks', 'plugins', name)
File.symlink(path, symlink)
pod name, :path => File.join(symlink, 'ios')
end
end
# Prevent Cocoapods from embedding a second Flutter framework and causing an error with the new Xcode build system.
install! 'cocoapods', :disable_input_output_paths => true
post_install do |installer|
installer.pods_project.targets.each do |target|
target.build_configurations.each do |config|
config.build_settings['ENABLE_BITCODE'] = 'NO'
end
end
end
```
|
1.0
|
Unknown object version -- 53 - I am using **CocoaPods 1.9.2** (and Flutter 1.17.1).
* [X] I've read and understood the [*CONTRIBUTING* guidelines and have done my best effort to follow](https://github.com/CocoaPods/CocoaPods/blob/master/CONTRIBUTING.md).
# Report
## What did you do?
I have a `ios/Runner.xcodeproj/project.pbxproj` file with an `objectVersion = 53;` -- this was written by XCode.
`pod update` reports **Unknown object version.**
```
RuntimeError - [Xcodeproj] Unknown object version.
/Library/Ruby/Gems/2.6.0/gems/xcodeproj-1.15.0/lib/xcodeproj/project.rb:227:in `initialize_from_file'
/Library/Ruby/Gems/2.6.0/gems/xcodeproj-1.15.0/lib/xcodeproj/project.rb:112:in `open'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer/analyzer.rb:1177:in `block (2 levels) in inspect_targets_to_integrate'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer/analyzer.rb:1176:in `each'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer/analyzer.rb:1176:in `block in inspect_targets_to_integrate'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/user_interface.rb:64:in `section'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer/analyzer.rb:1171:in `inspect_targets_to_integrate'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer/analyzer.rb:106:in `analyze'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer.rb:410:in `analyze'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer.rb:235:in `block in resolve_dependencies'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/user_interface.rb:64:in `section'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer.rb:234:in `resolve_dependencies'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/installer.rb:156:in `install!'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/command/update.rb:63:in `run'
/Library/Ruby/Gems/2.6.0/gems/claide-1.0.3/lib/claide/command.rb:334:in `run'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/lib/cocoapods/command.rb:52:in `run'
/Library/Ruby/Gems/2.6.0/gems/cocoapods-1.9.2/bin/pod:55:in `<top (required)>'
/usr/local/bin/pod:23:in `load'
/usr/local/bin/pod:23:in `<main>'
```
## What did you expect to happen?
I did not expect a _Runtime Error_.
## What happened instead?
`pod` is broken.
A workaround is to hack the `objectVersion` -- changing from 53 to 52 does the trick. However, then `pod` makes a few other changes to the file that make me uncomfortable. Here is the `diff`:
```
diff --git a/ios/Runner.xcodeproj/project.pbxproj b/ios/Runner.xcodeproj/project.pbxproj
index 0f4a643..5c5ced2 100644
--- a/ios/Runner.xcodeproj/project.pbxproj
+++ b/ios/Runner.xcodeproj/project.pbxproj
@@ -3,7 +3,7 @@
archiveVersion = 1;
classes = {
};
- objectVersion = 53;
+ objectVersion = 52;
objects = {
/* Begin PBXBuildFile section */
@@ -231,10 +231,10 @@
buildActionMask = 2147483647;
files = (
);
- inputPaths = (
+ inputFileListPaths = (
);
name = "[CP] Embed Pods Frameworks";
- outputPaths = (
+ outputFileListPaths = (
);
runOnlyForDeploymentPostprocessing = 0;
shellPath = /bin/sh;
@@ -314,10 +314,10 @@
buildActionMask = 2147483647;
files = (
);
- inputPaths = (
+ inputFileListPaths = (
);
name = "[CP] Copy Pods Resources";
- outputPaths = (
+ outputFileListPaths = (
);
runOnlyForDeploymentPostprocessing = 0;
shellPath = /bin/sh;
```
(_Is it normal for `pod` to rewrite this file?_ I would have guessed that it was read-only, and only XCode should modify it.)
## CocoaPods Environment
### Stack
```
CocoaPods : 1.9.2
Ruby : ruby 2.6.3p62 (2019-04-16 revision 67580) [universal.x86_64-darwin19]
RubyGems : 3.0.3
Host : Mac OS X 10.15.4 (19E287)
Xcode : 11.5 (11E608c)
Git : git version 2.24.3 (Apple Git-128)
Ruby lib dir : /System/Library/Frameworks/Ruby.framework/Versions/2.6/usr/lib
Repositories : master - git - https://github.com/CocoaPods/Specs.git @ 7a2a87b70aac3e71c0002b6ae4d4b6dc6090ae81
trunk - CDN - https://cdn.cocoapods.org/
```
### Installation Source
```
Executable Path: /usr/local/bin/pod
```
### Plugins
```
cocoapods-deintegrate : 1.0.4
cocoapods-plugins : 1.0.0
cocoapods-search : 1.0.0
cocoapods-stats : 1.1.0
cocoapods-trunk : 1.4.1
cocoapods-try : 1.1.0
```
### Podfile
```ruby
# Uncomment this line to define a global platform for your project
platform :ios, '10.1'
# CocoaPods analytics sends network stats synchronously affecting flutter build latency.
ENV['COCOAPODS_DISABLE_STATS'] = 'true'
project 'Runner', {
'Debug' => :debug,
'Profile' => :release,
'Release' => :release,
}
def parse_KV_file(file, separator='=')
file_abs_path = File.expand_path(file)
if !File.exists? file_abs_path
return [];
end
generated_key_values = {}
skip_line_start_symbols = ["#", "/"]
File.foreach(file_abs_path) do |line|
next if skip_line_start_symbols.any? { |symbol| line =~ /^\s*#{symbol}/ }
plugin = line.split(pattern=separator)
if plugin.length == 2
podname = plugin[0].strip()
path = plugin[1].strip()
podpath = File.expand_path("#{path}", file_abs_path)
generated_key_values[podname] = podpath
else
puts "Invalid plugin specification: #{line}"
end
end
generated_key_values
end
target 'Runner' do
use_frameworks!
use_modular_headers!
# Flutter Pod
copied_flutter_dir = File.join(__dir__, 'Flutter')
copied_framework_path = File.join(copied_flutter_dir, 'Flutter.framework')
copied_podspec_path = File.join(copied_flutter_dir, 'Flutter.podspec')
unless File.exist?(copied_framework_path) && File.exist?(copied_podspec_path)
# Copy Flutter.framework and Flutter.podspec to Flutter/ to have something to link against if the xcode backend script has not run yet.
# That script will copy the correct debug/profile/release version of the framework based on the currently selected Xcode configuration.
# CocoaPods will not embed the framework on pod install (before any build phases can generate) if the dylib does not exist.
generated_xcode_build_settings_path = File.join(copied_flutter_dir, 'Generated.xcconfig')
unless File.exist?(generated_xcode_build_settings_path)
raise "Generated.xcconfig must exist. If you're running pod install manually, make sure flutter pub get is executed first"
end
generated_xcode_build_settings = parse_KV_file(generated_xcode_build_settings_path)
cached_framework_dir = generated_xcode_build_settings['FLUTTER_FRAMEWORK_DIR'];
unless File.exist?(copied_framework_path)
FileUtils.cp_r(File.join(cached_framework_dir, 'Flutter.framework'), copied_flutter_dir)
end
unless File.exist?(copied_podspec_path)
FileUtils.cp(File.join(cached_framework_dir, 'Flutter.podspec'), copied_flutter_dir)
end
end
# Keep pod path relative so it can be checked into Podfile.lock.
pod 'Flutter', :path => 'Flutter'
# Plugin Pods
# Prepare symlinks folder. We use symlinks to avoid having Podfile.lock
# referring to absolute paths on developers' machines.
system('rm -rf .symlinks')
system('mkdir -p .symlinks/plugins')
plugin_pods = parse_KV_file('../.flutter-plugins')
plugin_pods.each do |name, path|
symlink = File.join('.symlinks', 'plugins', name)
File.symlink(path, symlink)
pod name, :path => File.join(symlink, 'ios')
end
end
# Prevent Cocoapods from embedding a second Flutter framework and causing an error with the new Xcode build system.
install! 'cocoapods', :disable_input_output_paths => true
post_install do |installer|
installer.pods_project.targets.each do |target|
target.build_configurations.each do |config|
config.build_settings['ENABLE_BITCODE'] = 'NO'
end
end
end
```
|
defect
|
unknown object version i am using cocoapods and flutter i ve read and understood the report what did you do i have a ios runner xcodeproj project pbxproj file with an objectversion this was written by xcode pod update reports unknown object version runtimeerror unknown object version library ruby gems gems xcodeproj lib xcodeproj project rb in initialize from file library ruby gems gems xcodeproj lib xcodeproj project rb in open library ruby gems gems cocoapods lib cocoapods installer analyzer rb in block levels in inspect targets to integrate library ruby gems gems cocoapods lib cocoapods installer analyzer rb in each library ruby gems gems cocoapods lib cocoapods installer analyzer rb in block in inspect targets to integrate library ruby gems gems cocoapods lib cocoapods user interface rb in section library ruby gems gems cocoapods lib cocoapods installer analyzer rb in inspect targets to integrate library ruby gems gems cocoapods lib cocoapods installer analyzer rb in analyze library ruby gems gems cocoapods lib cocoapods installer rb in analyze library ruby gems gems cocoapods lib cocoapods installer rb in block in resolve dependencies library ruby gems gems cocoapods lib cocoapods user interface rb in section library ruby gems gems cocoapods lib cocoapods installer rb in resolve dependencies library ruby gems gems cocoapods lib cocoapods installer rb in install library ruby gems gems cocoapods lib cocoapods command update rb in run library ruby gems gems claide lib claide command rb in run library ruby gems gems cocoapods lib cocoapods command rb in run library ruby gems gems cocoapods bin pod in usr local bin pod in load usr local bin pod in what did you expect to happen i did not expect a runtime error what happened instead pod is broken a workaround is to hack the objectversion changing from to does the trick however then pod makes a few other changes to the file that make me uncomfortable here is the diff diff git a ios runner xcodeproj project pbxproj b ios runner xcodeproj project pbxproj index a ios runner xcodeproj project pbxproj b ios runner xcodeproj project pbxproj archiveversion classes objectversion objectversion objects begin pbxbuildfile section buildactionmask files inputpaths inputfilelistpaths name embed pods frameworks outputpaths outputfilelistpaths runonlyfordeploymentpostprocessing shellpath bin sh buildactionmask files inputpaths inputfilelistpaths name copy pods resources outputpaths outputfilelistpaths runonlyfordeploymentpostprocessing shellpath bin sh is it normal for pod to rewrite this file i would have guessed that it was read only and only xcode should modify it cocoapods environment stack cocoapods ruby ruby revision rubygems host mac os x xcode git git version apple git ruby lib dir system library frameworks ruby framework versions usr lib repositories master git trunk cdn installation source executable path usr local bin pod plugins cocoapods deintegrate cocoapods plugins cocoapods search cocoapods stats cocoapods trunk cocoapods try podfile ruby uncomment this line to define a global platform for your project platform ios cocoapods analytics sends network stats synchronously affecting flutter build latency env true project runner debug debug profile release release release def parse kv file file separator file abs path file expand path file if file exists file abs path return end generated key values skip line start symbols file foreach file abs path do line next if skip line start symbols any symbol line s symbol plugin line split pattern separator if plugin length podname plugin strip path plugin strip podpath file expand path path file abs path generated key values podpath else puts invalid plugin specification line end end generated key values end target runner do use frameworks use modular headers flutter pod copied flutter dir file join dir flutter copied framework path file join copied flutter dir flutter framework copied podspec path file join copied flutter dir flutter podspec unless file exist copied framework path file exist copied podspec path copy flutter framework and flutter podspec to flutter to have something to link against if the xcode backend script has not run yet that script will copy the correct debug profile release version of the framework based on the currently selected xcode configuration cocoapods will not embed the framework on pod install before any build phases can generate if the dylib does not exist generated xcode build settings path file join copied flutter dir generated xcconfig unless file exist generated xcode build settings path raise generated xcconfig must exist if you re running pod install manually make sure flutter pub get is executed first end generated xcode build settings parse kv file generated xcode build settings path cached framework dir generated xcode build settings unless file exist copied framework path fileutils cp r file join cached framework dir flutter framework copied flutter dir end unless file exist copied podspec path fileutils cp file join cached framework dir flutter podspec copied flutter dir end end keep pod path relative so it can be checked into podfile lock pod flutter path flutter plugin pods prepare symlinks folder we use symlinks to avoid having podfile lock referring to absolute paths on developers machines system rm rf symlinks system mkdir p symlinks plugins plugin pods parse kv file flutter plugins plugin pods each do name path symlink file join symlinks plugins name file symlink path symlink pod name path file join symlink ios end end prevent cocoapods from embedding a second flutter framework and causing an error with the new xcode build system install cocoapods disable input output paths true post install do installer installer pods project targets each do target target build configurations each do config config build settings no end end end
| 1
|
34,678
| 14,485,191,002
|
IssuesEvent
|
2020-12-10 17:16:38
|
Azure/azure-sdk-for-java
|
https://api.github.com/repos/Azure/azure-sdk-for-java
|
closed
|
[FEATURE REQ] make costmanagement compatible with last sdk version.
|
Cost Management Mgmt Service Attention customer-reported
|
**Is your feature request related to a problem? Please describe.**
I am using **com.microsoft.azure/azure@1.31.1** and I need to fetch cost information.
I have been looking everywhere in the sdk until I found **com.microsoft.azure.costmanagement.v2018_05_31/azure-mgmt-costmanagement@1.0.0-beta**.
I tried using it but I get:
`Caused by: java.lang.SecurityException: class "com.microsoft.azure.credentials.ApplicationTokenCredentials"'s signer information does not match signer information of other classes in the same package`
**Describe the solution you'd like**
Could a compatible version be made?
**Describe alternatives you've considered**
Stop using de sdk and doing all by hand/rest.
**Additional context**
None
**Information Checklist**
Kindly make sure that you have added all the following information above and checkoff the required fields otherwise we will treat the issuer as an incomplete report
- [ x] Description Added
- [ x] Expected solution specified
|
1.0
|
[FEATURE REQ] make costmanagement compatible with last sdk version. - **Is your feature request related to a problem? Please describe.**
I am using **com.microsoft.azure/azure@1.31.1** and I need to fetch cost information.
I have been looking everywhere in the sdk until I found **com.microsoft.azure.costmanagement.v2018_05_31/azure-mgmt-costmanagement@1.0.0-beta**.
I tried using it but I get:
`Caused by: java.lang.SecurityException: class "com.microsoft.azure.credentials.ApplicationTokenCredentials"'s signer information does not match signer information of other classes in the same package`
**Describe the solution you'd like**
Could a compatible version be made?
**Describe alternatives you've considered**
Stop using de sdk and doing all by hand/rest.
**Additional context**
None
**Information Checklist**
Kindly make sure that you have added all the following information above and checkoff the required fields otherwise we will treat the issuer as an incomplete report
- [ x] Description Added
- [ x] Expected solution specified
|
non_defect
|
make costmanagement compatible with last sdk version is your feature request related to a problem please describe i am using com microsoft azure azure and i need to fetch cost information i have been looking everywhere in the sdk until i found com microsoft azure costmanagement azure mgmt costmanagement beta i tried using it but i get caused by java lang securityexception class com microsoft azure credentials applicationtokencredentials s signer information does not match signer information of other classes in the same package describe the solution you d like could a compatible version be made describe alternatives you ve considered stop using de sdk and doing all by hand rest additional context none information checklist kindly make sure that you have added all the following information above and checkoff the required fields otherwise we will treat the issuer as an incomplete report description added expected solution specified
| 0
|
62,190
| 17,023,869,331
|
IssuesEvent
|
2021-07-03 04:16:43
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
coordinates shifting in the new interface
|
Component: website Priority: blocker Resolution: duplicate Type: defect
|
**[Submitted to the original trac issue database at 10.16pm, Monday, 22nd July 2013]**
The new interface (with big black on white icons upper-right), return wrong coordinates when Long link is asked.
After clicking at the position, I ask for a long link, to get coordinates, as I done before. But after clicking on the long link, the view is shifted right, and given coordinates are shifted left, so I can't use it to add geolocalisation on pictures on wikipedia commons, or give adress coordinates to contacts.
|
1.0
|
coordinates shifting in the new interface - **[Submitted to the original trac issue database at 10.16pm, Monday, 22nd July 2013]**
The new interface (with big black on white icons upper-right), return wrong coordinates when Long link is asked.
After clicking at the position, I ask for a long link, to get coordinates, as I done before. But after clicking on the long link, the view is shifted right, and given coordinates are shifted left, so I can't use it to add geolocalisation on pictures on wikipedia commons, or give adress coordinates to contacts.
|
defect
|
coordinates shifting in the new interface the new interface with big black on white icons upper right return wrong coordinates when long link is asked after clicking at the position i ask for a long link to get coordinates as i done before but after clicking on the long link the view is shifted right and given coordinates are shifted left so i can t use it to add geolocalisation on pictures on wikipedia commons or give adress coordinates to contacts
| 1
|
22,006
| 3,588,040,979
|
IssuesEvent
|
2016-01-30 19:06:46
|
bigbluebutton/bigbluebutton
|
https://api.github.com/repos/bigbluebutton/bigbluebutton
|
closed
|
Spaces in meetingid creates a bug for listeners feature
|
API Defect Normal Priority
|
Originally reported on Google Code with ID 942
```
See issue description
http://groups.google.com/group/bigbluebutton-dev/browse_thread/thread/5cc76d134f3da237#
If I create a meeting with spaces in the meeting ID, the Listeners
window pane will never get populated with attendees who are really
listening.
My current work around is to make sure there are no spaces in meeting
ID.
I suspect something related to communication b/t BBB and client
regarding listeners isn't escaping the spaces of the meeting ID.
```
Reported by `ffdixon` on 2011-05-27 11:10:01
|
1.0
|
Spaces in meetingid creates a bug for listeners feature - Originally reported on Google Code with ID 942
```
See issue description
http://groups.google.com/group/bigbluebutton-dev/browse_thread/thread/5cc76d134f3da237#
If I create a meeting with spaces in the meeting ID, the Listeners
window pane will never get populated with attendees who are really
listening.
My current work around is to make sure there are no spaces in meeting
ID.
I suspect something related to communication b/t BBB and client
regarding listeners isn't escaping the spaces of the meeting ID.
```
Reported by `ffdixon` on 2011-05-27 11:10:01
|
defect
|
spaces in meetingid creates a bug for listeners feature originally reported on google code with id see issue description if i create a meeting with spaces in the meeting id the listeners window pane will never get populated with attendees who are really listening my current work around is to make sure there are no spaces in meeting id i suspect something related to communication b t bbb and client regarding listeners isn t escaping the spaces of the meeting id reported by ffdixon on
| 1
|
300,981
| 22,707,594,128
|
IssuesEvent
|
2022-07-05 15:55:46
|
nine03/project-board
|
https://api.github.com/repos/nine03/project-board
|
closed
|
Github 프로젝트와 이슈 정리하기
|
documentation
|
Github 프로젝트를 세팅하고, 카드를 만들어 정리하자.
* [x] 프로젝트 베타 만들기
* [x] 카드 목록 만들기 - 강의 커리큘럼 참고
* [x] 이슈로 적절히 바꾸기
|
1.0
|
Github 프로젝트와 이슈 정리하기 - Github 프로젝트를 세팅하고, 카드를 만들어 정리하자.
* [x] 프로젝트 베타 만들기
* [x] 카드 목록 만들기 - 강의 커리큘럼 참고
* [x] 이슈로 적절히 바꾸기
|
non_defect
|
github 프로젝트와 이슈 정리하기 github 프로젝트를 세팅하고 카드를 만들어 정리하자 프로젝트 베타 만들기 카드 목록 만들기 강의 커리큘럼 참고 이슈로 적절히 바꾸기
| 0
|
74,255
| 25,028,695,521
|
IssuesEvent
|
2022-11-04 10:23:19
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
opened
|
UBSAN: array-index-out-of-bounds in module/zcommon/zfs_prop.c:861:24
|
Type: Defect
|
### System information
Type | Version/Name
--- | ---
Distribution Name | Ubuntu
Distribution Version | jammy 22.04
Kernel Version | 5.15.0-52-generic
Architecture | amd64
OpenZFS Version | zfs-2.1.99-1545_gd093d2c04 zfs-kmod-2.1.99-1545_gd093d2c04 (master as of Nov 4th 2022)
### Describe the problem you're observing
UBSan error in kernel space.
### Describe how to reproduce the problem
ZTS sanity
### Include any warning/errors/backtraces from the system logs
```
[ 1876.792222] UBSAN: array-index-out-of-bounds in /tmp/zfs-build-ubuntu-6BMluWMH/BUILD/zfs-kmod-2.1.99/_kmod_build_5.15.0-52-generic/../zfs-2.1.99/module/zcommon/zfs_prop.c:861:24
[ 1876.793153] index -1 is out of range for type 'zprop_desc_t [96]'
[ 1876.793531] CPU: 0 PID: 405702 Comm: txg_sync Tainted: P OE 5.15.0-52-generic #58-Ubuntu
[ 1876.793534] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.16.0-debian-1.16.0-4 04/01/2014
[ 1876.793534] Call Trace:
[ 1876.793536] <TASK>
[ 1876.793538] show_stack+0x52/0x5c
[ 1876.793542] dump_stack_lvl+0x4a/0x63
[ 1876.793547] dump_stack+0x10/0x16
[ 1876.793548] ubsan_epilogue+0x9/0x49
[ 1876.793549] __ubsan_handle_out_of_bounds.cold+0x44/0x49
[ 1876.793551] ? propname_match+0x70/0x70 [zfs]
[ 1876.793648] zfs_prop_get_type+0x3a/0x40 [zfs]
[ 1876.793706] dsl_prop_known_index+0x33/0x80 [zfs]
[ 1876.793792] dsl_prop_get_dd+0x17d/0x310 [zfs]
[ 1876.793877] dsl_prop_get_ds+0x1a6/0x2a0 [zfs]
[ 1876.793961] ? zap_remove_norm+0x7e/0xb0 [zfs]
[ 1876.794057] dsl_prop_get_int_ds+0x1e/0x30 [zfs]
[ 1876.794142] dsl_prop_set_sync_impl+0x436/0xa20 [zfs]
[ 1876.794226] ? dmu_buf_rele+0xe/0x20 [zfs]
[ 1876.794299] ? dsl_dir_rele+0x3e/0xc0 [zfs]
[ 1876.794383] dsl_props_set_sync_impl+0x6f/0x140 [zfs]
[ 1876.794468] dsl_props_set_sync+0x5e/0xb0 [zfs]
[ 1876.794551] ? zcp_synctask_inherit_prop_sync+0x20/0x20 [zfs]
[ 1876.794646] zcp_synctask_inherit_prop_sync+0x12/0x20 [zfs]
[ 1876.794741] zcp_sync_task+0x5f/0xd0 [zfs]
[ 1876.794837] ? zcp_synctask_inherit_prop+0x110/0x110 [zfs]
[ 1876.794939] zcp_synctask_inherit_prop+0xcf/0x110 [zfs]
[ 1876.795049] zcp_synctask_wrapper+0xa3/0x180 [zfs]
[ 1876.795550] ? dsl_dir_phys+0x10/0x10 [zfs]
[ 1876.795653] luaD_precall+0xdd/0x290 [zfs]
[ 1876.795718] luaV_execute+0xce4/0x1370 [zfs]
[ 1876.795785] luaD_call+0x119/0x130 [zfs]
[ 1876.795849] f_call+0x16/0x20 [zfs]
[ 1876.795911] luaD_rawrunprotected+0x62/0xa0 [zfs]
[ 1876.795975] ? lua_setmetatable+0x100/0x100 [zfs]
[ 1876.796037] ? luaD_rawrunprotected+0x50/0xa0 [zfs]
[ 1876.796101] luaD_pcall+0x3a/0xa0 [zfs]
[ 1876.796164] lua_pcallk+0x8c/0x130 [zfs]
[ 1876.796226] zcp_eval_impl+0xb9/0x450 [zfs]
[ 1876.796328] zcp_eval_sync+0x2c/0x60 [zfs]
[ 1876.796430] dsl_sync_task_sync+0xc5/0x140 [zfs]
[ 1876.796526] dsl_pool_sync+0x4ab/0x660 [zfs]
[ 1876.796619] spa_sync_iterate_to_convergence+0xdd/0x300 [zfs]
[ 1876.796717] spa_sync+0x2e0/0x830 [zfs]
[ 1876.796814] txg_sync_thread+0x260/0x380 [zfs]
[ 1876.796913] ? txg_dispatch_callbacks+0x100/0x100 [zfs]
[ 1876.797013] thread_generic_wrapper+0x6c/0xa0 [spl]
[ 1876.797021] ? spl_assert.constprop.0+0x30/0x30 [spl]
[ 1876.797027] kthread+0x127/0x150
[ 1876.797030] ? set_kthread_struct+0x50/0x50
[ 1876.797032] ret_from_fork+0x1f/0x30
[ 1876.797035] </TASK>
```
|
1.0
|
UBSAN: array-index-out-of-bounds in module/zcommon/zfs_prop.c:861:24 - ### System information
Type | Version/Name
--- | ---
Distribution Name | Ubuntu
Distribution Version | jammy 22.04
Kernel Version | 5.15.0-52-generic
Architecture | amd64
OpenZFS Version | zfs-2.1.99-1545_gd093d2c04 zfs-kmod-2.1.99-1545_gd093d2c04 (master as of Nov 4th 2022)
### Describe the problem you're observing
UBSan error in kernel space.
### Describe how to reproduce the problem
ZTS sanity
### Include any warning/errors/backtraces from the system logs
```
[ 1876.792222] UBSAN: array-index-out-of-bounds in /tmp/zfs-build-ubuntu-6BMluWMH/BUILD/zfs-kmod-2.1.99/_kmod_build_5.15.0-52-generic/../zfs-2.1.99/module/zcommon/zfs_prop.c:861:24
[ 1876.793153] index -1 is out of range for type 'zprop_desc_t [96]'
[ 1876.793531] CPU: 0 PID: 405702 Comm: txg_sync Tainted: P OE 5.15.0-52-generic #58-Ubuntu
[ 1876.793534] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.16.0-debian-1.16.0-4 04/01/2014
[ 1876.793534] Call Trace:
[ 1876.793536] <TASK>
[ 1876.793538] show_stack+0x52/0x5c
[ 1876.793542] dump_stack_lvl+0x4a/0x63
[ 1876.793547] dump_stack+0x10/0x16
[ 1876.793548] ubsan_epilogue+0x9/0x49
[ 1876.793549] __ubsan_handle_out_of_bounds.cold+0x44/0x49
[ 1876.793551] ? propname_match+0x70/0x70 [zfs]
[ 1876.793648] zfs_prop_get_type+0x3a/0x40 [zfs]
[ 1876.793706] dsl_prop_known_index+0x33/0x80 [zfs]
[ 1876.793792] dsl_prop_get_dd+0x17d/0x310 [zfs]
[ 1876.793877] dsl_prop_get_ds+0x1a6/0x2a0 [zfs]
[ 1876.793961] ? zap_remove_norm+0x7e/0xb0 [zfs]
[ 1876.794057] dsl_prop_get_int_ds+0x1e/0x30 [zfs]
[ 1876.794142] dsl_prop_set_sync_impl+0x436/0xa20 [zfs]
[ 1876.794226] ? dmu_buf_rele+0xe/0x20 [zfs]
[ 1876.794299] ? dsl_dir_rele+0x3e/0xc0 [zfs]
[ 1876.794383] dsl_props_set_sync_impl+0x6f/0x140 [zfs]
[ 1876.794468] dsl_props_set_sync+0x5e/0xb0 [zfs]
[ 1876.794551] ? zcp_synctask_inherit_prop_sync+0x20/0x20 [zfs]
[ 1876.794646] zcp_synctask_inherit_prop_sync+0x12/0x20 [zfs]
[ 1876.794741] zcp_sync_task+0x5f/0xd0 [zfs]
[ 1876.794837] ? zcp_synctask_inherit_prop+0x110/0x110 [zfs]
[ 1876.794939] zcp_synctask_inherit_prop+0xcf/0x110 [zfs]
[ 1876.795049] zcp_synctask_wrapper+0xa3/0x180 [zfs]
[ 1876.795550] ? dsl_dir_phys+0x10/0x10 [zfs]
[ 1876.795653] luaD_precall+0xdd/0x290 [zfs]
[ 1876.795718] luaV_execute+0xce4/0x1370 [zfs]
[ 1876.795785] luaD_call+0x119/0x130 [zfs]
[ 1876.795849] f_call+0x16/0x20 [zfs]
[ 1876.795911] luaD_rawrunprotected+0x62/0xa0 [zfs]
[ 1876.795975] ? lua_setmetatable+0x100/0x100 [zfs]
[ 1876.796037] ? luaD_rawrunprotected+0x50/0xa0 [zfs]
[ 1876.796101] luaD_pcall+0x3a/0xa0 [zfs]
[ 1876.796164] lua_pcallk+0x8c/0x130 [zfs]
[ 1876.796226] zcp_eval_impl+0xb9/0x450 [zfs]
[ 1876.796328] zcp_eval_sync+0x2c/0x60 [zfs]
[ 1876.796430] dsl_sync_task_sync+0xc5/0x140 [zfs]
[ 1876.796526] dsl_pool_sync+0x4ab/0x660 [zfs]
[ 1876.796619] spa_sync_iterate_to_convergence+0xdd/0x300 [zfs]
[ 1876.796717] spa_sync+0x2e0/0x830 [zfs]
[ 1876.796814] txg_sync_thread+0x260/0x380 [zfs]
[ 1876.796913] ? txg_dispatch_callbacks+0x100/0x100 [zfs]
[ 1876.797013] thread_generic_wrapper+0x6c/0xa0 [spl]
[ 1876.797021] ? spl_assert.constprop.0+0x30/0x30 [spl]
[ 1876.797027] kthread+0x127/0x150
[ 1876.797030] ? set_kthread_struct+0x50/0x50
[ 1876.797032] ret_from_fork+0x1f/0x30
[ 1876.797035] </TASK>
```
|
defect
|
ubsan array index out of bounds in module zcommon zfs prop c system information type version name distribution name ubuntu distribution version jammy kernel version generic architecture openzfs version zfs zfs kmod master as of nov describe the problem you re observing ubsan error in kernel space describe how to reproduce the problem zts sanity include any warning errors backtraces from the system logs ubsan array index out of bounds in tmp zfs build ubuntu build zfs kmod kmod build generic zfs module zcommon zfs prop c index is out of range for type zprop desc t cpu pid comm txg sync tainted p oe generic ubuntu hardware name qemu standard pc piix bios debian call trace show stack dump stack lvl dump stack ubsan epilogue ubsan handle out of bounds cold propname match zfs prop get type dsl prop known index dsl prop get dd dsl prop get ds zap remove norm dsl prop get int ds dsl prop set sync impl dmu buf rele dsl dir rele dsl props set sync impl dsl props set sync zcp synctask inherit prop sync zcp synctask inherit prop sync zcp sync task zcp synctask inherit prop zcp synctask inherit prop zcp synctask wrapper dsl dir phys luad precall luav execute luad call f call luad rawrunprotected lua setmetatable luad rawrunprotected luad pcall lua pcallk zcp eval impl zcp eval sync dsl sync task sync dsl pool sync spa sync iterate to convergence spa sync txg sync thread txg dispatch callbacks thread generic wrapper spl assert constprop kthread set kthread struct ret from fork
| 1
|
78,926
| 15,094,899,120
|
IssuesEvent
|
2021-02-07 08:43:43
|
joomla/joomla-cms
|
https://api.github.com/repos/joomla/joomla-cms
|
closed
|
[4b6] Override created message twice
|
No Code Attached Yet good first issue
|
### Steps to reproduce the issue
Create an override in Joomla 4 admin
### Expected result
"Override created in /templates/cassiopeia/html/mod_banners"
### Actual result
Two messages - one is enough!
<img width="1031" alt="Screenshot 2021-01-08 at 14 28 36" src="https://user-images.githubusercontent.com/400092/104026432-ce7e8900-51bd-11eb-9fec-3d288783b081.png">
|
1.0
|
[4b6] Override created message twice - ### Steps to reproduce the issue
Create an override in Joomla 4 admin
### Expected result
"Override created in /templates/cassiopeia/html/mod_banners"
### Actual result
Two messages - one is enough!
<img width="1031" alt="Screenshot 2021-01-08 at 14 28 36" src="https://user-images.githubusercontent.com/400092/104026432-ce7e8900-51bd-11eb-9fec-3d288783b081.png">
|
non_defect
|
override created message twice steps to reproduce the issue create an override in joomla admin expected result override created in templates cassiopeia html mod banners actual result two messages one is enough img width alt screenshot at src
| 0
|
460,751
| 13,217,740,979
|
IssuesEvent
|
2020-08-17 07:24:45
|
nimblehq/rails-templates
|
https://api.github.com/repos/nimblehq/rails-templates
|
closed
|
Update nimble branding from `nimbl3` to `nimblehq`
|
priority : normal type : improvement
|
1. The screenshot bucket should be renamed
2. Our eslint package needs to be renamed - This will be related with another repo - `@nimbl3/eslint-config-nimbl3`
|
1.0
|
Update nimble branding from `nimbl3` to `nimblehq` - 1. The screenshot bucket should be renamed
2. Our eslint package needs to be renamed - This will be related with another repo - `@nimbl3/eslint-config-nimbl3`
|
non_defect
|
update nimble branding from to nimblehq the screenshot bucket should be renamed our eslint package needs to be renamed this will be related with another repo eslint config
| 0
|
178,871
| 21,509,573,061
|
IssuesEvent
|
2022-04-28 01:55:58
|
eldorplus/conception-website
|
https://api.github.com/repos/eldorplus/conception-website
|
closed
|
WS-2019-0019 (Medium) detected in braces-1.8.5.tgz, braces-0.1.5.tgz - autoclosed
|
security vulnerability
|
## WS-2019-0019 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>braces-1.8.5.tgz</b>, <b>braces-0.1.5.tgz</b></p></summary>
<p>
<details><summary><b>braces-1.8.5.tgz</b></p></summary>
<p>Fastest brace expansion for node.js, with the most complete support for the Bash 4.3 braces specification.</p>
<p>Library home page: <a href="https://registry.npmjs.org/braces/-/braces-1.8.5.tgz">https://registry.npmjs.org/braces/-/braces-1.8.5.tgz</a></p>
<p>Path to dependency file: /conception-website/package.json</p>
<p>Path to vulnerable library: /tmp/git/conception-website/node_modules/test-exclude/node_modules/braces/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.7.1.tgz (Root Library)
- chokidar-1.7.0.tgz
- anymatch-1.3.2.tgz
- micromatch-2.3.11.tgz
- :x: **braces-1.8.5.tgz** (Vulnerable Library)
</details>
<details><summary><b>braces-0.1.5.tgz</b></p></summary>
<p>Fastest brace expansion lib. Typically used with file paths, but can be used with any string. Expands comma-separated values (e.g. `foo/{a,b,c}/bar`) and alphabetical or numerical ranges (e.g. `{1..9}`)</p>
<p>Library home page: <a href="https://registry.npmjs.org/braces/-/braces-0.1.5.tgz">https://registry.npmjs.org/braces/-/braces-0.1.5.tgz</a></p>
<p>Path to dependency file: /conception-website/package.json</p>
<p>Path to vulnerable library: /tmp/git/conception-website/node_modules/expand-braces/node_modules/braces/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.7.1.tgz (Root Library)
- expand-braces-0.1.2.tgz
- :x: **braces-0.1.5.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/eldorplus/conception-website/commits/aafabe1f890f6614128b3e3a46fcacc878d945e9">aafabe1f890f6614128b3e3a46fcacc878d945e9</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Version of braces prior to 2.3.1 are vulnerable to Regular Expression Denial of Service (ReDoS). Untrusted input may cause catastrophic backtracking while matching regular expressions. This can cause the application to be unresponsive leading to Denial of Service.
<p>Publish Date: 2019-03-25
<p>URL: <a href=https://github.com/micromatch/braces/commit/abdafb0cae1e0c00f184abbadc692f4eaa98f451>WS-2019-0019</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/786">https://www.npmjs.com/advisories/786</a></p>
<p>Release Date: 2019-02-21</p>
<p>Fix Resolution: 2.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2019-0019 (Medium) detected in braces-1.8.5.tgz, braces-0.1.5.tgz - autoclosed - ## WS-2019-0019 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>braces-1.8.5.tgz</b>, <b>braces-0.1.5.tgz</b></p></summary>
<p>
<details><summary><b>braces-1.8.5.tgz</b></p></summary>
<p>Fastest brace expansion for node.js, with the most complete support for the Bash 4.3 braces specification.</p>
<p>Library home page: <a href="https://registry.npmjs.org/braces/-/braces-1.8.5.tgz">https://registry.npmjs.org/braces/-/braces-1.8.5.tgz</a></p>
<p>Path to dependency file: /conception-website/package.json</p>
<p>Path to vulnerable library: /tmp/git/conception-website/node_modules/test-exclude/node_modules/braces/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.7.1.tgz (Root Library)
- chokidar-1.7.0.tgz
- anymatch-1.3.2.tgz
- micromatch-2.3.11.tgz
- :x: **braces-1.8.5.tgz** (Vulnerable Library)
</details>
<details><summary><b>braces-0.1.5.tgz</b></p></summary>
<p>Fastest brace expansion lib. Typically used with file paths, but can be used with any string. Expands comma-separated values (e.g. `foo/{a,b,c}/bar`) and alphabetical or numerical ranges (e.g. `{1..9}`)</p>
<p>Library home page: <a href="https://registry.npmjs.org/braces/-/braces-0.1.5.tgz">https://registry.npmjs.org/braces/-/braces-0.1.5.tgz</a></p>
<p>Path to dependency file: /conception-website/package.json</p>
<p>Path to vulnerable library: /tmp/git/conception-website/node_modules/expand-braces/node_modules/braces/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.7.1.tgz (Root Library)
- expand-braces-0.1.2.tgz
- :x: **braces-0.1.5.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/eldorplus/conception-website/commits/aafabe1f890f6614128b3e3a46fcacc878d945e9">aafabe1f890f6614128b3e3a46fcacc878d945e9</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Version of braces prior to 2.3.1 are vulnerable to Regular Expression Denial of Service (ReDoS). Untrusted input may cause catastrophic backtracking while matching regular expressions. This can cause the application to be unresponsive leading to Denial of Service.
<p>Publish Date: 2019-03-25
<p>URL: <a href=https://github.com/micromatch/braces/commit/abdafb0cae1e0c00f184abbadc692f4eaa98f451>WS-2019-0019</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/786">https://www.npmjs.com/advisories/786</a></p>
<p>Release Date: 2019-02-21</p>
<p>Fix Resolution: 2.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
ws medium detected in braces tgz braces tgz autoclosed ws medium severity vulnerability vulnerable libraries braces tgz braces tgz braces tgz fastest brace expansion for node js with the most complete support for the bash braces specification library home page a href path to dependency file conception website package json path to vulnerable library tmp git conception website node modules test exclude node modules braces package json dependency hierarchy karma tgz root library chokidar tgz anymatch tgz micromatch tgz x braces tgz vulnerable library braces tgz fastest brace expansion lib typically used with file paths but can be used with any string expands comma separated values e g foo a b c bar and alphabetical or numerical ranges e g library home page a href path to dependency file conception website package json path to vulnerable library tmp git conception website node modules expand braces node modules braces package json dependency hierarchy karma tgz root library expand braces tgz x braces tgz vulnerable library found in head commit a href vulnerability details version of braces prior to are vulnerable to regular expression denial of service redos untrusted input may cause catastrophic backtracking while matching regular expressions this can cause the application to be unresponsive leading to denial of service publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
64,776
| 8,758,832,182
|
IssuesEvent
|
2018-12-15 09:13:59
|
usvc/accounts
|
https://api.github.com/repos/usvc/accounts
|
opened
|
STICKY: Documentation improvements
|
documentation
|
**DO NOT CLOSE THIS ISSUE**
Reference this issue for all documentation additions/improvements outside the scope of a feature.
|
1.0
|
STICKY: Documentation improvements - **DO NOT CLOSE THIS ISSUE**
Reference this issue for all documentation additions/improvements outside the scope of a feature.
|
non_defect
|
sticky documentation improvements do not close this issue reference this issue for all documentation additions improvements outside the scope of a feature
| 0
|
75,926
| 26,155,035,006
|
IssuesEvent
|
2022-12-30 19:59:16
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Voice message timeline is inaccurate on Element Web
|
T-Defect
|
### Steps to reproduce
1. Tap and hold microphone button on Element iOS or Android to record a short voice message and post it in the room
2. The message is posted correctly by the voice curve rendering varies quite a bit on Web vs iOS vs Android
3. Play back the message and note where the audio starts in relation to the timeline indicator.
### Outcome
#### What did you expect?
1. The curves should be largely the same on all platforms
2. The curve should match when audio is audible
#### What happened instead?
Including some screenshots to illustrate the difference in the rendered voice curves as well as a red arrow roughly where I can start hearing audio from the message. As far as I can tell, Android has the most accurate curve rendering and iOS has the most accurate timeline position indicator for when I actually hear something.
## Element Web

## Element iOS

## Element Android

g)
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
_No response_
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
No
|
1.0
|
Voice message timeline is inaccurate on Element Web - ### Steps to reproduce
1. Tap and hold microphone button on Element iOS or Android to record a short voice message and post it in the room
2. The message is posted correctly by the voice curve rendering varies quite a bit on Web vs iOS vs Android
3. Play back the message and note where the audio starts in relation to the timeline indicator.
### Outcome
#### What did you expect?
1. The curves should be largely the same on all platforms
2. The curve should match when audio is audible
#### What happened instead?
Including some screenshots to illustrate the difference in the rendered voice curves as well as a red arrow roughly where I can start hearing audio from the message. As far as I can tell, Android has the most accurate curve rendering and iOS has the most accurate timeline position indicator for when I actually hear something.
## Element Web

## Element iOS

## Element Android

g)
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
_No response_
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
No
|
defect
|
voice message timeline is inaccurate on element web steps to reproduce tap and hold microphone button on element ios or android to record a short voice message and post it in the room the message is posted correctly by the voice curve rendering varies quite a bit on web vs ios vs android play back the message and note where the audio starts in relation to the timeline indicator outcome what did you expect the curves should be largely the same on all platforms the curve should match when audio is audible what happened instead including some screenshots to illustrate the difference in the rendered voice curves as well as a red arrow roughly where i can start hearing audio from the message as far as i can tell android has the most accurate curve rendering and ios has the most accurate timeline position indicator for when i actually hear something element web element ios element android g operating system no response browser information no response url for webapp no response application version no response homeserver no response will you send logs no
| 1
|
71,941
| 23,863,515,152
|
IssuesEvent
|
2022-09-07 09:05:11
|
SeleniumHQ/selenium
|
https://api.github.com/repos/SeleniumHQ/selenium
|
opened
|
[🐛 Bug]: '--lang=en' or '--accept-lang=en' option not working with Node.js on Mac OS
|
I-defect needs-triaging
|
### What happened?
Hi ! ✌️
I try to create a web end to end automation framework for my company and I need to be able to run tests in multiple browsers and in multiple languages.
Current targets are:
- browsers: chrome and firefox
- languages: deutsh, english, french, italian, spanish (more to come later)
I am doing a Proof Of Concept project to see how it could be achieved with either Selenium or Puppeteer
I am trying to implement a simple scenario:
- Open Google home page
- Verify in the popup that the title is in the right language
Examples:
```
browser | langISO639 | countryCodeISO3166 | expectedTitle
${'chrome'} | ${'de'} | ${'DE'} | ${'Bevor Sie zu Google weitergehen'}
${'chrome'} | ${'en'} | ${'GB'} | ${'Before you continue to Google'}
${'chrome'} | ${'es'} | ${'ES'} | ${'Antes de ir a Google'}
${'chrome'} | ${'fr'} | ${'FR'} | ${'Avant d\'accéder à Google'}
${'firefox'} | ${'de'} | ${'DE'} | ${'Bevor Sie zu Google weitergehen'}
${'firefox'} | ${'en'} | ${'GB'} | ${'Before you continue to Google'}
${'firefox'} | ${'es'} | ${'ES'} | ${'Antes de ir a Google'}
${'firefox'} | ${'fr'} | ${'FR'} | ${'Avant d\'accéder à Google'}
```
My problem is that english test case is always failed and i've tried all workarounds i've found around the web (github issues, stack overflow, ...)
PS: I am using webdriver-manager to run a local grid but It is also failed with chromedriver and geckodriver local installations (from NPM)
### How can we reproduce the issue?
```shell
Node.js test script with selenium
const fs = require('fs')
const path = require('path')
const mkdirp = require('mkdirp')
const rimraf = require('rimraf')
const { Builder, until, By } = require('selenium-webdriver')
const Chrome = require('selenium-webdriver/chrome')
const Firefox = require('selenium-webdriver/firefox')
require('chromedriver')
require('geckodriver')
jest.setTimeout(1000*60*10)
async function getWebDriver(browser, lang, country) {
if (browser === 'chrome') {
const chromeOptions = new Chrome.Options()
.addArguments(
`--accept-lang=${lang}-${country}`,
'--browser-test',
'--bwsi',
`--default-country-code=${country}`,
'--disable-default-apps',
'--disable-extensions',
'--disable-gpu',
'--disable-logging',
'--disable-web-security',
'--dom-automation',
'--enable-automation',
'--force-headless-for-tests',
'--guest',
'--headless',
'--incognito',
`--lang=${lang}`,
'--no-sandbox',
'--window-size=1440,900',
)
.setUserPreferences({ ['intl.accept_languages']: lang, ['translate']: { enabled: true } })
const driver = new Builder()
.forBrowser('chrome')
.setChromeOptions(chromeOptions)
.usingServer('http://localhost:4444/wd/hub')
.build()
await sleep(3)
return driver
} else {
const firefoxOptions = new Firefox.Options()
.headless()
.setPreference('intl.accept_languages', lang)
const driver = new Builder()
.forBrowser('firefox')
.setFirefoxOptions(firefoxOptions)
.usingServer('http://localhost:4444/wd/hub')
.build()
await sleep(3)
return driver
}
}
async function sleep(seconds) {
return new Promise((resolve) => {
setTimeout(() => {
resolve()
}, 1000 * seconds)
})
}
describe.each`
browser | langISO639 | countryCodeISO3166 | expectedTitle
${'chrome'} | ${'de'} | ${'DE'} | ${'Bevor Sie zu Google weitergehen'}
${'chrome'} | ${'en'} | ${'GB'} | ${'Before you continue to Google'}
${'chrome'} | ${'es'} | ${'ES'} | ${'Antes de ir a Google'}
${'chrome'} | ${'fr'} | ${'FR'} | ${'Avant d\'accéder à Google'}
${'firefox'} | ${'de'} | ${'DE'} | ${'Bevor Sie zu Google weitergehen'}
${'firefox'} | ${'en'} | ${'GB'} | ${'Before you continue to Google'}
${'firefox'} | ${'es'} | ${'ES'} | ${'Antes de ir a Google'}
${'firefox'} | ${'fr'} | ${'FR'} | ${'Avant d\'accéder à Google'}
`('open Google on $browser browser with language $langISO639', ({ browser, langISO639, countryCodeISO3166, expectedTitle }) => {
let envLangBefore = process.env['LANG'] || undefined
beforeAll(async () => {
await rimraf.sync(path.join(__dirname, 'test-screenshots'))
await mkdirp.sync(path.join(__dirname, 'test-screenshots'))
process.env['LANG'] = langISO639
})
afterAll(() => {
process.env['LANG'] = envLangBefore
})
let driver
it('opens web driver instance', async () => {
driver = await getWebDriver(browser, langISO639, countryCodeISO3166)
})
it('reach goole web app', async () => {
await driver.get('https://www.google.com/')
await driver.navigate().refresh()
})
it('waits for translated text or take a screenshot and throw', async () => {
try {
await driver.wait(until.elementLocated(By.xpath(`//h1[text()="${expectedTitle}"]`)), 3000)
} catch (e) {
const image = await driver.takeScreenshot()
fs.writeFileSync(path.join(__dirname, `test-screenshots/${browser}_${langISO639}.png`), image, 'base64')
throw e
}
})
it('closes web driver instance', async () => {
await driver.quit()
})
})
```
Node.js test script with puppeteer
```
const fs = require('fs')
const path = require('path')
const mkdirp = require('mkdirp')
const puppeteer = require('puppeteer')
const rimraf = require('rimraf')
jest.setTimeout(1000*60*10)
async function getWebDriver(lang, country) {
const driver = await puppeteer.launch({
args: [
`--accept-lang=${lang}-${country}`,
'--browser-test',
'--bwsi',
`--default-country-code=${country}`,
'--disable-default-apps',
'--disable-extensions',
'--disable-gpu',
'--disable-logging',
'--disable-web-security',
'--dom-automation',
'--enable-automation',
'--force-headless-for-tests',
'--guest',
'--headless',
'--incognito',
`--lang=${lang}`,
'--no-sandbox',
'--window-size=1440,900',
]
})
await sleep(3)
return driver
}
async function sleep(seconds) {
return new Promise((resolve) => {
setTimeout(() => {
resolve()
}, 1000 * seconds)
})
}
describe.each`
langISO639 | countryCodeISO3166 | expectedTitle
${'de'} | ${'DE'} | ${'Bevor Sie zu Google weitergehen'}
${'en'} | ${'GB'} | ${'Before you continue to Google'}
${'es'} | ${'ES'} | ${'Antes de ir a Google'}
${'fr'} | ${'FR'} | ${'Avant d\'accéder à Google'}
`('open Puppeteer with language $langISO639', ({ langISO639, countryCodeISO3166, expectedTitle }) => {
let envLangBefore = process.env['LANG'] || undefined
beforeAll(async () => {
await rimraf.sync(path.join(__dirname, 'test-screenshots'))
await mkdirp.sync(path.join(__dirname, 'test-screenshots'))
process.env['LANG'] = langISO639
})
afterAll(() => {
process.env['LANG'] = envLangBefore
})
/** @type {puppeteer.Browser} */
let browser
/** @type {puppeteer.Page} */
let page
it('opens web driver instance', async () => {
browser = await getWebDriver(langISO639, countryCodeISO3166)
page = await browser.newPage()
})
it('reach goole web app', async () => {
await page.goto('https://www.google.com/')
await page.reload()
})
it('waits for translated text or take a screenshot and throw', async () => {
try {
await page.waitForXPath(`//h1[text()="${expectedTitle}"]`)
} catch (e) {
await page.screenshot({ path: path.join(__dirname, `test-screenshots/puppeteer_${langISO639}.png`) })
throw e
}
})
it('closes web driver instance', async () => {
await browser.close()
})
})
```
Node.js project package.json example
```
{
"name": "sample-en-issue-with-chromium",
"scripts": {
"test": "jest"
},
"devDependencies": {
"@types/jest": "^29.0.0",
"@types/selenium-webdriver": "^4.1.3",
"eslint": "^8.23.0",
"eslint-config-node": "^4.1.0",
"eslint-plugin-import": "^2.26.0",
"eslint-plugin-jest": "^27.0.1",
"jest": "^29.0.1",
"jest-runner-groups": "^2.2.0",
"webdriver-manager": "^12.1.8"
},
"dependencies": {
"chromedriver": "^104.0.0",
"geckodriver": "^3.0.2",
"mkdirp": "^1.0.4",
"puppeteer": "^17.1.1",
"rimraf": "^3.0.2",
"selenium-webdriver": "^4.4.0"
}
}
```
```
### Relevant log output
```shell
With Selenium
FAIL framework/automate-web/WebDriverManager.test.js (45.459 s)
open Google on chrome browser with language de
✓ opens web driver instance (3003 ms)
✓ reach goole web app (1587 ms)
✓ waits for translated text or take a screenshot and throw (527 ms)
✓ closes web driver instance (54 ms)
open Google on chrome browser with language en
✓ opens web driver instance (3002 ms)
✓ reach goole web app (1364 ms)
✕ waits for translated text or take a screenshot and throw (3506 ms)
✓ closes web driver instance (55 ms)
open Google on chrome browser with language es
✓ opens web driver instance (3002 ms)
✓ reach goole web app (1371 ms)
✓ waits for translated text or take a screenshot and throw (485 ms)
✓ closes web driver instance (58 ms)
open Google on chrome browser with language fr
✓ opens web driver instance (3001 ms)
✓ reach goole web app (1414 ms)
✓ waits for translated text or take a screenshot and throw (490 ms)
✓ closes web driver instance (55 ms)
open Google on firefox browser with language de
✓ opens web driver instance (3001 ms)
✓ reach goole web app (909 ms)
✓ waits for translated text or take a screenshot and throw (22 ms)
✓ closes web driver instance (404 ms)
open Google on firefox browser with language en
✓ opens web driver instance (3002 ms)
✓ reach goole web app (786 ms)
✕ waits for translated text or take a screenshot and throw (3202 ms)
✓ closes web driver instance (409 ms)
open Google on firefox browser with language es
✓ opens web driver instance (3001 ms)
✓ reach goole web app (2718 ms)
✓ waits for translated text or take a screenshot and throw (17 ms)
✓ closes web driver instance (408 ms)
open Google on firefox browser with language fr
✓ opens web driver instance (3002 ms)
✓ reach goole web app (788 ms)
✓ waits for translated text or take a screenshot and throw (19 ms)
✓ closes web driver instance (406 ms)
```
With Puppeteer
```
FAIL framework/automate-web/test-screenshots/Puppeteer-multiple-langs.test.js (63.579 s)
open Puppeteer with language de
✓ opens web driver instance (19901 ms)
✓ reach goole web app (852 ms)
✓ waits for translated text or take a screenshot and throw (8 ms)
✓ closes web driver instance (8 ms)
open Puppeteer with language en
✓ opens web driver instance (3319 ms)
✓ reach goole web app (679 ms)
✕ waits for translated text or take a screenshot and throw (30071 ms)
✓ closes web driver instance (7 ms)
open Puppeteer with language es
✓ opens web driver instance (3324 ms)
✓ reach goole web app (825 ms)
✓ waits for translated text or take a screenshot and throw (6 ms)
✓ closes web driver instance (11 ms)
open Puppeteer with language fr
✓ opens web driver instance (3334 ms)
✓ reach goole web app (705 ms)
✓ waits for translated text or take a screenshot and throw (7 ms)
✓ closes web driver instance (6 ms)
```
```
### Operating System
Mac OS Monterey 12.5.1
### Selenium version
"selenium-webdriver": "^4.4.0",
### What are the browser(s) and version(s) where you see this issue?
Firefox 104.0.1, Chrome version as of 2022-09-06
### What are the browser driver(s) and version(s) where you see this issue?
from NPM: {"chromedriver": "^104.0.0", "geckodriver": "^3.0.2"}
### Are you using Selenium Grid?
yes with npm module webdriver-manager / no
|
1.0
|
[🐛 Bug]: '--lang=en' or '--accept-lang=en' option not working with Node.js on Mac OS - ### What happened?
Hi ! ✌️
I try to create a web end to end automation framework for my company and I need to be able to run tests in multiple browsers and in multiple languages.
Current targets are:
- browsers: chrome and firefox
- languages: deutsh, english, french, italian, spanish (more to come later)
I am doing a Proof Of Concept project to see how it could be achieved with either Selenium or Puppeteer
I am trying to implement a simple scenario:
- Open Google home page
- Verify in the popup that the title is in the right language
Examples:
```
browser | langISO639 | countryCodeISO3166 | expectedTitle
${'chrome'} | ${'de'} | ${'DE'} | ${'Bevor Sie zu Google weitergehen'}
${'chrome'} | ${'en'} | ${'GB'} | ${'Before you continue to Google'}
${'chrome'} | ${'es'} | ${'ES'} | ${'Antes de ir a Google'}
${'chrome'} | ${'fr'} | ${'FR'} | ${'Avant d\'accéder à Google'}
${'firefox'} | ${'de'} | ${'DE'} | ${'Bevor Sie zu Google weitergehen'}
${'firefox'} | ${'en'} | ${'GB'} | ${'Before you continue to Google'}
${'firefox'} | ${'es'} | ${'ES'} | ${'Antes de ir a Google'}
${'firefox'} | ${'fr'} | ${'FR'} | ${'Avant d\'accéder à Google'}
```
My problem is that english test case is always failed and i've tried all workarounds i've found around the web (github issues, stack overflow, ...)
PS: I am using webdriver-manager to run a local grid but It is also failed with chromedriver and geckodriver local installations (from NPM)
### How can we reproduce the issue?
```shell
Node.js test script with selenium
const fs = require('fs')
const path = require('path')
const mkdirp = require('mkdirp')
const rimraf = require('rimraf')
const { Builder, until, By } = require('selenium-webdriver')
const Chrome = require('selenium-webdriver/chrome')
const Firefox = require('selenium-webdriver/firefox')
require('chromedriver')
require('geckodriver')
jest.setTimeout(1000*60*10)
async function getWebDriver(browser, lang, country) {
if (browser === 'chrome') {
const chromeOptions = new Chrome.Options()
.addArguments(
`--accept-lang=${lang}-${country}`,
'--browser-test',
'--bwsi',
`--default-country-code=${country}`,
'--disable-default-apps',
'--disable-extensions',
'--disable-gpu',
'--disable-logging',
'--disable-web-security',
'--dom-automation',
'--enable-automation',
'--force-headless-for-tests',
'--guest',
'--headless',
'--incognito',
`--lang=${lang}`,
'--no-sandbox',
'--window-size=1440,900',
)
.setUserPreferences({ ['intl.accept_languages']: lang, ['translate']: { enabled: true } })
const driver = new Builder()
.forBrowser('chrome')
.setChromeOptions(chromeOptions)
.usingServer('http://localhost:4444/wd/hub')
.build()
await sleep(3)
return driver
} else {
const firefoxOptions = new Firefox.Options()
.headless()
.setPreference('intl.accept_languages', lang)
const driver = new Builder()
.forBrowser('firefox')
.setFirefoxOptions(firefoxOptions)
.usingServer('http://localhost:4444/wd/hub')
.build()
await sleep(3)
return driver
}
}
async function sleep(seconds) {
return new Promise((resolve) => {
setTimeout(() => {
resolve()
}, 1000 * seconds)
})
}
describe.each`
browser | langISO639 | countryCodeISO3166 | expectedTitle
${'chrome'} | ${'de'} | ${'DE'} | ${'Bevor Sie zu Google weitergehen'}
${'chrome'} | ${'en'} | ${'GB'} | ${'Before you continue to Google'}
${'chrome'} | ${'es'} | ${'ES'} | ${'Antes de ir a Google'}
${'chrome'} | ${'fr'} | ${'FR'} | ${'Avant d\'accéder à Google'}
${'firefox'} | ${'de'} | ${'DE'} | ${'Bevor Sie zu Google weitergehen'}
${'firefox'} | ${'en'} | ${'GB'} | ${'Before you continue to Google'}
${'firefox'} | ${'es'} | ${'ES'} | ${'Antes de ir a Google'}
${'firefox'} | ${'fr'} | ${'FR'} | ${'Avant d\'accéder à Google'}
`('open Google on $browser browser with language $langISO639', ({ browser, langISO639, countryCodeISO3166, expectedTitle }) => {
let envLangBefore = process.env['LANG'] || undefined
beforeAll(async () => {
await rimraf.sync(path.join(__dirname, 'test-screenshots'))
await mkdirp.sync(path.join(__dirname, 'test-screenshots'))
process.env['LANG'] = langISO639
})
afterAll(() => {
process.env['LANG'] = envLangBefore
})
let driver
it('opens web driver instance', async () => {
driver = await getWebDriver(browser, langISO639, countryCodeISO3166)
})
it('reach goole web app', async () => {
await driver.get('https://www.google.com/')
await driver.navigate().refresh()
})
it('waits for translated text or take a screenshot and throw', async () => {
try {
await driver.wait(until.elementLocated(By.xpath(`//h1[text()="${expectedTitle}"]`)), 3000)
} catch (e) {
const image = await driver.takeScreenshot()
fs.writeFileSync(path.join(__dirname, `test-screenshots/${browser}_${langISO639}.png`), image, 'base64')
throw e
}
})
it('closes web driver instance', async () => {
await driver.quit()
})
})
```
Node.js test script with puppeteer
```
const fs = require('fs')
const path = require('path')
const mkdirp = require('mkdirp')
const puppeteer = require('puppeteer')
const rimraf = require('rimraf')
jest.setTimeout(1000*60*10)
async function getWebDriver(lang, country) {
const driver = await puppeteer.launch({
args: [
`--accept-lang=${lang}-${country}`,
'--browser-test',
'--bwsi',
`--default-country-code=${country}`,
'--disable-default-apps',
'--disable-extensions',
'--disable-gpu',
'--disable-logging',
'--disable-web-security',
'--dom-automation',
'--enable-automation',
'--force-headless-for-tests',
'--guest',
'--headless',
'--incognito',
`--lang=${lang}`,
'--no-sandbox',
'--window-size=1440,900',
]
})
await sleep(3)
return driver
}
async function sleep(seconds) {
return new Promise((resolve) => {
setTimeout(() => {
resolve()
}, 1000 * seconds)
})
}
describe.each`
langISO639 | countryCodeISO3166 | expectedTitle
${'de'} | ${'DE'} | ${'Bevor Sie zu Google weitergehen'}
${'en'} | ${'GB'} | ${'Before you continue to Google'}
${'es'} | ${'ES'} | ${'Antes de ir a Google'}
${'fr'} | ${'FR'} | ${'Avant d\'accéder à Google'}
`('open Puppeteer with language $langISO639', ({ langISO639, countryCodeISO3166, expectedTitle }) => {
let envLangBefore = process.env['LANG'] || undefined
beforeAll(async () => {
await rimraf.sync(path.join(__dirname, 'test-screenshots'))
await mkdirp.sync(path.join(__dirname, 'test-screenshots'))
process.env['LANG'] = langISO639
})
afterAll(() => {
process.env['LANG'] = envLangBefore
})
/** @type {puppeteer.Browser} */
let browser
/** @type {puppeteer.Page} */
let page
it('opens web driver instance', async () => {
browser = await getWebDriver(langISO639, countryCodeISO3166)
page = await browser.newPage()
})
it('reach goole web app', async () => {
await page.goto('https://www.google.com/')
await page.reload()
})
it('waits for translated text or take a screenshot and throw', async () => {
try {
await page.waitForXPath(`//h1[text()="${expectedTitle}"]`)
} catch (e) {
await page.screenshot({ path: path.join(__dirname, `test-screenshots/puppeteer_${langISO639}.png`) })
throw e
}
})
it('closes web driver instance', async () => {
await browser.close()
})
})
```
Node.js project package.json example
```
{
"name": "sample-en-issue-with-chromium",
"scripts": {
"test": "jest"
},
"devDependencies": {
"@types/jest": "^29.0.0",
"@types/selenium-webdriver": "^4.1.3",
"eslint": "^8.23.0",
"eslint-config-node": "^4.1.0",
"eslint-plugin-import": "^2.26.0",
"eslint-plugin-jest": "^27.0.1",
"jest": "^29.0.1",
"jest-runner-groups": "^2.2.0",
"webdriver-manager": "^12.1.8"
},
"dependencies": {
"chromedriver": "^104.0.0",
"geckodriver": "^3.0.2",
"mkdirp": "^1.0.4",
"puppeteer": "^17.1.1",
"rimraf": "^3.0.2",
"selenium-webdriver": "^4.4.0"
}
}
```
```
### Relevant log output
```shell
With Selenium
FAIL framework/automate-web/WebDriverManager.test.js (45.459 s)
open Google on chrome browser with language de
✓ opens web driver instance (3003 ms)
✓ reach goole web app (1587 ms)
✓ waits for translated text or take a screenshot and throw (527 ms)
✓ closes web driver instance (54 ms)
open Google on chrome browser with language en
✓ opens web driver instance (3002 ms)
✓ reach goole web app (1364 ms)
✕ waits for translated text or take a screenshot and throw (3506 ms)
✓ closes web driver instance (55 ms)
open Google on chrome browser with language es
✓ opens web driver instance (3002 ms)
✓ reach goole web app (1371 ms)
✓ waits for translated text or take a screenshot and throw (485 ms)
✓ closes web driver instance (58 ms)
open Google on chrome browser with language fr
✓ opens web driver instance (3001 ms)
✓ reach goole web app (1414 ms)
✓ waits for translated text or take a screenshot and throw (490 ms)
✓ closes web driver instance (55 ms)
open Google on firefox browser with language de
✓ opens web driver instance (3001 ms)
✓ reach goole web app (909 ms)
✓ waits for translated text or take a screenshot and throw (22 ms)
✓ closes web driver instance (404 ms)
open Google on firefox browser with language en
✓ opens web driver instance (3002 ms)
✓ reach goole web app (786 ms)
✕ waits for translated text or take a screenshot and throw (3202 ms)
✓ closes web driver instance (409 ms)
open Google on firefox browser with language es
✓ opens web driver instance (3001 ms)
✓ reach goole web app (2718 ms)
✓ waits for translated text or take a screenshot and throw (17 ms)
✓ closes web driver instance (408 ms)
open Google on firefox browser with language fr
✓ opens web driver instance (3002 ms)
✓ reach goole web app (788 ms)
✓ waits for translated text or take a screenshot and throw (19 ms)
✓ closes web driver instance (406 ms)
```
With Puppeteer
```
FAIL framework/automate-web/test-screenshots/Puppeteer-multiple-langs.test.js (63.579 s)
open Puppeteer with language de
✓ opens web driver instance (19901 ms)
✓ reach goole web app (852 ms)
✓ waits for translated text or take a screenshot and throw (8 ms)
✓ closes web driver instance (8 ms)
open Puppeteer with language en
✓ opens web driver instance (3319 ms)
✓ reach goole web app (679 ms)
✕ waits for translated text or take a screenshot and throw (30071 ms)
✓ closes web driver instance (7 ms)
open Puppeteer with language es
✓ opens web driver instance (3324 ms)
✓ reach goole web app (825 ms)
✓ waits for translated text or take a screenshot and throw (6 ms)
✓ closes web driver instance (11 ms)
open Puppeteer with language fr
✓ opens web driver instance (3334 ms)
✓ reach goole web app (705 ms)
✓ waits for translated text or take a screenshot and throw (7 ms)
✓ closes web driver instance (6 ms)
```
```
### Operating System
Mac OS Monterey 12.5.1
### Selenium version
"selenium-webdriver": "^4.4.0",
### What are the browser(s) and version(s) where you see this issue?
Firefox 104.0.1, Chrome version as of 2022-09-06
### What are the browser driver(s) and version(s) where you see this issue?
from NPM: {"chromedriver": "^104.0.0", "geckodriver": "^3.0.2"}
### Are you using Selenium Grid?
yes with npm module webdriver-manager / no
|
defect
|
lang en or accept lang en option not working with node js on mac os what happened hi ✌️ i try to create a web end to end automation framework for my company and i need to be able to run tests in multiple browsers and in multiple languages current targets are browsers chrome and firefox languages deutsh english french italian spanish more to come later i am doing a proof of concept project to see how it could be achieved with either selenium or puppeteer i am trying to implement a simple scenario open google home page verify in the popup that the title is in the right language examples browser expectedtitle chrome de de bevor sie zu google weitergehen chrome en gb before you continue to google chrome es es antes de ir a google chrome fr fr avant d accéder à google firefox de de bevor sie zu google weitergehen firefox en gb before you continue to google firefox es es antes de ir a google firefox fr fr avant d accéder à google my problem is that english test case is always failed and i ve tried all workarounds i ve found around the web github issues stack overflow ps i am using webdriver manager to run a local grid but it is also failed with chromedriver and geckodriver local installations from npm how can we reproduce the issue shell node js test script with selenium const fs require fs const path require path const mkdirp require mkdirp const rimraf require rimraf const builder until by require selenium webdriver const chrome require selenium webdriver chrome const firefox require selenium webdriver firefox require chromedriver require geckodriver jest settimeout async function getwebdriver browser lang country if browser chrome const chromeoptions new chrome options addarguments accept lang lang country browser test bwsi default country code country disable default apps disable extensions disable gpu disable logging disable web security dom automation enable automation force headless for tests guest headless incognito lang lang no sandbox window size setuserpreferences lang enabled true const driver new builder forbrowser chrome setchromeoptions chromeoptions usingserver build await sleep return driver else const firefoxoptions new firefox options headless setpreference intl accept languages lang const driver new builder forbrowser firefox setfirefoxoptions firefoxoptions usingserver build await sleep return driver async function sleep seconds return new promise resolve settimeout resolve seconds describe each browser expectedtitle chrome de de bevor sie zu google weitergehen chrome en gb before you continue to google chrome es es antes de ir a google chrome fr fr avant d accéder à google firefox de de bevor sie zu google weitergehen firefox en gb before you continue to google firefox es es antes de ir a google firefox fr fr avant d accéder à google open google on browser browser with language browser expectedtitle let envlangbefore process env undefined beforeall async await rimraf sync path join dirname test screenshots await mkdirp sync path join dirname test screenshots process env afterall process env envlangbefore let driver it opens web driver instance async driver await getwebdriver browser it reach goole web app async await driver get await driver navigate refresh it waits for translated text or take a screenshot and throw async try await driver wait until elementlocated by xpath catch e const image await driver takescreenshot fs writefilesync path join dirname test screenshots browser png image throw e it closes web driver instance async await driver quit node js test script with puppeteer const fs require fs const path require path const mkdirp require mkdirp const puppeteer require puppeteer const rimraf require rimraf jest settimeout async function getwebdriver lang country const driver await puppeteer launch args accept lang lang country browser test bwsi default country code country disable default apps disable extensions disable gpu disable logging disable web security dom automation enable automation force headless for tests guest headless incognito lang lang no sandbox window size await sleep return driver async function sleep seconds return new promise resolve settimeout resolve seconds describe each expectedtitle de de bevor sie zu google weitergehen en gb before you continue to google es es antes de ir a google fr fr avant d accéder à google open puppeteer with language expectedtitle let envlangbefore process env undefined beforeall async await rimraf sync path join dirname test screenshots await mkdirp sync path join dirname test screenshots process env afterall process env envlangbefore type puppeteer browser let browser type puppeteer page let page it opens web driver instance async browser await getwebdriver page await browser newpage it reach goole web app async await page goto await page reload it waits for translated text or take a screenshot and throw async try await page waitforxpath catch e await page screenshot path path join dirname test screenshots puppeteer png throw e it closes web driver instance async await browser close node js project package json example name sample en issue with chromium scripts test jest devdependencies types jest types selenium webdriver eslint eslint config node eslint plugin import eslint plugin jest jest jest runner groups webdriver manager dependencies chromedriver geckodriver mkdirp puppeteer rimraf selenium webdriver relevant log output shell with selenium fail framework automate web webdrivermanager test js s open google on chrome browser with language de ✓ opens web driver instance ms ✓ reach goole web app ms ✓ waits for translated text or take a screenshot and throw ms ✓ closes web driver instance ms open google on chrome browser with language en ✓ opens web driver instance ms ✓ reach goole web app ms ✕ waits for translated text or take a screenshot and throw ms ✓ closes web driver instance ms open google on chrome browser with language es ✓ opens web driver instance ms ✓ reach goole web app ms ✓ waits for translated text or take a screenshot and throw ms ✓ closes web driver instance ms open google on chrome browser with language fr ✓ opens web driver instance ms ✓ reach goole web app ms ✓ waits for translated text or take a screenshot and throw ms ✓ closes web driver instance ms open google on firefox browser with language de ✓ opens web driver instance ms ✓ reach goole web app ms ✓ waits for translated text or take a screenshot and throw ms ✓ closes web driver instance ms open google on firefox browser with language en ✓ opens web driver instance ms ✓ reach goole web app ms ✕ waits for translated text or take a screenshot and throw ms ✓ closes web driver instance ms open google on firefox browser with language es ✓ opens web driver instance ms ✓ reach goole web app ms ✓ waits for translated text or take a screenshot and throw ms ✓ closes web driver instance ms open google on firefox browser with language fr ✓ opens web driver instance ms ✓ reach goole web app ms ✓ waits for translated text or take a screenshot and throw ms ✓ closes web driver instance ms with puppeteer fail framework automate web test screenshots puppeteer multiple langs test js s open puppeteer with language de ✓ opens web driver instance ms ✓ reach goole web app ms ✓ waits for translated text or take a screenshot and throw ms ✓ closes web driver instance ms open puppeteer with language en ✓ opens web driver instance ms ✓ reach goole web app ms ✕ waits for translated text or take a screenshot and throw ms ✓ closes web driver instance ms open puppeteer with language es ✓ opens web driver instance ms ✓ reach goole web app ms ✓ waits for translated text or take a screenshot and throw ms ✓ closes web driver instance ms open puppeteer with language fr ✓ opens web driver instance ms ✓ reach goole web app ms ✓ waits for translated text or take a screenshot and throw ms ✓ closes web driver instance ms operating system mac os monterey selenium version selenium webdriver what are the browser s and version s where you see this issue firefox chrome version as of what are the browser driver s and version s where you see this issue from npm chromedriver geckodriver are you using selenium grid yes with npm module webdriver manager no
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.