Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
21,449
| 29,487,795,596
|
IssuesEvent
|
2023-06-02 11:07:57
|
openfoodfacts/openfoodfacts-server
|
https://api.github.com/repos/openfoodfacts/openfoodfacts-server
|
opened
|
Compute and display glycemic index
|
✔ task ingredients processing
|
While it is possible to enter the glycemic index in the nutrients' form, we could try to **compute and display the glycemic index**, based on the ingredients.
Either we can find an open data database, either we could start to collect the values from different scientific articles (eg. for [commonly consumed Thai fruits](http://ifrj.upm.edu.my/28%20(04)%202021/DONE%20-%2015%20-%20IFRJ20583.R1.pdf)).
Some resources [to be completed and reviewed]:
* https://academic.oup.com/ajcn/article/114/5/1625/6320814
* https://glycemicindex.com/
* https://foodstruct.com/glycemic-index-chart
Some important facts:
* glycemic index varies from different species: different banana or lentil species can have a glycemic index which can significantly vary (eg. [from 23 to 30 for bananas as stated in this article](http://ifrj.upm.edu.my/28%20(04)%202021/DONE%20-%2015%20-%20IFRJ20583.R1.pdf))
|
1.0
|
Compute and display glycemic index - While it is possible to enter the glycemic index in the nutrients' form, we could try to **compute and display the glycemic index**, based on the ingredients.
Either we can find an open data database, either we could start to collect the values from different scientific articles (eg. for [commonly consumed Thai fruits](http://ifrj.upm.edu.my/28%20(04)%202021/DONE%20-%2015%20-%20IFRJ20583.R1.pdf)).
Some resources [to be completed and reviewed]:
* https://academic.oup.com/ajcn/article/114/5/1625/6320814
* https://glycemicindex.com/
* https://foodstruct.com/glycemic-index-chart
Some important facts:
* glycemic index varies from different species: different banana or lentil species can have a glycemic index which can significantly vary (eg. [from 23 to 30 for bananas as stated in this article](http://ifrj.upm.edu.my/28%20(04)%202021/DONE%20-%2015%20-%20IFRJ20583.R1.pdf))
|
process
|
compute and display glycemic index while it is possible to enter the glycemic index in the nutrients form we could try to compute and display the glycemic index based on the ingredients either we can find an open data database either we could start to collect the values from different scientific articles eg for some resources some important facts glycemic index varies from different species different banana or lentil species can have a glycemic index which can significantly vary eg
| 1
|
21,689
| 30,185,622,360
|
IssuesEvent
|
2023-07-04 11:54:39
|
h4sh5/npm-auto-scanner
|
https://api.github.com/repos/h4sh5/npm-auto-scanner
|
opened
|
virtual-domain-proxy 2.0.0 has 2 guarddog issues
|
npm-install-script npm-silent-process-execution
|
```{"npm-install-script":[{"code":" \"postinstall\": \"node ./scripts/postinstall.js\"","location":"package/package.json:13","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" childProcess = (0, child_process_1.spawn)('node', [path_1.default.join(__dirname, \"./logging.js\")], {\n detached: true,\n stdio: 'ignore',\n });","location":"package/dist/server/index.js:58","message":"This package is silently executing another executable"}]}```
|
1.0
|
virtual-domain-proxy 2.0.0 has 2 guarddog issues - ```{"npm-install-script":[{"code":" \"postinstall\": \"node ./scripts/postinstall.js\"","location":"package/package.json:13","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" childProcess = (0, child_process_1.spawn)('node', [path_1.default.join(__dirname, \"./logging.js\")], {\n detached: true,\n stdio: 'ignore',\n });","location":"package/dist/server/index.js:58","message":"This package is silently executing another executable"}]}```
|
process
|
virtual domain proxy has guarddog issues npm install script npm silent process execution n detached true n stdio ignore n location package dist server index js message this package is silently executing another executable
| 1
|
651,699
| 21,485,436,885
|
IssuesEvent
|
2022-04-26 22:36:07
|
ProjectG-Plugins/CrossplatForms
|
https://api.github.com/repos/ProjectG-Plugins/CrossplatForms
|
opened
|
Convert configs from GeyserHub
|
enhancement priority: high
|
### What feature do you want to see added?
Automatically convert configs from GeyserHub or a command to convert the file.
If anyone is interested in attempting this you'll need to learn or know how to use Configurate's Transformations (see [1](https://github.com/SpongePowered/Configurate/wiki/Transformations-and-Visitors) and [2](https://github.com/SpongePowered/Configurate/blob/master/examples/src/main/java/org/spongepowered/configurate/examples/Transformations.java)). And everything that GeyserHub and CrossplatForms has in common will have to be translated. Start with [MenuConfig](https://github.com/ProjectG-Plugins/CrossplatForms/blob/main/core/src/main/java/dev/projectg/crossplatforms/interfacing/java/MenuConfig.java) and [FormConfig](https://github.com/ProjectG-Plugins/CrossplatForms/blob/main/core/src/main/java/dev/projectg/crossplatforms/interfacing/bedrock/FormConfig.java) and traverse all the fields.
### Are there any alternatives?
Convert it by hand.
|
1.0
|
Convert configs from GeyserHub - ### What feature do you want to see added?
Automatically convert configs from GeyserHub or a command to convert the file.
If anyone is interested in attempting this you'll need to learn or know how to use Configurate's Transformations (see [1](https://github.com/SpongePowered/Configurate/wiki/Transformations-and-Visitors) and [2](https://github.com/SpongePowered/Configurate/blob/master/examples/src/main/java/org/spongepowered/configurate/examples/Transformations.java)). And everything that GeyserHub and CrossplatForms has in common will have to be translated. Start with [MenuConfig](https://github.com/ProjectG-Plugins/CrossplatForms/blob/main/core/src/main/java/dev/projectg/crossplatforms/interfacing/java/MenuConfig.java) and [FormConfig](https://github.com/ProjectG-Plugins/CrossplatForms/blob/main/core/src/main/java/dev/projectg/crossplatforms/interfacing/bedrock/FormConfig.java) and traverse all the fields.
### Are there any alternatives?
Convert it by hand.
|
non_process
|
convert configs from geyserhub what feature do you want to see added automatically convert configs from geyserhub or a command to convert the file if anyone is interested in attempting this you ll need to learn or know how to use configurate s transformations see and and everything that geyserhub and crossplatforms has in common will have to be translated start with and and traverse all the fields are there any alternatives convert it by hand
| 0
|
15,866
| 6,048,680,330
|
IssuesEvent
|
2017-06-12 17:01:01
|
meteor/meteor
|
https://api.github.com/repos/meteor/meteor
|
closed
|
Arbitrary architecture argument for packages
|
feature Project:Isobuild
|
api.use accepts an architecture as its second argument. Is it possible to let it accept arbitrary global constants that are defined somewhere on the server?
|
1.0
|
Arbitrary architecture argument for packages - api.use accepts an architecture as its second argument. Is it possible to let it accept arbitrary global constants that are defined somewhere on the server?
|
non_process
|
arbitrary architecture argument for packages api use accepts an architecture as its second argument is it possible to let it accept arbitrary global constants that are defined somewhere on the server
| 0
|
472,028
| 13,614,950,154
|
IssuesEvent
|
2020-09-23 13:52:44
|
googleapis/elixir-google-api
|
https://api.github.com/repos/googleapis/elixir-google-api
|
closed
|
Synthesis failed for Webmaster
|
autosynth failure priority: p1 type: bug
|
Hello! Autosynth couldn't regenerate Webmaster. :broken_heart:
Here's the output from running `synth.py`:
```
2020-09-22 06:51:05,019 autosynth [INFO] > logs will be written to: /tmpfs/src/logs/elixir-google-api
2020-09-22 06:51:05,489 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore
2020-09-22 06:51:05,492 autosynth [DEBUG] > Running: git config user.name yoshi-automation
2020-09-22 06:51:05,494 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com
2020-09-22 06:51:05,497 autosynth [DEBUG] > Running: git config push.default simple
2020-09-22 06:51:05,500 autosynth [DEBUG] > Running: git branch -f autosynth-webmaster
2020-09-22 06:51:05,503 autosynth [DEBUG] > Running: git checkout autosynth-webmaster
Switched to branch 'autosynth-webmaster'
2020-09-22 06:51:05,720 autosynth [INFO] > Running synthtool
2020-09-22 06:51:05,720 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/webmaster/synth.metadata', 'synth.py', '--']
2020-09-22 06:51:05,720 autosynth [DEBUG] > log_file_path: /tmpfs/src/logs/elixir-google-api/Webmaster/sponge_log.log
2020-09-22 06:51:05,723 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata clients/webmaster/synth.metadata synth.py -- Webmaster
2020-09-22 06:51:05,958 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/elixir-google-api/synth.py.
On branch autosynth-webmaster
nothing to commit, working tree clean
2020-09-22 06:51:08,116 synthtool [DEBUG] > Running: docker run --rm -v/tmpfs/tmp/tmp4bqxef5c/repo:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh Webmaster
DEBUG:synthtool:Running: docker run --rm -v/tmpfs/tmp/tmp4bqxef5c/repo:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh Webmaster
/workspace /workspace
[33mThe mix.lock file was generated with a newer version of Hex. Update your client by running `mix local.hex` to avoid losing data.[0m
Resolving Hex dependencies...
Dependency resolution completed:
Unchanged:
certifi 2.5.1
google_api_discovery 0.7.0
google_gax 0.3.2
hackney 1.15.2
idna 6.0.0
jason 1.2.1
metrics 1.0.1
mime 1.3.1
mimerl 1.2.0
oauth2 0.9.4
parse_trans 3.3.0
poison 3.1.0
ssl_verify_fun 1.1.5
temp 0.4.7
tesla 1.3.3
unicode_util_compat 0.4.1
* Getting google_api_discovery (Hex package)
* Getting tesla (Hex package)
* Getting oauth2 (Hex package)
* Getting temp (Hex package)
* Getting jason (Hex package)
* Getting poison (Hex package)
* Getting hackney (Hex package)
* Getting certifi (Hex package)
* Getting idna (Hex package)
* Getting metrics (Hex package)
* Getting mimerl (Hex package)
* Getting ssl_verify_fun (Hex package)
* Getting unicode_util_compat (Hex package)
* Getting parse_trans (Hex package)
* Getting mime (Hex package)
* Getting google_gax (Hex package)
[33mThe mix.lock file was generated with a newer version of Hex. Update your client by running `mix local.hex` to avoid losing data.[0m
==> temp
Compiling 3 files (.ex)
Generated temp app
===> Compiling parse_trans
===> Compiling mimerl
===> Compiling metrics
===> Compiling unicode_util_compat
===> Compiling idna
==> jason
Compiling 8 files (.ex)
Generated jason app
warning: String.strip/1 is deprecated. Use String.trim/1 instead
/workspace/deps/poison/mix.exs:4
==> poison
Compiling 4 files (.ex)
warning: Integer.to_char_list/2 is deprecated. Use Integer.to_charlist/2 instead
lib/poison/encoder.ex:173
Generated poison app
==> ssl_verify_fun
Compiling 7 files (.erl)
Generated ssl_verify_fun app
===> Compiling certifi
===> Compiling hackney
==> oauth2
Compiling 13 files (.ex)
Generated oauth2 app
==> mime
Compiling 2 files (.ex)
Generated mime app
==> tesla
Compiling 26 files (.ex)
Generated tesla app
==> google_gax
Compiling 5 files (.ex)
Generated google_gax app
==> google_api_discovery
Compiling 21 files (.ex)
Generated google_api_discovery app
==> google_apis
Compiling 28 files (.ex)
Generated google_apis app
13:51:37.855 [info] FETCHING: https://www.googleapis.com/discovery/v1/apis/webmasters/v3/rest
{:error, "Error received status: 404 from discovery endpoint"}
** (File.Error) could not read file "/workspace/specifications/gdd/Webmaster-v3.json": no such file or directory
(elixir) lib/file.ex:353: File.read!/1
lib/google_apis/generator/elixir_generator/token.ex:79: GoogleApis.Generator.ElixirGenerator.Token.build/1
lib/google_apis/generator/elixir_generator.ex:40: GoogleApis.Generator.ElixirGenerator.generate_client/1
lib/mix/tasks/google_apis.generate.ex:33: anonymous fn/1 in Mix.Tasks.GoogleApis.Generate.builder/1
(elixir) lib/enum.ex:783: Enum."-each/2-lists^foreach/1-0-"/2
(elixir) lib/enum.ex:783: Enum.each/2
lib/mix/tasks/google_apis.generate.ex:31: Mix.Tasks.GoogleApis.Generate.builder/1
(mix) lib/mix/task.ex:331: Mix.Task.run_task/3
fixing file permissions
2020-09-22 06:51:38,856 synthtool [ERROR] > Failed executing docker run --rm -v/tmpfs/tmp/tmp4bqxef5c/repo:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh Webmaster:
None
ERROR:synthtool:Failed executing docker run --rm -v/tmpfs/tmp/tmp4bqxef5c/repo:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh Webmaster:
None
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/home/kbuilder/.cache/synthtool/elixir-google-api/synth.py", line 59, in <module>
shell.run(command, cwd=repository, hide_output=False)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['docker', 'run', '--rm', '-v/tmpfs/tmp/tmp4bqxef5c/repo:/workspace', '-v/var/run/docker.sock:/var/run/docker.sock', '-e', 'USER_GROUP=1000:1000', '-w', '/workspace', 'gcr.io/cloud-devrel-public-resources/elixir19', 'scripts/generate_client.sh', 'Webmaster']' returned non-zero exit status 1.
2020-09-22 06:51:39,270 autosynth [ERROR] > Synthesis failed
2020-09-22 06:51:39,270 autosynth [DEBUG] > Running: git clean -fdx
Removing __pycache__/
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 692, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 541, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 632, in _inner_main
).synthesize(synth_log_path / "sponge_log.log")
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/webmaster/synth.metadata', 'synth.py', '--', 'Webmaster']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/e8b30f3c-0681-4835-ab4e-cf39fd073af3/targets/github%2Fsynthtool;config=default/tests;query=elixir-google-api;failed=false).
|
1.0
|
Synthesis failed for Webmaster - Hello! Autosynth couldn't regenerate Webmaster. :broken_heart:
Here's the output from running `synth.py`:
```
2020-09-22 06:51:05,019 autosynth [INFO] > logs will be written to: /tmpfs/src/logs/elixir-google-api
2020-09-22 06:51:05,489 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore
2020-09-22 06:51:05,492 autosynth [DEBUG] > Running: git config user.name yoshi-automation
2020-09-22 06:51:05,494 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com
2020-09-22 06:51:05,497 autosynth [DEBUG] > Running: git config push.default simple
2020-09-22 06:51:05,500 autosynth [DEBUG] > Running: git branch -f autosynth-webmaster
2020-09-22 06:51:05,503 autosynth [DEBUG] > Running: git checkout autosynth-webmaster
Switched to branch 'autosynth-webmaster'
2020-09-22 06:51:05,720 autosynth [INFO] > Running synthtool
2020-09-22 06:51:05,720 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/webmaster/synth.metadata', 'synth.py', '--']
2020-09-22 06:51:05,720 autosynth [DEBUG] > log_file_path: /tmpfs/src/logs/elixir-google-api/Webmaster/sponge_log.log
2020-09-22 06:51:05,723 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata clients/webmaster/synth.metadata synth.py -- Webmaster
2020-09-22 06:51:05,958 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/elixir-google-api/synth.py.
On branch autosynth-webmaster
nothing to commit, working tree clean
2020-09-22 06:51:08,116 synthtool [DEBUG] > Running: docker run --rm -v/tmpfs/tmp/tmp4bqxef5c/repo:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh Webmaster
DEBUG:synthtool:Running: docker run --rm -v/tmpfs/tmp/tmp4bqxef5c/repo:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh Webmaster
/workspace /workspace
[33mThe mix.lock file was generated with a newer version of Hex. Update your client by running `mix local.hex` to avoid losing data.[0m
Resolving Hex dependencies...
Dependency resolution completed:
Unchanged:
certifi 2.5.1
google_api_discovery 0.7.0
google_gax 0.3.2
hackney 1.15.2
idna 6.0.0
jason 1.2.1
metrics 1.0.1
mime 1.3.1
mimerl 1.2.0
oauth2 0.9.4
parse_trans 3.3.0
poison 3.1.0
ssl_verify_fun 1.1.5
temp 0.4.7
tesla 1.3.3
unicode_util_compat 0.4.1
* Getting google_api_discovery (Hex package)
* Getting tesla (Hex package)
* Getting oauth2 (Hex package)
* Getting temp (Hex package)
* Getting jason (Hex package)
* Getting poison (Hex package)
* Getting hackney (Hex package)
* Getting certifi (Hex package)
* Getting idna (Hex package)
* Getting metrics (Hex package)
* Getting mimerl (Hex package)
* Getting ssl_verify_fun (Hex package)
* Getting unicode_util_compat (Hex package)
* Getting parse_trans (Hex package)
* Getting mime (Hex package)
* Getting google_gax (Hex package)
[33mThe mix.lock file was generated with a newer version of Hex. Update your client by running `mix local.hex` to avoid losing data.[0m
==> temp
Compiling 3 files (.ex)
Generated temp app
===> Compiling parse_trans
===> Compiling mimerl
===> Compiling metrics
===> Compiling unicode_util_compat
===> Compiling idna
==> jason
Compiling 8 files (.ex)
Generated jason app
warning: String.strip/1 is deprecated. Use String.trim/1 instead
/workspace/deps/poison/mix.exs:4
==> poison
Compiling 4 files (.ex)
warning: Integer.to_char_list/2 is deprecated. Use Integer.to_charlist/2 instead
lib/poison/encoder.ex:173
Generated poison app
==> ssl_verify_fun
Compiling 7 files (.erl)
Generated ssl_verify_fun app
===> Compiling certifi
===> Compiling hackney
==> oauth2
Compiling 13 files (.ex)
Generated oauth2 app
==> mime
Compiling 2 files (.ex)
Generated mime app
==> tesla
Compiling 26 files (.ex)
Generated tesla app
==> google_gax
Compiling 5 files (.ex)
Generated google_gax app
==> google_api_discovery
Compiling 21 files (.ex)
Generated google_api_discovery app
==> google_apis
Compiling 28 files (.ex)
Generated google_apis app
13:51:37.855 [info] FETCHING: https://www.googleapis.com/discovery/v1/apis/webmasters/v3/rest
{:error, "Error received status: 404 from discovery endpoint"}
** (File.Error) could not read file "/workspace/specifications/gdd/Webmaster-v3.json": no such file or directory
(elixir) lib/file.ex:353: File.read!/1
lib/google_apis/generator/elixir_generator/token.ex:79: GoogleApis.Generator.ElixirGenerator.Token.build/1
lib/google_apis/generator/elixir_generator.ex:40: GoogleApis.Generator.ElixirGenerator.generate_client/1
lib/mix/tasks/google_apis.generate.ex:33: anonymous fn/1 in Mix.Tasks.GoogleApis.Generate.builder/1
(elixir) lib/enum.ex:783: Enum."-each/2-lists^foreach/1-0-"/2
(elixir) lib/enum.ex:783: Enum.each/2
lib/mix/tasks/google_apis.generate.ex:31: Mix.Tasks.GoogleApis.Generate.builder/1
(mix) lib/mix/task.ex:331: Mix.Task.run_task/3
fixing file permissions
2020-09-22 06:51:38,856 synthtool [ERROR] > Failed executing docker run --rm -v/tmpfs/tmp/tmp4bqxef5c/repo:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh Webmaster:
None
ERROR:synthtool:Failed executing docker run --rm -v/tmpfs/tmp/tmp4bqxef5c/repo:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh Webmaster:
None
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/home/kbuilder/.cache/synthtool/elixir-google-api/synth.py", line 59, in <module>
shell.run(command, cwd=repository, hide_output=False)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['docker', 'run', '--rm', '-v/tmpfs/tmp/tmp4bqxef5c/repo:/workspace', '-v/var/run/docker.sock:/var/run/docker.sock', '-e', 'USER_GROUP=1000:1000', '-w', '/workspace', 'gcr.io/cloud-devrel-public-resources/elixir19', 'scripts/generate_client.sh', 'Webmaster']' returned non-zero exit status 1.
2020-09-22 06:51:39,270 autosynth [ERROR] > Synthesis failed
2020-09-22 06:51:39,270 autosynth [DEBUG] > Running: git clean -fdx
Removing __pycache__/
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 692, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 541, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 632, in _inner_main
).synthesize(synth_log_path / "sponge_log.log")
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/webmaster/synth.metadata', 'synth.py', '--', 'Webmaster']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/e8b30f3c-0681-4835-ab4e-cf39fd073af3/targets/github%2Fsynthtool;config=default/tests;query=elixir-google-api;failed=false).
|
non_process
|
synthesis failed for webmaster hello autosynth couldn t regenerate webmaster broken heart here s the output from running synth py autosynth logs will be written to tmpfs src logs elixir google api autosynth running git config global core excludesfile home kbuilder autosynth gitignore autosynth running git config user name yoshi automation autosynth running git config user email yoshi automation google com autosynth running git config push default simple autosynth running git branch f autosynth webmaster autosynth running git checkout autosynth webmaster switched to branch autosynth webmaster autosynth running synthtool autosynth autosynth log file path tmpfs src logs elixir google api webmaster sponge log log autosynth running tmpfs src github synthtool env bin m synthtool metadata clients webmaster synth metadata synth py webmaster synthtool executing home kbuilder cache synthtool elixir google api synth py on branch autosynth webmaster nothing to commit working tree clean synthtool running docker run rm v tmpfs tmp repo workspace v var run docker sock var run docker sock e user group w workspace gcr io cloud devrel public resources scripts generate client sh webmaster debug synthtool running docker run rm v tmpfs tmp repo workspace v var run docker sock var run docker sock e user group w workspace gcr io cloud devrel public resources scripts generate client sh webmaster workspace workspace mix lock file was generated with a newer version of hex update your client by running mix local hex to avoid losing data resolving hex dependencies dependency resolution completed unchanged certifi google api discovery google gax hackney idna jason metrics mime mimerl parse trans poison ssl verify fun temp tesla unicode util compat getting google api discovery hex package getting tesla hex package getting hex package getting temp hex package getting jason hex package getting poison hex package getting hackney hex package getting certifi hex package getting idna hex package getting metrics hex package getting mimerl hex package getting ssl verify fun hex package getting unicode util compat hex package getting parse trans hex package getting mime hex package getting google gax hex package mix lock file was generated with a newer version of hex update your client by running mix local hex to avoid losing data temp compiling files ex generated temp app compiling parse trans compiling mimerl compiling metrics compiling unicode util compat compiling idna jason compiling files ex generated jason app warning string strip is deprecated use string trim instead workspace deps poison mix exs poison compiling files ex warning integer to char list is deprecated use integer to charlist instead lib poison encoder ex generated poison app ssl verify fun compiling files erl generated ssl verify fun app compiling certifi compiling hackney compiling files ex generated app mime compiling files ex generated mime app tesla compiling files ex generated tesla app google gax compiling files ex generated google gax app google api discovery compiling files ex generated google api discovery app google apis compiling files ex generated google apis app fetching error error received status from discovery endpoint file error could not read file workspace specifications gdd webmaster json no such file or directory elixir lib file ex file read lib google apis generator elixir generator token ex googleapis generator elixirgenerator token build lib google apis generator elixir generator ex googleapis generator elixirgenerator generate client lib mix tasks google apis generate ex anonymous fn in mix tasks googleapis generate builder elixir lib enum ex enum each lists foreach elixir lib enum ex enum each lib mix tasks google apis generate ex mix tasks googleapis generate builder mix lib mix task ex mix task run task fixing file permissions synthtool failed executing docker run rm v tmpfs tmp repo workspace v var run docker sock var run docker sock e user group w workspace gcr io cloud devrel public resources scripts generate client sh webmaster none error synthtool failed executing docker run rm v tmpfs tmp repo workspace v var run docker sock var run docker sock e user group w workspace gcr io cloud devrel public resources scripts generate client sh webmaster none traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file home kbuilder cache synthtool elixir google api synth py line in shell run command cwd repository hide output false file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status autosynth synthesis failed autosynth running git clean fdx removing pycache traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize synth log path sponge log log file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
| 0
|
2,946
| 2,534,568,476
|
IssuesEvent
|
2015-01-25 03:23:51
|
xela144/CANaconda
|
https://api.github.com/repos/xela144/CANaconda
|
closed
|
Move the field unit text to outside the textbox
|
enhancement priority[low]
|
I like how the field range is displayed in grey in the textbox, and it makes that it's not visible once you enter a value, but the units should always be visible. So add a label to the right of the textfield that always displays the units.
|
1.0
|
Move the field unit text to outside the textbox - I like how the field range is displayed in grey in the textbox, and it makes that it's not visible once you enter a value, but the units should always be visible. So add a label to the right of the textfield that always displays the units.
|
non_process
|
move the field unit text to outside the textbox i like how the field range is displayed in grey in the textbox and it makes that it s not visible once you enter a value but the units should always be visible so add a label to the right of the textfield that always displays the units
| 0
|
15,202
| 19,025,834,161
|
IssuesEvent
|
2021-11-24 03:17:38
|
NationalSecurityAgency/ghidra
|
https://api.github.com/repos/NationalSecurityAgency/ghidra
|
closed
|
MIPS64 reverse instruction is found to be wrong, how to fix the disassembly?
|
Feature: Processor/MIPS
|
ida:
```
text:0000000000019B08 A0 FF BD 67 daddiu $sp, -0x60 # Doubleword Add Immediate Unsigned
.text:0000000000019B0C 48 00 BC FF sd $gp, 0x60+var_18($sp) # Store Doubleword
.text:0000000000019B10 09 00 1C 3C lui $gp, 9 # Load Upper Immediate
.text:0000000000019B14 2D E0 99 03 daddu $gp, $t9 # Doubleword Add Unsigned
.text:0000000000019B18 48 85 9C 67 daddiu $gp, (strncpy_ptr - 0xA2050) # Doubleword Add Immediate Unsigned
.text:0000000000019B1C 10 85 99 DF dla $t9, strdup # Load 64-bit address
.text:0000000000019B20 50 00 BE FF sd $fp, 0x60+var_10($sp) # Store Doubleword
.text:0000000000019B24 38 00 B2 FF sd $s2, 0x60+var_28($sp) # Store Doubleword
```
ghidra:
```
00119b08 a0 ff bd 67 daddiu sp,sp,-0x60
00119b0c 48 00 bc ff sd gp,local_18(sp)
00119b10 09 00 1c 3c lui gp,0x9
00119b14 2d e0 99 03 daddu gp,gp,t9
00119b18 48 85 9c 67 daddiu gp,gp,-0x7ab8
00119b1c 10 85 99 df ld t9,-0x7af0(gp)
00119b20 50 00 be ff sd s8,local_10(sp)
00119b24 38 00 b2 ff sd s2,local_28(sp)
```
0x000019B1C this offset ghidra is wrong ,ida is right.
ghidra cannot find function "strdup" ?
how to fix the disassembly?
|
1.0
|
MIPS64 reverse instruction is found to be wrong, how to fix the disassembly? -
ida:
```
text:0000000000019B08 A0 FF BD 67 daddiu $sp, -0x60 # Doubleword Add Immediate Unsigned
.text:0000000000019B0C 48 00 BC FF sd $gp, 0x60+var_18($sp) # Store Doubleword
.text:0000000000019B10 09 00 1C 3C lui $gp, 9 # Load Upper Immediate
.text:0000000000019B14 2D E0 99 03 daddu $gp, $t9 # Doubleword Add Unsigned
.text:0000000000019B18 48 85 9C 67 daddiu $gp, (strncpy_ptr - 0xA2050) # Doubleword Add Immediate Unsigned
.text:0000000000019B1C 10 85 99 DF dla $t9, strdup # Load 64-bit address
.text:0000000000019B20 50 00 BE FF sd $fp, 0x60+var_10($sp) # Store Doubleword
.text:0000000000019B24 38 00 B2 FF sd $s2, 0x60+var_28($sp) # Store Doubleword
```
ghidra:
```
00119b08 a0 ff bd 67 daddiu sp,sp,-0x60
00119b0c 48 00 bc ff sd gp,local_18(sp)
00119b10 09 00 1c 3c lui gp,0x9
00119b14 2d e0 99 03 daddu gp,gp,t9
00119b18 48 85 9c 67 daddiu gp,gp,-0x7ab8
00119b1c 10 85 99 df ld t9,-0x7af0(gp)
00119b20 50 00 be ff sd s8,local_10(sp)
00119b24 38 00 b2 ff sd s2,local_28(sp)
```
0x000019B1C this offset ghidra is wrong ,ida is right.
ghidra cannot find function "strdup" ?
how to fix the disassembly?
|
process
|
reverse instruction is found to be wrong how to fix the disassembly ida text ff bd daddiu sp doubleword add immediate unsigned text bc ff sd gp var sp store doubleword text lui gp load upper immediate text daddu gp doubleword add unsigned text daddiu gp strncpy ptr doubleword add immediate unsigned text df dla strdup load bit address text be ff sd fp var sp store doubleword text ff sd var sp store doubleword ghidra ff bd daddiu sp sp bc ff sd gp local sp lui gp daddu gp gp daddiu gp gp df ld gp be ff sd local sp ff sd local sp this offset ghidra is wrong ida is right ghidra cannot find function strdup how to fix the disassembly
| 1
|
165,547
| 6,278,021,356
|
IssuesEvent
|
2017-07-18 13:35:22
|
CS2103JUN2017-T3/main
|
https://api.github.com/repos/CS2103JUN2017-T3/main
|
closed
|
As a user I can view my task history
|
priority.medium type.story
|
So that I can look back at the tasks I have completed so far
|
1.0
|
As a user I can view my task history - So that I can look back at the tasks I have completed so far
|
non_process
|
as a user i can view my task history so that i can look back at the tasks i have completed so far
| 0
|
12,594
| 14,992,390,485
|
IssuesEvent
|
2021-01-29 09:48:05
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
[Linux/ARM] Access permission issue for System.Diagnostics.Process.GetProcessesByName()/GetProcesses()
|
area-System.Diagnostics.Process bug in pr up-for-grabs
|
### Description
I ran the test code in an environment with smack applied(Tizen, 32-bit).
I called two method(```System.Diagnostics.Process.GetProcessesByName()```, ```System.Diagnostics.Process.GetProcesses()```).
* Sample code:
```
static void Main(string[] args)
{
string name = System.Diagnostics.Process.GetCurrentProcess().ProcessName;
Console.WriteLine(System.Diagnostics.Process.GetProcessesByName(name));
}
```
It seems to be a problem that occurs when the test app accesses the processes under ```/proc/```.
The following exception occurred.
* Exception:
```
Unhandled exception. System.UnauthorizedAccessException: Access to the path '/proc/1/stat' is denied.
---> System.IO.IOException: Permission denied
--- End of inner exception stack trace ---
at Interop.ThrowExceptionForIoErrno(ErrorInfo errorInfo, String path, Boolean isDirectory, Func`2 errorRewriter)
at Microsoft.Win32.SafeHandles.SafeFileHandle.Open(String path, OpenFlags flags, Int32 mode)
at System.IO.FileStream.OpenHandle(FileMode mode, FileShare share, FileOptions options)
at System.IO.FileStream..ctor(String path, FileMode mode, FileAccess access, FileShare share, Int32 bufferSize, FileOptions options)
at System.IO.FileStream..ctor(String path, FileMode mode, FileAccess access, FileShare share, Int32 bufferSize, Boolean useAsync)
at Interop.procfs.TryReadFile(String filePath, ReusableTextReader reusableReader, String& fileContents)
at Interop.procfs.TryParseStatFile(String statFilePath, ParsedStat& result, ReusableTextReader reusableReader)
at System.Diagnostics.Process.GetProcessesByName(String processName, String machineName)
at System.Diagnostics.Process.GetProcessesByName(String processName)
at ProcessTest.Program.Main(String[] args)
```
* Analysis:
```at Interop.procfs.TryReadFile(String filePath, ReusableTextReader reusableReader, String& fileContents)```
When I checked the part where the exception occurred, only ```IOException``` is handling the exception.
However, ```UnauthorizedAccessException``` is occurring when calling the ```FileStrem``` method.
Also, the document of the associated API is not mentioned.
So I think not only ```IOException``` but also ```UnauthorizedAccessException``` should be added.
|
1.0
|
[Linux/ARM] Access permission issue for System.Diagnostics.Process.GetProcessesByName()/GetProcesses() -
### Description
I ran the test code in an environment with smack applied(Tizen, 32-bit).
I called two method(```System.Diagnostics.Process.GetProcessesByName()```, ```System.Diagnostics.Process.GetProcesses()```).
* Sample code:
```
static void Main(string[] args)
{
string name = System.Diagnostics.Process.GetCurrentProcess().ProcessName;
Console.WriteLine(System.Diagnostics.Process.GetProcessesByName(name));
}
```
It seems to be a problem that occurs when the test app accesses the processes under ```/proc/```.
The following exception occurred.
* Exception:
```
Unhandled exception. System.UnauthorizedAccessException: Access to the path '/proc/1/stat' is denied.
---> System.IO.IOException: Permission denied
--- End of inner exception stack trace ---
at Interop.ThrowExceptionForIoErrno(ErrorInfo errorInfo, String path, Boolean isDirectory, Func`2 errorRewriter)
at Microsoft.Win32.SafeHandles.SafeFileHandle.Open(String path, OpenFlags flags, Int32 mode)
at System.IO.FileStream.OpenHandle(FileMode mode, FileShare share, FileOptions options)
at System.IO.FileStream..ctor(String path, FileMode mode, FileAccess access, FileShare share, Int32 bufferSize, FileOptions options)
at System.IO.FileStream..ctor(String path, FileMode mode, FileAccess access, FileShare share, Int32 bufferSize, Boolean useAsync)
at Interop.procfs.TryReadFile(String filePath, ReusableTextReader reusableReader, String& fileContents)
at Interop.procfs.TryParseStatFile(String statFilePath, ParsedStat& result, ReusableTextReader reusableReader)
at System.Diagnostics.Process.GetProcessesByName(String processName, String machineName)
at System.Diagnostics.Process.GetProcessesByName(String processName)
at ProcessTest.Program.Main(String[] args)
```
* Analysis:
```at Interop.procfs.TryReadFile(String filePath, ReusableTextReader reusableReader, String& fileContents)```
When I checked the part where the exception occurred, only ```IOException``` is handling the exception.
However, ```UnauthorizedAccessException``` is occurring when calling the ```FileStrem``` method.
Also, the document of the associated API is not mentioned.
So I think not only ```IOException``` but also ```UnauthorizedAccessException``` should be added.
|
process
|
access permission issue for system diagnostics process getprocessesbyname getprocesses description i ran the test code in an environment with smack applied tizen bit i called two method system diagnostics process getprocessesbyname system diagnostics process getprocesses sample code static void main string args string name system diagnostics process getcurrentprocess processname console writeline system diagnostics process getprocessesbyname name it seems to be a problem that occurs when the test app accesses the processes under proc the following exception occurred exception unhandled exception system unauthorizedaccessexception access to the path proc stat is denied system io ioexception permission denied end of inner exception stack trace at interop throwexceptionforioerrno errorinfo errorinfo string path boolean isdirectory func errorrewriter at microsoft safehandles safefilehandle open string path openflags flags mode at system io filestream openhandle filemode mode fileshare share fileoptions options at system io filestream ctor string path filemode mode fileaccess access fileshare share buffersize fileoptions options at system io filestream ctor string path filemode mode fileaccess access fileshare share buffersize boolean useasync at interop procfs tryreadfile string filepath reusabletextreader reusablereader string filecontents at interop procfs tryparsestatfile string statfilepath parsedstat result reusabletextreader reusablereader at system diagnostics process getprocessesbyname string processname string machinename at system diagnostics process getprocessesbyname string processname at processtest program main string args analysis at interop procfs tryreadfile string filepath reusabletextreader reusablereader string filecontents when i checked the part where the exception occurred only ioexception is handling the exception however unauthorizedaccessexception is occurring when calling the filestrem method also the document of the associated api is not mentioned so i think not only ioexception but also unauthorizedaccessexception should be added
| 1
|
22,143
| 30,684,599,790
|
IssuesEvent
|
2023-07-26 11:28:38
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Extension Based Hybrid Worker on Linux limits CPU to 5% which is nowhere mentioned - Agent Based Hybrid Worker didn't have this limit
|
automation/svc triaged assigned-to-author doc-enhancement process-automation/subsvc Pri2
|
Hi colleagues,
please at least mention the default limit to 5 % CPU, as the former Agent-Based Hybrid Worker didn't have this limit under Linux.
When upgrading to the new worker, existing CPU intensive runbooks therefore run extrem slow and also might get errors due to the extremely boxed resources.
see entry "CPUQuota=5%" in files:
- /var/lib/waagent/Microsoft.Azure.Automation.HybridWorker.HybridWorkerForLinux-0.1.0.11/HybridWorkerAgent/DaemonScripts/service_scripts/hwd.systemd
- /var/lib/waagent/Microsoft.Azure.Automation.HybridWorker.HybridWorkerForLinux-0.1.0.11/HybridWorkerAgent/DaemonScripts/service_temp/hwd.systemd
In my environment, I've actually updated the value from 5% to 95% - now, that I'm aware of it. :-)
Thanks & Best Regards,
Johann
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 2854aee3-7933-ee94-bb03-3a6a4c540088
* Version Independent ID: c902ddb9-46bf-c53c-46c4-e1c3d2cec7fd
* Content: [Deploy an extension-based Windows or Linux User Hybrid Runbook Worker in Azure Automation (Preview)](https://docs.microsoft.com/en-us/azure/automation/extension-based-hybrid-runbook-worker-install?tabs=windows)
* Content Source: [articles/automation/extension-based-hybrid-runbook-worker-install.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/extension-based-hybrid-runbook-worker-install.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SGSneha
* Microsoft Alias: **v-ssudhir**
|
1.0
|
Extension Based Hybrid Worker on Linux limits CPU to 5% which is nowhere mentioned - Agent Based Hybrid Worker didn't have this limit - Hi colleagues,
please at least mention the default limit to 5 % CPU, as the former Agent-Based Hybrid Worker didn't have this limit under Linux.
When upgrading to the new worker, existing CPU intensive runbooks therefore run extrem slow and also might get errors due to the extremely boxed resources.
see entry "CPUQuota=5%" in files:
- /var/lib/waagent/Microsoft.Azure.Automation.HybridWorker.HybridWorkerForLinux-0.1.0.11/HybridWorkerAgent/DaemonScripts/service_scripts/hwd.systemd
- /var/lib/waagent/Microsoft.Azure.Automation.HybridWorker.HybridWorkerForLinux-0.1.0.11/HybridWorkerAgent/DaemonScripts/service_temp/hwd.systemd
In my environment, I've actually updated the value from 5% to 95% - now, that I'm aware of it. :-)
Thanks & Best Regards,
Johann
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 2854aee3-7933-ee94-bb03-3a6a4c540088
* Version Independent ID: c902ddb9-46bf-c53c-46c4-e1c3d2cec7fd
* Content: [Deploy an extension-based Windows or Linux User Hybrid Runbook Worker in Azure Automation (Preview)](https://docs.microsoft.com/en-us/azure/automation/extension-based-hybrid-runbook-worker-install?tabs=windows)
* Content Source: [articles/automation/extension-based-hybrid-runbook-worker-install.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/extension-based-hybrid-runbook-worker-install.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SGSneha
* Microsoft Alias: **v-ssudhir**
|
process
|
extension based hybrid worker on linux limits cpu to which is nowhere mentioned agent based hybrid worker didn t have this limit hi colleagues please at least mention the default limit to cpu as the former agent based hybrid worker didn t have this limit under linux when upgrading to the new worker existing cpu intensive runbooks therefore run extrem slow and also might get errors due to the extremely boxed resources see entry cpuquota in files var lib waagent microsoft azure automation hybridworker hybridworkerforlinux hybridworkeragent daemonscripts service scripts hwd systemd var lib waagent microsoft azure automation hybridworker hybridworkerforlinux hybridworkeragent daemonscripts service temp hwd systemd in my environment i ve actually updated the value from to now that i m aware of it thanks best regards johann document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login sgsneha microsoft alias v ssudhir
| 1
|
2,891
| 5,871,928,141
|
IssuesEvent
|
2017-05-15 09:59:54
|
ProgrammingLife2017/Desoxyribonucleinezuur
|
https://api.github.com/repos/ProgrammingLife2017/Desoxyribonucleinezuur
|
closed
|
Planning meeting
|
feedback priority: A process / tools time:2
|
Have a meeting / brainstorming session about the organization (who does what) and the implementation (how do we do it)
|
1.0
|
Planning meeting - Have a meeting / brainstorming session about the organization (who does what) and the implementation (how do we do it)
|
process
|
planning meeting have a meeting brainstorming session about the organization who does what and the implementation how do we do it
| 1
|
6,408
| 9,488,277,619
|
IssuesEvent
|
2019-04-22 19:07:48
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Manual Deployment, Step 2 confusing or inaccurate
|
assigned-to-author automation/svc doc-enhancement process-automation/subsvc triaged
|
In step 2 of the manual installation steps, it states "Add the Automation solution". What solution is this referring to? Is this Automation & Control or Automation Hybrid Worker or Azure Automation service? In the past this referred to the Automation & Control solution. I've noticed that this solution fails to install. This step needs more clarity on how exactly you link Log Analytics and Azure Automation such that you can successfully get the HRW installed on the remote system.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 7b29372c-7bd9-7da2-4cff-9afbb432bccf
* Version Independent ID: 66ce101d-d21b-3fdf-be70-7f9cadc1570e
* Content: [Azure Automation Windows Hybrid Runbook Worker](https://docs.microsoft.com/en-us/azure/automation/automation-windows-hrw-install)
* Content Source: [articles/automation/automation-windows-hrw-install.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-windows-hrw-install.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @georgewallace
* Microsoft Alias: **gwallace**
|
1.0
|
Manual Deployment, Step 2 confusing or inaccurate - In step 2 of the manual installation steps, it states "Add the Automation solution". What solution is this referring to? Is this Automation & Control or Automation Hybrid Worker or Azure Automation service? In the past this referred to the Automation & Control solution. I've noticed that this solution fails to install. This step needs more clarity on how exactly you link Log Analytics and Azure Automation such that you can successfully get the HRW installed on the remote system.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 7b29372c-7bd9-7da2-4cff-9afbb432bccf
* Version Independent ID: 66ce101d-d21b-3fdf-be70-7f9cadc1570e
* Content: [Azure Automation Windows Hybrid Runbook Worker](https://docs.microsoft.com/en-us/azure/automation/automation-windows-hrw-install)
* Content Source: [articles/automation/automation-windows-hrw-install.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-windows-hrw-install.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @georgewallace
* Microsoft Alias: **gwallace**
|
process
|
manual deployment step confusing or inaccurate in step of the manual installation steps it states add the automation solution what solution is this referring to is this automation control or automation hybrid worker or azure automation service in the past this referred to the automation control solution i ve noticed that this solution fails to install this step needs more clarity on how exactly you link log analytics and azure automation such that you can successfully get the hrw installed on the remote system document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login georgewallace microsoft alias gwallace
| 1
|
279,924
| 24,265,114,553
|
IssuesEvent
|
2022-09-28 05:09:24
|
WPChill/download-monitor
|
https://api.github.com/repos/WPChill/download-monitor
|
closed
|
Versions don't add the meta to the count
|
Bug needs testing
|
**Describe the bug**
Need to add a filter same as for Download for backwards compatibility. Has commit https://github.com/WPChill/download-monitor/commit/a8b7be50401177618ef14dd13984c5078a6f2c16
|
1.0
|
Versions don't add the meta to the count - **Describe the bug**
Need to add a filter same as for Download for backwards compatibility. Has commit https://github.com/WPChill/download-monitor/commit/a8b7be50401177618ef14dd13984c5078a6f2c16
|
non_process
|
versions don t add the meta to the count describe the bug need to add a filter same as for download for backwards compatibility has commit
| 0
|
221,094
| 16,994,948,835
|
IssuesEvent
|
2021-07-01 04:32:45
|
faridlesosibirsk/TrainingRepository
|
https://api.github.com/repos/faridlesosibirsk/TrainingRepository
|
opened
|
20. Перемещение файлов
|
documentation
|
Написать заголовок и цель. Ответ сохранить в файл "20. Перемещение файлов.md".
|
1.0
|
20. Перемещение файлов - Написать заголовок и цель. Ответ сохранить в файл "20. Перемещение файлов.md".
|
non_process
|
перемещение файлов написать заголовок и цель ответ сохранить в файл перемещение файлов md
| 0
|
3,315
| 6,419,221,963
|
IssuesEvent
|
2017-08-08 20:43:48
|
TomorrowPartners/tomorrow-web
|
https://api.github.com/repos/TomorrowPartners/tomorrow-web
|
closed
|
Wireframe based on content
|
process_in progress
|
Based on #6. Here we go!
[Content for direction #1](https://docs.google.com/document/d/1q7xW9cDkjgN9PB4S_T0cG2szixjliL8q1127kuXAgWM/edit#)
[Content for direction #2](https://docs.google.com/document/d/1irOjuxS46GxVTsvZ5RG7udPn2gi1mcIYABwUhDRGlH0/edit)
Note: feel free to explore combinations of these, but just make sure that you're clear on your north star for that particular sketch/exploration/wireframe.
|
1.0
|
Wireframe based on content - Based on #6. Here we go!
[Content for direction #1](https://docs.google.com/document/d/1q7xW9cDkjgN9PB4S_T0cG2szixjliL8q1127kuXAgWM/edit#)
[Content for direction #2](https://docs.google.com/document/d/1irOjuxS46GxVTsvZ5RG7udPn2gi1mcIYABwUhDRGlH0/edit)
Note: feel free to explore combinations of these, but just make sure that you're clear on your north star for that particular sketch/exploration/wireframe.
|
process
|
wireframe based on content based on here we go note feel free to explore combinations of these but just make sure that you re clear on your north star for that particular sketch exploration wireframe
| 1
|
2,786
| 5,718,274,449
|
IssuesEvent
|
2017-04-19 19:08:46
|
imminfo/immdeep
|
https://api.github.com/repos/imminfo/immdeep
|
opened
|
Compare autoencoders
|
preprocessing
|
Fix the number of epochs = 150. Compare accuracy/loss of models with different latent dimensions and architectures.
- LSTM / GRU
- Bidirectional versions
- VAE
- (additional task) RNN with 1D convolutions
|
1.0
|
Compare autoencoders - Fix the number of epochs = 150. Compare accuracy/loss of models with different latent dimensions and architectures.
- LSTM / GRU
- Bidirectional versions
- VAE
- (additional task) RNN with 1D convolutions
|
process
|
compare autoencoders fix the number of epochs compare accuracy loss of models with different latent dimensions and architectures lstm gru bidirectional versions vae additional task rnn with convolutions
| 1
|
96,449
| 12,130,260,565
|
IssuesEvent
|
2020-04-23 01:00:00
|
solex2006/SELIProject
|
https://api.github.com/repos/solex2006/SELIProject
|
opened
|
Abbreviations
|
2 - Ready Feature Design Notes :notebook: P3 - Normal role:Student
|
This is a "not-end" requirement. What I want to mean is that this should be considered for any new page you create that Student can access. Also should correct the old pages.
I will use the label **Feature Design Notes** to this cases
***************
A mechanism for identifying the expanded form or meaning of abbreviations is available.
- [ ] Providing the expansion or explanation of an abbreviation **for the first occurrence** of the abbreviation, If the abbreviation **has only one meaning within the Web page**. Using one of the following techniques:
- Providing the first use of an abbreviation immediately before or after the expanded form
to make the expanded form of an abbreviation available by associating the expanded form with its abbreviation the first time it occurs within a Web page
- Linking to definitions
To make the definition of a word, phrase, or abbreviation available by providing the definition, either within the same Web page or in a different Web page, and establishing a link between the item and its definition.
- Providing definitions for abbreviations by using the `abbr` element
to provide expansions or definitions for abbreviations by using the abbr element
- [ ] Providing the expansion or explanation of an abbreviation **for all occurrence** of the abbreviation, If the abbreviation **has only one meaning within the Web page**. Using one of the following techniques:
- Linking to definitions
- Providing a glossary
- Using the link element to link to a glossary
- Providing a function to search an online dictionary
- Providing definitions for abbreviations by using the `abbr` element
- [ ] Providing the expansion or explanation of an abbreviation **for all occurrence** of the abbreviation, If the abbreviation **means different things within the Web page**. Using one of the following techniques:
- Linking to definitions
- Providing definitions for abbreviations by using the `abbr` element
### Examples
#### Example: Providing the first use of an abbreviation immediately before or after the expanded form
"The United Nations High Commissioner for Human Rights (UNHCR) was established in 1950 to provide protection and assistance to refugees."
"The WAI (Web Accessibility Initiative) demonstrates the W3C commitment to accessibility."
#### Example: Using abbr element to expand abbreviations.
```html
<p>Sugar is commonly sold in 5 <abbr title="pound">lb.</abbr> bags.</p>
<p>Welcome to the <abbr title="World Wide Web">WWW</abbr>!</p>
```
#### Example: Using abbr element to define abbreviations.
```html
<p>
Tasini <abbr title="and others">et al.</abbr> <abbr title="versus">v.</abbr>
The New York Times <abbr title="and others">et al.</abbr> is the landmark lawsuit
brought by members of the National Writers Union against ......
</p>
```
#### Example 4: Using the abbr element to expand an initialism
```html
<p>
<abbr title="World Wide Web">WWW</abbr>
</p>
```
# Test Procedure
### Providing the first use of an abbreviation immediately before or after the expanded form
For each abbreviation in the content,
1. Search for the first use of that abbreviation in the authored component.
2. Check that the first use is immediately preceded or followed by the expanded form of the abbreviation.
3. Check that the expanded form is the correct expanded form for the use of the abbreviation.
Expected Results: Checks 2 and 3 are true.
### Linking to definitions
For each word, phrase, or abbreviation to be defined:
1. Check that at least the first instance of the item is a link.
1. Check that each link navigates to the definition of the item.
Expected Results: all are true.
### using the abbr element
1. Check that an expansion or definition is provided for each abbreviation via abbr.
Expected Results: Check 1 is true.
### Glossary
1. Check that either
1. The glossary is included in the Web page, or
1. A mechanism is available to locate the glossary.
1. Check that each word, phrase, or abbreviation to be defined is defined in the glossary
1. Check that the glossary contains only one definition for each item.
Expected Results: All three checks above are true.
### Online dictionary
For each word, phrase, or abbreviation to be defined:
1. Check that a mechanism exists within the Web page to search for the word, phrase, or abbreviation via an on-line dictionary.
1. Check that the result of the search of the dictionary for the word, phrase, or abbreviation is the correct definition.
Expected Results: Checks 1 and 2 are true.
# :busts_in_silhouette: Benefits
This Success Criterion may help people who:
- have difficulty decoding words;
- rely on screen magnifiers (magnification may reduce contextual cues);
- have limited memory;
- have difficulty using context to aid understanding.
### Using the link element to link to a glossary
For any set of words and their definitions that are meant to serve as a glossary:
1. Check that the head section of the Web page that contains words, phrases or abbreviations defined in a glossary contains a `link` element
1. Check that the `link` element has attribute `rel="glossary"`
1. Check that the `href` attribute of the link element refers to the glossary page.
Expected Results: all are true.
|
1.0
|
Abbreviations - This is a "not-end" requirement. What I want to mean is that this should be considered for any new page you create that Student can access. Also should correct the old pages.
I will use the label **Feature Design Notes** to this cases
***************
A mechanism for identifying the expanded form or meaning of abbreviations is available.
- [ ] Providing the expansion or explanation of an abbreviation **for the first occurrence** of the abbreviation, If the abbreviation **has only one meaning within the Web page**. Using one of the following techniques:
- Providing the first use of an abbreviation immediately before or after the expanded form
to make the expanded form of an abbreviation available by associating the expanded form with its abbreviation the first time it occurs within a Web page
- Linking to definitions
To make the definition of a word, phrase, or abbreviation available by providing the definition, either within the same Web page or in a different Web page, and establishing a link between the item and its definition.
- Providing definitions for abbreviations by using the `abbr` element
to provide expansions or definitions for abbreviations by using the abbr element
- [ ] Providing the expansion or explanation of an abbreviation **for all occurrence** of the abbreviation, If the abbreviation **has only one meaning within the Web page**. Using one of the following techniques:
- Linking to definitions
- Providing a glossary
- Using the link element to link to a glossary
- Providing a function to search an online dictionary
- Providing definitions for abbreviations by using the `abbr` element
- [ ] Providing the expansion or explanation of an abbreviation **for all occurrence** of the abbreviation, If the abbreviation **means different things within the Web page**. Using one of the following techniques:
- Linking to definitions
- Providing definitions for abbreviations by using the `abbr` element
### Examples
#### Example: Providing the first use of an abbreviation immediately before or after the expanded form
"The United Nations High Commissioner for Human Rights (UNHCR) was established in 1950 to provide protection and assistance to refugees."
"The WAI (Web Accessibility Initiative) demonstrates the W3C commitment to accessibility."
#### Example: Using abbr element to expand abbreviations.
```html
<p>Sugar is commonly sold in 5 <abbr title="pound">lb.</abbr> bags.</p>
<p>Welcome to the <abbr title="World Wide Web">WWW</abbr>!</p>
```
#### Example: Using abbr element to define abbreviations.
```html
<p>
Tasini <abbr title="and others">et al.</abbr> <abbr title="versus">v.</abbr>
The New York Times <abbr title="and others">et al.</abbr> is the landmark lawsuit
brought by members of the National Writers Union against ......
</p>
```
#### Example 4: Using the abbr element to expand an initialism
```html
<p>
<abbr title="World Wide Web">WWW</abbr>
</p>
```
# Test Procedure
### Providing the first use of an abbreviation immediately before or after the expanded form
For each abbreviation in the content,
1. Search for the first use of that abbreviation in the authored component.
2. Check that the first use is immediately preceded or followed by the expanded form of the abbreviation.
3. Check that the expanded form is the correct expanded form for the use of the abbreviation.
Expected Results: Checks 2 and 3 are true.
### Linking to definitions
For each word, phrase, or abbreviation to be defined:
1. Check that at least the first instance of the item is a link.
1. Check that each link navigates to the definition of the item.
Expected Results: all are true.
### using the abbr element
1. Check that an expansion or definition is provided for each abbreviation via abbr.
Expected Results: Check 1 is true.
### Glossary
1. Check that either
1. The glossary is included in the Web page, or
1. A mechanism is available to locate the glossary.
1. Check that each word, phrase, or abbreviation to be defined is defined in the glossary
1. Check that the glossary contains only one definition for each item.
Expected Results: All three checks above are true.
### Online dictionary
For each word, phrase, or abbreviation to be defined:
1. Check that a mechanism exists within the Web page to search for the word, phrase, or abbreviation via an on-line dictionary.
1. Check that the result of the search of the dictionary for the word, phrase, or abbreviation is the correct definition.
Expected Results: Checks 1 and 2 are true.
# :busts_in_silhouette: Benefits
This Success Criterion may help people who:
- have difficulty decoding words;
- rely on screen magnifiers (magnification may reduce contextual cues);
- have limited memory;
- have difficulty using context to aid understanding.
### Using the link element to link to a glossary
For any set of words and their definitions that are meant to serve as a glossary:
1. Check that the head section of the Web page that contains words, phrases or abbreviations defined in a glossary contains a `link` element
1. Check that the `link` element has attribute `rel="glossary"`
1. Check that the `href` attribute of the link element refers to the glossary page.
Expected Results: all are true.
|
non_process
|
abbreviations this is a not end requirement what i want to mean is that this should be considered for any new page you create that student can access also should correct the old pages i will use the label feature design notes to this cases a mechanism for identifying the expanded form or meaning of abbreviations is available providing the expansion or explanation of an abbreviation for the first occurrence of the abbreviation if the abbreviation has only one meaning within the web page using one of the following techniques providing the first use of an abbreviation immediately before or after the expanded form to make the expanded form of an abbreviation available by associating the expanded form with its abbreviation the first time it occurs within a web page linking to definitions to make the definition of a word phrase or abbreviation available by providing the definition either within the same web page or in a different web page and establishing a link between the item and its definition providing definitions for abbreviations by using the abbr element to provide expansions or definitions for abbreviations by using the abbr element providing the expansion or explanation of an abbreviation for all occurrence of the abbreviation if the abbreviation has only one meaning within the web page using one of the following techniques linking to definitions providing a glossary using the link element to link to a glossary providing a function to search an online dictionary providing definitions for abbreviations by using the abbr element providing the expansion or explanation of an abbreviation for all occurrence of the abbreviation if the abbreviation means different things within the web page using one of the following techniques linking to definitions providing definitions for abbreviations by using the abbr element examples example providing the first use of an abbreviation immediately before or after the expanded form the united nations high commissioner for human rights unhcr was established in to provide protection and assistance to refugees the wai web accessibility initiative demonstrates the commitment to accessibility example using abbr element to expand abbreviations html sugar is commonly sold in lb bags welcome to the www example using abbr element to define abbreviations html tasini et al v the new york times et al is the landmark lawsuit brought by members of the national writers union against example using the abbr element to expand an initialism html www test procedure providing the first use of an abbreviation immediately before or after the expanded form for each abbreviation in the content search for the first use of that abbreviation in the authored component check that the first use is immediately preceded or followed by the expanded form of the abbreviation check that the expanded form is the correct expanded form for the use of the abbreviation expected results checks and are true linking to definitions for each word phrase or abbreviation to be defined check that at least the first instance of the item is a link check that each link navigates to the definition of the item expected results all are true using the abbr element check that an expansion or definition is provided for each abbreviation via abbr expected results check is true glossary check that either the glossary is included in the web page or a mechanism is available to locate the glossary check that each word phrase or abbreviation to be defined is defined in the glossary check that the glossary contains only one definition for each item expected results all three checks above are true online dictionary for each word phrase or abbreviation to be defined check that a mechanism exists within the web page to search for the word phrase or abbreviation via an on line dictionary check that the result of the search of the dictionary for the word phrase or abbreviation is the correct definition expected results checks and are true busts in silhouette benefits this success criterion may help people who have difficulty decoding words rely on screen magnifiers magnification may reduce contextual cues have limited memory have difficulty using context to aid understanding using the link element to link to a glossary for any set of words and their definitions that are meant to serve as a glossary check that the head section of the web page that contains words phrases or abbreviations defined in a glossary contains a link element check that the link element has attribute rel glossary check that the href attribute of the link element refers to the glossary page expected results all are true
| 0
|
269,827
| 20,507,756,156
|
IssuesEvent
|
2022-03-01 00:54:15
|
ChrisMenkhus/SpaceCaseDev
|
https://api.github.com/repos/ChrisMenkhus/SpaceCaseDev
|
opened
|
High Level Components Should Be Documented
|
documentation
|
Add comments to functions to better explain what they do
|
1.0
|
High Level Components Should Be Documented - Add comments to functions to better explain what they do
|
non_process
|
high level components should be documented add comments to functions to better explain what they do
| 0
|
13,450
| 15,896,626,110
|
IssuesEvent
|
2021-04-11 18:05:40
|
hasura/ask-me-anything
|
https://api.github.com/repos/hasura/ask-me-anything
|
opened
|
What numeric data types does Hasura support in relation to the underlying database(s)?
|
data type question series next-up-for-ama processing-for-shortvid question
|
## Postgres has the following Numeric Types:
<table border="1" class="colwidths-given docutils">
<colgroup>
<col width="24%">
<col width="20%">
<col width="45%">
<col width="11%">
</colgroup>
<thead valign="bottom">
<tr class="row-odd"><th class="head">Name</th>
<th class="head">Aliases</th>
<th class="head">Description</th>
<th class="head">Hasura Type</th>
</tr>
</thead>
<tbody valign="top">
<tr class="row-even"><td>bigint</td>
<td>int8</td>
<td>signed eight-byte integer</td>
<td><a class="reference internal" href="#string">String</a></td>
</tr>
<tr class="row-odd"><td>bigserial</td>
<td>serial8</td>
<td>autoincrementing eight-byte integer</td>
<td><a class="reference internal" href="#string">String</a></td>
</tr>
<tr class="row-even"><td>bit [ (n) ]</td>
<td> </td>
<td>fixed-length bit string</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>bit varying [ (n) ]</td>
<td>varbit [ (n) ]</td>
<td>variable-length bit string</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>boolean</td>
<td>bool</td>
<td>logical Boolean (true/false)</td>
<td><a class="reference internal" href="#bool">Bool</a></td>
</tr>
<tr class="row-odd"><td>box</td>
<td> </td>
<td>rectangular box on a plane</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>bytea</td>
<td> </td>
<td>binary data (“byte array”)</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>character [ (n) ]</td>
<td>char [ (n) ]</td>
<td>fixed-length character string</td>
<td><a class="reference internal" href="#char">Char</a></td>
</tr>
<tr class="row-even"><td>character varying [ (n) ]</td>
<td>varchar [ (n) ]</td>
<td>variable-length character string</td>
<td><a class="reference internal" href="#string">String</a></td>
</tr>
<tr class="row-odd"><td>cidr</td>
<td> </td>
<td>IPv4 or IPv6 network address</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>circle</td>
<td> </td>
<td>circle on a plane</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>date</td>
<td> </td>
<td>calendar date (year, month, day)</td>
<td><a class="reference internal" href="#date">Date</a></td>
</tr>
<tr class="row-even"><td>double precision</td>
<td>float8</td>
<td>double precision floating-point number (8 bytes)</td>
<td><a class="reference internal" href="#float">Float</a></td>
</tr>
<tr class="row-odd"><td>inet</td>
<td> </td>
<td>IPv4 or IPv6 host address</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>integer</td>
<td>int, int4</td>
<td>signed four-byte integer</td>
<td><a class="reference internal" href="#int">Int</a></td>
</tr>
<tr class="row-odd"><td>interval [ fields ] [ (p) ]</td>
<td> </td>
<td>time span</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>json</td>
<td> </td>
<td>textual JSON data</td>
<td><a class="reference internal" href="#json">JSON</a></td>
</tr>
<tr class="row-odd"><td>jsonb</td>
<td> </td>
<td>binary JSON data, decomposed</td>
<td><a class="reference internal" href="#jsonb">JSONB</a></td>
</tr>
<tr class="row-even"><td>line</td>
<td> </td>
<td>infinite line on a plane</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>lseg</td>
<td> </td>
<td>line segment on a plane</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>ltree</td>
<td> </td>
<td>labels of data stored in a hierarchical tree-like structure</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>geometry</td>
<td> </td>
<td>PostGIS Geometry type</td>
<td><a class="reference internal" href="#geometry">Geometry</a></td>
</tr>
<tr class="row-even"><td>geography</td>
<td> </td>
<td>PostGIS Geography type</td>
<td><a class="reference internal" href="#geography">Geography</a></td>
</tr>
<tr class="row-odd"><td>macaddr</td>
<td> </td>
<td>MAC (Media Access Control) address</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>macaddr8</td>
<td> </td>
<td>MAC (Media Access Control) address (EUI-64 format)</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>money</td>
<td> </td>
<td>currency amount</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>numeric [ (p, s) ]</td>
<td>decimal [ (p, s) ]</td>
<td>exact numeric of selectable precision</td>
<td><a class="reference internal" href="#numeric">Numeric</a></td>
</tr>
<tr class="row-odd"><td>path</td>
<td> </td>
<td>geometric path on a plane</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>pg_lsn</td>
<td> </td>
<td>PostgreSQL Log Sequence Number</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>point</td>
<td> </td>
<td>geometric point on a plane</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>polygon</td>
<td> </td>
<td>closed geometric path on a plane</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>real</td>
<td>float4</td>
<td>single precision floating-point number (4 bytes)</td>
<td><a class="reference internal" href="#float">Float</a></td>
</tr>
<tr class="row-even"><td>smallint</td>
<td>int2</td>
<td>signed two-byte integer</td>
<td><a class="reference internal" href="#int">Int</a></td>
</tr>
<tr class="row-odd"><td>smallserial</td>
<td>serial2</td>
<td>autoincrementing two-byte integer</td>
<td><a class="reference internal" href="#int">Int</a></td>
</tr>
<tr class="row-even"><td>serial</td>
<td>serial4</td>
<td>autoincrementing four-byte integer</td>
<td><a class="reference internal" href="#int">Int</a></td>
</tr>
<tr class="row-odd"><td>text</td>
<td> </td>
<td>variable-length character string</td>
<td><a class="reference internal" href="#string">String</a></td>
</tr>
<tr class="row-even"><td>time [ (p) ] [ without time zone ]</td>
<td> </td>
<td>time of day (no time zone)</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>time [ (p) ] with time zone</td>
<td>timetz</td>
<td>time of day, including time zone</td>
<td><a class="reference internal" href="#timetz">Timetz</a></td>
</tr>
<tr class="row-even"><td>timestamp [ (p) ] [ without time zone ]</td>
<td> </td>
<td>date and time (no time zone)</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>timestamp [ (p) ] with time zone</td>
<td>timestamptz</td>
<td>date and time, including time zone</td>
<td><a class="reference internal" href="#timestamptz">Timestamptz</a></td>
</tr>
<tr class="row-even"><td>tsquery</td>
<td> </td>
<td>text search query</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>tsvector</td>
<td> </td>
<td>text search document</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>txid_snapshot</td>
<td> </td>
<td>user-level transaction ID snapshot</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>uuid</td>
<td> </td>
<td>universally unique identifier</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>xml</td>
<td> </td>
<td>XML data</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
</tbody>
</table>
|
1.0
|
What numeric data types does Hasura support in relation to the underlying database(s)? - ## Postgres has the following Numeric Types:
<table border="1" class="colwidths-given docutils">
<colgroup>
<col width="24%">
<col width="20%">
<col width="45%">
<col width="11%">
</colgroup>
<thead valign="bottom">
<tr class="row-odd"><th class="head">Name</th>
<th class="head">Aliases</th>
<th class="head">Description</th>
<th class="head">Hasura Type</th>
</tr>
</thead>
<tbody valign="top">
<tr class="row-even"><td>bigint</td>
<td>int8</td>
<td>signed eight-byte integer</td>
<td><a class="reference internal" href="#string">String</a></td>
</tr>
<tr class="row-odd"><td>bigserial</td>
<td>serial8</td>
<td>autoincrementing eight-byte integer</td>
<td><a class="reference internal" href="#string">String</a></td>
</tr>
<tr class="row-even"><td>bit [ (n) ]</td>
<td> </td>
<td>fixed-length bit string</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>bit varying [ (n) ]</td>
<td>varbit [ (n) ]</td>
<td>variable-length bit string</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>boolean</td>
<td>bool</td>
<td>logical Boolean (true/false)</td>
<td><a class="reference internal" href="#bool">Bool</a></td>
</tr>
<tr class="row-odd"><td>box</td>
<td> </td>
<td>rectangular box on a plane</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>bytea</td>
<td> </td>
<td>binary data (“byte array”)</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>character [ (n) ]</td>
<td>char [ (n) ]</td>
<td>fixed-length character string</td>
<td><a class="reference internal" href="#char">Char</a></td>
</tr>
<tr class="row-even"><td>character varying [ (n) ]</td>
<td>varchar [ (n) ]</td>
<td>variable-length character string</td>
<td><a class="reference internal" href="#string">String</a></td>
</tr>
<tr class="row-odd"><td>cidr</td>
<td> </td>
<td>IPv4 or IPv6 network address</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>circle</td>
<td> </td>
<td>circle on a plane</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>date</td>
<td> </td>
<td>calendar date (year, month, day)</td>
<td><a class="reference internal" href="#date">Date</a></td>
</tr>
<tr class="row-even"><td>double precision</td>
<td>float8</td>
<td>double precision floating-point number (8 bytes)</td>
<td><a class="reference internal" href="#float">Float</a></td>
</tr>
<tr class="row-odd"><td>inet</td>
<td> </td>
<td>IPv4 or IPv6 host address</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>integer</td>
<td>int, int4</td>
<td>signed four-byte integer</td>
<td><a class="reference internal" href="#int">Int</a></td>
</tr>
<tr class="row-odd"><td>interval [ fields ] [ (p) ]</td>
<td> </td>
<td>time span</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>json</td>
<td> </td>
<td>textual JSON data</td>
<td><a class="reference internal" href="#json">JSON</a></td>
</tr>
<tr class="row-odd"><td>jsonb</td>
<td> </td>
<td>binary JSON data, decomposed</td>
<td><a class="reference internal" href="#jsonb">JSONB</a></td>
</tr>
<tr class="row-even"><td>line</td>
<td> </td>
<td>infinite line on a plane</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>lseg</td>
<td> </td>
<td>line segment on a plane</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>ltree</td>
<td> </td>
<td>labels of data stored in a hierarchical tree-like structure</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>geometry</td>
<td> </td>
<td>PostGIS Geometry type</td>
<td><a class="reference internal" href="#geometry">Geometry</a></td>
</tr>
<tr class="row-even"><td>geography</td>
<td> </td>
<td>PostGIS Geography type</td>
<td><a class="reference internal" href="#geography">Geography</a></td>
</tr>
<tr class="row-odd"><td>macaddr</td>
<td> </td>
<td>MAC (Media Access Control) address</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>macaddr8</td>
<td> </td>
<td>MAC (Media Access Control) address (EUI-64 format)</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>money</td>
<td> </td>
<td>currency amount</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>numeric [ (p, s) ]</td>
<td>decimal [ (p, s) ]</td>
<td>exact numeric of selectable precision</td>
<td><a class="reference internal" href="#numeric">Numeric</a></td>
</tr>
<tr class="row-odd"><td>path</td>
<td> </td>
<td>geometric path on a plane</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>pg_lsn</td>
<td> </td>
<td>PostgreSQL Log Sequence Number</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>point</td>
<td> </td>
<td>geometric point on a plane</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>polygon</td>
<td> </td>
<td>closed geometric path on a plane</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>real</td>
<td>float4</td>
<td>single precision floating-point number (4 bytes)</td>
<td><a class="reference internal" href="#float">Float</a></td>
</tr>
<tr class="row-even"><td>smallint</td>
<td>int2</td>
<td>signed two-byte integer</td>
<td><a class="reference internal" href="#int">Int</a></td>
</tr>
<tr class="row-odd"><td>smallserial</td>
<td>serial2</td>
<td>autoincrementing two-byte integer</td>
<td><a class="reference internal" href="#int">Int</a></td>
</tr>
<tr class="row-even"><td>serial</td>
<td>serial4</td>
<td>autoincrementing four-byte integer</td>
<td><a class="reference internal" href="#int">Int</a></td>
</tr>
<tr class="row-odd"><td>text</td>
<td> </td>
<td>variable-length character string</td>
<td><a class="reference internal" href="#string">String</a></td>
</tr>
<tr class="row-even"><td>time [ (p) ] [ without time zone ]</td>
<td> </td>
<td>time of day (no time zone)</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>time [ (p) ] with time zone</td>
<td>timetz</td>
<td>time of day, including time zone</td>
<td><a class="reference internal" href="#timetz">Timetz</a></td>
</tr>
<tr class="row-even"><td>timestamp [ (p) ] [ without time zone ]</td>
<td> </td>
<td>date and time (no time zone)</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>timestamp [ (p) ] with time zone</td>
<td>timestamptz</td>
<td>date and time, including time zone</td>
<td><a class="reference internal" href="#timestamptz">Timestamptz</a></td>
</tr>
<tr class="row-even"><td>tsquery</td>
<td> </td>
<td>text search query</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>tsvector</td>
<td> </td>
<td>text search document</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>txid_snapshot</td>
<td> </td>
<td>user-level transaction ID snapshot</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-odd"><td>uuid</td>
<td> </td>
<td>universally unique identifier</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
<tr class="row-even"><td>xml</td>
<td> </td>
<td>XML data</td>
<td><a class="reference internal" href="#implicit">Implicit</a></td>
</tr>
</tbody>
</table>
|
process
|
what numeric data types does hasura support in relation to the underlying database s postgres has the following numeric types name aliases description hasura type bigint signed eight byte integer string bigserial autoincrementing eight byte integer string bit nbsp fixed length bit string implicit bit varying varbit variable length bit string implicit boolean bool logical boolean true false bool box nbsp rectangular box on a plane implicit bytea nbsp binary data “byte array” implicit character char fixed length character string char character varying varchar variable length character string string cidr nbsp or network address implicit circle nbsp circle on a plane implicit date nbsp calendar date year month day date double precision double precision floating point number bytes float inet nbsp or host address implicit integer int signed four byte integer int interval nbsp time span implicit json nbsp textual json data json jsonb nbsp binary json data decomposed jsonb line nbsp infinite line on a plane implicit lseg nbsp line segment on a plane implicit ltree nbsp labels of data stored in a hierarchical tree like structure implicit geometry nbsp postgis geometry type geometry geography nbsp postgis geography type geography macaddr nbsp mac media access control address implicit nbsp mac media access control address eui format implicit money nbsp currency amount implicit numeric decimal exact numeric of selectable precision numeric path nbsp geometric path on a plane implicit pg lsn nbsp postgresql log sequence number implicit point nbsp geometric point on a plane implicit polygon nbsp closed geometric path on a plane implicit real single precision floating point number bytes float smallint signed two byte integer int smallserial autoincrementing two byte integer int serial autoincrementing four byte integer int text nbsp variable length character string string time nbsp time of day no time zone implicit time with time zone timetz time of day including time zone timetz timestamp nbsp date and time no time zone implicit timestamp with time zone timestamptz date and time including time zone timestamptz tsquery nbsp text search query implicit tsvector nbsp text search document implicit txid snapshot nbsp user level transaction id snapshot implicit uuid nbsp universally unique identifier implicit xml nbsp xml data implicit
| 1
|
61,106
| 14,615,287,882
|
IssuesEvent
|
2020-12-22 11:17:24
|
rsoreq/NodeGoat
|
https://api.github.com/repos/rsoreq/NodeGoat
|
opened
|
CVE-2020-7656 (Medium) detected in jquery-1.4.4.min.js
|
security vulnerability
|
## CVE-2020-7656 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.4.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js</a></p>
<p>Path to dependency file: NodeGoat/node_modules/selenium-webdriver/lib/test/data/droppableItems.html</p>
<p>Path to vulnerable library: NodeGoat/node_modules/selenium-webdriver/lib/test/data/js/jquery-1.4.4.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.4.4.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rsoreq/NodeGoat/commit/30db8811fba7f9bab1b860cabc70c1795eac29e9">30db8811fba7f9bab1b860cabc70c1795eac29e9</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed.
<p>Publish Date: 2020-05-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/rails/jquery-rails/commit/8f601cbfa08749ee5bbd2bffb6e509db9d753568">https://github.com/rails/jquery-rails/commit/8f601cbfa08749ee5bbd2bffb6e509db9d753568</a></p>
<p>Release Date: 2020-05-19</p>
<p>Fix Resolution: jquery-rails - 2.2.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.4.4","isTransitiveDependency":false,"dependencyTree":"jquery:1.4.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jquery-rails - 2.2.0"}],"vulnerabilityIdentifier":"CVE-2020-7656","vulnerabilityDetails":"jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove \"\u003cscript\u003e\" HTML tags that contain a whitespace character, i.e: \"\u003c/script \u003e\", which results in the enclosed script logic to be executed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-7656 (Medium) detected in jquery-1.4.4.min.js - ## CVE-2020-7656 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.4.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.4/jquery.min.js</a></p>
<p>Path to dependency file: NodeGoat/node_modules/selenium-webdriver/lib/test/data/droppableItems.html</p>
<p>Path to vulnerable library: NodeGoat/node_modules/selenium-webdriver/lib/test/data/js/jquery-1.4.4.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.4.4.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rsoreq/NodeGoat/commit/30db8811fba7f9bab1b860cabc70c1795eac29e9">30db8811fba7f9bab1b860cabc70c1795eac29e9</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed.
<p>Publish Date: 2020-05-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/rails/jquery-rails/commit/8f601cbfa08749ee5bbd2bffb6e509db9d753568">https://github.com/rails/jquery-rails/commit/8f601cbfa08749ee5bbd2bffb6e509db9d753568</a></p>
<p>Release Date: 2020-05-19</p>
<p>Fix Resolution: jquery-rails - 2.2.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.4.4","isTransitiveDependency":false,"dependencyTree":"jquery:1.4.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jquery-rails - 2.2.0"}],"vulnerabilityIdentifier":"CVE-2020-7656","vulnerabilityDetails":"jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove \"\u003cscript\u003e\" HTML tags that contain a whitespace character, i.e: \"\u003c/script \u003e\", which results in the enclosed script logic to be executed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file nodegoat node modules selenium webdriver lib test data droppableitems html path to vulnerable library nodegoat node modules selenium webdriver lib test data js jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery prior to allows cross site scripting attacks via the load method the load method fails to recognize and remove html tags that contain a whitespace character i e which results in the enclosed script logic to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery rails isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails jquery prior to allows cross site scripting attacks via the load method the load method fails to recognize and remove html tags that contain a whitespace character i e script which results in the enclosed script logic to be executed vulnerabilityurl
| 0
|
26,285
| 4,213,423,558
|
IssuesEvent
|
2016-06-29 19:02:24
|
ODM2/ODM2PythonAPI
|
https://api.github.com/repos/ODM2/ODM2PythonAPI
|
closed
|
Consolidate List of API Functions
|
enhancement ready for testing
|
After looking at the existing code, we have decided on a major consolidation of the functions in the API to avoid complexity and repetition. For a description of the planned signatures (inputs and outputs) for each of the functions, [click here](https://github.com/ODM2/ODM2PythonAPI/blob/master/doc/APIFunctionList.md).
Consolidated List of ODM2 API Low-level Get Functions
---
- [x] getActions()
- [x] getAffilations()
- [x] getDatasets()
- [x] getEquipment()
- [x] getMethods()
- [x] getModels()
- [x] getOrganizations()
- [x] getPeople()
- [x] getProcessingLevels()
- [x] getRelatedActions()
- [x] getRelatedModels()
- [x] getRelatedSamplingFeatures()
- [x] getResults()
- [x] getResultValues()
- [x] getSamplingFeatures()
- [x] getSimulations()
- [x] getUnits()
- [x] getVariables()
Consolidated List of ODM2 API Low-level Create Functions
---
- [x] createVariable()
- [x] createMethod()
- [x] createProcessingLevel()
- [x] createSamplingFeature()
- [x] createUnit()
- [x] createOrganization()
- [x] createPerson()
- [x] createAffiliation()
- [x] createDataset()
- [x] createDatasetResults()
- [x] createAction()
- [x] createRelatedAction()
- [x] createResult()
- [x] createResultValues()
- [x] createSamplingFeature()
- [x] createSpatialReference()
- [x] createModel()
- [x] createRelatedModel()
- [x] createSimulation()
For Reference - Functions that are being removed
---
getVariableById - Covered by getVariables (passing in a VariableID)
getVariableByCode - covered by getVariables (passing in a VariableCode)
getResultById - covered by getResults (passing in a ResultID)
getMethodById - covered by getMethods (passing in a MethodID)
getMethodByCode - coverd by getMethods (passing in a MethodCode)
getProcessingLevelById - covered by getProcessingLevels (passing in a ProcessingLevelID)
getProcessingLevelByCode - covered by getProcessingLevels (passing in a ProcessingLevelCode)
getSamplingFeatureById - Covered by getSamplingFeatures (passing in a SamplingFeatureID)
getSamplingFeatureByCode - Covered by getSamplingFeatures (Passing in a SamplingFeatureCode)
getSamplingFeaturesByType - Covered by getSamplingFeatures (Passing in a SamplingFeatureType)
getSamplingFeatureByGeometry - Covered by getSamplingFeatures (passing in a GeometryType)
getGeometryTest - I don't know what this is and don't think we need it
getUnitById - Covered by getUnits (passing in a UnitsID)
getUnitByName - Covered by getUnits (passing in a UnitsName)
getOrganizationById - Covered by getOrganizations (passing in an OrganizationID)
getOrganizationByCode - Covered by getOrganizations (passing in an OrganizationCode)
getPersonById - Covered by getPeople (passing in a PersonID)
getPersonByName - Covered by getPeople (passing in a name)
getAffiliationByPersonAndOrg - why do we need this? Covered by getAffiliations?
getAffiliationsByPerson - Covered by getAffiliations (passing in a PersonID)
getResultByActionID - Covered by getResults (passing in an ActionID)
getResultByID - Covered by getResults
getResultAndGeomByID - Covered by getResults (passing in a ResultID)
getResultAndGeomByActionID - coverec by getResults (passing in an ActionID)
getResultValidDateTime - covered by getting the result and querying the metadata
getDatasetByCode - Covered by getDatasets (passing in a DatasetCode)
getAllDataQuality - what is this for?
getAllEquipment - covered by getEquipment
getCitations - why do we need this independently?
getTimeSeriesResults - covered by getResults
getTimeSeriesResultByResultId - Covered by getResults (passing in a ResultID)
getTimeSeriesResultbyCode - Results don't have a code, so not sure what this would do anyway
getTimeSeriesResultValues - covered by getResultValues
getTimeSeriesResultValuesByResultId - covered by getResultValues (passing in a ResultID)
getTimeSeriesResultValuesByCode - results don't have a code so this wouldnt work anyway
getTimeSeriesResultValuesByTime - covered by getResultValues (passing in a ResultID and a time period)
getAllSites - Covered by getSamplingFeatures (passing a type of "Site")
getSiteBySFId - covered by getSamplingFeatures (passing in a SamplingFeatureID)
getSiteBySFCode - covered by getSamplingFeatures (passing in a SamplingFeatureCode)
getSpatialReferenceByCode - don't think we need this
getAllDeploymentAction - covered by getActions (passing in an ActionType)
getDeploymentActionById - covered by getActions (passing in an ActionID)
getDeploymentActionByCode - Actions do not have code, so this one wouldn't work anyway
getAllModels - covered by getModels
getModelByCode - covered by getModels (passing in a ModelCode)
getAllSimulations - covered by getSimulations
getSimulationByName - covered by getSimulations (passing in a Simulation Name)
getSimulationByActionID - covered by getSimulations (passing in an ActionID)
getRelatedModelsByID - covered by getRelatedModels (Passing in a ModelID)
getRelatedModelsByCode - covered by getRelatedModels (passing in a ModelCode)
getResultsBySimulationID - covered by GetResults (passing in an ActionID)
createTimeSeriesResult - covered by createResult
createTimeSeriesResultValues - covered by createResultValues
createSite - covered by createSamplingFeature
createDeploymentAction - covered by createAction
|
1.0
|
Consolidate List of API Functions - After looking at the existing code, we have decided on a major consolidation of the functions in the API to avoid complexity and repetition. For a description of the planned signatures (inputs and outputs) for each of the functions, [click here](https://github.com/ODM2/ODM2PythonAPI/blob/master/doc/APIFunctionList.md).
Consolidated List of ODM2 API Low-level Get Functions
---
- [x] getActions()
- [x] getAffilations()
- [x] getDatasets()
- [x] getEquipment()
- [x] getMethods()
- [x] getModels()
- [x] getOrganizations()
- [x] getPeople()
- [x] getProcessingLevels()
- [x] getRelatedActions()
- [x] getRelatedModels()
- [x] getRelatedSamplingFeatures()
- [x] getResults()
- [x] getResultValues()
- [x] getSamplingFeatures()
- [x] getSimulations()
- [x] getUnits()
- [x] getVariables()
Consolidated List of ODM2 API Low-level Create Functions
---
- [x] createVariable()
- [x] createMethod()
- [x] createProcessingLevel()
- [x] createSamplingFeature()
- [x] createUnit()
- [x] createOrganization()
- [x] createPerson()
- [x] createAffiliation()
- [x] createDataset()
- [x] createDatasetResults()
- [x] createAction()
- [x] createRelatedAction()
- [x] createResult()
- [x] createResultValues()
- [x] createSamplingFeature()
- [x] createSpatialReference()
- [x] createModel()
- [x] createRelatedModel()
- [x] createSimulation()
For Reference - Functions that are being removed
---
getVariableById - Covered by getVariables (passing in a VariableID)
getVariableByCode - covered by getVariables (passing in a VariableCode)
getResultById - covered by getResults (passing in a ResultID)
getMethodById - covered by getMethods (passing in a MethodID)
getMethodByCode - coverd by getMethods (passing in a MethodCode)
getProcessingLevelById - covered by getProcessingLevels (passing in a ProcessingLevelID)
getProcessingLevelByCode - covered by getProcessingLevels (passing in a ProcessingLevelCode)
getSamplingFeatureById - Covered by getSamplingFeatures (passing in a SamplingFeatureID)
getSamplingFeatureByCode - Covered by getSamplingFeatures (Passing in a SamplingFeatureCode)
getSamplingFeaturesByType - Covered by getSamplingFeatures (Passing in a SamplingFeatureType)
getSamplingFeatureByGeometry - Covered by getSamplingFeatures (passing in a GeometryType)
getGeometryTest - I don't know what this is and don't think we need it
getUnitById - Covered by getUnits (passing in a UnitsID)
getUnitByName - Covered by getUnits (passing in a UnitsName)
getOrganizationById - Covered by getOrganizations (passing in an OrganizationID)
getOrganizationByCode - Covered by getOrganizations (passing in an OrganizationCode)
getPersonById - Covered by getPeople (passing in a PersonID)
getPersonByName - Covered by getPeople (passing in a name)
getAffiliationByPersonAndOrg - why do we need this? Covered by getAffiliations?
getAffiliationsByPerson - Covered by getAffiliations (passing in a PersonID)
getResultByActionID - Covered by getResults (passing in an ActionID)
getResultByID - Covered by getResults
getResultAndGeomByID - Covered by getResults (passing in a ResultID)
getResultAndGeomByActionID - coverec by getResults (passing in an ActionID)
getResultValidDateTime - covered by getting the result and querying the metadata
getDatasetByCode - Covered by getDatasets (passing in a DatasetCode)
getAllDataQuality - what is this for?
getAllEquipment - covered by getEquipment
getCitations - why do we need this independently?
getTimeSeriesResults - covered by getResults
getTimeSeriesResultByResultId - Covered by getResults (passing in a ResultID)
getTimeSeriesResultbyCode - Results don't have a code, so not sure what this would do anyway
getTimeSeriesResultValues - covered by getResultValues
getTimeSeriesResultValuesByResultId - covered by getResultValues (passing in a ResultID)
getTimeSeriesResultValuesByCode - results don't have a code so this wouldnt work anyway
getTimeSeriesResultValuesByTime - covered by getResultValues (passing in a ResultID and a time period)
getAllSites - Covered by getSamplingFeatures (passing a type of "Site")
getSiteBySFId - covered by getSamplingFeatures (passing in a SamplingFeatureID)
getSiteBySFCode - covered by getSamplingFeatures (passing in a SamplingFeatureCode)
getSpatialReferenceByCode - don't think we need this
getAllDeploymentAction - covered by getActions (passing in an ActionType)
getDeploymentActionById - covered by getActions (passing in an ActionID)
getDeploymentActionByCode - Actions do not have code, so this one wouldn't work anyway
getAllModels - covered by getModels
getModelByCode - covered by getModels (passing in a ModelCode)
getAllSimulations - covered by getSimulations
getSimulationByName - covered by getSimulations (passing in a Simulation Name)
getSimulationByActionID - covered by getSimulations (passing in an ActionID)
getRelatedModelsByID - covered by getRelatedModels (Passing in a ModelID)
getRelatedModelsByCode - covered by getRelatedModels (passing in a ModelCode)
getResultsBySimulationID - covered by GetResults (passing in an ActionID)
createTimeSeriesResult - covered by createResult
createTimeSeriesResultValues - covered by createResultValues
createSite - covered by createSamplingFeature
createDeploymentAction - covered by createAction
|
non_process
|
consolidate list of api functions after looking at the existing code we have decided on a major consolidation of the functions in the api to avoid complexity and repetition for a description of the planned signatures inputs and outputs for each of the functions consolidated list of api low level get functions getactions getaffilations getdatasets getequipment getmethods getmodels getorganizations getpeople getprocessinglevels getrelatedactions getrelatedmodels getrelatedsamplingfeatures getresults getresultvalues getsamplingfeatures getsimulations getunits getvariables consolidated list of api low level create functions createvariable createmethod createprocessinglevel createsamplingfeature createunit createorganization createperson createaffiliation createdataset createdatasetresults createaction createrelatedaction createresult createresultvalues createsamplingfeature createspatialreference createmodel createrelatedmodel createsimulation for reference functions that are being removed getvariablebyid covered by getvariables passing in a variableid getvariablebycode covered by getvariables passing in a variablecode getresultbyid covered by getresults passing in a resultid getmethodbyid covered by getmethods passing in a methodid getmethodbycode coverd by getmethods passing in a methodcode getprocessinglevelbyid covered by getprocessinglevels passing in a processinglevelid getprocessinglevelbycode covered by getprocessinglevels passing in a processinglevelcode getsamplingfeaturebyid covered by getsamplingfeatures passing in a samplingfeatureid getsamplingfeaturebycode covered by getsamplingfeatures passing in a samplingfeaturecode getsamplingfeaturesbytype covered by getsamplingfeatures passing in a samplingfeaturetype getsamplingfeaturebygeometry covered by getsamplingfeatures passing in a geometrytype getgeometrytest i don t know what this is and don t think we need it getunitbyid covered by getunits passing in a unitsid getunitbyname covered by getunits passing in a unitsname getorganizationbyid covered by getorganizations passing in an organizationid getorganizationbycode covered by getorganizations passing in an organizationcode getpersonbyid covered by getpeople passing in a personid getpersonbyname covered by getpeople passing in a name getaffiliationbypersonandorg why do we need this covered by getaffiliations getaffiliationsbyperson covered by getaffiliations passing in a personid getresultbyactionid covered by getresults passing in an actionid getresultbyid covered by getresults getresultandgeombyid covered by getresults passing in a resultid getresultandgeombyactionid coverec by getresults passing in an actionid getresultvaliddatetime covered by getting the result and querying the metadata getdatasetbycode covered by getdatasets passing in a datasetcode getalldataquality what is this for getallequipment covered by getequipment getcitations why do we need this independently gettimeseriesresults covered by getresults gettimeseriesresultbyresultid covered by getresults passing in a resultid gettimeseriesresultbycode results don t have a code so not sure what this would do anyway gettimeseriesresultvalues covered by getresultvalues gettimeseriesresultvaluesbyresultid covered by getresultvalues passing in a resultid gettimeseriesresultvaluesbycode results don t have a code so this wouldnt work anyway gettimeseriesresultvaluesbytime covered by getresultvalues passing in a resultid and a time period getallsites covered by getsamplingfeatures passing a type of site getsitebysfid covered by getsamplingfeatures passing in a samplingfeatureid getsitebysfcode covered by getsamplingfeatures passing in a samplingfeaturecode getspatialreferencebycode don t think we need this getalldeploymentaction covered by getactions passing in an actiontype getdeploymentactionbyid covered by getactions passing in an actionid getdeploymentactionbycode actions do not have code so this one wouldn t work anyway getallmodels covered by getmodels getmodelbycode covered by getmodels passing in a modelcode getallsimulations covered by getsimulations getsimulationbyname covered by getsimulations passing in a simulation name getsimulationbyactionid covered by getsimulations passing in an actionid getrelatedmodelsbyid covered by getrelatedmodels passing in a modelid getrelatedmodelsbycode covered by getrelatedmodels passing in a modelcode getresultsbysimulationid covered by getresults passing in an actionid createtimeseriesresult covered by createresult createtimeseriesresultvalues covered by createresultvalues createsite covered by createsamplingfeature createdeploymentaction covered by createaction
| 0
|
267,536
| 23,306,488,890
|
IssuesEvent
|
2022-08-08 01:59:29
|
void-linux/void-packages
|
https://api.github.com/repos/void-linux/void-packages
|
opened
|
godot 3.5: Font rendering in editor broken.
|
bug needs-testing
|
### Is this a new report?
Yes
### System Info
Void 5.18.16_1 x86_64 GenuineIntel uptodate FFFFFFFFFF
### Package(s) Affected
godot-3.5_1
### Does a report exist for this bug with the project's home (upstream) and/or another distro?
_No response_
### Expected behaviour
Font rendering in the editor should work.
### Actual behaviour
Font rendering in the editor is broken:

Godot writes the following error message:
```
ERROR: Unknown font format.
at: _load (scene/resources/dynamic_font.cpp:178)
```
here is the relevant line in godot's source: https://github.com/godotengine/godot/blob/3.5-stable/scene/resources/dynamic_font.cpp#L178
### Steps to reproduce
1. Install Godot 3.5
2. Open the editor
|
1.0
|
godot 3.5: Font rendering in editor broken. - ### Is this a new report?
Yes
### System Info
Void 5.18.16_1 x86_64 GenuineIntel uptodate FFFFFFFFFF
### Package(s) Affected
godot-3.5_1
### Does a report exist for this bug with the project's home (upstream) and/or another distro?
_No response_
### Expected behaviour
Font rendering in the editor should work.
### Actual behaviour
Font rendering in the editor is broken:

Godot writes the following error message:
```
ERROR: Unknown font format.
at: _load (scene/resources/dynamic_font.cpp:178)
```
here is the relevant line in godot's source: https://github.com/godotengine/godot/blob/3.5-stable/scene/resources/dynamic_font.cpp#L178
### Steps to reproduce
1. Install Godot 3.5
2. Open the editor
|
non_process
|
godot font rendering in editor broken is this a new report yes system info void genuineintel uptodate ffffffffff package s affected godot does a report exist for this bug with the project s home upstream and or another distro no response expected behaviour font rendering in the editor should work actual behaviour font rendering in the editor is broken godot writes the following error message error unknown font format at load scene resources dynamic font cpp here is the relevant line in godot s source steps to reproduce install godot open the editor
| 0
|
15,869
| 20,036,522,963
|
IssuesEvent
|
2022-02-02 12:30:39
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[iOS] Study activities > Incorrect start/end date and time displayed for mobile participants having different time zone
|
Bug P1 iOS Process: Fixed
|
**Scenario 1:** Incorrect start/end date and time displayed for mobile participants when participant having different timezone enrolls into study
Steps:
1. Keep iOS mobile to different timezone having DST (Daylight savings) eg. EDT
2. Enroll into any study
3. Observe the start/end date and time for scheduling activities
Actual: Incorrect start/end date and time displayed for mobile participants having different timezone and study reminder notification displayed incorrectly
Expected: start/end date and time should display as per Study Builder's configured
Issue observed for all scheduling activities
Normal Timezone without daylight saving Eg. IST

EDT Timezone screenshot:

**Scenario 2:** Activities appear 1 hour before scheduled post changing timezone
Steps:
1. Enroll into any study
2. Change the timezone with DST eg. EDT
3. Observe the start/end date and time for scheduling activities
Actual: Activities appear 1 hour before scheduled post changing timezone and study reminder notification displayed incorrectly
Expected: start/end date and time should display as per Study Builder's configured
|
1.0
|
[iOS] Study activities > Incorrect start/end date and time displayed for mobile participants having different time zone - **Scenario 1:** Incorrect start/end date and time displayed for mobile participants when participant having different timezone enrolls into study
Steps:
1. Keep iOS mobile to different timezone having DST (Daylight savings) eg. EDT
2. Enroll into any study
3. Observe the start/end date and time for scheduling activities
Actual: Incorrect start/end date and time displayed for mobile participants having different timezone and study reminder notification displayed incorrectly
Expected: start/end date and time should display as per Study Builder's configured
Issue observed for all scheduling activities
Normal Timezone without daylight saving Eg. IST

EDT Timezone screenshot:

**Scenario 2:** Activities appear 1 hour before scheduled post changing timezone
Steps:
1. Enroll into any study
2. Change the timezone with DST eg. EDT
3. Observe the start/end date and time for scheduling activities
Actual: Activities appear 1 hour before scheduled post changing timezone and study reminder notification displayed incorrectly
Expected: start/end date and time should display as per Study Builder's configured
|
process
|
study activities incorrect start end date and time displayed for mobile participants having different time zone scenario incorrect start end date and time displayed for mobile participants when participant having different timezone enrolls into study steps keep ios mobile to different timezone having dst daylight savings eg edt enroll into any study observe the start end date and time for scheduling activities actual incorrect start end date and time displayed for mobile participants having different timezone and study reminder notification displayed incorrectly expected start end date and time should display as per study builder s configured issue observed for all scheduling activities normal timezone without daylight saving eg ist edt timezone screenshot scenario activities appear hour before scheduled post changing timezone steps enroll into any study change the timezone with dst eg edt observe the start end date and time for scheduling activities actual activities appear hour before scheduled post changing timezone and study reminder notification displayed incorrectly expected start end date and time should display as per study builder s configured
| 1
|
22,329
| 30,913,745,443
|
IssuesEvent
|
2023-08-05 02:48:09
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
pih 1.48036 has 2 GuardDog issues
|
guarddog typosquatting silent-process-execution
|
https://pypi.org/project/pih
https://inspector.pypi.io/project/pih
```{
"dependency": "pih",
"version": "1.48036",
"result": {
"issues": 2,
"errors": {},
"results": {
"typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pid, pip",
"silent-process-execution": [
{
"location": "pih-1.48036/pih/tools.py:781",
"code": " result = subprocess.run(command, stdin=subprocess.DEVNULL, stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmphuwc_tvc/pih"
}
}```
|
1.0
|
pih 1.48036 has 2 GuardDog issues - https://pypi.org/project/pih
https://inspector.pypi.io/project/pih
```{
"dependency": "pih",
"version": "1.48036",
"result": {
"issues": 2,
"errors": {},
"results": {
"typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pid, pip",
"silent-process-execution": [
{
"location": "pih-1.48036/pih/tools.py:781",
"code": " result = subprocess.run(command, stdin=subprocess.DEVNULL, stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmphuwc_tvc/pih"
}
}```
|
process
|
pih has guarddog issues dependency pih version result issues errors results typosquatting this package closely ressembles the following package names and might be a typosquatting attempt pid pip silent process execution location pih pih tools py code result subprocess run command stdin subprocess devnull stdout subprocess devnull stderr subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp tmphuwc tvc pih
| 1
|
21,010
| 27,945,635,493
|
IssuesEvent
|
2023-03-24 02:39:48
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
QGIS crashes on MacOS 13.0 when using `fix geometries`
|
Processing Regression Bug Crash/Data Corruption macOS Linux
|
### What is the bug or the crash?
QGIS crashes the moment I use `fix geometries` on a set of polygons. To be specific, the vector data format that I am using is `.geojson` and `.shp`. I also tried to use `fix geometries` on a single polygon, but that also crashes. Currently I am using QGIS 3.28 on MacOS 13.0.
### Steps to reproduce the issue
The only step I followed is to load the `.geojson` or `.shp` file and use `fix geometries` in the processing toolbox.
### Versions
QGIS version | 3.28.0-Firenze | QGIS code revision | ed3ad0430f
-- | -- | -- | --
Qt version | 5.15.2
Python version | 3.9.5
GDAL/OGR version | 3.3.2
PROJ version | 8.1.1
EPSG Registry database version | v10.028 (2021-07-07)
GEOS version | 3.9.1-CAPI-1.14.2
SQLite version | 3.35.2
PDAL version | 2.3.0
PostgreSQL client version | unknown
SpatiaLite version | 5.0.1
QWT version | 6.1.6
QScintilla2 version | 2.11.5
OS version | macOS 13.0
| | |
Active Python plugins
quick_map_services | 0.19.32
processing | 2.12.99
sagaprovider | 2.12.99
grassprovider | 2.12.99
db_manager | 0.1.20
MetaSearch | 0.3.6
QGIS version
3.28.0-Firenze
QGIS code revision
[ed3ad0430f](https://github.com/qgis/QGIS/commit/ed3ad0430f)
Qt version
5.15.2
Python version
3.9.5
GDAL/OGR version
3.3.2
PROJ version
8.1.1
EPSG Registry database version
v10.028 (2021-07-07)
GEOS version
3.9.1-CAPI-1.14.2
SQLite version
3.35.2
PDAL version
2.3.0
PostgreSQL client version
unknown
SpatiaLite version
5.0.1
QWT version
6.1.6
QScintilla2 version
2.11.5
OS version
macOS 13.0
Active Python plugins
quick_map_services
0.19.32
processing
2.12.99
sagaprovider
2.12.99
grassprovider
2.12.99
db_manager
0.1.20
MetaSearch
0.3.6
### Supported QGIS version
- [ ] I'm running a supported QGIS version according to the roadmap.
### New profile
- [x] I tried with a new QGIS profile
### Additional context
_No response_
|
1.0
|
QGIS crashes on MacOS 13.0 when using `fix geometries` - ### What is the bug or the crash?
QGIS crashes the moment I use `fix geometries` on a set of polygons. To be specific, the vector data format that I am using is `.geojson` and `.shp`. I also tried to use `fix geometries` on a single polygon, but that also crashes. Currently I am using QGIS 3.28 on MacOS 13.0.
### Steps to reproduce the issue
The only step I followed is to load the `.geojson` or `.shp` file and use `fix geometries` in the processing toolbox.
### Versions
QGIS version | 3.28.0-Firenze | QGIS code revision | ed3ad0430f
-- | -- | -- | --
Qt version | 5.15.2
Python version | 3.9.5
GDAL/OGR version | 3.3.2
PROJ version | 8.1.1
EPSG Registry database version | v10.028 (2021-07-07)
GEOS version | 3.9.1-CAPI-1.14.2
SQLite version | 3.35.2
PDAL version | 2.3.0
PostgreSQL client version | unknown
SpatiaLite version | 5.0.1
QWT version | 6.1.6
QScintilla2 version | 2.11.5
OS version | macOS 13.0
| | |
Active Python plugins
quick_map_services | 0.19.32
processing | 2.12.99
sagaprovider | 2.12.99
grassprovider | 2.12.99
db_manager | 0.1.20
MetaSearch | 0.3.6
QGIS version
3.28.0-Firenze
QGIS code revision
[ed3ad0430f](https://github.com/qgis/QGIS/commit/ed3ad0430f)
Qt version
5.15.2
Python version
3.9.5
GDAL/OGR version
3.3.2
PROJ version
8.1.1
EPSG Registry database version
v10.028 (2021-07-07)
GEOS version
3.9.1-CAPI-1.14.2
SQLite version
3.35.2
PDAL version
2.3.0
PostgreSQL client version
unknown
SpatiaLite version
5.0.1
QWT version
6.1.6
QScintilla2 version
2.11.5
OS version
macOS 13.0
Active Python plugins
quick_map_services
0.19.32
processing
2.12.99
sagaprovider
2.12.99
grassprovider
2.12.99
db_manager
0.1.20
MetaSearch
0.3.6
### Supported QGIS version
- [ ] I'm running a supported QGIS version according to the roadmap.
### New profile
- [x] I tried with a new QGIS profile
### Additional context
_No response_
|
process
|
qgis crashes on macos when using fix geometries what is the bug or the crash qgis crashes the moment i use fix geometries on a set of polygons to be specific the vector data format that i am using is geojson and shp i also tried to use fix geometries on a single polygon but that also crashes currently i am using qgis on macos steps to reproduce the issue the only step i followed is to load the geojson or shp file and use fix geometries in the processing toolbox versions qgis version firenze qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version unknown spatialite version qwt version version os version macos active python plugins quick map services processing sagaprovider grassprovider db manager metasearch qgis version firenze qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version unknown spatialite version qwt version version os version macos active python plugins quick map services processing sagaprovider grassprovider db manager metasearch supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
| 1
|
56,537
| 11,594,338,228
|
IssuesEvent
|
2020-02-24 15:10:30
|
Regalis11/Barotrauma
|
https://api.github.com/repos/Regalis11/Barotrauma
|
reopened
|
Fulgurium Battery Cell is just a regular Battery Cell outside of Stun Baton
|
Code Design Feature request
|
- [x] I have searched the issue tracker to check if the issue has already been reported.
**Description**
Fulgurium Battery Cell is just a regular Battery Cell outside of Stun Baton. Perhaps higher life span or stronger electronic effects?
Handheld Sonar: Rapid pings?
Sonar Beacon: Further receiving range for Navigation Terminals and Handheld Sonars?
Underwater Scooter: Increased Speed?
Onboard/Shuttle Battery: Slower battery drain the more you put into them, or increased capacity?
Flashlight: Wider and Brighter ray? (I would put it on my head more if so)
Headset: Further transmission and receiving range.
**Version**
~0.9.703
|
1.0
|
Fulgurium Battery Cell is just a regular Battery Cell outside of Stun Baton - - [x] I have searched the issue tracker to check if the issue has already been reported.
**Description**
Fulgurium Battery Cell is just a regular Battery Cell outside of Stun Baton. Perhaps higher life span or stronger electronic effects?
Handheld Sonar: Rapid pings?
Sonar Beacon: Further receiving range for Navigation Terminals and Handheld Sonars?
Underwater Scooter: Increased Speed?
Onboard/Shuttle Battery: Slower battery drain the more you put into them, or increased capacity?
Flashlight: Wider and Brighter ray? (I would put it on my head more if so)
Headset: Further transmission and receiving range.
**Version**
~0.9.703
|
non_process
|
fulgurium battery cell is just a regular battery cell outside of stun baton i have searched the issue tracker to check if the issue has already been reported description fulgurium battery cell is just a regular battery cell outside of stun baton perhaps higher life span or stronger electronic effects handheld sonar rapid pings sonar beacon further receiving range for navigation terminals and handheld sonars underwater scooter increased speed onboard shuttle battery slower battery drain the more you put into them or increased capacity flashlight wider and brighter ray i would put it on my head more if so headset further transmission and receiving range version
| 0
|
539,607
| 15,791,817,667
|
IssuesEvent
|
2021-04-02 05:37:46
|
capitalone/DataProfiler
|
https://api.github.com/repos/capitalone/DataProfiler
|
closed
|
Multi-threaded column profiling
|
Medium Priority New Feature
|
**Is your feature request related to a problem? Please describe.**
Multithread the profiling of the columns:

|
1.0
|
Multi-threaded column profiling - **Is your feature request related to a problem? Please describe.**
Multithread the profiling of the columns:

|
non_process
|
multi threaded column profiling is your feature request related to a problem please describe multithread the profiling of the columns
| 0
|
1,487
| 4,059,115,678
|
IssuesEvent
|
2016-05-25 08:22:50
|
e-government-ua/iBP
|
https://api.github.com/repos/e-government-ua/iBP
|
closed
|
Черкаська область - Катеринопільський район - запит на отримання публічної інформації - розкриття
|
In process of testing
|
Координатор:
Колодич Олена - Координатор IGov в Черкаській області. тел.+380674704730,
**дуже велике прохання координатора** -
коли будь-які листи будуть надсилатися на контактних осіб (тестування, питання тощо)
ставити ії в копію elena.kolodich@privatbank.ua
та називати листи **IGov - район/місто -(назва послуги)**
по Катеринополю - основное контактирующее лицо - Шарига Ольга Александровна - +380672245259
Контактна особа:
Задвернюк Валентина Дмитрівна (0969783942); ekonomikakater@ukr.net
|
1.0
|
Черкаська область - Катеринопільський район - запит на отримання публічної інформації - розкриття - Координатор:
Колодич Олена - Координатор IGov в Черкаській області. тел.+380674704730,
**дуже велике прохання координатора** -
коли будь-які листи будуть надсилатися на контактних осіб (тестування, питання тощо)
ставити ії в копію elena.kolodich@privatbank.ua
та називати листи **IGov - район/місто -(назва послуги)**
по Катеринополю - основное контактирующее лицо - Шарига Ольга Александровна - +380672245259
Контактна особа:
Задвернюк Валентина Дмитрівна (0969783942); ekonomikakater@ukr.net
|
process
|
черкаська область катеринопільський район запит на отримання публічної інформації розкриття координатор колодич олена координатор igov в черкаській області тел дуже велике прохання координатора коли будь які листи будуть надсилатися на контактних осіб тестування питання тощо ставити ії в копію elena kolodich privatbank ua та називати листи igov район місто назва послуги по катеринополю основное контактирующее лицо шарига ольга александровна контактна особа задвернюк валентина дмитрівна ekonomikakater ukr net
| 1
|
12,827
| 15,211,828,193
|
IssuesEvent
|
2021-02-17 09:36:38
|
emacs-ess/ESS
|
https://api.github.com/repos/emacs-ess/ESS
|
closed
|
Start delimiter of `ess-command` is not stripped with multiline inputs
|
process:command
|
Commands containing new lines:
```elisp
(ess-command "{ 1\n2 }\n")
```
currently produce this output:
```r
+ ess-output-delimiter226-START
[1] 2
```
This makes me realise that another benefit of the delimiters is to reliably strip the `+` characters (created by the parser with multiline inputs) from the command outputs.
|
1.0
|
Start delimiter of `ess-command` is not stripped with multiline inputs - Commands containing new lines:
```elisp
(ess-command "{ 1\n2 }\n")
```
currently produce this output:
```r
+ ess-output-delimiter226-START
[1] 2
```
This makes me realise that another benefit of the delimiters is to reliably strip the `+` characters (created by the parser with multiline inputs) from the command outputs.
|
process
|
start delimiter of ess command is not stripped with multiline inputs commands containing new lines elisp ess command n currently produce this output r ess output start this makes me realise that another benefit of the delimiters is to reliably strip the characters created by the parser with multiline inputs from the command outputs
| 1
|
265,049
| 23,145,538,758
|
IssuesEvent
|
2022-07-29 00:04:11
|
MPMG-DCC-UFMG/F01
|
https://api.github.com/repos/MPMG-DCC-UFMG/F01
|
closed
|
Teste de generalizacao para a tag Servidores - Proventos de aposentadoria - Divisa Alegre
|
generalization test development template-Síntese tecnologia informatica subtag-Proventos de Aposentadoria tag-Servidores
|
DoD: Realizar o teste de Generalização do validador da tag Servidores - Proventos de aposentadoria para o Município de Divisa Alegre.
|
1.0
|
Teste de generalizacao para a tag Servidores - Proventos de aposentadoria - Divisa Alegre - DoD: Realizar o teste de Generalização do validador da tag Servidores - Proventos de aposentadoria para o Município de Divisa Alegre.
|
non_process
|
teste de generalizacao para a tag servidores proventos de aposentadoria divisa alegre dod realizar o teste de generalização do validador da tag servidores proventos de aposentadoria para o município de divisa alegre
| 0
|
421,379
| 28,314,652,465
|
IssuesEvent
|
2023-04-10 18:28:13
|
bounswe/bounswe2023group6
|
https://api.github.com/repos/bounswe/bounswe2023group6
|
closed
|
Forming Individual Contribution Report for Milestone report 1
|
type: documentation priority: high status: inprogress area: wiki area: milestone
|
### Problem
For the milestone report 1, everyone needs to prepare Individual Contribution Report which will include information about our contributions.
### Solution
I will prepare my individual contribution report as follow(which is given by the course):
* Member: Info about myself (name, group)
* Responsibilities: Overall description of the responsibilities that are assigned to me
* Main contributions: Overall description of my contributions to the project until Milestone. Provide the links to your work and related issues in your project repository,
* Management related significant issues: My issues that contribute to the management of our software project (Please only lists the most-significant ones, we DO track all your issues separately.)
* Additional information (optional): Further relevant information related to my contributions
### Documentation
Solution will be [here](https://github.com/bounswe/bounswe2023group6/wiki/Individual-Contribution-Report-Erkam-Kavak)
### Additional notes
_No response_
### Reviewers
Halis Ayberk Erdem
### Deadline
10.03.2023
|
1.0
|
Forming Individual Contribution Report for Milestone report 1 - ### Problem
For the milestone report 1, everyone needs to prepare Individual Contribution Report which will include information about our contributions.
### Solution
I will prepare my individual contribution report as follow(which is given by the course):
* Member: Info about myself (name, group)
* Responsibilities: Overall description of the responsibilities that are assigned to me
* Main contributions: Overall description of my contributions to the project until Milestone. Provide the links to your work and related issues in your project repository,
* Management related significant issues: My issues that contribute to the management of our software project (Please only lists the most-significant ones, we DO track all your issues separately.)
* Additional information (optional): Further relevant information related to my contributions
### Documentation
Solution will be [here](https://github.com/bounswe/bounswe2023group6/wiki/Individual-Contribution-Report-Erkam-Kavak)
### Additional notes
_No response_
### Reviewers
Halis Ayberk Erdem
### Deadline
10.03.2023
|
non_process
|
forming individual contribution report for milestone report problem for the milestone report everyone needs to prepare individual contribution report which will include information about our contributions solution i will prepare my individual contribution report as follow which is given by the course member info about myself name group responsibilities overall description of the responsibilities that are assigned to me main contributions overall description of my contributions to the project until milestone provide the links to your work and related issues in your project repository management related significant issues my issues that contribute to the management of our software project please only lists the most significant ones we do track all your issues separately additional information optional further relevant information related to my contributions documentation solution will be additional notes no response reviewers halis ayberk erdem deadline
| 0
|
218,293
| 7,331,044,006
|
IssuesEvent
|
2018-03-05 12:06:17
|
NCEAS/metacat
|
https://api.github.com/repos/NCEAS/metacat
|
closed
|
Metacat performance issue in Sanparks skin
|
Category: metacat Component: Bugzilla-Id Priority: Normal Status: Resolved Tracker: Bug
|
---
Author Name: **Jing Tao** (Jing Tao)
Original Redmine Issue: 3174, https://projects.ecoinformatics.org/ecoinfo/issues/3174
Original Date: 2008-03-13
Original Assignee: Jing Tao
---
Matt and Mike reported it would like about 4 or 5 minutes to do a search in sanparks skin of production server. We should fixed before 1.8.1 release.
|
1.0
|
Metacat performance issue in Sanparks skin - ---
Author Name: **Jing Tao** (Jing Tao)
Original Redmine Issue: 3174, https://projects.ecoinformatics.org/ecoinfo/issues/3174
Original Date: 2008-03-13
Original Assignee: Jing Tao
---
Matt and Mike reported it would like about 4 or 5 minutes to do a search in sanparks skin of production server. We should fixed before 1.8.1 release.
|
non_process
|
metacat performance issue in sanparks skin author name jing tao jing tao original redmine issue original date original assignee jing tao matt and mike reported it would like about or minutes to do a search in sanparks skin of production server we should fixed before release
| 0
|
100,752
| 21,510,202,327
|
IssuesEvent
|
2022-04-28 03:02:45
|
RobertsLab/resources
|
https://api.github.com/repos/RobertsLab/resources
|
opened
|
Problem with stringtie2 not matching gene names from gff file (TagSeq)
|
code
|
Hi! I am having a problem with generating a gene count matrix from some TagSeq data. It seems that I am having issues properly formatting the .gff3 file from the reference genome that I am using (Pocillopora acuta). Most of the genes in my count matrix are named as "STRG###", which indicates that stringtie2 did not find a gene "match" and is therefore calling that transcript as a splice variant. This is happening for ~80% of my genes, so I know this is an issue with the analysis and formatting.
Has anyone has this issue before? Is tehre a particular format of the gene information column in a .gff3 file that I should be using to make sure it can match the transcript to the proper gene id?
|
1.0
|
Problem with stringtie2 not matching gene names from gff file (TagSeq) - Hi! I am having a problem with generating a gene count matrix from some TagSeq data. It seems that I am having issues properly formatting the .gff3 file from the reference genome that I am using (Pocillopora acuta). Most of the genes in my count matrix are named as "STRG###", which indicates that stringtie2 did not find a gene "match" and is therefore calling that transcript as a splice variant. This is happening for ~80% of my genes, so I know this is an issue with the analysis and formatting.
Has anyone has this issue before? Is tehre a particular format of the gene information column in a .gff3 file that I should be using to make sure it can match the transcript to the proper gene id?
|
non_process
|
problem with not matching gene names from gff file tagseq hi i am having a problem with generating a gene count matrix from some tagseq data it seems that i am having issues properly formatting the file from the reference genome that i am using pocillopora acuta most of the genes in my count matrix are named as strg which indicates that did not find a gene match and is therefore calling that transcript as a splice variant this is happening for of my genes so i know this is an issue with the analysis and formatting has anyone has this issue before is tehre a particular format of the gene information column in a file that i should be using to make sure it can match the transcript to the proper gene id
| 0
|
9,757
| 12,740,432,436
|
IssuesEvent
|
2020-06-26 02:32:30
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
regression: data truncation with stream.writev in IPC channel
|
child_process help wanted stream
|
<!--
Thank you for reporting a possible bug in Node.js.
Please fill in as much of the template below as you can.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify the affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you can.
-->
* **Version**: all supported versions
* **Platform**: all UNIX variants
* **Subsystem**: stream, child_process
<!-- Please provide more details below this comment. -->
I was working on a potential solution for the `data truncation in child standard streams upon process exit` problem, and came across this scenario:
```js
const cp = require('child_process')
const KB = 1024
const MB = KB * KB
if (process.argv[2] === 'child') {
const data = Buffer.alloc(KB).fill('x').toString()
for (var i = 0; i < MB; i++)
console.log(data)
// process.exit(0)
} else {
const child = cp.spawn(process.execPath, [__filename, 'child'])
var count = 0
child.stdout.on('data', c => count += c.length)
child.stdout.on('end', () => console.log(count))
child.stderr.on('data', (d) => console.log(d.toString()))
}
```
- create a 1KB buffer in a spawned child.
- write it to the parent 1MB times.
- parent should get `1074790400` bytes (1GB +1MB new line chars that console.log adds)
actual data obtained in UNIX is much less, and is arbitrary in each run. Windows works fine.
The behavior is well understood in the presence of `process.exit` but in the absence of it, the process is supposed to drain all the data while there are active requests in the event loop. That is not happening here.
Tracing back, I landed at d85d1205932d583f72eefc849a3e655e26e5e305 which introduced vectored writes in pipe streams. With this in place, when the channel experiences congestion, it throws UV_ENOBUFS which is passed to the calling code unhandled:
```error
Error: write ENOBUFS
at exports._errnoException (util.js:1026:11)
at Socket._writeGeneric (net.js:715:26)
at Socket._writev (net.js:729:8)
at doWrite (_stream_writable.js:324:12)
at clearBuffer (_stream_writable.js:413:5)
at onwrite (_stream_writable.js:365:7)
at WriteWrap.afterWrite [as oncomplete] (net.js:816:12)
```
Further, commit 1d6b729cea909769ac0bcb3aa68f5fe567c4ffb7 inadvertently caused this error from not being visible.
Subsequently there are lot of changes in stream_base.cc and stream_wrap.cc around `DoTryWrite` and `Writev` which makes it difficult to pin-point any particular point where this can be rectified.
Looking at `Writev` implementation, it does not handle when the data size is more than `INT_MAX`. In my opinion this should be fixed, as vectored write is an internal mechanism that should hide itself from the JS API. Also, looks like it does not leverage libuv's `write-when-you-can` capability (`uv__write`)?
/cc @nodejs/streams @nodejs/child_process
Happy to do any further assistance / testing.
|
1.0
|
regression: data truncation with stream.writev in IPC channel - <!--
Thank you for reporting a possible bug in Node.js.
Please fill in as much of the template below as you can.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify the affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you can.
-->
* **Version**: all supported versions
* **Platform**: all UNIX variants
* **Subsystem**: stream, child_process
<!-- Please provide more details below this comment. -->
I was working on a potential solution for the `data truncation in child standard streams upon process exit` problem, and came across this scenario:
```js
const cp = require('child_process')
const KB = 1024
const MB = KB * KB
if (process.argv[2] === 'child') {
const data = Buffer.alloc(KB).fill('x').toString()
for (var i = 0; i < MB; i++)
console.log(data)
// process.exit(0)
} else {
const child = cp.spawn(process.execPath, [__filename, 'child'])
var count = 0
child.stdout.on('data', c => count += c.length)
child.stdout.on('end', () => console.log(count))
child.stderr.on('data', (d) => console.log(d.toString()))
}
```
- create a 1KB buffer in a spawned child.
- write it to the parent 1MB times.
- parent should get `1074790400` bytes (1GB +1MB new line chars that console.log adds)
actual data obtained in UNIX is much less, and is arbitrary in each run. Windows works fine.
The behavior is well understood in the presence of `process.exit` but in the absence of it, the process is supposed to drain all the data while there are active requests in the event loop. That is not happening here.
Tracing back, I landed at d85d1205932d583f72eefc849a3e655e26e5e305 which introduced vectored writes in pipe streams. With this in place, when the channel experiences congestion, it throws UV_ENOBUFS which is passed to the calling code unhandled:
```error
Error: write ENOBUFS
at exports._errnoException (util.js:1026:11)
at Socket._writeGeneric (net.js:715:26)
at Socket._writev (net.js:729:8)
at doWrite (_stream_writable.js:324:12)
at clearBuffer (_stream_writable.js:413:5)
at onwrite (_stream_writable.js:365:7)
at WriteWrap.afterWrite [as oncomplete] (net.js:816:12)
```
Further, commit 1d6b729cea909769ac0bcb3aa68f5fe567c4ffb7 inadvertently caused this error from not being visible.
Subsequently there are lot of changes in stream_base.cc and stream_wrap.cc around `DoTryWrite` and `Writev` which makes it difficult to pin-point any particular point where this can be rectified.
Looking at `Writev` implementation, it does not handle when the data size is more than `INT_MAX`. In my opinion this should be fixed, as vectored write is an internal mechanism that should hide itself from the JS API. Also, looks like it does not leverage libuv's `write-when-you-can` capability (`uv__write`)?
/cc @nodejs/streams @nodejs/child_process
Happy to do any further assistance / testing.
|
process
|
regression data truncation with stream writev in ipc channel thank you for reporting a possible bug in node js please fill in as much of the template below as you can version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify the affected core module name if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you can version all supported versions platform all unix variants subsystem stream child process i was working on a potential solution for the data truncation in child standard streams upon process exit problem and came across this scenario js const cp require child process const kb const mb kb kb if process argv child const data buffer alloc kb fill x tostring for var i i mb i console log data process exit else const child cp spawn process execpath var count child stdout on data c count c length child stdout on end console log count child stderr on data d console log d tostring create a buffer in a spawned child write it to the parent times parent should get bytes new line chars that console log adds actual data obtained in unix is much less and is arbitrary in each run windows works fine the behavior is well understood in the presence of process exit but in the absence of it the process is supposed to drain all the data while there are active requests in the event loop that is not happening here tracing back i landed at which introduced vectored writes in pipe streams with this in place when the channel experiences congestion it throws uv enobufs which is passed to the calling code unhandled error error write enobufs at exports errnoexception util js at socket writegeneric net js at socket writev net js at dowrite stream writable js at clearbuffer stream writable js at onwrite stream writable js at writewrap afterwrite net js further commit inadvertently caused this error from not being visible subsequently there are lot of changes in stream base cc and stream wrap cc around dotrywrite and writev which makes it difficult to pin point any particular point where this can be rectified looking at writev implementation it does not handle when the data size is more than int max in my opinion this should be fixed as vectored write is an internal mechanism that should hide itself from the js api also looks like it does not leverage libuv s write when you can capability uv write cc nodejs streams nodejs child process happy to do any further assistance testing
| 1
|
100,676
| 30,753,161,433
|
IssuesEvent
|
2023-07-28 21:32:39
|
NixOS/nixpkgs
|
https://api.github.com/repos/NixOS/nixpkgs
|
closed
|
Build failure: keychain on 23.11 unstable
|
0.kind: build failure
|
### Steps To Reproduce
Steps to reproduce the behavior:
1. build a 23.11 unstable derivation (flake-enabled in configuration.nix, but not using any flakes yet)
2. `nix-env -iA nixos.keychain`
Edit: I updated both personal and root nix-channel to nixos-unstable and rebuilt a new derivation with keychain included in configuration.nix, and that completes without error.
### Build log
```
> nix-env -iA nixos.keychain
warning: name collision in input Nix expressions, skipping '/home/bgibson/.nix-defexpr/channels_root/nixos'
suggestion: remove 'nixos' from either the root channels or the user channels
installing 'keychain-2.8.5'
these 4 paths will be fetched (1.86 MiB download, 10.93 MiB unpacked):
/nix/store/4qrbfa42xyicny7vlgbbqxz07kcsjbx2-findutils-4.9.0
/nix/store/r2p66d3sh6syxiq7gxv8l4h7d2y733jr-keychain-2.8.5
/nix/store/s1xnsm86dy195fr4i4n984ivakrvjv9d-openssh-9.3p1
/nix/store/pi05gf5p4p1p1j41p12sha27zgjfs767-procps-3.3.17
copying path '/nix/store/4qrbfa42xyicny7vlgbbqxz07kcsjbx2-findutils-4.9.0' from 'https://cache.nixos.org'...
copying path '/nix/store/s1xnsm86dy195fr4i4n984ivakrvjv9d-openssh-9.3p1' from 'https://cache.nixos.org'...
copying path '/nix/store/pi05gf5p4p1p1j41p12sha27zgjfs767-procps-3.3.17' from 'https://cache.nixos.org'...
copying path '/nix/store/r2p66d3sh6syxiq7gxv8l4h7d2y733jr-keychain-2.8.5' from 'https://cache.nixos.org'...
error: this derivation has bad 'meta.outputsToInstall'
```
### Additional context
Installing in nix-shell works:
```
> nix-shell -p keychain
this path will be fetched (0.02 MiB download, 0.06 MiB unpacked):
/nix/store/vl6w8gzd2q1fvbkch9wpnjwwd370vq64-keychain-2.8.5
copying path '/nix/store/vl6w8gzd2q1fvbkch9wpnjwwd370vq64-keychain-2.8.5' from 'https://cache.nixos.org'...
[nix-shell:~]$ keychain --help
* keychain 2.8.5 ~ http://www.funtoo.org
```
### Notify maintainers
@SuperSandro2000
(@sigma listed on [search.nixos.org](https://search.nixos.org/packages?channel=unstable&show=keychain&from=0&size=50&sort=relevance&type=packages&query=keychain) as maintainer, but seems inactive on this repo)
### Metadata
Please run `nix-shell -p nix-info --run "nix-info -m"` and paste the result.
```console
> nix-shell -p nix-info --run "nix-info -m"
- system: `"x86_64-linux"`
- host os: `Linux 6.3.13, NixOS, 23.11 (Tapir), 23.11pre507929.c9cf0708f00`
- multi-user?: `yes`
- sandbox: `yes`
- version: `nix-env (Nix) 2.15.1`
- channels(root): `"nixos"`
- channels(bgibson): `"nixos-23.05"`
- nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos`
```
Not sure why root channel is showing `nixos`, but...
```
> nix-channel --list
nixos https://channels.nixos.org/nixos-23.05
> sudo nix-channel --list
[sudo] password:
nixos https://nixos.org/channels/nixos-unstable
```
|
1.0
|
Build failure: keychain on 23.11 unstable - ### Steps To Reproduce
Steps to reproduce the behavior:
1. build a 23.11 unstable derivation (flake-enabled in configuration.nix, but not using any flakes yet)
2. `nix-env -iA nixos.keychain`
Edit: I updated both personal and root nix-channel to nixos-unstable and rebuilt a new derivation with keychain included in configuration.nix, and that completes without error.
### Build log
```
> nix-env -iA nixos.keychain
warning: name collision in input Nix expressions, skipping '/home/bgibson/.nix-defexpr/channels_root/nixos'
suggestion: remove 'nixos' from either the root channels or the user channels
installing 'keychain-2.8.5'
these 4 paths will be fetched (1.86 MiB download, 10.93 MiB unpacked):
/nix/store/4qrbfa42xyicny7vlgbbqxz07kcsjbx2-findutils-4.9.0
/nix/store/r2p66d3sh6syxiq7gxv8l4h7d2y733jr-keychain-2.8.5
/nix/store/s1xnsm86dy195fr4i4n984ivakrvjv9d-openssh-9.3p1
/nix/store/pi05gf5p4p1p1j41p12sha27zgjfs767-procps-3.3.17
copying path '/nix/store/4qrbfa42xyicny7vlgbbqxz07kcsjbx2-findutils-4.9.0' from 'https://cache.nixos.org'...
copying path '/nix/store/s1xnsm86dy195fr4i4n984ivakrvjv9d-openssh-9.3p1' from 'https://cache.nixos.org'...
copying path '/nix/store/pi05gf5p4p1p1j41p12sha27zgjfs767-procps-3.3.17' from 'https://cache.nixos.org'...
copying path '/nix/store/r2p66d3sh6syxiq7gxv8l4h7d2y733jr-keychain-2.8.5' from 'https://cache.nixos.org'...
error: this derivation has bad 'meta.outputsToInstall'
```
### Additional context
Installing in nix-shell works:
```
> nix-shell -p keychain
this path will be fetched (0.02 MiB download, 0.06 MiB unpacked):
/nix/store/vl6w8gzd2q1fvbkch9wpnjwwd370vq64-keychain-2.8.5
copying path '/nix/store/vl6w8gzd2q1fvbkch9wpnjwwd370vq64-keychain-2.8.5' from 'https://cache.nixos.org'...
[nix-shell:~]$ keychain --help
* keychain 2.8.5 ~ http://www.funtoo.org
```
### Notify maintainers
@SuperSandro2000
(@sigma listed on [search.nixos.org](https://search.nixos.org/packages?channel=unstable&show=keychain&from=0&size=50&sort=relevance&type=packages&query=keychain) as maintainer, but seems inactive on this repo)
### Metadata
Please run `nix-shell -p nix-info --run "nix-info -m"` and paste the result.
```console
> nix-shell -p nix-info --run "nix-info -m"
- system: `"x86_64-linux"`
- host os: `Linux 6.3.13, NixOS, 23.11 (Tapir), 23.11pre507929.c9cf0708f00`
- multi-user?: `yes`
- sandbox: `yes`
- version: `nix-env (Nix) 2.15.1`
- channels(root): `"nixos"`
- channels(bgibson): `"nixos-23.05"`
- nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos`
```
Not sure why root channel is showing `nixos`, but...
```
> nix-channel --list
nixos https://channels.nixos.org/nixos-23.05
> sudo nix-channel --list
[sudo] password:
nixos https://nixos.org/channels/nixos-unstable
```
|
non_process
|
build failure keychain on unstable steps to reproduce steps to reproduce the behavior build a unstable derivation flake enabled in configuration nix but not using any flakes yet nix env ia nixos keychain edit i updated both personal and root nix channel to nixos unstable and rebuilt a new derivation with keychain included in configuration nix and that completes without error build log nix env ia nixos keychain warning name collision in input nix expressions skipping home bgibson nix defexpr channels root nixos suggestion remove nixos from either the root channels or the user channels installing keychain these paths will be fetched mib download mib unpacked nix store findutils nix store keychain nix store openssh nix store procps copying path nix store findutils from copying path nix store openssh from copying path nix store procps from copying path nix store keychain from error this derivation has bad meta outputstoinstall additional context installing in nix shell works nix shell p keychain this path will be fetched mib download mib unpacked nix store keychain copying path nix store keychain from keychain help keychain notify maintainers sigma listed on as maintainer but seems inactive on this repo metadata please run nix shell p nix info run nix info m and paste the result console nix shell p nix info run nix info m system linux host os linux nixos tapir multi user yes sandbox yes version nix env nix channels root nixos channels bgibson nixos nixpkgs nix var nix profiles per user root channels nixos not sure why root channel is showing nixos but nix channel list nixos sudo nix channel list password nixos
| 0
|
943
| 3,409,566,933
|
IssuesEvent
|
2015-12-04 16:16:46
|
hammerlab/pileup.js
|
https://api.github.com/repos/hammerlab/pileup.js
|
closed
|
Move modules into packages
|
process
|
Source tree should probably look something like:
```
src/
formats/
viz/
components/
util/
```
|
1.0
|
Move modules into packages - Source tree should probably look something like:
```
src/
formats/
viz/
components/
util/
```
|
process
|
move modules into packages source tree should probably look something like src formats viz components util
| 1
|
847
| 2,517,144,033
|
IssuesEvent
|
2015-01-16 12:10:42
|
ajency/Foodstree
|
https://api.github.com/repos/ajency/Foodstree
|
closed
|
Subtitle for grouping seller's information is required on edit seller page
|
bug Pushed to test site
|
Steps:
1.Login as admin and click on seller
2. Edit any of the seller
Current behaviour: The seller information currently is shown below the subtitle 'Pincode Info'
Expected Behaviour: A subtitle is required to group the seller's info fields

|
1.0
|
Subtitle for grouping seller's information is required on edit seller page - Steps:
1.Login as admin and click on seller
2. Edit any of the seller
Current behaviour: The seller information currently is shown below the subtitle 'Pincode Info'
Expected Behaviour: A subtitle is required to group the seller's info fields

|
non_process
|
subtitle for grouping seller s information is required on edit seller page steps login as admin and click on seller edit any of the seller current behaviour the seller information currently is shown below the subtitle pincode info expected behaviour a subtitle is required to group the seller s info fields
| 0
|
14,527
| 17,630,649,452
|
IssuesEvent
|
2021-08-19 07:30:47
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
Add Tiger Fist
|
suggested title in process
|
Please add as much of the following info as you can:
Title: Tiger Fist
Type (film/tv show): TV Show
Film or show in which it appears: _Gravity Falls_
Is the parent film/show streaming anywhere? [Disney+](https://www.disneyplus.com/series/gravity-falls/HZxayxzMJqed)
About when in the parent film/show does it appear? Season 1, Episode 4, ~15sec
Actual footage of the film/show can be seen (yes/no)? Yes






|
1.0
|
Add Tiger Fist - Please add as much of the following info as you can:
Title: Tiger Fist
Type (film/tv show): TV Show
Film or show in which it appears: _Gravity Falls_
Is the parent film/show streaming anywhere? [Disney+](https://www.disneyplus.com/series/gravity-falls/HZxayxzMJqed)
About when in the parent film/show does it appear? Season 1, Episode 4, ~15sec
Actual footage of the film/show can be seen (yes/no)? Yes






|
process
|
add tiger fist please add as much of the following info as you can title tiger fist type film tv show tv show film or show in which it appears gravity falls is the parent film show streaming anywhere about when in the parent film show does it appear season episode actual footage of the film show can be seen yes no yes
| 1
|
470,265
| 13,534,722,009
|
IssuesEvent
|
2020-09-16 06:17:54
|
pingcap/tidb-lightning
|
https://api.github.com/repos/pingcap/tidb-lightning
|
closed
|
Support importing data in Apache Parquet format
|
5.0 request difficulty/3-hard feature/accepted priority/P0
|
# Description
https://parquet.apache.org/
[AWS Aurora exported snapshot](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/USER_ExportSnapshot.html#USER_ExportSnapshot.data-types) are encoded in Parquet format. We should investigate how to restore from this serialization to allow quick Aurora → TiDB data migration.
# Category
* Feature
# Value
## Value description
(TBD)
## Value score
* (TBD) / 5
## Workload estimation
* (TBD) person-day
|
1.0
|
Support importing data in Apache Parquet format - # Description
https://parquet.apache.org/
[AWS Aurora exported snapshot](https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/USER_ExportSnapshot.html#USER_ExportSnapshot.data-types) are encoded in Parquet format. We should investigate how to restore from this serialization to allow quick Aurora → TiDB data migration.
# Category
* Feature
# Value
## Value description
(TBD)
## Value score
* (TBD) / 5
## Workload estimation
* (TBD) person-day
|
non_process
|
support importing data in apache parquet format description are encoded in parquet format we should investigate how to restore from this serialization to allow quick aurora → tidb data migration category feature value value description tbd value score tbd workload estimation tbd person day
| 0
|
87,430
| 15,774,419,618
|
IssuesEvent
|
2021-04-01 01:02:04
|
RG4421/spark-tpcds-benchmark
|
https://api.github.com/repos/RG4421/spark-tpcds-benchmark
|
opened
|
CVE-2021-21409 (Medium) detected in netty-all-4.1.50.Final.jar
|
security vulnerability
|
## CVE-2021-21409 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-all-4.1.50.Final.jar</b></p></summary>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: spark-tpcds-benchmark/spark-tpcds-benchmark-runner/build.gradle</p>
<p>Path to vulnerable library: 20210226193317_RPTMIF/downloadResource_EEVKWP/20210226193449/netty-all-4.1.50.Final.jar,/tmp/ws-ua_20210226193317_RPTMIF/downloadResource_EEVKWP/20210226193449/netty-all-4.1.50.Final.jar</p>
<p>
Dependency Hierarchy:
- :x: **netty-all-4.1.50.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers & clients. In Netty (io.netty:netty-codec-http2) before version 4.1.61.Final there is a vulnerability that enables request smuggling. The content-length header is not correctly validated if the request only uses a single Http2HeaderFrame with the endStream set to to true. This could lead to request smuggling if the request is proxied to a remote peer and translated to HTTP/1.1. This is a followup of GHSA-wm47-8v5p-wjpj/CVE-2021-21295 which did miss to fix this one case. This was fixed as part of 4.1.61.Final.
<p>Publish Date: 2021-03-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21409>CVE-2021-21409</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/netty/netty/security/advisories/GHSA-f256-j965-7f32">https://github.com/netty/netty/security/advisories/GHSA-f256-j965-7f32</a></p>
<p>Release Date: 2021-03-30</p>
<p>Fix Resolution: io.netty:netty-codec-http2:4.1.61.Final</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.netty","packageName":"netty-all","packageVersion":"4.1.50.Final","packageFilePaths":["/spark-tpcds-benchmark-runner/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"io.netty:netty-all:4.1.50.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.netty:netty-codec-http2:4.1.61.Final"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2021-21409","vulnerabilityDetails":"Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers \u0026 clients. In Netty (io.netty:netty-codec-http2) before version 4.1.61.Final there is a vulnerability that enables request smuggling. The content-length header is not correctly validated if the request only uses a single Http2HeaderFrame with the endStream set to to true. This could lead to request smuggling if the request is proxied to a remote peer and translated to HTTP/1.1. This is a followup of GHSA-wm47-8v5p-wjpj/CVE-2021-21295 which did miss to fix this one case. This was fixed as part of 4.1.61.Final.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21409","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-21409 (Medium) detected in netty-all-4.1.50.Final.jar - ## CVE-2021-21409 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-all-4.1.50.Final.jar</b></p></summary>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: spark-tpcds-benchmark/spark-tpcds-benchmark-runner/build.gradle</p>
<p>Path to vulnerable library: 20210226193317_RPTMIF/downloadResource_EEVKWP/20210226193449/netty-all-4.1.50.Final.jar,/tmp/ws-ua_20210226193317_RPTMIF/downloadResource_EEVKWP/20210226193449/netty-all-4.1.50.Final.jar</p>
<p>
Dependency Hierarchy:
- :x: **netty-all-4.1.50.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers & clients. In Netty (io.netty:netty-codec-http2) before version 4.1.61.Final there is a vulnerability that enables request smuggling. The content-length header is not correctly validated if the request only uses a single Http2HeaderFrame with the endStream set to to true. This could lead to request smuggling if the request is proxied to a remote peer and translated to HTTP/1.1. This is a followup of GHSA-wm47-8v5p-wjpj/CVE-2021-21295 which did miss to fix this one case. This was fixed as part of 4.1.61.Final.
<p>Publish Date: 2021-03-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21409>CVE-2021-21409</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/netty/netty/security/advisories/GHSA-f256-j965-7f32">https://github.com/netty/netty/security/advisories/GHSA-f256-j965-7f32</a></p>
<p>Release Date: 2021-03-30</p>
<p>Fix Resolution: io.netty:netty-codec-http2:4.1.61.Final</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.netty","packageName":"netty-all","packageVersion":"4.1.50.Final","packageFilePaths":["/spark-tpcds-benchmark-runner/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"io.netty:netty-all:4.1.50.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.netty:netty-codec-http2:4.1.61.Final"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2021-21409","vulnerabilityDetails":"Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers \u0026 clients. In Netty (io.netty:netty-codec-http2) before version 4.1.61.Final there is a vulnerability that enables request smuggling. The content-length header is not correctly validated if the request only uses a single Http2HeaderFrame with the endStream set to to true. This could lead to request smuggling if the request is proxied to a remote peer and translated to HTTP/1.1. This is a followup of GHSA-wm47-8v5p-wjpj/CVE-2021-21295 which did miss to fix this one case. This was fixed as part of 4.1.61.Final.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21409","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve medium detected in netty all final jar cve medium severity vulnerability vulnerable library netty all final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients path to dependency file spark tpcds benchmark spark tpcds benchmark runner build gradle path to vulnerable library rptmif downloadresource eevkwp netty all final jar tmp ws ua rptmif downloadresource eevkwp netty all final jar dependency hierarchy x netty all final jar vulnerable library found in base branch develop vulnerability details netty is an open source asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers clients in netty io netty netty codec before version final there is a vulnerability that enables request smuggling the content length header is not correctly validated if the request only uses a single with the endstream set to to true this could lead to request smuggling if the request is proxied to a remote peer and translated to http this is a followup of ghsa wjpj cve which did miss to fix this one case this was fixed as part of final publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty codec final check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree io netty netty all final isminimumfixversionavailable true minimumfixversion io netty netty codec final basebranches vulnerabilityidentifier cve vulnerabilitydetails netty is an open source asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers clients in netty io netty netty codec before version final there is a vulnerability that enables request smuggling the content length header is not correctly validated if the request only uses a single with the endstream set to to true this could lead to request smuggling if the request is proxied to a remote peer and translated to http this is a followup of ghsa wjpj cve which did miss to fix this one case this was fixed as part of final vulnerabilityurl
| 0
|
50,224
| 12,484,695,198
|
IssuesEvent
|
2020-05-30 15:59:51
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
Return to signing jars?
|
.Building & Releasing
|
We dropped signed jars in a recent release, mainly because of how long it takes to verify the jar and the fact that it seems like no one actually verifies them in production.
We should probably at least provide a checksum however.
Does anyone feel strongly about signed jars?
|
1.0
|
Return to signing jars? - We dropped signed jars in a recent release, mainly because of how long it takes to verify the jar and the fact that it seems like no one actually verifies them in production.
We should probably at least provide a checksum however.
Does anyone feel strongly about signed jars?
|
non_process
|
return to signing jars we dropped signed jars in a recent release mainly because of how long it takes to verify the jar and the fact that it seems like no one actually verifies them in production we should probably at least provide a checksum however does anyone feel strongly about signed jars
| 0
|
7,978
| 11,167,790,185
|
IssuesEvent
|
2019-12-27 18:41:34
|
openopps/openopps-platform
|
https://api.github.com/repos/openopps/openopps-platform
|
closed
|
Updates to submit modal on opportunity application
|
Apply Process Approved Requirements Ready
|
Who: opportunity applicants
What: let them know if they have not updated their biography or skills
Why: so more applicants will input biography data
Acceptance Criteria:
Current Screen Shot

For the last bullet that currently reads, "I've updated my profile, including my biography and skills (optional, but recommended).
1) The system should check against the USAJOBS biography field and the Open Opps skills and flag if either of them has not been completed (at least one skill)
2) If both are complete, display the green check
3) if either isn't complete, display a yellow warning triangle to indicate. However, they can still submit as normal and this display will not prevent them from applying
4) modify the language to read, "I've updated my profile, including my biography and/or skills (optional, but recommended).
|
1.0
|
Updates to submit modal on opportunity application - Who: opportunity applicants
What: let them know if they have not updated their biography or skills
Why: so more applicants will input biography data
Acceptance Criteria:
Current Screen Shot

For the last bullet that currently reads, "I've updated my profile, including my biography and skills (optional, but recommended).
1) The system should check against the USAJOBS biography field and the Open Opps skills and flag if either of them has not been completed (at least one skill)
2) If both are complete, display the green check
3) if either isn't complete, display a yellow warning triangle to indicate. However, they can still submit as normal and this display will not prevent them from applying
4) modify the language to read, "I've updated my profile, including my biography and/or skills (optional, but recommended).
|
process
|
updates to submit modal on opportunity application who opportunity applicants what let them know if they have not updated their biography or skills why so more applicants will input biography data acceptance criteria current screen shot for the last bullet that currently reads i ve updated my profile including my biography and skills optional but recommended the system should check against the usajobs biography field and the open opps skills and flag if either of them has not been completed at least one skill if both are complete display the green check if either isn t complete display a yellow warning triangle to indicate however they can still submit as normal and this display will not prevent them from applying modify the language to read i ve updated my profile including my biography and or skills optional but recommended
| 1
|
8,011
| 11,202,738,651
|
IssuesEvent
|
2020-01-04 14:51:07
|
allinurl/goaccess
|
https://api.github.com/repos/allinurl/goaccess
|
closed
|
No real-time output when running goaccess locally from a remote access log
|
log-processing websocket-server
|
Sorry to bother you. I do appreciate the tremendous amount of work and talent you're putting into this. I have the following problem.
I have installed GoAccess on a local machine. I'm trying to monitor the logs on a remote shared server over ssh. I'm doing this:
ssh user@ssh.example.com cat '/remoteserver/logs/access.log' | goaccess -p /usr/local/etc/goaccess.conf -o /localserver/Documents/report.html --real-time.html
This produces a full html output, however, it doesn't refresh.
Lsof -i tells me goacces listens to port 7890.
I have disabled the firewall on my router.
There are no errors in the browser console.
I know there have been many questions about realtime reporting, and I have read them all, but I can't get this to work. Once again, sorry to bother you.
|
1.0
|
No real-time output when running goaccess locally from a remote access log - Sorry to bother you. I do appreciate the tremendous amount of work and talent you're putting into this. I have the following problem.
I have installed GoAccess on a local machine. I'm trying to monitor the logs on a remote shared server over ssh. I'm doing this:
ssh user@ssh.example.com cat '/remoteserver/logs/access.log' | goaccess -p /usr/local/etc/goaccess.conf -o /localserver/Documents/report.html --real-time.html
This produces a full html output, however, it doesn't refresh.
Lsof -i tells me goacces listens to port 7890.
I have disabled the firewall on my router.
There are no errors in the browser console.
I know there have been many questions about realtime reporting, and I have read them all, but I can't get this to work. Once again, sorry to bother you.
|
process
|
no real time output when running goaccess locally from a remote access log sorry to bother you i do appreciate the tremendous amount of work and talent you re putting into this i have the following problem i have installed goaccess on a local machine i m trying to monitor the logs on a remote shared server over ssh i m doing this ssh user ssh example com cat remoteserver logs access log goaccess p usr local etc goaccess conf o localserver documents report html real time html this produces a full html output however it doesn t refresh lsof i tells me goacces listens to port i have disabled the firewall on my router there are no errors in the browser console i know there have been many questions about realtime reporting and i have read them all but i can t get this to work once again sorry to bother you
| 1
|
6,289
| 9,299,234,556
|
IssuesEvent
|
2019-03-23 01:16:42
|
shirou/gopsutil
|
https://api.github.com/repos/shirou/gopsutil
|
closed
|
[Darwin][process_darwin.go] convertCPUTimes does not handle cpu time string more than 1 hour.
|
os:darwin package:process
|
**Describe the bug**
Function convertCPUTimes only handle cpu time string within 1 hour range.
if the string is xx:xx.xx format then the function work great.
But if the string is xx:xx:xx.xx format then the function will fail.
Mac process would exceed hour time very often.

**To Reproduce**
```go
// +build darwin
package main
import "fmt"
import (
"strconv"
"strings"
)
const (
ClockTicks = 100 // C.sysconf(C._SC_CLK_TCK)
)
func convertCPUTimes(s string) (ret float64, err error) {
var t int
var _tmp string
if strings.Contains(s, ":") {
_t := strings.Split(s, ":")
hour, err := strconv.Atoi(_t[0])
if err != nil {
return ret, err
}
t += hour * 60 * 100
_tmp = _t[1]
} else {
_tmp = s
}
_t := strings.Split(_tmp, ".")
if err != nil {
return ret, err
}
h, err := strconv.Atoi(_t[0])
t += h * 100
h, err = strconv.Atoi(_t[1])
t += h
return float64(t) / ClockTicks, nil
}
func main() {
timeVal, err := convertCPUTimes("01:59:23.20")
if err != nil {
fmt.Printf("err=%s\n", err)
return
}
fmt.Printf("utime=%f\n", timeVal)
}
```
**Expected behavior**
7163.20000 should be the correct output value for the demo code.
**Environment (please complete the following information):**
- [ ] Windows: [paste the result of `ver`]
- [ ] Linux: [paste contents of `/etc/os-release` and the result of `uname -a`]
- [X] Mac OS: [paste the result of `sw_vers` and `uname -a`
- [ ] FreeBSD: [paste the result of `freebsd-version -k -r -u` and `uname -a`]
- [ ] OpenBSD: [paste the result of `uname -a`]
**Additional context**
[Cross-compiling? Paste the command you are using to cross-compile and the result of the corresponding `go env`]
|
1.0
|
[Darwin][process_darwin.go] convertCPUTimes does not handle cpu time string more than 1 hour. - **Describe the bug**
Function convertCPUTimes only handle cpu time string within 1 hour range.
if the string is xx:xx.xx format then the function work great.
But if the string is xx:xx:xx.xx format then the function will fail.
Mac process would exceed hour time very often.

**To Reproduce**
```go
// +build darwin
package main
import "fmt"
import (
"strconv"
"strings"
)
const (
ClockTicks = 100 // C.sysconf(C._SC_CLK_TCK)
)
func convertCPUTimes(s string) (ret float64, err error) {
var t int
var _tmp string
if strings.Contains(s, ":") {
_t := strings.Split(s, ":")
hour, err := strconv.Atoi(_t[0])
if err != nil {
return ret, err
}
t += hour * 60 * 100
_tmp = _t[1]
} else {
_tmp = s
}
_t := strings.Split(_tmp, ".")
if err != nil {
return ret, err
}
h, err := strconv.Atoi(_t[0])
t += h * 100
h, err = strconv.Atoi(_t[1])
t += h
return float64(t) / ClockTicks, nil
}
func main() {
timeVal, err := convertCPUTimes("01:59:23.20")
if err != nil {
fmt.Printf("err=%s\n", err)
return
}
fmt.Printf("utime=%f\n", timeVal)
}
```
**Expected behavior**
7163.20000 should be the correct output value for the demo code.
**Environment (please complete the following information):**
- [ ] Windows: [paste the result of `ver`]
- [ ] Linux: [paste contents of `/etc/os-release` and the result of `uname -a`]
- [X] Mac OS: [paste the result of `sw_vers` and `uname -a`
- [ ] FreeBSD: [paste the result of `freebsd-version -k -r -u` and `uname -a`]
- [ ] OpenBSD: [paste the result of `uname -a`]
**Additional context**
[Cross-compiling? Paste the command you are using to cross-compile and the result of the corresponding `go env`]
|
process
|
convertcputimes does not handle cpu time string more than hour describe the bug function convertcputimes only handle cpu time string within hour range if the string is xx xx xx format then the function work great but if the string is xx xx xx xx format then the function will fail mac process would exceed hour time very often to reproduce go build darwin package main import fmt import strconv strings const clockticks c sysconf c sc clk tck func convertcputimes s string ret err error var t int var tmp string if strings contains s t strings split s hour err strconv atoi t if err nil return ret err t hour tmp t else tmp s t strings split tmp if err nil return ret err h err strconv atoi t t h h err strconv atoi t t h return t clockticks nil func main timeval err convertcputimes if err nil fmt printf err s n err return fmt printf utime f n timeval expected behavior should be the correct output value for the demo code environment please complete the following information windows linux mac os paste the result of sw vers and uname a freebsd openbsd additional context
| 1
|
660,382
| 21,964,159,483
|
IssuesEvent
|
2022-05-24 18:28:59
|
ramp4-pcar4/ramp4-pcar4
|
https://api.github.com/repos/ramp4-pcar4/ramp4-pcar4
|
closed
|
Custom Layer Sample
|
effort: far away flavour: docs flavour: feature priority: low type: perfective
|
The layers architecture has been restructured to support "custom layers" that can be added to R4MP, similar to how a custom `fixture` can be added. It has not yet been tested / proven to work.
This is probably not a must for 1.0 release.
## Code Updates
In `geo/layer/layers.ts`, function `updateBaseToInstance` needs to be fleshed out to port all the possible things from the `LayerBase` interface to the result. See the similar function in `api/fixture.ts` to see how this is done.
Also, after a successful test, look into making some of the `LayerBase` items optional (i.e. begin with a `?`), mainly the stuff used by layers that support attributes. It could be forcing an author to define dud functions that do nothing is not the best approach.
## Test Custom Layer
Create an external layer class that satisfies the `LayerBase` signatures and has the internal plumbing to support an ESRI layer. E.g. a basic FeatureLayer that just has one hardcoded point, one hardcoded field, etc. So any config with the custom layertype will load a layer with that one point.
Using custom loading script, add the new layer definition, then provide a config with a layer that has the custom layer type.
## Document Custom Layer
After successful run of the test layer, update the last section of `docs/geo/layer.md` with instructions on how to create a custom layer.
|
1.0
|
Custom Layer Sample - The layers architecture has been restructured to support "custom layers" that can be added to R4MP, similar to how a custom `fixture` can be added. It has not yet been tested / proven to work.
This is probably not a must for 1.0 release.
## Code Updates
In `geo/layer/layers.ts`, function `updateBaseToInstance` needs to be fleshed out to port all the possible things from the `LayerBase` interface to the result. See the similar function in `api/fixture.ts` to see how this is done.
Also, after a successful test, look into making some of the `LayerBase` items optional (i.e. begin with a `?`), mainly the stuff used by layers that support attributes. It could be forcing an author to define dud functions that do nothing is not the best approach.
## Test Custom Layer
Create an external layer class that satisfies the `LayerBase` signatures and has the internal plumbing to support an ESRI layer. E.g. a basic FeatureLayer that just has one hardcoded point, one hardcoded field, etc. So any config with the custom layertype will load a layer with that one point.
Using custom loading script, add the new layer definition, then provide a config with a layer that has the custom layer type.
## Document Custom Layer
After successful run of the test layer, update the last section of `docs/geo/layer.md` with instructions on how to create a custom layer.
|
non_process
|
custom layer sample the layers architecture has been restructured to support custom layers that can be added to similar to how a custom fixture can be added it has not yet been tested proven to work this is probably not a must for release code updates in geo layer layers ts function updatebasetoinstance needs to be fleshed out to port all the possible things from the layerbase interface to the result see the similar function in api fixture ts to see how this is done also after a successful test look into making some of the layerbase items optional i e begin with a mainly the stuff used by layers that support attributes it could be forcing an author to define dud functions that do nothing is not the best approach test custom layer create an external layer class that satisfies the layerbase signatures and has the internal plumbing to support an esri layer e g a basic featurelayer that just has one hardcoded point one hardcoded field etc so any config with the custom layertype will load a layer with that one point using custom loading script add the new layer definition then provide a config with a layer that has the custom layer type document custom layer after successful run of the test layer update the last section of docs geo layer md with instructions on how to create a custom layer
| 0
|
727
| 3,213,371,430
|
IssuesEvent
|
2015-10-06 19:34:54
|
nationalparkservice/places-data
|
https://api.github.com/repos/nationalparkservice/places-data
|
closed
|
Add tags for ferry route
|
pending-other-process schema
|
http://wiki.openstreetmap.org/wiki/Tag:route%3Dferry
This should be in the superclass of "Trail" class of "Motorized Trail"
|
1.0
|
Add tags for ferry route - http://wiki.openstreetmap.org/wiki/Tag:route%3Dferry
This should be in the superclass of "Trail" class of "Motorized Trail"
|
process
|
add tags for ferry route this should be in the superclass of trail class of motorized trail
| 1
|
4,938
| 7,795,878,398
|
IssuesEvent
|
2018-06-08 09:35:51
|
StrikeNP/trac_test
|
https://api.github.com/repos/StrikeNP/trac_test
|
closed
|
In the call to PlotCreator, what is VarName used for? (Trac #173)
|
Migrated from Trac post_processing senkbeil@uwm.edu task
|
I just realized that VarName is being passed in to the MATLAB script; what is its intended purpose? Can we get rid of it? I'm not using it.
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/173
```json
{
"status": "closed",
"changetime": "2009-09-02T20:44:39",
"description": "I just realized that VarName is being passed in to the MATLAB script; what is its intended purpose? Can we get rid of it? I'm not using it.",
"reporter": "nielsenb@uwm.edu",
"cc": "nielsenb@uwm.edu",
"resolution": "Verified by V. Larson",
"_ts": "1251924279000000",
"component": "post_processing",
"summary": "In the call to PlotCreator, what is VarName used for?",
"priority": "trivial",
"keywords": "",
"time": "2009-08-05T22:44:19",
"milestone": "Plotgen 3.0",
"owner": "senkbeil@uwm.edu",
"type": "task"
}
```
|
1.0
|
In the call to PlotCreator, what is VarName used for? (Trac #173) - I just realized that VarName is being passed in to the MATLAB script; what is its intended purpose? Can we get rid of it? I'm not using it.
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/173
```json
{
"status": "closed",
"changetime": "2009-09-02T20:44:39",
"description": "I just realized that VarName is being passed in to the MATLAB script; what is its intended purpose? Can we get rid of it? I'm not using it.",
"reporter": "nielsenb@uwm.edu",
"cc": "nielsenb@uwm.edu",
"resolution": "Verified by V. Larson",
"_ts": "1251924279000000",
"component": "post_processing",
"summary": "In the call to PlotCreator, what is VarName used for?",
"priority": "trivial",
"keywords": "",
"time": "2009-08-05T22:44:19",
"milestone": "Plotgen 3.0",
"owner": "senkbeil@uwm.edu",
"type": "task"
}
```
|
process
|
in the call to plotcreator what is varname used for trac i just realized that varname is being passed in to the matlab script what is its intended purpose can we get rid of it i m not using it attachments migrated from json status closed changetime description i just realized that varname is being passed in to the matlab script what is its intended purpose can we get rid of it i m not using it reporter nielsenb uwm edu cc nielsenb uwm edu resolution verified by v larson ts component post processing summary in the call to plotcreator what is varname used for priority trivial keywords time milestone plotgen owner senkbeil uwm edu type task
| 1
|
14,094
| 16,983,796,099
|
IssuesEvent
|
2021-06-30 12:13:47
|
deepset-ai/haystack
|
https://api.github.com/repos/deepset-ai/haystack
|
closed
|
Allow for batch indexing when using Pipelines
|
good first issue topic:preprocessing
|
Currently, the `BaseConverter` and `PreProcessor` classes only work on single files when used as nodes in a Pipeline. This means that we write documents one at a time to the DocumentStore which can significantly slow down performance.
We should fix this by allowing the `run()` methods of `BaseConverter` and `PreProcessor` to processes batches of documents rather than just single docs.
To do:
- [ ] Get `BaseConverter` to work on batches
- [ ] Get `PreProcessor` to work on batches (#1163)
- [ ] Ensure indexing pipeline is working on batch level
- [ ] Add indexing pipeline example to preprocessing tutorial
|
1.0
|
Allow for batch indexing when using Pipelines - Currently, the `BaseConverter` and `PreProcessor` classes only work on single files when used as nodes in a Pipeline. This means that we write documents one at a time to the DocumentStore which can significantly slow down performance.
We should fix this by allowing the `run()` methods of `BaseConverter` and `PreProcessor` to processes batches of documents rather than just single docs.
To do:
- [ ] Get `BaseConverter` to work on batches
- [ ] Get `PreProcessor` to work on batches (#1163)
- [ ] Ensure indexing pipeline is working on batch level
- [ ] Add indexing pipeline example to preprocessing tutorial
|
process
|
allow for batch indexing when using pipelines currently the baseconverter and preprocessor classes only work on single files when used as nodes in a pipeline this means that we write documents one at a time to the documentstore which can significantly slow down performance we should fix this by allowing the run methods of baseconverter and preprocessor to processes batches of documents rather than just single docs to do get baseconverter to work on batches get preprocessor to work on batches ensure indexing pipeline is working on batch level add indexing pipeline example to preprocessing tutorial
| 1
|
1,123
| 3,594,530,169
|
IssuesEvent
|
2016-02-02 00:09:32
|
DarkEnergyScienceCollaboration/ComputingInfrastructure
|
https://api.github.com/repos/DarkEnergyScienceCollaboration/ComputingInfrastructure
|
opened
|
Perform reprocessing at NERSC
|
CI Reprocessing Stress Test Workflow Engine
|
We need to confirm that NERSC can handle our HTC workflows. Image level reprocessing has been identified as a good test of this. Should engage reprocessing with Workflow Engine (or think about it). @boutigny @tony-johnson @wenaus @nugent68 @salmanhabib are known interested parties.
|
1.0
|
Perform reprocessing at NERSC - We need to confirm that NERSC can handle our HTC workflows. Image level reprocessing has been identified as a good test of this. Should engage reprocessing with Workflow Engine (or think about it). @boutigny @tony-johnson @wenaus @nugent68 @salmanhabib are known interested parties.
|
process
|
perform reprocessing at nersc we need to confirm that nersc can handle our htc workflows image level reprocessing has been identified as a good test of this should engage reprocessing with workflow engine or think about it boutigny tony johnson wenaus salmanhabib are known interested parties
| 1
|
6,610
| 9,694,879,388
|
IssuesEvent
|
2019-05-24 20:21:07
|
kerubistan/kerub
|
https://api.github.com/repos/kerubistan/kerub
|
closed
|
get rid of the assumption that there can be only one allocation of a virtual storage
|
component:data processing priority: high
|
read only virtual storage can have any number of allocations
|
1.0
|
get rid of the assumption that there can be only one allocation of a virtual storage - read only virtual storage can have any number of allocations
|
process
|
get rid of the assumption that there can be only one allocation of a virtual storage read only virtual storage can have any number of allocations
| 1
|
104,186
| 22,601,389,155
|
IssuesEvent
|
2022-06-29 09:26:12
|
UnitTestBot/UTBotJava
|
https://api.github.com/repos/UnitTestBot/UTBotJava
|
closed
|
Invalid import of nested enums from default package
|
bug codegen top focus
|
**Description**
Invalid `import` statements are generated for inner enums declared in the default (unnamed) package. It seems that in this specific case (default package) `import` should not be generated, instead fully qualified references should be used in the test body.
**To Reproduce**
This seems to be a rare case, as default package is almost never used in real projects.
Suppose we have a class in the default ([JLS-7.4.2](https://docs.oracle.com/javase/specs/jls/se8/html/jls-7.html#jls-7.4.2)) package. In this class, an inner `enum` is declared. When UTBotJava generates the test for a method involving this enum, the generated test file contains the invalid import from the unnamed package, which results in compilation error.
```
import org.jetbrains.annotations.NotNull;
public class InnerClassExamples {
public enum InnerEnum {
ONE,
TWO
}
public int decode(@NotNull InnerEnum key) {
switch (key) {
case ONE: return 1;
case TWO: return 2;
default: return 0;
}
}
}
```
Generate the test for `decode` function.
**Expected behavior**
Fully qualified name of `InnerClassExamples.InnerEnum` should be used in the code.
```
import org.junit.jupiter.api.Test;
import org.junit.jupiter.api.DisplayName;
import static org.junit.jupiter.api.Assertions.assertEquals;
public class InnerClassExamplesTest {
///region Test suites for executable InnerClassExamples.decode
///region SUCCESSFUL EXECUTIONS for method decode(InnerClassExamples.InnerEnum)
/**
* <pre>
* Test activates switch case: {@code default }, returns from: {@code return 1; }
* </pre>
*/
@Test
@DisplayName("decode: switch(key) case: Default -> return 1")
public void testDecode_Return1() throws Throwable {
InnerClassExamples innerClassExamples = new InnerClassExamples();
InnerClassExamples.InnerEnum innerEnum = InnerClassExamples.InnerEnum.ONE; // N.B.: fully qualified name
int actual = innerClassExamples.decode(innerEnum);
assertEquals(1, actual);
}
/**
* <pre>
* Test activates switch case: {@code default }, returns from: {@code return 2; }
* </pre>
*/
@Test
@DisplayName("decode: switch(key) case: Default -> return 2")
public void testDecode_Return2() throws Throwable {
InnerClassExamples innerClassExamples = new InnerClassExamples();
InnerClassExamples.InnerEnum innerEnum = InnerClassExamples.InnerEnum.TWO; // N.B.: fully qualified name
int actual = innerClassExamples.decode(innerEnum);
assertEquals(2, actual);
}
///endregion
///endregion
}
```
**Actual behavior**
Invalid `import` statement from the default package is generated.
```
import org.junit.jupiter.api.Test;
import org.junit.jupiter.api.DisplayName;
import InnerClassExamples.InnerEnum; // N.B.: This line is wrong, the code does not compile
import static org.junit.jupiter.api.Assertions.assertEquals;
public class InnerClassExamplesTest {
///region Test suites for executable InnerClassExamples.decode
///region SUCCESSFUL EXECUTIONS for method decode(InnerClassExamples.InnerEnum)
/**
* <pre>
* Test activates switch case: {@code default }, returns from: {@code return 1; }
* </pre>
*/
@Test
@DisplayName("decode: switch(key) case: Default -> return 1")
public void testDecode_Return1() throws Throwable {
InnerClassExamples innerClassExamples = new InnerClassExamples();
InnerEnum innerEnum = InnerEnum.ONE;
int actual = innerClassExamples.decode(innerEnum);
assertEquals(1, actual);
}
/**
* <pre>
* Test activates switch case: {@code default }, returns from: {@code return 2; }
* </pre>
*/
@Test
@DisplayName("decode: switch(key) case: Default -> return 2")
public void testDecode_Return2() throws Throwable {
InnerClassExamples innerClassExamples = new InnerClassExamples();
InnerEnum innerEnum = InnerEnum.TWO;
int actual = innerClassExamples.decode(innerEnum);
assertEquals(2, actual);
}
///endregion
///endregion
}
```
**Visual proofs (screenshots, logs, images)**
See the "Actual behavior" section.
**Environment**
Sample tests were generated with JUnit5 (not parameterized), no mocking.
**Additional context**
Importing from unnamed packages is explicitly [forbidden](https://docs.oracle.com/javase/specs/jls/se11/html/jls-7.html#jls-7.5.1) in Java 11 and does not seem to work in Java 8 as well.
|
1.0
|
Invalid import of nested enums from default package - **Description**
Invalid `import` statements are generated for inner enums declared in the default (unnamed) package. It seems that in this specific case (default package) `import` should not be generated, instead fully qualified references should be used in the test body.
**To Reproduce**
This seems to be a rare case, as default package is almost never used in real projects.
Suppose we have a class in the default ([JLS-7.4.2](https://docs.oracle.com/javase/specs/jls/se8/html/jls-7.html#jls-7.4.2)) package. In this class, an inner `enum` is declared. When UTBotJava generates the test for a method involving this enum, the generated test file contains the invalid import from the unnamed package, which results in compilation error.
```
import org.jetbrains.annotations.NotNull;
public class InnerClassExamples {
public enum InnerEnum {
ONE,
TWO
}
public int decode(@NotNull InnerEnum key) {
switch (key) {
case ONE: return 1;
case TWO: return 2;
default: return 0;
}
}
}
```
Generate the test for `decode` function.
**Expected behavior**
Fully qualified name of `InnerClassExamples.InnerEnum` should be used in the code.
```
import org.junit.jupiter.api.Test;
import org.junit.jupiter.api.DisplayName;
import static org.junit.jupiter.api.Assertions.assertEquals;
public class InnerClassExamplesTest {
///region Test suites for executable InnerClassExamples.decode
///region SUCCESSFUL EXECUTIONS for method decode(InnerClassExamples.InnerEnum)
/**
* <pre>
* Test activates switch case: {@code default }, returns from: {@code return 1; }
* </pre>
*/
@Test
@DisplayName("decode: switch(key) case: Default -> return 1")
public void testDecode_Return1() throws Throwable {
InnerClassExamples innerClassExamples = new InnerClassExamples();
InnerClassExamples.InnerEnum innerEnum = InnerClassExamples.InnerEnum.ONE; // N.B.: fully qualified name
int actual = innerClassExamples.decode(innerEnum);
assertEquals(1, actual);
}
/**
* <pre>
* Test activates switch case: {@code default }, returns from: {@code return 2; }
* </pre>
*/
@Test
@DisplayName("decode: switch(key) case: Default -> return 2")
public void testDecode_Return2() throws Throwable {
InnerClassExamples innerClassExamples = new InnerClassExamples();
InnerClassExamples.InnerEnum innerEnum = InnerClassExamples.InnerEnum.TWO; // N.B.: fully qualified name
int actual = innerClassExamples.decode(innerEnum);
assertEquals(2, actual);
}
///endregion
///endregion
}
```
**Actual behavior**
Invalid `import` statement from the default package is generated.
```
import org.junit.jupiter.api.Test;
import org.junit.jupiter.api.DisplayName;
import InnerClassExamples.InnerEnum; // N.B.: This line is wrong, the code does not compile
import static org.junit.jupiter.api.Assertions.assertEquals;
public class InnerClassExamplesTest {
///region Test suites for executable InnerClassExamples.decode
///region SUCCESSFUL EXECUTIONS for method decode(InnerClassExamples.InnerEnum)
/**
* <pre>
* Test activates switch case: {@code default }, returns from: {@code return 1; }
* </pre>
*/
@Test
@DisplayName("decode: switch(key) case: Default -> return 1")
public void testDecode_Return1() throws Throwable {
InnerClassExamples innerClassExamples = new InnerClassExamples();
InnerEnum innerEnum = InnerEnum.ONE;
int actual = innerClassExamples.decode(innerEnum);
assertEquals(1, actual);
}
/**
* <pre>
* Test activates switch case: {@code default }, returns from: {@code return 2; }
* </pre>
*/
@Test
@DisplayName("decode: switch(key) case: Default -> return 2")
public void testDecode_Return2() throws Throwable {
InnerClassExamples innerClassExamples = new InnerClassExamples();
InnerEnum innerEnum = InnerEnum.TWO;
int actual = innerClassExamples.decode(innerEnum);
assertEquals(2, actual);
}
///endregion
///endregion
}
```
**Visual proofs (screenshots, logs, images)**
See the "Actual behavior" section.
**Environment**
Sample tests were generated with JUnit5 (not parameterized), no mocking.
**Additional context**
Importing from unnamed packages is explicitly [forbidden](https://docs.oracle.com/javase/specs/jls/se11/html/jls-7.html#jls-7.5.1) in Java 11 and does not seem to work in Java 8 as well.
|
non_process
|
invalid import of nested enums from default package description invalid import statements are generated for inner enums declared in the default unnamed package it seems that in this specific case default package import should not be generated instead fully qualified references should be used in the test body to reproduce this seems to be a rare case as default package is almost never used in real projects suppose we have a class in the default package in this class an inner enum is declared when utbotjava generates the test for a method involving this enum the generated test file contains the invalid import from the unnamed package which results in compilation error import org jetbrains annotations notnull public class innerclassexamples public enum innerenum one two public int decode notnull innerenum key switch key case one return case two return default return generate the test for decode function expected behavior fully qualified name of innerclassexamples innerenum should be used in the code import org junit jupiter api test import org junit jupiter api displayname import static org junit jupiter api assertions assertequals public class innerclassexamplestest region test suites for executable innerclassexamples decode region successful executions for method decode innerclassexamples innerenum test activates switch case code default returns from code return test displayname decode switch key case default return public void testdecode throws throwable innerclassexamples innerclassexamples new innerclassexamples innerclassexamples innerenum innerenum innerclassexamples innerenum one n b fully qualified name int actual innerclassexamples decode innerenum assertequals actual test activates switch case code default returns from code return test displayname decode switch key case default return public void testdecode throws throwable innerclassexamples innerclassexamples new innerclassexamples innerclassexamples innerenum innerenum innerclassexamples innerenum two n b fully qualified name int actual innerclassexamples decode innerenum assertequals actual endregion endregion actual behavior invalid import statement from the default package is generated import org junit jupiter api test import org junit jupiter api displayname import innerclassexamples innerenum n b this line is wrong the code does not compile import static org junit jupiter api assertions assertequals public class innerclassexamplestest region test suites for executable innerclassexamples decode region successful executions for method decode innerclassexamples innerenum test activates switch case code default returns from code return test displayname decode switch key case default return public void testdecode throws throwable innerclassexamples innerclassexamples new innerclassexamples innerenum innerenum innerenum one int actual innerclassexamples decode innerenum assertequals actual test activates switch case code default returns from code return test displayname decode switch key case default return public void testdecode throws throwable innerclassexamples innerclassexamples new innerclassexamples innerenum innerenum innerenum two int actual innerclassexamples decode innerenum assertequals actual endregion endregion visual proofs screenshots logs images see the actual behavior section environment sample tests were generated with not parameterized no mocking additional context importing from unnamed packages is explicitly in java and does not seem to work in java as well
| 0
|
415
| 2,852,365,086
|
IssuesEvent
|
2015-06-01 13:17:00
|
genomizer/genomizer-server
|
https://api.github.com/repos/genomizer/genomizer-server
|
opened
|
Change the upload API
|
enhancement Low priority Processing
|
This is something that I haven't had time to finish this year. Right now, the upload API is a bit bonkers: the user does a POST request that looks approx. like `POST /file \n\n {fileName : "foo", otherMetadata ...}` and then gets back a URL that is used for the actual upload (the usual `multipart/form-data` POST). This makes the code unnecessarily complex: we have to distinguish between files that are "Done" and "In Progress" in the DB, which was the source of a bunch of bugs. Ideally, the file would be uploaded in a single step with a `multipart/form-data` POST request that'd contain both JSON metadata and the file itself. This requires changing all clients, though, so we punted on it for the time being. Maybe next year's group will finally fix this issue.
|
1.0
|
Change the upload API - This is something that I haven't had time to finish this year. Right now, the upload API is a bit bonkers: the user does a POST request that looks approx. like `POST /file \n\n {fileName : "foo", otherMetadata ...}` and then gets back a URL that is used for the actual upload (the usual `multipart/form-data` POST). This makes the code unnecessarily complex: we have to distinguish between files that are "Done" and "In Progress" in the DB, which was the source of a bunch of bugs. Ideally, the file would be uploaded in a single step with a `multipart/form-data` POST request that'd contain both JSON metadata and the file itself. This requires changing all clients, though, so we punted on it for the time being. Maybe next year's group will finally fix this issue.
|
process
|
change the upload api this is something that i haven t had time to finish this year right now the upload api is a bit bonkers the user does a post request that looks approx like post file n n filename foo othermetadata and then gets back a url that is used for the actual upload the usual multipart form data post this makes the code unnecessarily complex we have to distinguish between files that are done and in progress in the db which was the source of a bunch of bugs ideally the file would be uploaded in a single step with a multipart form data post request that d contain both json metadata and the file itself this requires changing all clients though so we punted on it for the time being maybe next year s group will finally fix this issue
| 1
|
269
| 2,699,150,978
|
IssuesEvent
|
2015-04-03 14:48:40
|
appsgate2015/appsgate
|
https://api.github.com/repos/appsgate2015/appsgate
|
closed
|
L'état du Domicube ne remonte pas dans le client
|
P1 PROCESSING
|
cf. Mail de Joëlle:
. Le DomiCube est present mais s'obstine à rester ds l'etat "unknown" que tu le bouges ou pas
Observation Thibaud:
Et je reproduis le bug, j’ai le cube qui est là et qui reste dans l’état unknown.
Par contre, dans les logs du client, je vois bien les messages arriver qd on change de face.
Donc le msg part bien du cube, il est traité par le serveur(service domicube qui transmet au serveur appsgate), il est remonté au client qui n’affiche pas le nouvel état.
Sur les logs serveurs:
des NPE sur TraceMAN,
public synchronized void coreEventNotify(long timeStamp, String srcId, String varName, String value) {
-> value est parfois à null, et la NPE provoque des effets de bord
|
1.0
|
L'état du Domicube ne remonte pas dans le client - cf. Mail de Joëlle:
. Le DomiCube est present mais s'obstine à rester ds l'etat "unknown" que tu le bouges ou pas
Observation Thibaud:
Et je reproduis le bug, j’ai le cube qui est là et qui reste dans l’état unknown.
Par contre, dans les logs du client, je vois bien les messages arriver qd on change de face.
Donc le msg part bien du cube, il est traité par le serveur(service domicube qui transmet au serveur appsgate), il est remonté au client qui n’affiche pas le nouvel état.
Sur les logs serveurs:
des NPE sur TraceMAN,
public synchronized void coreEventNotify(long timeStamp, String srcId, String varName, String value) {
-> value est parfois à null, et la NPE provoque des effets de bord
|
process
|
l état du domicube ne remonte pas dans le client cf mail de joëlle le domicube est present mais s obstine à rester ds l etat unknown que tu le bouges ou pas observation thibaud et je reproduis le bug j’ai le cube qui est là et qui reste dans l’état unknown par contre dans les logs du client je vois bien les messages arriver qd on change de face donc le msg part bien du cube il est traité par le serveur service domicube qui transmet au serveur appsgate il est remonté au client qui n’affiche pas le nouvel état sur les logs serveurs des npe sur traceman public synchronized void coreeventnotify long timestamp string srcid string varname string value value est parfois à null et la npe provoque des effets de bord
| 1
|
11,158
| 13,957,693,763
|
IssuesEvent
|
2020-10-24 08:11:04
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
RO: Missing resources in Geoportal
|
Geoportal Harvesting process RO - Romania
|
Collected from the Geoportal Workshop online survey answers: (see attachment)
Missing download icon for SDS (Spatial Data Services).
The icon appear only for Network Services (ATOM or WFS).
Metadata available online at http://gmlid.eu/RO/GeoEcoMar/EUXINUS/MD
|
1.0
|
RO: Missing resources in Geoportal - Collected from the Geoportal Workshop online survey answers: (see attachment)
Missing download icon for SDS (Spatial Data Services).
The icon appear only for Network Services (ATOM or WFS).
Metadata available online at http://gmlid.eu/RO/GeoEcoMar/EUXINUS/MD
|
process
|
ro missing resources in geoportal collected from the geoportal workshop online survey answers see attachment missing download icon for sds spatial data services the icon appear only for network services atom or wfs metadata available online at
| 1
|
30,539
| 11,839,201,978
|
IssuesEvent
|
2020-03-23 16:47:25
|
Mohib-hub/TwxAzureDataLakeConnector
|
https://api.github.com/repos/Mohib-hub/TwxAzureDataLakeConnector
|
opened
|
WS-2019-0318 (Medium) detected in handlebars-4.1.0.tgz
|
security vulnerability
|
## WS-2019-0318 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.0.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.1.0.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.1.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/TwxAzureDataLakeConnector/src/package.json</p>
<p>Path to vulnerable library: /tmp/archiveExtraction_da66323a-d17c-4593-a6c9-6dbbe5d33fb7/20200323164443_55650/ws-scm_depth_0/TwxAzureDataLakeConnector/dist/ptc-adls-connector-1-0-0/ptc-adls-connector/node_modules/nyc/node_modules/handlebars/package.json,/tmp/archiveExtraction_da66323a-d17c-4593-a6c9-6dbbe5d33fb7/20200323164443_55650/ws-scm_depth_0/TwxAzureDataLakeConnector/dist/ptc-adls-connector-1-0-0/ptc-adls-connector/node_modules/nyc/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- nyc-13.3.0.tgz (Root Library)
- istanbul-reports-2.1.1.tgz
- :x: **handlebars-4.1.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Mohib-hub/TwxAzureDataLakeConnector/commit/ec2d5e599c3035a1557d1a928e4a5702861ed5ed">ec2d5e599c3035a1557d1a928e4a5702861ed5ed</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Denial of Service vulnerability found in handlebars 4.x before 4.4.5.While processing specially-crafted templates, the parser may be forced into endless loop. Attackers may exhaust system resources.
<p>Publish Date: 2019-12-01
<p>URL: <a href=https://github.com/wycats/handlebars.js/commit/8d5530ee2c3ea9f0aee3fde310b9f36887d00b8b>WS-2019-0318</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1300">https://www.npmjs.com/advisories/1300</a></p>
<p>Release Date: 2019-12-01</p>
<p>Fix Resolution: handlebars - 4.4.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"4.1.0","isTransitiveDependency":true,"dependencyTree":"nyc:13.3.0;istanbul-reports:2.1.1;handlebars:4.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.4.5"}],"vulnerabilityIdentifier":"WS-2019-0318","vulnerabilityDetails":"A Denial of Service vulnerability found in handlebars 4.x before 4.4.5.While processing specially-crafted templates, the parser may be forced into endless loop. Attackers may exhaust system resources.","vulnerabilityUrl":"https://github.com/wycats/handlebars.js/commit/8d5530ee2c3ea9f0aee3fde310b9f36887d00b8b","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> -->
|
True
|
WS-2019-0318 (Medium) detected in handlebars-4.1.0.tgz - ## WS-2019-0318 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.0.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.1.0.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.1.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/TwxAzureDataLakeConnector/src/package.json</p>
<p>Path to vulnerable library: /tmp/archiveExtraction_da66323a-d17c-4593-a6c9-6dbbe5d33fb7/20200323164443_55650/ws-scm_depth_0/TwxAzureDataLakeConnector/dist/ptc-adls-connector-1-0-0/ptc-adls-connector/node_modules/nyc/node_modules/handlebars/package.json,/tmp/archiveExtraction_da66323a-d17c-4593-a6c9-6dbbe5d33fb7/20200323164443_55650/ws-scm_depth_0/TwxAzureDataLakeConnector/dist/ptc-adls-connector-1-0-0/ptc-adls-connector/node_modules/nyc/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- nyc-13.3.0.tgz (Root Library)
- istanbul-reports-2.1.1.tgz
- :x: **handlebars-4.1.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Mohib-hub/TwxAzureDataLakeConnector/commit/ec2d5e599c3035a1557d1a928e4a5702861ed5ed">ec2d5e599c3035a1557d1a928e4a5702861ed5ed</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Denial of Service vulnerability found in handlebars 4.x before 4.4.5.While processing specially-crafted templates, the parser may be forced into endless loop. Attackers may exhaust system resources.
<p>Publish Date: 2019-12-01
<p>URL: <a href=https://github.com/wycats/handlebars.js/commit/8d5530ee2c3ea9f0aee3fde310b9f36887d00b8b>WS-2019-0318</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1300">https://www.npmjs.com/advisories/1300</a></p>
<p>Release Date: 2019-12-01</p>
<p>Fix Resolution: handlebars - 4.4.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"4.1.0","isTransitiveDependency":true,"dependencyTree":"nyc:13.3.0;istanbul-reports:2.1.1;handlebars:4.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.4.5"}],"vulnerabilityIdentifier":"WS-2019-0318","vulnerabilityDetails":"A Denial of Service vulnerability found in handlebars 4.x before 4.4.5.While processing specially-crafted templates, the parser may be forced into endless loop. Attackers may exhaust system resources.","vulnerabilityUrl":"https://github.com/wycats/handlebars.js/commit/8d5530ee2c3ea9f0aee3fde310b9f36887d00b8b","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> -->
|
non_process
|
ws medium detected in handlebars tgz ws medium severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file tmp ws scm twxazuredatalakeconnector src package json path to vulnerable library tmp archiveextraction ws scm depth twxazuredatalakeconnector dist ptc adls connector ptc adls connector node modules nyc node modules handlebars package json tmp archiveextraction ws scm depth twxazuredatalakeconnector dist ptc adls connector ptc adls connector node modules nyc node modules handlebars package json dependency hierarchy nyc tgz root library istanbul reports tgz x handlebars tgz vulnerable library found in head commit a href vulnerability details a denial of service vulnerability found in handlebars x before while processing specially crafted templates the parser may be forced into endless loop attackers may exhaust system resources publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution handlebars isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails a denial of service vulnerability found in handlebars x before while processing specially crafted templates the parser may be forced into endless loop attackers may exhaust system resources vulnerabilityurl
| 0
|
382,357
| 11,304,974,020
|
IssuesEvent
|
2020-01-18 01:30:38
|
rathena/rathena
|
https://api.github.com/repos/rathena/rathena
|
closed
|
General Egnigem Cenia Card HP/SP regen timer is reset when receiving buffs
|
component:core mode:prerenewal mode:renewal priority:low status:confirmed type:bug
|
* **rAthena Hash**: Latest
* **Client Date**: 2017-06-14bRagexeRE
* **Server Mode**: Pre-Renewal
* **Description of Issue**:
* Result: When wearing General Egnigem Cenia card, HP and SP are regen every 10 seconds. However, when we receive any type of buff (self skill, support skill from someone else, or item), then the timer gets reset, and we have to wait for a full 10 seconds after the last buff has been applied to get the HP/SP regen again. Said differently, you don't have the HP/SP regen if you get any buff every 9 seconds or less.
* Expected Result: HP/SP regen should kick in every 10 seconds, no matter what.
* How to Reproduce: Wear General Egnigem Cenia card, and spam any of the following every less than 10 seconds. The list is non-exhaustive, so there might be other things that resets the HP/SP regen timer:
```
Self targeted skills:
- Two-Hand Quicken
- Counter-Attack
- Parrying
- Concentration
- Magnum Break
- Energy Coat
Have a support character spamming these skills on the GEC wearer:
- Any endow weapon from Sage/Prof
- Increase Agility
- Blessing
- Assumptio
- Magnificat
Using buff items:
- Concentration Potion, Awakening Potion, Berserk Potion
- Cursed Water
- Stats food
```
|
1.0
|
General Egnigem Cenia Card HP/SP regen timer is reset when receiving buffs - * **rAthena Hash**: Latest
* **Client Date**: 2017-06-14bRagexeRE
* **Server Mode**: Pre-Renewal
* **Description of Issue**:
* Result: When wearing General Egnigem Cenia card, HP and SP are regen every 10 seconds. However, when we receive any type of buff (self skill, support skill from someone else, or item), then the timer gets reset, and we have to wait for a full 10 seconds after the last buff has been applied to get the HP/SP regen again. Said differently, you don't have the HP/SP regen if you get any buff every 9 seconds or less.
* Expected Result: HP/SP regen should kick in every 10 seconds, no matter what.
* How to Reproduce: Wear General Egnigem Cenia card, and spam any of the following every less than 10 seconds. The list is non-exhaustive, so there might be other things that resets the HP/SP regen timer:
```
Self targeted skills:
- Two-Hand Quicken
- Counter-Attack
- Parrying
- Concentration
- Magnum Break
- Energy Coat
Have a support character spamming these skills on the GEC wearer:
- Any endow weapon from Sage/Prof
- Increase Agility
- Blessing
- Assumptio
- Magnificat
Using buff items:
- Concentration Potion, Awakening Potion, Berserk Potion
- Cursed Water
- Stats food
```
|
non_process
|
general egnigem cenia card hp sp regen timer is reset when receiving buffs rathena hash latest client date server mode pre renewal description of issue result when wearing general egnigem cenia card hp and sp are regen every seconds however when we receive any type of buff self skill support skill from someone else or item then the timer gets reset and we have to wait for a full seconds after the last buff has been applied to get the hp sp regen again said differently you don t have the hp sp regen if you get any buff every seconds or less expected result hp sp regen should kick in every seconds no matter what how to reproduce wear general egnigem cenia card and spam any of the following every less than seconds the list is non exhaustive so there might be other things that resets the hp sp regen timer self targeted skills two hand quicken counter attack parrying concentration magnum break energy coat have a support character spamming these skills on the gec wearer any endow weapon from sage prof increase agility blessing assumptio magnificat using buff items concentration potion awakening potion berserk potion cursed water stats food
| 0
|
683,306
| 23,376,276,892
|
IssuesEvent
|
2022-08-11 03:40:12
|
kubevela/kubevela
|
https://api.github.com/repos/kubevela/kubevela
|
closed
|
[Feature] add policy definitions for advanced internal policy
|
type/enhancement good first issue help wanted priority/nice-to-have effort/small area/policy
|
**Is your feature request related to a problem? Please describe.**
<!--
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
-->
Previous PR(https://github.com/kubevela/kubevela/pull/3894) adds topology and override policy definition for schema.
Advanced policies' are also welcomed, including apply-once, garbage-collect, shared-resouece.
**Describe the solution you'd like**
<!--
A clear and concise description of what you want to happen.
-->
**Describe alternatives you've considered**
<!--
A clear and concise description of any alternative solutions or features you've considered.
-->
**Additional context**
<!--
Add any other context or screenshots about the feature request here.
-->
|
1.0
|
[Feature] add policy definitions for advanced internal policy - **Is your feature request related to a problem? Please describe.**
<!--
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
-->
Previous PR(https://github.com/kubevela/kubevela/pull/3894) adds topology and override policy definition for schema.
Advanced policies' are also welcomed, including apply-once, garbage-collect, shared-resouece.
**Describe the solution you'd like**
<!--
A clear and concise description of what you want to happen.
-->
**Describe alternatives you've considered**
<!--
A clear and concise description of any alternative solutions or features you've considered.
-->
**Additional context**
<!--
Add any other context or screenshots about the feature request here.
-->
|
non_process
|
add policy definitions for advanced internal policy is your feature request related to a problem please describe a clear and concise description of what the problem is ex i m always frustrated when previous pr adds topology and override policy definition for schema advanced policies are also welcomed including apply once garbage collect shared resouece describe the solution you d like a clear and concise description of what you want to happen describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
| 0
|
5,974
| 8,793,573,851
|
IssuesEvent
|
2018-12-21 20:31:01
|
docker/docker.github.io
|
https://api.github.com/repos/docker/docker.github.io
|
closed
|
Create diagrams for scanning and jobrunner
|
area/dtr area/ee process/images & screenshots
|
Awesome sketches from @venalen
## Scanning:
A:

B:

(**Note**: Follow up with @toli, b/c he asked for the .sketch file to create the diagram, but unsure if he completed it)
-----
## Jobrunner:


|
1.0
|
Create diagrams for scanning and jobrunner - Awesome sketches from @venalen
## Scanning:
A:

B:

(**Note**: Follow up with @toli, b/c he asked for the .sketch file to create the diagram, but unsure if he completed it)
-----
## Jobrunner:


|
process
|
create diagrams for scanning and jobrunner awesome sketches from venalen scanning a b note follow up with toli b c he asked for the sketch file to create the diagram but unsure if he completed it jobrunner
| 1
|
56,208
| 3,078,505,569
|
IssuesEvent
|
2015-08-21 10:41:29
|
pavel-pimenov/flylinkdc-r5xx
|
https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx
|
closed
|
1.заваливается, 2 непонятки с автопробросом портов
|
bug Component-Logic Component-UI imported invalid Priority-Critical Stability
|
_From [cep...@mail.ru](https://code.google.com/u/108682685495662585618/) on June 07, 2011 18:14:14_
500-7250. винхп про сп3
1.в в списке для закачки оч долго висело 2 файла (скачан, ожидание). надоело ждать-нажал удалить. порга завалилась,создав репорт(прилагается)
2.при автоматическом пробросе портов через день перестает работать поиск и не скачиваются файл-листы(на всех подключенных хабах). непосредственно на роутере пробросил порты,выставил их в проге принудительно-все заработало.
**Attachment:** [crash-stack-r500-build-7250.dmp](http://code.google.com/p/flylinkdc/issues/detail?id=490)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=490_
|
1.0
|
1.заваливается, 2 непонятки с автопробросом портов - _From [cep...@mail.ru](https://code.google.com/u/108682685495662585618/) on June 07, 2011 18:14:14_
500-7250. винхп про сп3
1.в в списке для закачки оч долго висело 2 файла (скачан, ожидание). надоело ждать-нажал удалить. порга завалилась,создав репорт(прилагается)
2.при автоматическом пробросе портов через день перестает работать поиск и не скачиваются файл-листы(на всех подключенных хабах). непосредственно на роутере пробросил порты,выставил их в проге принудительно-все заработало.
**Attachment:** [crash-stack-r500-build-7250.dmp](http://code.google.com/p/flylinkdc/issues/detail?id=490)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=490_
|
non_process
|
заваливается непонятки с автопробросом портов from on june винхп про в в списке для закачки оч долго висело файла скачан ожидание надоело ждать нажал удалить порга завалилась создав репорт прилагается при автоматическом пробросе портов через день перестает работать поиск и не скачиваются файл листы на всех подключенных хабах непосредственно на роутере пробросил порты выставил их в проге принудительно все заработало attachment original issue
| 0
|
18,812
| 24,712,881,836
|
IssuesEvent
|
2022-10-20 03:19:09
|
TabooLib/chemdah
|
https://api.github.com/repos/TabooLib/chemdah
|
closed
|
使用滚轮切换回复内容 开启时出现的错误
|
bug processed
|
```
[16:06:18 ERROR]: Could not pass event PlayerItemHeldEvent to Chemdah v0.2.39
java.lang.NoClassDefFoundError: net/minecraft/server/v1_16_R3/PacketPlayOutHeldItemSlot
at ink.ptms.chemdah.core.conversation.theme.ThemeChat.onItemHeld(ThemeChat.kt:98) ~[Chemdah-0.2.39.jar:?]
at java.lang.invoke.MethodHandle.invokeWithArguments(MethodHandle.java:732) ~[?:?]
at ink.ptms.chemdah.taboolib.library.reflex.JavaClassMethod.invoke(JavaClassMethod.kt:31) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.common.platform.event.EventBus.invoke(EventBus.kt:132) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.common.platform.event.EventBus.invoke$default(EventBus.kt:130) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.common.platform.event.EventBus$registerBukkit$2.invoke(EventBus.kt:77) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.common.platform.event.EventBus$registerBukkit$2.invoke(EventBus.kt:76) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.common.platform.function.ListenerKt$registerBukkitListener$1.invoke(Listener.kt:21) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.common.platform.function.ListenerKt$registerBukkitListener$1.invoke(Listener.kt:21) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.platform.BukkitListener$registerListener$listener$1.invoke(BukkitListener.kt:35) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.platform.BukkitListener$registerListener$listener$1.invoke(BukkitListener.kt:35) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.platform.BukkitListener$BukkitListener.execute(BukkitListener.kt:72) ~[Chemdah-0.2.39.jar:?]
at co.aikar.timings.TimedEventExecutor.execute(TimedEventExecutor.java:76) ~[purpur-api-1.18.2-R0.1-SNAPSHOT.jar:git-Purpur-1632]
at org.bukkit.plugin.RegisteredListener.callEvent(RegisteredListener.java:70) ~[purpur-api-1.18.2-R0.1-SNAPSHOT.jar:?]
at org.bukkit.plugin.SimplePluginManager.callEvent(SimplePluginManager.java:671) ~[purpur-api-1.18.2-R0.1-SNAPSHOT.jar:?]
at net.minecraft.server.network.ServerGamePacketListenerImpl.handleSetCarriedItem(ServerGamePacketListenerImpl.java:2142) ~[?:?]
at net.minecraft.network.protocol.game.ServerboundSetCarriedItemPacket.handle(ServerboundSetCarriedItemPacket.java:24) ~[?:?]
at net.minecraft.network.protocol.game.ServerboundSetCarriedItemPacket.a(ServerboundSetCarriedItemPacket.java:10) ~[?:?]
at net.minecraft.network.protocol.PacketUtils.lambda$ensureRunningOnSameThread$1(PacketUtils.java:51) ~[?:?]
at net.minecraft.server.TickTask.run(TickTask.java:18) ~[purpur-1.18.2.jar:git-Purpur-1632]
at net.minecraft.util.thread.BlockableEventLoop.doRunTask(BlockableEventLoop.java:153) ~[?:?]
at net.minecraft.util.thread.ReentrantBlockableEventLoop.doRunTask(ReentrantBlockableEventLoop.java:24) ~[?:?]
at net.minecraft.server.MinecraftServer.doRunTask(MinecraftServer.java:1427) ~[purpur-1.18.2.jar:git-Purpur-1632]
at net.minecraft.server.MinecraftServer.d(MinecraftServer.java:188) ~[purpur-1.18.2.jar:git-Purpur-1632]
at net.minecraft.util.thread.BlockableEventLoop.pollTask(BlockableEventLoop.java:126) ~[?:?]
at net.minecraft.server.MinecraftServer.pollTaskInternal(MinecraftServer.java:1404) ~[purpur-1.18.2.jar:git-Purpur-1632]
at net.minecraft.server.MinecraftServer.pollTask(MinecraftServer.java:1397) ~[purpur-1.18.2.jar:git-Purpur-1632]
at net.minecraft.util.thread.BlockableEventLoop.managedBlock(BlockableEventLoop.java:136) ~[?:?]
at net.minecraft.server.MinecraftServer.waitUntilNextTick(MinecraftServer.java:1375) ~[purpur-1.18.2.jar:git-Purpur-1632]
at net.minecraft.server.MinecraftServer.runServer(MinecraftServer.java:1257) ~[purpur-1.18.2.jar:git-Purpur-1632]
at net.minecraft.server.MinecraftServer.lambda$spin$1(MinecraftServer.java:321) ~[purpur-1.18.2.jar:git-Purpur-1632]
at java.lang.Thread.run(Thread.java:891) ~[?:?]
Caused by: java.lang.ClassNotFoundException: net.minecraft.server.v1_16_R3.PacketPlayOutHeldItemSlot
at org.bukkit.plugin.java.PluginClassLoader.loadClass0(PluginClassLoader.java:153) ~[purpur-api-1.18.2-R0.1-SNAPSHOT.jar:?]
at org.bukkit.plugin.java.PluginClassLoader.loadClass(PluginClassLoader.java:105) ~[purpur-api-1.18.2-R0.1-SNAPSHOT.jar:?]
at java.lang.ClassLoader.loadClass(ClassLoader.java:520) ~[?:?]
... 32 more
```
使用的服务器核心: `Current: git-Purpur-1632 (MC: 1.18.2)`
|
1.0
|
使用滚轮切换回复内容 开启时出现的错误 - ```
[16:06:18 ERROR]: Could not pass event PlayerItemHeldEvent to Chemdah v0.2.39
java.lang.NoClassDefFoundError: net/minecraft/server/v1_16_R3/PacketPlayOutHeldItemSlot
at ink.ptms.chemdah.core.conversation.theme.ThemeChat.onItemHeld(ThemeChat.kt:98) ~[Chemdah-0.2.39.jar:?]
at java.lang.invoke.MethodHandle.invokeWithArguments(MethodHandle.java:732) ~[?:?]
at ink.ptms.chemdah.taboolib.library.reflex.JavaClassMethod.invoke(JavaClassMethod.kt:31) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.common.platform.event.EventBus.invoke(EventBus.kt:132) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.common.platform.event.EventBus.invoke$default(EventBus.kt:130) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.common.platform.event.EventBus$registerBukkit$2.invoke(EventBus.kt:77) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.common.platform.event.EventBus$registerBukkit$2.invoke(EventBus.kt:76) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.common.platform.function.ListenerKt$registerBukkitListener$1.invoke(Listener.kt:21) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.common.platform.function.ListenerKt$registerBukkitListener$1.invoke(Listener.kt:21) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.platform.BukkitListener$registerListener$listener$1.invoke(BukkitListener.kt:35) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.platform.BukkitListener$registerListener$listener$1.invoke(BukkitListener.kt:35) ~[Chemdah-0.2.39.jar:?]
at ink.ptms.chemdah.taboolib.platform.BukkitListener$BukkitListener.execute(BukkitListener.kt:72) ~[Chemdah-0.2.39.jar:?]
at co.aikar.timings.TimedEventExecutor.execute(TimedEventExecutor.java:76) ~[purpur-api-1.18.2-R0.1-SNAPSHOT.jar:git-Purpur-1632]
at org.bukkit.plugin.RegisteredListener.callEvent(RegisteredListener.java:70) ~[purpur-api-1.18.2-R0.1-SNAPSHOT.jar:?]
at org.bukkit.plugin.SimplePluginManager.callEvent(SimplePluginManager.java:671) ~[purpur-api-1.18.2-R0.1-SNAPSHOT.jar:?]
at net.minecraft.server.network.ServerGamePacketListenerImpl.handleSetCarriedItem(ServerGamePacketListenerImpl.java:2142) ~[?:?]
at net.minecraft.network.protocol.game.ServerboundSetCarriedItemPacket.handle(ServerboundSetCarriedItemPacket.java:24) ~[?:?]
at net.minecraft.network.protocol.game.ServerboundSetCarriedItemPacket.a(ServerboundSetCarriedItemPacket.java:10) ~[?:?]
at net.minecraft.network.protocol.PacketUtils.lambda$ensureRunningOnSameThread$1(PacketUtils.java:51) ~[?:?]
at net.minecraft.server.TickTask.run(TickTask.java:18) ~[purpur-1.18.2.jar:git-Purpur-1632]
at net.minecraft.util.thread.BlockableEventLoop.doRunTask(BlockableEventLoop.java:153) ~[?:?]
at net.minecraft.util.thread.ReentrantBlockableEventLoop.doRunTask(ReentrantBlockableEventLoop.java:24) ~[?:?]
at net.minecraft.server.MinecraftServer.doRunTask(MinecraftServer.java:1427) ~[purpur-1.18.2.jar:git-Purpur-1632]
at net.minecraft.server.MinecraftServer.d(MinecraftServer.java:188) ~[purpur-1.18.2.jar:git-Purpur-1632]
at net.minecraft.util.thread.BlockableEventLoop.pollTask(BlockableEventLoop.java:126) ~[?:?]
at net.minecraft.server.MinecraftServer.pollTaskInternal(MinecraftServer.java:1404) ~[purpur-1.18.2.jar:git-Purpur-1632]
at net.minecraft.server.MinecraftServer.pollTask(MinecraftServer.java:1397) ~[purpur-1.18.2.jar:git-Purpur-1632]
at net.minecraft.util.thread.BlockableEventLoop.managedBlock(BlockableEventLoop.java:136) ~[?:?]
at net.minecraft.server.MinecraftServer.waitUntilNextTick(MinecraftServer.java:1375) ~[purpur-1.18.2.jar:git-Purpur-1632]
at net.minecraft.server.MinecraftServer.runServer(MinecraftServer.java:1257) ~[purpur-1.18.2.jar:git-Purpur-1632]
at net.minecraft.server.MinecraftServer.lambda$spin$1(MinecraftServer.java:321) ~[purpur-1.18.2.jar:git-Purpur-1632]
at java.lang.Thread.run(Thread.java:891) ~[?:?]
Caused by: java.lang.ClassNotFoundException: net.minecraft.server.v1_16_R3.PacketPlayOutHeldItemSlot
at org.bukkit.plugin.java.PluginClassLoader.loadClass0(PluginClassLoader.java:153) ~[purpur-api-1.18.2-R0.1-SNAPSHOT.jar:?]
at org.bukkit.plugin.java.PluginClassLoader.loadClass(PluginClassLoader.java:105) ~[purpur-api-1.18.2-R0.1-SNAPSHOT.jar:?]
at java.lang.ClassLoader.loadClass(ClassLoader.java:520) ~[?:?]
... 32 more
```
使用的服务器核心: `Current: git-Purpur-1632 (MC: 1.18.2)`
|
process
|
使用滚轮切换回复内容 开启时出现的错误 could not pass event playeritemheldevent to chemdah java lang noclassdeffounderror net minecraft server packetplayouthelditemslot at ink ptms chemdah core conversation theme themechat onitemheld themechat kt at java lang invoke methodhandle invokewitharguments methodhandle java at ink ptms chemdah taboolib library reflex javaclassmethod invoke javaclassmethod kt at ink ptms chemdah taboolib common platform event eventbus invoke eventbus kt at ink ptms chemdah taboolib common platform event eventbus invoke default eventbus kt at ink ptms chemdah taboolib common platform event eventbus registerbukkit invoke eventbus kt at ink ptms chemdah taboolib common platform event eventbus registerbukkit invoke eventbus kt at ink ptms chemdah taboolib common platform function listenerkt registerbukkitlistener invoke listener kt at ink ptms chemdah taboolib common platform function listenerkt registerbukkitlistener invoke listener kt at ink ptms chemdah taboolib platform bukkitlistener registerlistener listener invoke bukkitlistener kt at ink ptms chemdah taboolib platform bukkitlistener registerlistener listener invoke bukkitlistener kt at ink ptms chemdah taboolib platform bukkitlistener bukkitlistener execute bukkitlistener kt at co aikar timings timedeventexecutor execute timedeventexecutor java at org bukkit plugin registeredlistener callevent registeredlistener java at org bukkit plugin simplepluginmanager callevent simplepluginmanager java at net minecraft server network servergamepacketlistenerimpl handlesetcarrieditem servergamepacketlistenerimpl java at net minecraft network protocol game serverboundsetcarrieditempacket handle serverboundsetcarrieditempacket java at net minecraft network protocol game serverboundsetcarrieditempacket a serverboundsetcarrieditempacket java at net minecraft network protocol packetutils lambda ensurerunningonsamethread packetutils java at net minecraft server ticktask run ticktask java at net minecraft util thread blockableeventloop doruntask blockableeventloop java at net minecraft util thread reentrantblockableeventloop doruntask reentrantblockableeventloop java at net minecraft server minecraftserver doruntask minecraftserver java at net minecraft server minecraftserver d minecraftserver java at net minecraft util thread blockableeventloop polltask blockableeventloop java at net minecraft server minecraftserver polltaskinternal minecraftserver java at net minecraft server minecraftserver polltask minecraftserver java at net minecraft util thread blockableeventloop managedblock blockableeventloop java at net minecraft server minecraftserver waituntilnexttick minecraftserver java at net minecraft server minecraftserver runserver minecraftserver java at net minecraft server minecraftserver lambda spin minecraftserver java at java lang thread run thread java caused by java lang classnotfoundexception net minecraft server packetplayouthelditemslot at org bukkit plugin java pluginclassloader pluginclassloader java at org bukkit plugin java pluginclassloader loadclass pluginclassloader java at java lang classloader loadclass classloader java more 使用的服务器核心 current git purpur mc
| 1
|
224,872
| 7,473,573,076
|
IssuesEvent
|
2018-04-03 15:43:27
|
neuropoly/axondeepseg
|
https://api.github.com/repos/neuropoly/axondeepseg
|
closed
|
Implement continuous integration and integrity checking
|
priority:HIGH
|
to avoid this kind of issue: https://github.com/neuropoly/axondeepseg/issues/59
we could use continuous integration (CI) service such as [Travis](https://travis-ci.org/).
|
1.0
|
Implement continuous integration and integrity checking - to avoid this kind of issue: https://github.com/neuropoly/axondeepseg/issues/59
we could use continuous integration (CI) service such as [Travis](https://travis-ci.org/).
|
non_process
|
implement continuous integration and integrity checking to avoid this kind of issue we could use continuous integration ci service such as
| 0
|
6,485
| 9,554,722,250
|
IssuesEvent
|
2019-05-02 23:11:36
|
ncbo/bioportal-project
|
https://api.github.com/repos/ncbo/bioportal-project
|
closed
|
NEO: fails to load
|
in progress ontology processing problem
|
The [NEO ontology](http://bioportal.bioontology.org/ontologies/NEO) shows "Error Rdf" status in the BioPortal UI. Stack trace from parsing log file:
```
E, [2019-04-22T11:27:38.785943 #5770] ERROR -- : ["Exception: could not `LANG=C grep -v '_:genid' /tmp/file_nobnodes20190422-5770-1hw2ajk/data.nt > /tmp/file_nobnodes20190422-5770-1hw2ajk/data_no_bnodes.nt`:
/srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.5.0/bundler/gems/goo-0bc8c933f3f8/lib/goo/sparql/client.rb:66:in `bnodes_filter_file'
/srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.5.0/bundler/gems/goo-0bc8c933f3f8/lib/goo/sparql/client.rb:81:in `append_triples_no_bnodes'
/srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.5.0/bundler/gems/goo-0bc8c933f3f8/lib/goo/sparql/client.rb:129:in `put_triples'
/srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.5.0/bundler/gems/ontologies_linked_data-548e7b1e4fb8/lib/ontologies_linked_data/models/ontology_submission.rb:1529:in `delete_and_append'
/srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.5.0/bundler/gems/ontologies_linked_data-548e7b1e4fb8/lib/ontologies_linked_data/models/ontology_submission.rb:475:in `generate_rdf'
/srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.5.0/bundler/gems/ontologies_linked_data-548e7b1e4fb8/lib/ontologies_linked_data/models/ontology_submission.rb:973:in `process_submission'
/srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:177:in `process_submission'
bin/ncbo_ontology_process:98:in `block in <main>'
bin/ncbo_ontology_process:81:in `each'
bin/ncbo_ontology_process:81:in `<main>'"]
```
Oddly, the owlapi.xrdf file is well formed, but lacks any ontology data:
```
<?xml version="1.0"?>
<rdf:RDF xmlns="http://www.w3.org/2002/07/owl#"
xml:base="http://www.w3.org/2002/07/owl"
xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"
xmlns:owl="http://www.w3.org/2002/07/owl#"
xmlns:xml="http://www.w3.org/XML/1998/namespace"
xmlns:xsd="http://www.w3.org/2001/XMLSchema#"
xmlns:rdfs="http://www.w3.org/2000/01/rdf-schema#">
<Ontology/>
</rdf:RDF>
<!-- Generated by the OWL API (version 4.3.2) https://github.com/owlcs/owlapi -->
```
|
1.0
|
NEO: fails to load - The [NEO ontology](http://bioportal.bioontology.org/ontologies/NEO) shows "Error Rdf" status in the BioPortal UI. Stack trace from parsing log file:
```
E, [2019-04-22T11:27:38.785943 #5770] ERROR -- : ["Exception: could not `LANG=C grep -v '_:genid' /tmp/file_nobnodes20190422-5770-1hw2ajk/data.nt > /tmp/file_nobnodes20190422-5770-1hw2ajk/data_no_bnodes.nt`:
/srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.5.0/bundler/gems/goo-0bc8c933f3f8/lib/goo/sparql/client.rb:66:in `bnodes_filter_file'
/srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.5.0/bundler/gems/goo-0bc8c933f3f8/lib/goo/sparql/client.rb:81:in `append_triples_no_bnodes'
/srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.5.0/bundler/gems/goo-0bc8c933f3f8/lib/goo/sparql/client.rb:129:in `put_triples'
/srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.5.0/bundler/gems/ontologies_linked_data-548e7b1e4fb8/lib/ontologies_linked_data/models/ontology_submission.rb:1529:in `delete_and_append'
/srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.5.0/bundler/gems/ontologies_linked_data-548e7b1e4fb8/lib/ontologies_linked_data/models/ontology_submission.rb:475:in `generate_rdf'
/srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.5.0/bundler/gems/ontologies_linked_data-548e7b1e4fb8/lib/ontologies_linked_data/models/ontology_submission.rb:973:in `process_submission'
/srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:177:in `process_submission'
bin/ncbo_ontology_process:98:in `block in <main>'
bin/ncbo_ontology_process:81:in `each'
bin/ncbo_ontology_process:81:in `<main>'"]
```
Oddly, the owlapi.xrdf file is well formed, but lacks any ontology data:
```
<?xml version="1.0"?>
<rdf:RDF xmlns="http://www.w3.org/2002/07/owl#"
xml:base="http://www.w3.org/2002/07/owl"
xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"
xmlns:owl="http://www.w3.org/2002/07/owl#"
xmlns:xml="http://www.w3.org/XML/1998/namespace"
xmlns:xsd="http://www.w3.org/2001/XMLSchema#"
xmlns:rdfs="http://www.w3.org/2000/01/rdf-schema#">
<Ontology/>
</rdf:RDF>
<!-- Generated by the OWL API (version 4.3.2) https://github.com/owlcs/owlapi -->
```
|
process
|
neo fails to load the shows error rdf status in the bioportal ui stack trace from parsing log file e error exception could not lang c grep v genid tmp file data nt tmp file data no bnodes nt srv ncbo ncbo cron vendor bundle ruby bundler gems goo lib goo sparql client rb in bnodes filter file srv ncbo ncbo cron vendor bundle ruby bundler gems goo lib goo sparql client rb in append triples no bnodes srv ncbo ncbo cron vendor bundle ruby bundler gems goo lib goo sparql client rb in put triples srv ncbo ncbo cron vendor bundle ruby bundler gems ontologies linked data lib ontologies linked data models ontology submission rb in delete and append srv ncbo ncbo cron vendor bundle ruby bundler gems ontologies linked data lib ontologies linked data models ontology submission rb in generate rdf srv ncbo ncbo cron vendor bundle ruby bundler gems ontologies linked data lib ontologies linked data models ontology submission rb in process submission srv ncbo ncbo cron lib ncbo cron ontology submission parser rb in process submission bin ncbo ontology process in block in bin ncbo ontology process in each bin ncbo ontology process in oddly the owlapi xrdf file is well formed but lacks any ontology data rdf rdf xmlns xml base xmlns rdf xmlns owl xmlns xml xmlns xsd xmlns rdfs
| 1
|
129,039
| 10,560,941,669
|
IssuesEvent
|
2019-10-04 14:53:01
|
mozilla/iris_firefox
|
https://api.github.com/repos/mozilla/iris_firefox
|
opened
|
Fix default_search_code_yandex for treeherder Nightly
|
regression target-nightly test case
|
This is failing at the first code validation assert. There aren't any debug images here because it's a bunch of keyboard and copy to clipboard stuff. As such I can't tell if the search page was loaded or if it's being captcha's. I am going to have to add some sort of pattern check just to see what's on screen around the time of the failure.
|
1.0
|
Fix default_search_code_yandex for treeherder Nightly - This is failing at the first code validation assert. There aren't any debug images here because it's a bunch of keyboard and copy to clipboard stuff. As such I can't tell if the search page was loaded or if it's being captcha's. I am going to have to add some sort of pattern check just to see what's on screen around the time of the failure.
|
non_process
|
fix default search code yandex for treeherder nightly this is failing at the first code validation assert there aren t any debug images here because it s a bunch of keyboard and copy to clipboard stuff as such i can t tell if the search page was loaded or if it s being captcha s i am going to have to add some sort of pattern check just to see what s on screen around the time of the failure
| 0
|
1,260
| 5,348,482,978
|
IssuesEvent
|
2017-02-18 05:30:02
|
diofant/diofant
|
https://api.github.com/repos/diofant/diofant
|
opened
|
Use "new" style for string formatting
|
maintainability
|
I.e. ``"{0:s}".format("spam")`` instead of ``"%s" % "spam"``
|
True
|
Use "new" style for string formatting - I.e. ``"{0:s}".format("spam")`` instead of ``"%s" % "spam"``
|
non_process
|
use new style for string formatting i e s format spam instead of s spam
| 0
|
10,784
| 13,608,982,336
|
IssuesEvent
|
2020-09-23 03:55:53
|
googleapis/java-dataproc
|
https://api.github.com/repos/googleapis/java-dataproc
|
closed
|
Dependency Dashboard
|
api: dataproc type: process
|
This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/org.apache.maven.plugins-maven-project-info-reports-plugin-3.x -->build(deps): update dependency org.apache.maven.plugins:maven-project-info-reports-plugin to v3.1.1
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-dataproc-1.x -->chore(deps): update dependency com.google.cloud:google-cloud-dataproc to v1.1.0
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-storage-1.x -->deps: update dependency com.google.cloud:google-cloud-storage to v1.113.1
- [ ] <!-- rebase-all-open-prs -->**Check this option to rebase all the above open PRs at once**
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
1.0
|
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/org.apache.maven.plugins-maven-project-info-reports-plugin-3.x -->build(deps): update dependency org.apache.maven.plugins:maven-project-info-reports-plugin to v3.1.1
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-dataproc-1.x -->chore(deps): update dependency com.google.cloud:google-cloud-dataproc to v1.1.0
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-storage-1.x -->deps: update dependency com.google.cloud:google-cloud-storage to v1.113.1
- [ ] <!-- rebase-all-open-prs -->**Check this option to rebase all the above open PRs at once**
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
process
|
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any build deps update dependency org apache maven plugins maven project info reports plugin to chore deps update dependency com google cloud google cloud dataproc to deps update dependency com google cloud google cloud storage to check this option to rebase all the above open prs at once check this box to trigger a request for renovate to run again on this repository
| 1
|
174,163
| 27,586,500,326
|
IssuesEvent
|
2023-03-08 20:12:24
|
department-of-veterans-affairs/vets-design-system-documentation
|
https://api.github.com/repos/department-of-veterans-affairs/vets-design-system-documentation
|
opened
|
Process List css bug
|
bug vsp-design-system-team
|
# Bug Report
- [x] I’ve searched for any related issues and avoided creating a duplicate issue.
## What happened
I found a minor css bug. There are a few lines of css that could be added to make [these process steps](https://github.com/department-of-veterans-affairs/va-digital-services-platform-docs/blob/064760c4e33689e463a6e24efdcb713f07a004ef/sass/modules/_m-process-list.scss#L62) numbers have proper circles around them rather than ovals that they are now.
If we added the following lines of css to the `:before` element on this [sass file](https://github.com/department-of-veterans-affairs/va-digital-services-platform-docs/blob/064760c4e33689e463a6e24efdcb713f07a004ef/sass/modules/_m-process-list.scss#L62) at `line 62` it'd fix the issue.
```
height: 2em;
line-height: 1.575em;
```
<img width="1247" alt="Screenshot 2023-03-08 at 12 05 53 PM" src="https://user-images.githubusercontent.com/1631062/223836047-4e76b46a-1291-4a53-b768-7ee3a47d8c4a.png">
It looks like what's in production doesn't match what's in StoryBook.
## What I expected to happen
According [to the component](https://www.sketch.com/s/610156b6-f281-4497-81f3-64454fc72156/a/WKYjOpA#Inspect), it should show a circle around the number, but it shows an oval instead. After applying CSS updates it should be a nice round circle like this:

However, when looking at the component in [StoryBook](https://design.va.gov/storybook/?path=/docs/components-va-process-list--default), it is correct. And inspecting it, it shows that's it all set rems. Whereas the .scss file is all relative using ems.
<img width="1575" alt="Screenshot 2023-03-08 at 11 59 27 AM" src="https://user-images.githubusercontent.com/1631062/223833712-4946d8c5-409c-4545-9ace-08bc9e4908b7.png">
## Reproducing
- Formation version: Latest
- Device: Macbook Pro
- Browser: Chrome
Steps to reproduce:
1. View in Chrome: https://www.va.gov/manage-va-debt/request-debt-help-form-5655/introduction
2. Inspect code
## Urgency
How urgent is this request? Please select the approriate option below and/or provide details
- [ ] This bug is blocking work currently in progress
- [ ] This bug is affecting work currently in progress but we have a workaround
- [ ] This bug is blocking work planned within the next few sprints
- [x] This bug is not blocking any work
- [ ] Other
### Details
|
1.0
|
Process List css bug - # Bug Report
- [x] I’ve searched for any related issues and avoided creating a duplicate issue.
## What happened
I found a minor css bug. There are a few lines of css that could be added to make [these process steps](https://github.com/department-of-veterans-affairs/va-digital-services-platform-docs/blob/064760c4e33689e463a6e24efdcb713f07a004ef/sass/modules/_m-process-list.scss#L62) numbers have proper circles around them rather than ovals that they are now.
If we added the following lines of css to the `:before` element on this [sass file](https://github.com/department-of-veterans-affairs/va-digital-services-platform-docs/blob/064760c4e33689e463a6e24efdcb713f07a004ef/sass/modules/_m-process-list.scss#L62) at `line 62` it'd fix the issue.
```
height: 2em;
line-height: 1.575em;
```
<img width="1247" alt="Screenshot 2023-03-08 at 12 05 53 PM" src="https://user-images.githubusercontent.com/1631062/223836047-4e76b46a-1291-4a53-b768-7ee3a47d8c4a.png">
It looks like what's in production doesn't match what's in StoryBook.
## What I expected to happen
According [to the component](https://www.sketch.com/s/610156b6-f281-4497-81f3-64454fc72156/a/WKYjOpA#Inspect), it should show a circle around the number, but it shows an oval instead. After applying CSS updates it should be a nice round circle like this:

However, when looking at the component in [StoryBook](https://design.va.gov/storybook/?path=/docs/components-va-process-list--default), it is correct. And inspecting it, it shows that's it all set rems. Whereas the .scss file is all relative using ems.
<img width="1575" alt="Screenshot 2023-03-08 at 11 59 27 AM" src="https://user-images.githubusercontent.com/1631062/223833712-4946d8c5-409c-4545-9ace-08bc9e4908b7.png">
## Reproducing
- Formation version: Latest
- Device: Macbook Pro
- Browser: Chrome
Steps to reproduce:
1. View in Chrome: https://www.va.gov/manage-va-debt/request-debt-help-form-5655/introduction
2. Inspect code
## Urgency
How urgent is this request? Please select the approriate option below and/or provide details
- [ ] This bug is blocking work currently in progress
- [ ] This bug is affecting work currently in progress but we have a workaround
- [ ] This bug is blocking work planned within the next few sprints
- [x] This bug is not blocking any work
- [ ] Other
### Details
|
non_process
|
process list css bug bug report i’ve searched for any related issues and avoided creating a duplicate issue what happened i found a minor css bug there are a few lines of css that could be added to make numbers have proper circles around them rather than ovals that they are now if we added the following lines of css to the before element on this at line it d fix the issue height line height img width alt screenshot at pm src it looks like what s in production doesn t match what s in storybook what i expected to happen according it should show a circle around the number but it shows an oval instead after applying css updates it should be a nice round circle like this however when looking at the component in it is correct and inspecting it it shows that s it all set rems whereas the scss file is all relative using ems img width alt screenshot at am src reproducing formation version latest device macbook pro browser chrome steps to reproduce view in chrome inspect code urgency how urgent is this request please select the approriate option below and or provide details this bug is blocking work currently in progress this bug is affecting work currently in progress but we have a workaround this bug is blocking work planned within the next few sprints this bug is not blocking any work other details
| 0
|
16,775
| 21,958,526,152
|
IssuesEvent
|
2022-05-24 14:03:11
|
GoogleCloudPlatform/anthos-samples
|
https://api.github.com/repos/GoogleCloudPlatform/anthos-samples
|
closed
|
CI flakiness
|
type: bug type: process priority: p2 samples
|
@Shabirmean and I noticed flaky behaviour occurring.
We believe that it's due to github actions/checkout@v3, where the perms get overwritten
Every once in a while we have to ssh into the runners and change the permissions, or else the CI errors out with
```
Error: fatal: --local can only be used inside a git repository
Deleting the contents of '/home/$USER/actions-runner/_work/anthos-samples/anthos-samples'
Error: Command failed: rm -rf "/home/$USER/actions-runner/_work/anthos-samples/anthos-samples/.git"
```
|
1.0
|
CI flakiness - @Shabirmean and I noticed flaky behaviour occurring.
We believe that it's due to github actions/checkout@v3, where the perms get overwritten
Every once in a while we have to ssh into the runners and change the permissions, or else the CI errors out with
```
Error: fatal: --local can only be used inside a git repository
Deleting the contents of '/home/$USER/actions-runner/_work/anthos-samples/anthos-samples'
Error: Command failed: rm -rf "/home/$USER/actions-runner/_work/anthos-samples/anthos-samples/.git"
```
|
process
|
ci flakiness shabirmean and i noticed flaky behaviour occurring we believe that it s due to github actions checkout where the perms get overwritten every once in a while we have to ssh into the runners and change the permissions or else the ci errors out with error fatal local can only be used inside a git repository deleting the contents of home user actions runner work anthos samples anthos samples error command failed rm rf home user actions runner work anthos samples anthos samples git
| 1
|
4,685
| 7,522,508,585
|
IssuesEvent
|
2018-04-12 20:40:11
|
googlegenomics/gcp-variant-transforms
|
https://api.github.com/repos/googlegenomics/gcp-variant-transforms
|
closed
|
Add documentation on how to run the pipeline in a particular zone/region
|
P1 process
|
There is both region [1] and zone arguments (and in two places! one in the pipelines API and one in the Dataflow API). We should document how to use these features.
[1] https://cloud.google.com/dataflow/docs/concepts/regional-endpoints#supported_regional_endpoints
|
1.0
|
Add documentation on how to run the pipeline in a particular zone/region - There is both region [1] and zone arguments (and in two places! one in the pipelines API and one in the Dataflow API). We should document how to use these features.
[1] https://cloud.google.com/dataflow/docs/concepts/regional-endpoints#supported_regional_endpoints
|
process
|
add documentation on how to run the pipeline in a particular zone region there is both region and zone arguments and in two places one in the pipelines api and one in the dataflow api we should document how to use these features
| 1
|
14,206
| 17,103,899,225
|
IssuesEvent
|
2021-07-09 14:54:43
|
3drepo/3drepobouncer
|
https://api.github.com/repos/3drepo/3drepobouncer
|
closed
|
DWG/DXF support
|
In Staging feature file processing
|
<!-- FEATURE TEMPLATE (delete as appropriate) -->
<!-- Remember to tag this issue as a feature! -->
### Description
Add support to dwg/dxf via Teigha library
### Goals
- [ ] DWG model file should be accepted and correct geometry generated
- [ ] DXF model file should be accepted and correct geometry generated
### Tasks
- [x] Add code to read dwg/dxf, use DGN Exporter to convert to dgn db.
- [x] Clean up code
- [x] Resolve crashing on clean up
- [ ] add test examples
|
1.0
|
DWG/DXF support - <!-- FEATURE TEMPLATE (delete as appropriate) -->
<!-- Remember to tag this issue as a feature! -->
### Description
Add support to dwg/dxf via Teigha library
### Goals
- [ ] DWG model file should be accepted and correct geometry generated
- [ ] DXF model file should be accepted and correct geometry generated
### Tasks
- [x] Add code to read dwg/dxf, use DGN Exporter to convert to dgn db.
- [x] Clean up code
- [x] Resolve crashing on clean up
- [ ] add test examples
|
process
|
dwg dxf support description add support to dwg dxf via teigha library goals dwg model file should be accepted and correct geometry generated dxf model file should be accepted and correct geometry generated tasks add code to read dwg dxf use dgn exporter to convert to dgn db clean up code resolve crashing on clean up add test examples
| 1
|
61,295
| 7,458,854,974
|
IssuesEvent
|
2018-03-30 12:41:38
|
dotnet/project-system
|
https://api.github.com/repos/dotnet/project-system
|
closed
|
Selecting the Application page on AppDesigner, show "PropPageDesignerRootComponent" in Properties window
|
Bug Feature-AppDesigner Up for Grabs
|

|
1.0
|
Selecting the Application page on AppDesigner, show "PropPageDesignerRootComponent" in Properties window - 
|
non_process
|
selecting the application page on appdesigner show proppagedesignerrootcomponent in properties window
| 0
|
14,775
| 18,051,321,816
|
IssuesEvent
|
2021-09-19 19:55:20
|
Leviatan-Analytics/LA-data-processing
|
https://api.github.com/repos/Leviatan-Analytics/LA-data-processing
|
closed
|
Create pro match analysis data schema [2]
|
Data Processing Week 3 Sprint 4
|
Develop schema for saving a pro match manual analysis.
|
1.0
|
Create pro match analysis data schema [2] - Develop schema for saving a pro match manual analysis.
|
process
|
create pro match analysis data schema develop schema for saving a pro match manual analysis
| 1
|
751,388
| 26,242,885,854
|
IssuesEvent
|
2023-01-05 13:02:33
|
matrixorigin/matrixone
|
https://api.github.com/repos/matrixorigin/matrixone
|
opened
|
[Feature Request]: add information function for the account;
|
priority/p0 kind/feature
|
### Is there an existing issue for the same feature request?
- [X] I have checked the existing issues.
### Is your feature request related to a problem?
```Markdown
mo cloud
```
### Describe the feature you'd like
1, select current_account();
print the account information:
column1:account_name;
column2:account_id;
column3:use_name;
column4:use_id;
column5:role_name;
column6:role_id;
2, select current_account_id();
3, select current_account_name();
4, select current_role_id();
5, select_current_role_name();
6, select_current_user_name();
7, select_current_user_id();
### Describe implementation you've considered
_No response_
### Documentation, Adoption, Use Case, Migration Strategy
_No response_
### Additional information
_No response_
|
1.0
|
[Feature Request]: add information function for the account; - ### Is there an existing issue for the same feature request?
- [X] I have checked the existing issues.
### Is your feature request related to a problem?
```Markdown
mo cloud
```
### Describe the feature you'd like
1, select current_account();
print the account information:
column1:account_name;
column2:account_id;
column3:use_name;
column4:use_id;
column5:role_name;
column6:role_id;
2, select current_account_id();
3, select current_account_name();
4, select current_role_id();
5, select_current_role_name();
6, select_current_user_name();
7, select_current_user_id();
### Describe implementation you've considered
_No response_
### Documentation, Adoption, Use Case, Migration Strategy
_No response_
### Additional information
_No response_
|
non_process
|
add information function for the account is there an existing issue for the same feature request i have checked the existing issues is your feature request related to a problem markdown mo cloud describe the feature you d like select current account print the account information account name account id use name use id role name role id select current account id select current account name select current role id select current role name select current user name select current user id describe implementation you ve considered no response documentation adoption use case migration strategy no response additional information no response
| 0
|
22,499
| 31,476,635,849
|
IssuesEvent
|
2023-08-30 11:11:14
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
skypilot-nightly 1.0.0.dev20230830 has 2 GuardDog issues
|
guarddog exec-base64 silent-process-execution
|
https://pypi.org/project/skypilot-nightly
https://inspector.pypi.io/project/skypilot-nightly
```{
"dependency": "skypilot-nightly",
"version": "1.0.0.dev20230830",
"result": {
"issues": 2,
"errors": {},
"results": {
"exec-base64": [
{
"location": "skypilot-nightly-1.0.0.dev20230830/sky/cloud_stores.py:116",
"code": " p = subprocess.run(command,\n stdout=subprocess.PIPE,\n shell=True,\n check=True,\n executable='/bin/bash')",
"message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"
}
],
"silent-process-execution": [
{
"location": "skypilot-nightly-1.0.0.dev20230830/sky/skylet/log_lib.py:219",
"code": " subprocess.Popen(\n daemon_cmd,\n start_new_session=True,\n # Suppress output\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n # Disa... )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmptug225o8/skypilot-nightly"
}
}```
|
1.0
|
skypilot-nightly 1.0.0.dev20230830 has 2 GuardDog issues - https://pypi.org/project/skypilot-nightly
https://inspector.pypi.io/project/skypilot-nightly
```{
"dependency": "skypilot-nightly",
"version": "1.0.0.dev20230830",
"result": {
"issues": 2,
"errors": {},
"results": {
"exec-base64": [
{
"location": "skypilot-nightly-1.0.0.dev20230830/sky/cloud_stores.py:116",
"code": " p = subprocess.run(command,\n stdout=subprocess.PIPE,\n shell=True,\n check=True,\n executable='/bin/bash')",
"message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"
}
],
"silent-process-execution": [
{
"location": "skypilot-nightly-1.0.0.dev20230830/sky/skylet/log_lib.py:219",
"code": " subprocess.Popen(\n daemon_cmd,\n start_new_session=True,\n # Suppress output\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n # Disa... )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmptug225o8/skypilot-nightly"
}
}```
|
process
|
skypilot nightly has guarddog issues dependency skypilot nightly version result issues errors results exec location skypilot nightly sky cloud stores py code p subprocess run command n stdout subprocess pipe n shell true n check true n executable bin bash message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n silent process execution location skypilot nightly sky skylet log lib py code subprocess popen n daemon cmd n start new session true n suppress output n stdout subprocess devnull n stderr subprocess devnull n disa message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp skypilot nightly
| 1
|
19,416
| 25,560,957,175
|
IssuesEvent
|
2022-11-30 10:43:22
|
aiidateam/aiida-core
|
https://api.github.com/repos/aiidateam/aiida-core
|
opened
|
Restore support for dynamic nested namespaces in inputs for process functions
|
type/bug topic/engine topic/processes
|
Before v2.0 the following was possible:
```python
@workfunction
def function(**kwargs):
return kwargs
inputs = {
'nested': {
'namespace': {
'int': orm.Int(1).store()
}
}
}
results, node = function.run_get_node(**inputs)
assert results == inputs
assert node.get_incoming().nested() == inputs
```
In v2.0 and up this raises a `ValueError`:
```
ValueError: Error occurred validating port 'inputs.level_1': Invalid type <class 'dict'> for dynamic port value: expected <class 'aiida.orm.nodes.data.data.Data'>
```
|
1.0
|
Restore support for dynamic nested namespaces in inputs for process functions - Before v2.0 the following was possible:
```python
@workfunction
def function(**kwargs):
return kwargs
inputs = {
'nested': {
'namespace': {
'int': orm.Int(1).store()
}
}
}
results, node = function.run_get_node(**inputs)
assert results == inputs
assert node.get_incoming().nested() == inputs
```
In v2.0 and up this raises a `ValueError`:
```
ValueError: Error occurred validating port 'inputs.level_1': Invalid type <class 'dict'> for dynamic port value: expected <class 'aiida.orm.nodes.data.data.Data'>
```
|
process
|
restore support for dynamic nested namespaces in inputs for process functions before the following was possible python workfunction def function kwargs return kwargs inputs nested namespace int orm int store results node function run get node inputs assert results inputs assert node get incoming nested inputs in and up this raises a valueerror valueerror error occurred validating port inputs level invalid type for dynamic port value expected
| 1
|
15,825
| 20,018,702,159
|
IssuesEvent
|
2022-02-01 14:33:49
|
elastic/beats
|
https://api.github.com/repos/elastic/beats
|
opened
|
Create Syslog Processor
|
enhancement :Processors Team:Security-External Integrations
|
## Overview
Syslog parsing in Beats is currently implemented as a dedicated input. It supports accepting data over UDP, TCP, and Unix sockets. Because it is implemented as an input, the parsing cannot be used by other inputs such as filestream, httpjson, or kafka.
The goal is to decouple the input from parsing by moving the parsing to a processor. This also provides the opportunity for to improve other areas of processing at the same time.
## Summary of Intended Changes
- Implement a parser for syslog
- Strongly follow RFC 3164, allow for flexible timestamps. RFC 3146 is also rather vague in areas, and a best effort parser will be necessary for some elements (tag and content fields, for instance)
- Strictly follow RFC 5424
- Automatically detect format by default
- Parsers will allow multiline to work with syslog
- Always retain event.original
- Populate ECS fields
- Deprecate syslog input
## Flexible Timestamp Parsing for RFC 3146
In addition to the RFC-defined timestamp format ("Mmm dd hh:mm:ss", e.g. Feb 5 17:32:18), the processor should also allow for parsing RFC 3339 timestamps. There tends to be low adherence to this part of the specfication due to lack of subsecond precision, year, and timezone. These deviations will be documented with examples.
## Other Notes
- Cisco does not follow RFCs. Cisco modules will use plain udp/tcp/tls inputs (without parser) and use grok or dissect to do best effort syslog header parser. Even if header parsing fails they should pick up parsing from the Cisco message ID (e.g. %SEC-6-IPACCESSLOGP)
## See Also
- https://datatracker.ietf.org/doc/html/rfc3164
- https://datatracker.ietf.org/doc/html/rfc5424
|
1.0
|
Create Syslog Processor - ## Overview
Syslog parsing in Beats is currently implemented as a dedicated input. It supports accepting data over UDP, TCP, and Unix sockets. Because it is implemented as an input, the parsing cannot be used by other inputs such as filestream, httpjson, or kafka.
The goal is to decouple the input from parsing by moving the parsing to a processor. This also provides the opportunity for to improve other areas of processing at the same time.
## Summary of Intended Changes
- Implement a parser for syslog
- Strongly follow RFC 3164, allow for flexible timestamps. RFC 3146 is also rather vague in areas, and a best effort parser will be necessary for some elements (tag and content fields, for instance)
- Strictly follow RFC 5424
- Automatically detect format by default
- Parsers will allow multiline to work with syslog
- Always retain event.original
- Populate ECS fields
- Deprecate syslog input
## Flexible Timestamp Parsing for RFC 3146
In addition to the RFC-defined timestamp format ("Mmm dd hh:mm:ss", e.g. Feb 5 17:32:18), the processor should also allow for parsing RFC 3339 timestamps. There tends to be low adherence to this part of the specfication due to lack of subsecond precision, year, and timezone. These deviations will be documented with examples.
## Other Notes
- Cisco does not follow RFCs. Cisco modules will use plain udp/tcp/tls inputs (without parser) and use grok or dissect to do best effort syslog header parser. Even if header parsing fails they should pick up parsing from the Cisco message ID (e.g. %SEC-6-IPACCESSLOGP)
## See Also
- https://datatracker.ietf.org/doc/html/rfc3164
- https://datatracker.ietf.org/doc/html/rfc5424
|
process
|
create syslog processor overview syslog parsing in beats is currently implemented as a dedicated input it supports accepting data over udp tcp and unix sockets because it is implemented as an input the parsing cannot be used by other inputs such as filestream httpjson or kafka the goal is to decouple the input from parsing by moving the parsing to a processor this also provides the opportunity for to improve other areas of processing at the same time summary of intended changes implement a parser for syslog strongly follow rfc allow for flexible timestamps rfc is also rather vague in areas and a best effort parser will be necessary for some elements tag and content fields for instance strictly follow rfc automatically detect format by default parsers will allow multiline to work with syslog always retain event original populate ecs fields deprecate syslog input flexible timestamp parsing for rfc in addition to the rfc defined timestamp format mmm dd hh mm ss e g feb the processor should also allow for parsing rfc timestamps there tends to be low adherence to this part of the specfication due to lack of subsecond precision year and timezone these deviations will be documented with examples other notes cisco does not follow rfcs cisco modules will use plain udp tcp tls inputs without parser and use grok or dissect to do best effort syslog header parser even if header parsing fails they should pick up parsing from the cisco message id e g sec ipaccesslogp see also
| 1
|
21,068
| 3,455,776,753
|
IssuesEvent
|
2015-12-17 21:40:53
|
netty/netty
|
https://api.github.com/repos/netty/netty
|
opened
|
HTTP/2 DefaultHttp2RemoteFlowController Stream writability notification broken
|
defect
|
`DefaultHttp2RemoteFlowController.ListenerWritabilityMonitor` no longer reliably detects when a stream's writability change occurs. `ListenerWritabilityMonitor` was implemented to avoid duplicating iteration over all streams when possible and instead was relying on the `PriorityStreamByteDistributor` to call `write` for each stream during its iteration process. However the new `StreamByteDistributor` classes do not do an iteration over all active streams and so this assumption is now invalid.
The impact is isolated to stream writability change notifications and no impact unless you explicitly add a listener.
|
1.0
|
HTTP/2 DefaultHttp2RemoteFlowController Stream writability notification broken - `DefaultHttp2RemoteFlowController.ListenerWritabilityMonitor` no longer reliably detects when a stream's writability change occurs. `ListenerWritabilityMonitor` was implemented to avoid duplicating iteration over all streams when possible and instead was relying on the `PriorityStreamByteDistributor` to call `write` for each stream during its iteration process. However the new `StreamByteDistributor` classes do not do an iteration over all active streams and so this assumption is now invalid.
The impact is isolated to stream writability change notifications and no impact unless you explicitly add a listener.
|
non_process
|
http stream writability notification broken listenerwritabilitymonitor no longer reliably detects when a stream s writability change occurs listenerwritabilitymonitor was implemented to avoid duplicating iteration over all streams when possible and instead was relying on the prioritystreambytedistributor to call write for each stream during its iteration process however the new streambytedistributor classes do not do an iteration over all active streams and so this assumption is now invalid the impact is isolated to stream writability change notifications and no impact unless you explicitly add a listener
| 0
|
109,913
| 4,415,350,500
|
IssuesEvent
|
2016-08-14 01:11:02
|
MinetestForFun/server-minetestforfun-skyblock
|
https://api.github.com/repos/MinetestForFun/server-minetestforfun-skyblock
|
opened
|
All passwords have been reset for accounts older than a month!
|
Priority: High
|
I have just verified ANY player that has not logged into the game for a month or longer has had there password reset!
:large_orange_diamond:
|
1.0
|
All passwords have been reset for accounts older than a month! - I have just verified ANY player that has not logged into the game for a month or longer has had there password reset!
:large_orange_diamond:
|
non_process
|
all passwords have been reset for accounts older than a month i have just verified any player that has not logged into the game for a month or longer has had there password reset large orange diamond
| 0
|
3,346
| 6,486,457,873
|
IssuesEvent
|
2017-08-19 19:51:58
|
OpenMined/Docs
|
https://api.github.com/repos/OpenMined/Docs
|
closed
|
Funneling new users into projects easier
|
enhancement help wanted process
|
I'm trying to wrap my mind around how we might guide new users exploring OpenMined for the first time to join a specific project. It's hard for people to grasp the entire ecosystem and so often they join Slack without much aim of where they can best fit in. If we formed teams, naturally this might fall into place... but for the time being, we need some means of directing people according to their skills and interests. I would propose the following system (i.e. Machine Learning):
### Machine learning
**Repositories:**
- [PySonar](https://github.com/OpenMined/PySonar) - Python client to blockchain application (Sonar)
- [PySyft](https://github.com/OpenMined/PySyft) - Python deep learning library
- [syft.js](https://github.com/OpenMined/syft.js) - Node.js deep learning library (for web and mobile browsers)
**Slack channels:** #sonar, #syft
This gives people a high-level overview of what repositories are important and what Slack channels contain the primary discussion on the matter. Other things we could do would be to include a link to the top contributors of a project (so they can be accessible for DM by an individual to get caught up). These individuals would act as pseudo "team leaders" basically implied from how much they've contributed to a project. We also may want to include a link to the issues pages directly so that people can see the actual work that needs to get done.
Open for thoughts here. I've already begun addressing this [in the following pull request](https://github.com/OpenMined/Docs/pull/35). Please feel free to contribute and discuss!
|
1.0
|
Funneling new users into projects easier - I'm trying to wrap my mind around how we might guide new users exploring OpenMined for the first time to join a specific project. It's hard for people to grasp the entire ecosystem and so often they join Slack without much aim of where they can best fit in. If we formed teams, naturally this might fall into place... but for the time being, we need some means of directing people according to their skills and interests. I would propose the following system (i.e. Machine Learning):
### Machine learning
**Repositories:**
- [PySonar](https://github.com/OpenMined/PySonar) - Python client to blockchain application (Sonar)
- [PySyft](https://github.com/OpenMined/PySyft) - Python deep learning library
- [syft.js](https://github.com/OpenMined/syft.js) - Node.js deep learning library (for web and mobile browsers)
**Slack channels:** #sonar, #syft
This gives people a high-level overview of what repositories are important and what Slack channels contain the primary discussion on the matter. Other things we could do would be to include a link to the top contributors of a project (so they can be accessible for DM by an individual to get caught up). These individuals would act as pseudo "team leaders" basically implied from how much they've contributed to a project. We also may want to include a link to the issues pages directly so that people can see the actual work that needs to get done.
Open for thoughts here. I've already begun addressing this [in the following pull request](https://github.com/OpenMined/Docs/pull/35). Please feel free to contribute and discuss!
|
process
|
funneling new users into projects easier i m trying to wrap my mind around how we might guide new users exploring openmined for the first time to join a specific project it s hard for people to grasp the entire ecosystem and so often they join slack without much aim of where they can best fit in if we formed teams naturally this might fall into place but for the time being we need some means of directing people according to their skills and interests i would propose the following system i e machine learning machine learning repositories python client to blockchain application sonar python deep learning library node js deep learning library for web and mobile browsers slack channels sonar syft this gives people a high level overview of what repositories are important and what slack channels contain the primary discussion on the matter other things we could do would be to include a link to the top contributors of a project so they can be accessible for dm by an individual to get caught up these individuals would act as pseudo team leaders basically implied from how much they ve contributed to a project we also may want to include a link to the issues pages directly so that people can see the actual work that needs to get done open for thoughts here i ve already begun addressing this please feel free to contribute and discuss
| 1
|
10,400
| 13,202,152,291
|
IssuesEvent
|
2020-08-14 11:40:54
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Messed up data types after running SAGA snap points to lines
|
Bug Feedback Processing
|
**Describe the bug**
I run SAGA 'snap points to lines' tool using the belowing files:
points layer: CSV files contain x/y coordinates
Snap feature: OSM street shapefile
The output seems have changed data type and returned symbols that don't make any sense (see screenshot below)

**QGIS and OS versions**
QGIS version: 3.14.0-Pi
QGIS code revision: 9f7028fd23
Qt version: 5.11.2
GDAL version: 3.0.4
GEOS version: 3.8.1-CAPI-1.13.3
PROJ version: Rel. 6.3.2, May 1st, 2020
|
1.0
|
Messed up data types after running SAGA snap points to lines - **Describe the bug**
I run SAGA 'snap points to lines' tool using the belowing files:
points layer: CSV files contain x/y coordinates
Snap feature: OSM street shapefile
The output seems have changed data type and returned symbols that don't make any sense (see screenshot below)

**QGIS and OS versions**
QGIS version: 3.14.0-Pi
QGIS code revision: 9f7028fd23
Qt version: 5.11.2
GDAL version: 3.0.4
GEOS version: 3.8.1-CAPI-1.13.3
PROJ version: Rel. 6.3.2, May 1st, 2020
|
process
|
messed up data types after running saga snap points to lines describe the bug i run saga snap points to lines tool using the belowing files points layer csv files contain x y coordinates snap feature osm street shapefile the output seems have changed data type and returned symbols that don t make any sense see screenshot below qgis and os versions qgis version pi qgis code revision qt version gdal version geos version capi proj version rel may
| 1
|
153
| 2,581,440,253
|
IssuesEvent
|
2015-02-14 02:01:34
|
tinkerpop/tinkerpop3
|
https://api.github.com/repos/tinkerpop/tinkerpop3
|
opened
|
Make Memory a single sideEffect in the Traversal
|
enhancement process
|
RIght now we add a new sideEffect entry for each key of the ComputerResult Memory. Don't do this -- group it and include other things like runtime(), iterations(), etc.
|
1.0
|
Make Memory a single sideEffect in the Traversal - RIght now we add a new sideEffect entry for each key of the ComputerResult Memory. Don't do this -- group it and include other things like runtime(), iterations(), etc.
|
process
|
make memory a single sideeffect in the traversal right now we add a new sideeffect entry for each key of the computerresult memory don t do this group it and include other things like runtime iterations etc
| 1
|
140,327
| 12,891,596,857
|
IssuesEvent
|
2020-07-13 18:01:13
|
vercel/next.js
|
https://api.github.com/repos/vercel/next.js
|
closed
|
Fast Refresh Full Reload warning is incomplete (and partly unclear)
|
kind: story type: documentation
|
# Bug report
## Describe the bug
Fast Refresh does not support unnamed function components ([closed issue](https://github.com/zeit/next.js/issues/12891)), yet the warning provided does not mention this.
## To Reproduce
`pages/index.js`:
```
import '../src/somefile'
export default () => null;
```
Then edit and save `somefile`.
### Works
```
import '../src/somefile'
const Page = () => null;
export default Page;
```
## Expected behavior
The warning should mention that Fast Refresh does not support unnamed function components.
## System information
- Version of Next.js: 9.4.1
## Additional context
I know that the warning was only introduced in 9.4.1 (and is a huge improvement on the error in 9.4.0), so it will most likely be subject to near future iterations. Here's how I read it:
> Fast Refresh will perform a full reload when you edit a file that's imported by modules outside of the React tree.
Probably my ignorance, but I was a bit confused about what "modules outside of the React tree" mean here. Does this mean the React components tree? Is `somefile` outside of the react tree? I mean, there is a dependency graph, but it is not a tree.
> You might have a file which renders a React component but also exports a value that is imported by a non-React component.
First, files don't 'render' a React component - files 'export' react components.
I don't really understand whether this is said within the context of the whole dependency graph or what? Does this mean that if I have a file that exports a react component, but also some function (or a type like Props) that is then imported in a test, that will break Fast Refresh
> Consider migrating the non-React component export to a separate file and importing it into both files.
What's a "non-React component`? An angular one?
> It is also possible you are using class components at the top-level of your application, which disables Fast Refresh.
This is also unclear. What constitutes the "top-level of your application"? Some options:
- A page's default export
- `pages/_app.js`
- `pages/_document.js`
> Fast Refresh requires at least one function component in your React tree.
That is clear and the reproduction case above comply with this.
Anyhow, while there is a place for improvement in my view to other parts of the warning, I believe my issue would be solved if the last line would say:
> Fast Refresh requires at least one _named_ function component in your React tree.
Yet I'm not sure this is correct. Is it any odd function component that needs to be named, or the one exported from a page?
|
1.0
|
Fast Refresh Full Reload warning is incomplete (and partly unclear) - # Bug report
## Describe the bug
Fast Refresh does not support unnamed function components ([closed issue](https://github.com/zeit/next.js/issues/12891)), yet the warning provided does not mention this.
## To Reproduce
`pages/index.js`:
```
import '../src/somefile'
export default () => null;
```
Then edit and save `somefile`.
### Works
```
import '../src/somefile'
const Page = () => null;
export default Page;
```
## Expected behavior
The warning should mention that Fast Refresh does not support unnamed function components.
## System information
- Version of Next.js: 9.4.1
## Additional context
I know that the warning was only introduced in 9.4.1 (and is a huge improvement on the error in 9.4.0), so it will most likely be subject to near future iterations. Here's how I read it:
> Fast Refresh will perform a full reload when you edit a file that's imported by modules outside of the React tree.
Probably my ignorance, but I was a bit confused about what "modules outside of the React tree" mean here. Does this mean the React components tree? Is `somefile` outside of the react tree? I mean, there is a dependency graph, but it is not a tree.
> You might have a file which renders a React component but also exports a value that is imported by a non-React component.
First, files don't 'render' a React component - files 'export' react components.
I don't really understand whether this is said within the context of the whole dependency graph or what? Does this mean that if I have a file that exports a react component, but also some function (or a type like Props) that is then imported in a test, that will break Fast Refresh
> Consider migrating the non-React component export to a separate file and importing it into both files.
What's a "non-React component`? An angular one?
> It is also possible you are using class components at the top-level of your application, which disables Fast Refresh.
This is also unclear. What constitutes the "top-level of your application"? Some options:
- A page's default export
- `pages/_app.js`
- `pages/_document.js`
> Fast Refresh requires at least one function component in your React tree.
That is clear and the reproduction case above comply with this.
Anyhow, while there is a place for improvement in my view to other parts of the warning, I believe my issue would be solved if the last line would say:
> Fast Refresh requires at least one _named_ function component in your React tree.
Yet I'm not sure this is correct. Is it any odd function component that needs to be named, or the one exported from a page?
|
non_process
|
fast refresh full reload warning is incomplete and partly unclear bug report describe the bug fast refresh does not support unnamed function components yet the warning provided does not mention this to reproduce pages index js import src somefile export default null then edit and save somefile works import src somefile const page null export default page expected behavior the warning should mention that fast refresh does not support unnamed function components system information version of next js additional context i know that the warning was only introduced in and is a huge improvement on the error in so it will most likely be subject to near future iterations here s how i read it fast refresh will perform a full reload when you edit a file that s imported by modules outside of the react tree probably my ignorance but i was a bit confused about what modules outside of the react tree mean here does this mean the react components tree is somefile outside of the react tree i mean there is a dependency graph but it is not a tree you might have a file which renders a react component but also exports a value that is imported by a non react component first files don t render a react component files export react components i don t really understand whether this is said within the context of the whole dependency graph or what does this mean that if i have a file that exports a react component but also some function or a type like props that is then imported in a test that will break fast refresh consider migrating the non react component export to a separate file and importing it into both files what s a non react component an angular one it is also possible you are using class components at the top level of your application which disables fast refresh this is also unclear what constitutes the top level of your application some options a page s default export pages app js pages document js fast refresh requires at least one function component in your react tree that is clear and the reproduction case above comply with this anyhow while there is a place for improvement in my view to other parts of the warning i believe my issue would be solved if the last line would say fast refresh requires at least one named function component in your react tree yet i m not sure this is correct is it any odd function component that needs to be named or the one exported from a page
| 0
|
22,725
| 32,044,845,464
|
IssuesEvent
|
2023-09-22 23:48:41
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
marimo 0.1.16 has 2 GuardDog issues
|
guarddog exec-base64 silent-process-execution
|
https://pypi.org/project/marimo
https://inspector.pypi.io/project/marimo
```{
"dependency": "marimo",
"version": "0.1.16",
"result": {
"issues": 2,
"errors": {},
"results": {
"exec-base64": [
{
"location": "marimo-0.1.16/marimo/_ast/test_codegen.py:96",
"code": " stringified = eval(\"\\n\".join(raw.split(\"\\n\")[1:5])).split(\"\\n\")",
"message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"
}
],
"silent-process-execution": [
{
"location": "marimo-0.1.16/marimo/_server/sessions.py:465",
"code": " self.lsp_process = subprocess.Popen(\n cmd.split(),\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n stdin=subprocess.DEVNULL,\n )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpzgo8csq9/marimo"
}
}```
|
1.0
|
marimo 0.1.16 has 2 GuardDog issues - https://pypi.org/project/marimo
https://inspector.pypi.io/project/marimo
```{
"dependency": "marimo",
"version": "0.1.16",
"result": {
"issues": 2,
"errors": {},
"results": {
"exec-base64": [
{
"location": "marimo-0.1.16/marimo/_ast/test_codegen.py:96",
"code": " stringified = eval(\"\\n\".join(raw.split(\"\\n\")[1:5])).split(\"\\n\")",
"message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"
}
],
"silent-process-execution": [
{
"location": "marimo-0.1.16/marimo/_server/sessions.py:465",
"code": " self.lsp_process = subprocess.Popen(\n cmd.split(),\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n stdin=subprocess.DEVNULL,\n )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpzgo8csq9/marimo"
}
}```
|
process
|
marimo has guarddog issues dependency marimo version result issues errors results exec location marimo marimo ast test codegen py code stringified eval n join raw split n split n message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n silent process execution location marimo marimo server sessions py code self lsp process subprocess popen n cmd split n stdout subprocess devnull n stderr subprocess devnull n stdin subprocess devnull n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp marimo
| 1
|
7,992
| 11,185,768,808
|
IssuesEvent
|
2020-01-01 05:53:53
|
arunkumar9t2/scabbard
|
https://api.github.com/repos/arunkumar9t2/scabbard
|
opened
|
Choose a logging library and provide verbose logging to debug issues
|
enhancement module:gradle-plugin module:processor
|
Looking at [Timber](http://jakewharton.github.io/timber/). Scabbard should expose flag via supported options and gradle plugin should expose API to receive user defined values and delegate it to processor.
|
1.0
|
Choose a logging library and provide verbose logging to debug issues - Looking at [Timber](http://jakewharton.github.io/timber/). Scabbard should expose flag via supported options and gradle plugin should expose API to receive user defined values and delegate it to processor.
|
process
|
choose a logging library and provide verbose logging to debug issues looking at scabbard should expose flag via supported options and gradle plugin should expose api to receive user defined values and delegate it to processor
| 1
|
15,724
| 19,897,803,964
|
IssuesEvent
|
2022-01-25 02:26:52
|
microsoft/vscode
|
https://api.github.com/repos/microsoft/vscode
|
closed
|
Problem with the terminal showing french accents (chcp active code page issue)
|
bug terminal-process
|
<!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions -->
<!-- 🔎 Search existing issues to avoid creating duplicates. -->
<!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ -->
<!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. -->
<!-- 🔧 Launch with `code --disable-extensions` to check. -->
Does this issue occur when all extensions are disabled?: Yes/No
Yes, its hard to test because I need the terminal to test my java program
<!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. -->
<!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. -->
- VS Code Version: 1.63.2
- OS Version: WINDOWS_NET x64 10.0.19044
Steps to Reproduce:
For java coding
1. chcp 1252 to change active code page to that, it will show É È À just fine from my println
2. It wont show any accents for txt files that I load into my program so it shows as a string in the terminal
3. https://i.gyazo.com/d218824fc9e932d0a8cdfccf0f0c83aa.png
4. Here's what im supposed to see https://i.gyazo.com/dba8471779ffcbd9ab219f7e4aebccc2.png
4. Changing chcp to 65001 will fix the accents for the txt file I load in so it shows accent properly, but now it literally deletes any accents in my println
5. Here's what I see when i do 65001 and what im supposed to see from my own print ln
6. https://i.gyazo.com/ac3ce48194e5cfa17cf0ae5b468ad201.png
7. Here's what im supposed to see https://i.gyazo.com/be05803f618169c3aadc89dc6176fcc0.png
Im at a lost, idk what to do. I had the issue before so I set my active code page to 1252 but now that ive started using exterior text files to use as strings, nothing works properly and I really want to use vscode and not the other IDEA
|
1.0
|
Problem with the terminal showing french accents (chcp active code page issue) - <!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions -->
<!-- 🔎 Search existing issues to avoid creating duplicates. -->
<!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ -->
<!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. -->
<!-- 🔧 Launch with `code --disable-extensions` to check. -->
Does this issue occur when all extensions are disabled?: Yes/No
Yes, its hard to test because I need the terminal to test my java program
<!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. -->
<!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. -->
- VS Code Version: 1.63.2
- OS Version: WINDOWS_NET x64 10.0.19044
Steps to Reproduce:
For java coding
1. chcp 1252 to change active code page to that, it will show É È À just fine from my println
2. It wont show any accents for txt files that I load into my program so it shows as a string in the terminal
3. https://i.gyazo.com/d218824fc9e932d0a8cdfccf0f0c83aa.png
4. Here's what im supposed to see https://i.gyazo.com/dba8471779ffcbd9ab219f7e4aebccc2.png
4. Changing chcp to 65001 will fix the accents for the txt file I load in so it shows accent properly, but now it literally deletes any accents in my println
5. Here's what I see when i do 65001 and what im supposed to see from my own print ln
6. https://i.gyazo.com/ac3ce48194e5cfa17cf0ae5b468ad201.png
7. Here's what im supposed to see https://i.gyazo.com/be05803f618169c3aadc89dc6176fcc0.png
Im at a lost, idk what to do. I had the issue before so I set my active code page to 1252 but now that ive started using exterior text files to use as strings, nothing works properly and I really want to use vscode and not the other IDEA
|
process
|
problem with the terminal showing french accents chcp active code page issue does this issue occur when all extensions are disabled yes no yes its hard to test because i need the terminal to test my java program report issue dialog can assist with this vs code version os version windows net steps to reproduce for java coding chcp to change active code page to that it will show é è à just fine from my println it wont show any accents for txt files that i load into my program so it shows as a string in the terminal here s what im supposed to see changing chcp to will fix the accents for the txt file i load in so it shows accent properly but now it literally deletes any accents in my println here s what i see when i do and what im supposed to see from my own print ln here s what im supposed to see im at a lost idk what to do i had the issue before so i set my active code page to but now that ive started using exterior text files to use as strings nothing works properly and i really want to use vscode and not the other idea
| 1
|
18,214
| 24,272,919,640
|
IssuesEvent
|
2022-09-28 11:47:47
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
Metrics filtering
|
question processor/filter
|
### Describe the issue you're reporting
Hi Guys ! I got metrics and I want to do a filter on "Data point attributes" in the collector. Is it possible ?
I already tried to use filter processor but this one can only filter on resource attributes.
```
filter:
metrics:
include:
match_type: regexp
metric_names:
- http\.server\.requests
resource_attributes:
- Key: service.name
Value: svcname
```
I haven't found a filter based on data point attributes and I need to remove specifics metrics from their uri attribues. E.G:
```
ScopeMetrics SchemaURL:
InstrumentationScope io.opentelemetry.micrometer-1.5
Metric #0
Descriptor:
-> Name: http.server.requests
-> Description:
-> Unit: ms
-> DataType: Histogram
-> AggregationTemporality: AGGREGATION_TEMPORALITY_CUMULATIVE
HistogramDataPoints #0
Data point attributes:
-> application: STRING(appname)
-> exception: STRING(None)
-> instance: STRING(someInstance)
-> method: STRING(GET)
-> outcome: STRING(CLIENT_ERROR)
-> status: STRING(401)
-> uri: STRING(NEED_TO_BE_FILTERED)
```
|
1.0
|
Metrics filtering - ### Describe the issue you're reporting
Hi Guys ! I got metrics and I want to do a filter on "Data point attributes" in the collector. Is it possible ?
I already tried to use filter processor but this one can only filter on resource attributes.
```
filter:
metrics:
include:
match_type: regexp
metric_names:
- http\.server\.requests
resource_attributes:
- Key: service.name
Value: svcname
```
I haven't found a filter based on data point attributes and I need to remove specifics metrics from their uri attribues. E.G:
```
ScopeMetrics SchemaURL:
InstrumentationScope io.opentelemetry.micrometer-1.5
Metric #0
Descriptor:
-> Name: http.server.requests
-> Description:
-> Unit: ms
-> DataType: Histogram
-> AggregationTemporality: AGGREGATION_TEMPORALITY_CUMULATIVE
HistogramDataPoints #0
Data point attributes:
-> application: STRING(appname)
-> exception: STRING(None)
-> instance: STRING(someInstance)
-> method: STRING(GET)
-> outcome: STRING(CLIENT_ERROR)
-> status: STRING(401)
-> uri: STRING(NEED_TO_BE_FILTERED)
```
|
process
|
metrics filtering describe the issue you re reporting hi guys i got metrics and i want to do a filter on data point attributes in the collector is it possible i already tried to use filter processor but this one can only filter on resource attributes filter metrics include match type regexp metric names http server requests resource attributes key service name value svcname i haven t found a filter based on data point attributes and i need to remove specifics metrics from their uri attribues e g scopemetrics schemaurl instrumentationscope io opentelemetry micrometer metric descriptor name http server requests description unit ms datatype histogram aggregationtemporality aggregation temporality cumulative histogramdatapoints data point attributes application string appname exception string none instance string someinstance method string get outcome string client error status string uri string need to be filtered
| 1
|
3,576
| 6,619,149,129
|
IssuesEvent
|
2017-09-21 10:59:37
|
w3c/w3process
|
https://api.github.com/repos/w3c/w3process
|
closed
|
Sections 3.4 and 6.2.6 have different statements about Voting rules in a Charter
|
Not agreed for merging (yet) Process2018Candidate
|
Transferred from https://www.w3.org/community/w3process/track/issues/161
State: Raised
Note, the second reference is 5.2.6 in Process 2017
|
1.0
|
Sections 3.4 and 6.2.6 have different statements about Voting rules in a Charter - Transferred from https://www.w3.org/community/w3process/track/issues/161
State: Raised
Note, the second reference is 5.2.6 in Process 2017
|
process
|
sections and have different statements about voting rules in a charter transferred from state raised note the second reference is in process
| 1
|
150,981
| 19,648,156,218
|
IssuesEvent
|
2022-01-10 01:04:06
|
tom9carthron1/continuous-delivery-aws-template
|
https://api.github.com/repos/tom9carthron1/continuous-delivery-aws-template
|
closed
|
WS-2022-0007 (Medium) detected in node-forge-0.10.0.tgz, node-forge-0.9.0.tgz - autoclosed
|
security vulnerability
|
## WS-2022-0007 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-forge-0.10.0.tgz</b>, <b>node-forge-0.9.0.tgz</b></p></summary>
<p>
<details><summary><b>node-forge-0.10.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /continuous-delivery-aws-template/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-forge-0.10.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>node-forge-0.9.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/selfsigned/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.11.0.tgz (Root Library)
- selfsigned-1.10.7.tgz
- :x: **node-forge-0.9.0.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In node-forge before 1.0.0 he regex used for the forge.util.parseUrl API would not properly parse certain inputs resulting in a parsed data structure that could lead to undesired behavior.
<p>Publish Date: 2022-01-08
<p>URL: <a href=https://github.com/digitalbazaar/forge/commit/db8016c805371e72b06d8e2edfe0ace0df934a5e>WS-2022-0007</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-gf8q-jrpm-jvxq">https://github.com/advisories/GHSA-gf8q-jrpm-jvxq</a></p>
<p>Release Date: 2022-01-08</p>
<p>Fix Resolution: node-forge - 1.0.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-forge","packageVersion":"0.10.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"node-forge:0.10.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"node-forge - 1.0.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"node-forge","packageVersion":"0.9.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"webpack-dev-server:3.11.0;selfsigned:1.10.7;node-forge:0.9.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"node-forge - 1.0.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2022-0007","vulnerabilityDetails":"In node-forge before 1.0.0 he regex used for the forge.util.parseUrl API would not properly parse certain inputs resulting in a parsed data structure that could lead to undesired behavior.","vulnerabilityUrl":"https://github.com/digitalbazaar/forge/commit/db8016c805371e72b06d8e2edfe0ace0df934a5e","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
True
|
WS-2022-0007 (Medium) detected in node-forge-0.10.0.tgz, node-forge-0.9.0.tgz - autoclosed - ## WS-2022-0007 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-forge-0.10.0.tgz</b>, <b>node-forge-0.9.0.tgz</b></p></summary>
<p>
<details><summary><b>node-forge-0.10.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /continuous-delivery-aws-template/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-forge-0.10.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>node-forge-0.9.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/selfsigned/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.11.0.tgz (Root Library)
- selfsigned-1.10.7.tgz
- :x: **node-forge-0.9.0.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In node-forge before 1.0.0 he regex used for the forge.util.parseUrl API would not properly parse certain inputs resulting in a parsed data structure that could lead to undesired behavior.
<p>Publish Date: 2022-01-08
<p>URL: <a href=https://github.com/digitalbazaar/forge/commit/db8016c805371e72b06d8e2edfe0ace0df934a5e>WS-2022-0007</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-gf8q-jrpm-jvxq">https://github.com/advisories/GHSA-gf8q-jrpm-jvxq</a></p>
<p>Release Date: 2022-01-08</p>
<p>Fix Resolution: node-forge - 1.0.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"node-forge","packageVersion":"0.10.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"node-forge:0.10.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"node-forge - 1.0.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"node-forge","packageVersion":"0.9.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"webpack-dev-server:3.11.0;selfsigned:1.10.7;node-forge:0.9.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"node-forge - 1.0.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2022-0007","vulnerabilityDetails":"In node-forge before 1.0.0 he regex used for the forge.util.parseUrl API would not properly parse certain inputs resulting in a parsed data structure that could lead to undesired behavior.","vulnerabilityUrl":"https://github.com/digitalbazaar/forge/commit/db8016c805371e72b06d8e2edfe0ace0df934a5e","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
ws medium detected in node forge tgz node forge tgz autoclosed ws medium severity vulnerability vulnerable libraries node forge tgz node forge tgz node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file package json path to vulnerable library continuous delivery aws template node modules node forge package json dependency hierarchy x node forge tgz vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file package json path to vulnerable library node modules selfsigned node modules node forge package json dependency hierarchy webpack dev server tgz root library selfsigned tgz x node forge tgz vulnerable library found in base branch master vulnerability details in node forge before he regex used for the forge util parseurl api would not properly parse certain inputs resulting in a parsed data structure that could lead to undesired behavior publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree node forge isminimumfixversionavailable true minimumfixversion node forge isbinary false packagetype javascript node js packagename node forge packageversion packagefilepaths istransitivedependency true dependencytree webpack dev server selfsigned node forge isminimumfixversionavailable true minimumfixversion node forge isbinary false basebranches vulnerabilityidentifier ws vulnerabilitydetails in node forge before he regex used for the forge util parseurl api would not properly parse certain inputs resulting in a parsed data structure that could lead to undesired behavior vulnerabilityurl
| 0
|
223,593
| 17,116,912,187
|
IssuesEvent
|
2021-07-11 14:48:38
|
GrayLian97/GrayLian97.github.io
|
https://api.github.com/repos/GrayLian97/GrayLian97.github.io
|
opened
|
01 Preface
|
documentation
|
#Preface
It is the first lesson of CMU 15-445, this class introduce the reason why we need database system , the history of database system, an example of database from scratch and some important parts of Database Systems.
##Definition
The definition of database is that it's a collection of data that's related to gathers in some way that's trying to model some aspect of the real world. The data of database is usually related together or have some common theme to them.
##Example : A digital music store
Database is used to keep track of various Artists and their albums. Both artists and albums have their information such as name, time, or something else.
The simplest database we could implement and manage is to store data as a brunch of value or CSV files.
|
1.0
|
01 Preface - #Preface
It is the first lesson of CMU 15-445, this class introduce the reason why we need database system , the history of database system, an example of database from scratch and some important parts of Database Systems.
##Definition
The definition of database is that it's a collection of data that's related to gathers in some way that's trying to model some aspect of the real world. The data of database is usually related together or have some common theme to them.
##Example : A digital music store
Database is used to keep track of various Artists and their albums. Both artists and albums have their information such as name, time, or something else.
The simplest database we could implement and manage is to store data as a brunch of value or CSV files.
|
non_process
|
preface preface it is the first lesson of cmu this class introduce the reason why we need database system the history of database system an example of database from scratch and some important parts of database systems definition the definition of database is that it s a collection of data that s related to gathers in some way that s trying to model some aspect of the real world the data of database is usually related together or have some common theme to them example a digital music store database is used to keep track of various artists and their albums both artists and albums have their information such as name time or something else the simplest database we could implement and manage is to store data as a brunch of value or csv files
| 0
|
11,545
| 14,429,097,947
|
IssuesEvent
|
2020-12-06 12:52:25
|
emacs-ess/ESS
|
https://api.github.com/repos/emacs-ess/ESS
|
closed
|
R process shown in two windows instead of one
|
process:windows
|
When I open an `.R` file and execute
```
M-x ess-eval-region-or-line-visibly-and-step
```
on a line of code when there’s no R process yet, the current window is split and a new R process is started.
Problem: The R process is shown in both windows, not just in the new window.
Environment:
- Emacs version: current master branch with native-comp.
- ESS installed from MELPA.
- Tested with `emacs -q`
|
1.0
|
R process shown in two windows instead of one - When I open an `.R` file and execute
```
M-x ess-eval-region-or-line-visibly-and-step
```
on a line of code when there’s no R process yet, the current window is split and a new R process is started.
Problem: The R process is shown in both windows, not just in the new window.
Environment:
- Emacs version: current master branch with native-comp.
- ESS installed from MELPA.
- Tested with `emacs -q`
|
process
|
r process shown in two windows instead of one when i open an r file and execute m x ess eval region or line visibly and step on a line of code when there’s no r process yet the current window is split and a new r process is started problem the r process is shown in both windows not just in the new window environment emacs version current master branch with native comp ess installed from melpa tested with emacs q
| 1
|
631,801
| 20,160,522,710
|
IssuesEvent
|
2022-02-09 20:59:48
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
opened
|
Filtering on nested questions causes error in Public and Embedding
|
Type:Bug Priority:P1 Querying/Parameters & Variables Querying/Nested Queries .Frontend Embedding/Public Embedding/Signed .Regression
|
**Describe the bug**
Filtering on nested questions causes error in Public and Embedding
There's no backend errors on public sharing and there's no request for the card, which makes me think it's just frontend.
**To Reproduce**
1. Native query > Sample Dataset > `select * from PRODUCTS` - save as "Q1"
2. Question > Saved Questions > "Q1" - save as "Q2" - add to dashboard
3. On dashboard, add All Options time filter and connect to card - save dashboard

4. Go to Sharing :arrow_upper_right: - enable Public Sharing and go to the URL

Log during step 3 and 4:
```
2022-02-09 21:55:23,147 WARN middleware.log :: GET /api/field/field%2CCREATED_AT%2C%5Bobject%20Object%5D 404 168.7 µs (0 DB calls)
"API endpoint does not exist."
2022-02-09 21:55:26,585 WARN middleware.log :: GET /api/field/field%2CCREATED_AT%2C%5Bobject%20Object%5D 404 141.1 µs (0 DB calls)
"API endpoint does not exist."
2022-02-09 21:55:33,685 DEBUG middleware.log :: PUT /api/dashboard/6/cards 200 12.7 ms (15 DB calls) App DB connections: 0/15 Jetty threads: 4/50 (3 idle, 0 queued) (110 total active threads) Queries in flight: 0 (0 queued)
2022-02-09 21:55:33,729 DEBUG middleware.log :: PUT /api/dashboard/6 200 3.9 ms (8 DB calls) App DB connections: 0/15 Jetty threads: 4/50 (3 idle, 0 queued) (110 total active threads) Queries in flight: 0 (0 queued)
2022-02-09 21:55:33,765 DEBUG middleware.log :: GET /api/dashboard/6 200 7.5 ms (10 DB calls) App DB connections: 3/15 Jetty threads: 4/50 (3 idle, 0 queued) (110 total active threads) Queries in flight: 0 (0 queued)
2022-02-09 21:55:34,031 WARN middleware.log :: GET /api/table//query_metadata 404 193.5 µs (0 DB calls)
"API endpoint does not exist."
2022-02-09 21:55:46,742 DEBUG middleware.log :: GET /api/public/dashboard/34ab699c-7edf-4680-aa54-5b6793eee4ea 200 5.3 ms (5 DB calls) App DB connections: 0/15 Jetty threads: 4/50 (2 idle, 0 queued) (110 total active threads) Queries in flight: 0 (0 queued)
2022-02-09 21:55:46,774 DEBUG middleware.log :: GET /api/session/properties 200 10.3 ms (1 DB calls) App DB connections: 0/15 Jetty threads: 4/50 (2 idle, 0 queued) (110 total active threads) Queries in flight: 0 (0 queued)
```
**Information about your Metabase Installation:**
0.42.0
**Additional context**
Likely the same as #20232
|
1.0
|
Filtering on nested questions causes error in Public and Embedding - **Describe the bug**
Filtering on nested questions causes error in Public and Embedding
There's no backend errors on public sharing and there's no request for the card, which makes me think it's just frontend.
**To Reproduce**
1. Native query > Sample Dataset > `select * from PRODUCTS` - save as "Q1"
2. Question > Saved Questions > "Q1" - save as "Q2" - add to dashboard
3. On dashboard, add All Options time filter and connect to card - save dashboard

4. Go to Sharing :arrow_upper_right: - enable Public Sharing and go to the URL

Log during step 3 and 4:
```
2022-02-09 21:55:23,147 WARN middleware.log :: GET /api/field/field%2CCREATED_AT%2C%5Bobject%20Object%5D 404 168.7 µs (0 DB calls)
"API endpoint does not exist."
2022-02-09 21:55:26,585 WARN middleware.log :: GET /api/field/field%2CCREATED_AT%2C%5Bobject%20Object%5D 404 141.1 µs (0 DB calls)
"API endpoint does not exist."
2022-02-09 21:55:33,685 DEBUG middleware.log :: PUT /api/dashboard/6/cards 200 12.7 ms (15 DB calls) App DB connections: 0/15 Jetty threads: 4/50 (3 idle, 0 queued) (110 total active threads) Queries in flight: 0 (0 queued)
2022-02-09 21:55:33,729 DEBUG middleware.log :: PUT /api/dashboard/6 200 3.9 ms (8 DB calls) App DB connections: 0/15 Jetty threads: 4/50 (3 idle, 0 queued) (110 total active threads) Queries in flight: 0 (0 queued)
2022-02-09 21:55:33,765 DEBUG middleware.log :: GET /api/dashboard/6 200 7.5 ms (10 DB calls) App DB connections: 3/15 Jetty threads: 4/50 (3 idle, 0 queued) (110 total active threads) Queries in flight: 0 (0 queued)
2022-02-09 21:55:34,031 WARN middleware.log :: GET /api/table//query_metadata 404 193.5 µs (0 DB calls)
"API endpoint does not exist."
2022-02-09 21:55:46,742 DEBUG middleware.log :: GET /api/public/dashboard/34ab699c-7edf-4680-aa54-5b6793eee4ea 200 5.3 ms (5 DB calls) App DB connections: 0/15 Jetty threads: 4/50 (2 idle, 0 queued) (110 total active threads) Queries in flight: 0 (0 queued)
2022-02-09 21:55:46,774 DEBUG middleware.log :: GET /api/session/properties 200 10.3 ms (1 DB calls) App DB connections: 0/15 Jetty threads: 4/50 (2 idle, 0 queued) (110 total active threads) Queries in flight: 0 (0 queued)
```
**Information about your Metabase Installation:**
0.42.0
**Additional context**
Likely the same as #20232
|
non_process
|
filtering on nested questions causes error in public and embedding describe the bug filtering on nested questions causes error in public and embedding there s no backend errors on public sharing and there s no request for the card which makes me think it s just frontend to reproduce native query sample dataset select from products save as question saved questions save as add to dashboard on dashboard add all options time filter and connect to card save dashboard go to sharing arrow upper right enable public sharing and go to the url log during step and warn middleware log get api field field at µs db calls api endpoint does not exist warn middleware log get api field field at µs db calls api endpoint does not exist debug middleware log put api dashboard cards ms db calls app db connections jetty threads idle queued total active threads queries in flight queued debug middleware log put api dashboard ms db calls app db connections jetty threads idle queued total active threads queries in flight queued debug middleware log get api dashboard ms db calls app db connections jetty threads idle queued total active threads queries in flight queued warn middleware log get api table query metadata µs db calls api endpoint does not exist debug middleware log get api public dashboard ms db calls app db connections jetty threads idle queued total active threads queries in flight queued debug middleware log get api session properties ms db calls app db connections jetty threads idle queued total active threads queries in flight queued information about your metabase installation additional context likely the same as
| 0
|
17,241
| 23,782,097,933
|
IssuesEvent
|
2022-09-02 06:27:53
|
arcticicestudio/nord-visual-studio-code
|
https://api.github.com/repos/arcticicestudio/nord-visual-studio-code
|
closed
|
Having kind of "tints" inside the code editor
|
status-pending context-ui scope-compatibility type-support scope-quality
|
## Description
Only with Nord theme, my editor displays kind of 'tints', shapes with different colors, on the right side, see prtscn

### Steps to Reproduce
In my environment, it will happen in every code file. Just installed 'Nord' and opened a code files
### Expected Behavior
The background to be consistent
## Example
See prtscn
## Environment and Versions
- What is the version of _Nord Visual Studio Code_ you are running? - 0.19
- What is the name and the version of your OS? - MacOS Monterey 12.4
- Have you tried to reproduce it on different OS environments and if yes is the behavior the same for all? No, don't have one
- Are you using any additional CLI arguments to run the project? No
- What is the version of the build tool (e.g. [npm](https://www.npmjs.com), [dep](https://golang.github.io/dep) or [Gradle](https://gradle.org)) you are running? 8.12.1
- Are you using any additional CLI arguments to start the build tool task/script other than defined by the project? No
|
True
|
Having kind of "tints" inside the code editor - ## Description
Only with Nord theme, my editor displays kind of 'tints', shapes with different colors, on the right side, see prtscn

### Steps to Reproduce
In my environment, it will happen in every code file. Just installed 'Nord' and opened a code files
### Expected Behavior
The background to be consistent
## Example
See prtscn
## Environment and Versions
- What is the version of _Nord Visual Studio Code_ you are running? - 0.19
- What is the name and the version of your OS? - MacOS Monterey 12.4
- Have you tried to reproduce it on different OS environments and if yes is the behavior the same for all? No, don't have one
- Are you using any additional CLI arguments to run the project? No
- What is the version of the build tool (e.g. [npm](https://www.npmjs.com), [dep](https://golang.github.io/dep) or [Gradle](https://gradle.org)) you are running? 8.12.1
- Are you using any additional CLI arguments to start the build tool task/script other than defined by the project? No
|
non_process
|
having kind of tints inside the code editor description only with nord theme my editor displays kind of tints shapes with different colors on the right side see prtscn steps to reproduce in my environment it will happen in every code file just installed nord and opened a code files expected behavior the background to be consistent example see prtscn environment and versions what is the version of nord visual studio code you are running what is the name and the version of your os macos monterey have you tried to reproduce it on different os environments and if yes is the behavior the same for all no don t have one are you using any additional cli arguments to run the project no what is the version of the build tool e g or you are running are you using any additional cli arguments to start the build tool task script other than defined by the project no
| 0
|
86,322
| 24,818,064,751
|
IssuesEvent
|
2022-10-25 14:29:15
|
elementor/elementor
|
https://api.github.com/repos/elementor/elementor
|
closed
|
🐞 Bug Report: Loop Builder - Stripe Button is throwing a 500 error when added to a loop template
|
bug compatibility/3rd_party solved_by_loop compatibility/assets component/loop-builder
|
### Prerequisites
- [X] I have searched for similar issues in both open and closed tickets and cannot find a duplicate.
- [X] The issue still exists against the latest stable version of Elementor.
### Description
I added a stripe button into a loop and it doesn't work.
### Steps to reproduce
1) Create a loop with Stripe Button included.
2) Open dev tools
3) Test using Stripe Test Secret Key.
### Isolating the problem
- [X] This bug happens with only Elementor plugin active (and Elementor Pro).
- [ ] This bug happens with a Blank WordPress theme active ([Hello theme](https://wordpress.org/themes/hello-elementor/)).
- [X] I can reproduce this bug consistently using the steps above.
### System Info
== Server Environment ==
Operating System: Linux
Software: nginx/1.20.1
MySQL version: mariadb.org binary distribution v10.3.36
PHP Version: 8.0.24
PHP Memory Limit: 256M
PHP Max Input Vars: 3000
PHP Max Post Size: 128M
GD Installed: Yes
ZIP Installed: Yes
Write Permissions: All right
Elementor Library: Connected
== WordPress Environment ==
Version: 6.0.3
Site URL: https://utahmountainair.staging.tempurl.host
Home URL: https://utahmountainair.staging.tempurl.host
WP Multisite: No
Max Upload Size: 128 MB
Memory limit: 40M
Max Memory limit: 256M
Permalink Structure: /%postname%/
Language: en-US
Timezone: 0
Admin Email: daniel@finidev.tech
Debug Mode: Inactive
== Theme ==
Name: Hello Elementor
Version: 2.6.1
Author: Elementor Team
Child Theme: No
== User ==
Role: administrator
WP Profile lang: en_US
User Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/16.0 Safari/605.1.15
== Active Plugins ==
Defender Pro
Version: 3.3.3
Author: WPMU DEV
Duplicate Page
Version: 4.4.9
Author: mndpsingh287
Easy Media Replace
Version: 0.1.3
Author: Nabil Lemsieh
Elementor
Version: 3.9.0-dev3
Author: Elementor.com
Elementor Beta (Developer Edition)
Version: 1.1.1
Author: Elementor.com
Elementor Pro
Version: 3.9.0-dev4
Author: Elementor.com
Hummingbird Pro
Version: 3.3.5
Author: WPMU DEV
JetEngine
Version: 3.0.5
Author: Crocoblock
JetEngine - Custom visibility conditions
Version: 1.1.0
Author: Crocoblock
Smush Pro
Version: 3.12.2
Author: WPMU DEV
Ultimate Dashboard
Version: 3.7.1.5
Author: David Vongries
Ultimate Dashboard PRO
Version: 3.6.4
Author: David Vongries
User Profile Picture
Version: 2.6.0
Author: Cozmoslabs
WPMU DEV Dashboard
Version: 4.11.17
Author: WPMU DEV
== Must-Use Plugins ==
Hosting
Version: 0.1
Author:
== Elements Usage ==
container : 1
button : 1
heading : 1
image : 1
text-editor : 1
footer : 1
heading : 7
icon : 3
icon-list : 2
image : 2
social-icons : 1
spacer : 2
text-editor : 1
theme-site-logo : 1
header : 1
button : 3
heading : 2
icon-list : 2
nav-menu : 1
theme-site-logo : 1
loop : 0
container : 1
heading : 2
image : 1
spacer : 2
stripe-button : 1
text-editor : 1
page : 9
button : 41
counter : 16
divider : 13
form : 6
google_maps : 1
heading : 104
html : 1
icon : 3
icon-box : 17
icon-list : 10
image : 72
image-box : 14
progress : 7
social-icons : 1
spacer : 1
text-editor : 63
toggle : 5
video : 1
section : 1
button : 9
counter : 4
form : 1
heading : 17
icon : 1
icon-list : 3
image : 24
image-box : 1
text-editor : 7
video : 1
widget : 0
form : 1
wp-page : 7
button : 11
container : 18
counter : 9
form : 2
heading : 35
html : 2
icon : 5
icon-list : 6
image : 40
jet-listing-dynamic-field : 2
jet-listing-grid : 3
loop-grid : 1
spacer : 5
text-editor : 12
video : 1
loop-item : 1
container : 1
heading : 2
image : 1
spacer : 2
stripe-button : 1
text-editor : 1
jet-listing-items : 0
button : 2
container : 7
heading : 2
icon-box : 1
image : 3
jet-listing-dynamic-field : 5
spacer : 2
stripe-button : 1
text-editor : 1
== Elementor Experiments ==
Optimized DOM Output: Active
Improved Asset Loading: Active
Improved CSS Loading: Active
Inline Font Icons: Active
Accessibility Improvements: Active
Additional Custom Breakpoints: Active
Import Export Website Kit: Active
Hide native WordPress widgets from search results: Active
admin_menu_rearrangement: Inactive by default
Flexbox Container: Active
Default to New Theme Builder: Active
Hello Theme Header & Footer: Active
Landing Pages: Active
Color Sampler: Active
Favorite Widgets: Active
Admin Top Bar: Active
Page Transitions: Active
Notes: Active
Loop: Active
Form Submissions: Active
Scroll Snap: Active
== Log ==
Log: showing 20 of 352022-10-03 08:08:06 [info] Elementor data updater process has been completed. [array (
'plugin' => 'Elementor',
'from' => '3.7.7',
'to' => '3.7.8',
)]
2022-10-23 18:10:54 [info] Elementor data updater process has been queued. [array (
'plugin' => 'Elementor',
'from' => '3.7.8',
'to' => '3.8.0-beta5',
)]
2022-10-23 18:10:54 [info] elementor::elementor_updater Started
2022-10-23 18:10:55 [info] Elementor/Upgrades - _on_each_version Start
2022-10-23 18:10:56 [info] Elementor/Upgrades - _on_each_version Finished
2022-10-23 18:10:56 [info] Elementor/Upgrades - _v_3_8_0_fix_php8_image_custom_size Start
2022-10-23 18:10:56 [info] Elementor/Upgrades - _v_3_8_0_fix_php8_image_custom_size Finished
2022-10-23 18:10:56 [info] Elementor data updater process has been completed. [array (
'plugin' => 'Elementor',
'from' => '3.7.8',
'to' => '3.8.0-beta5',
)]
2022-10-23 18:16:35 [info] Elementor data updater process has been queued. [array (
'plugin' => 'Elementor',
'from' => '3.8.0-beta5',
'to' => '3.9.0-dev3',
)]
2022-10-23 18:16:37 [info] elementor::elementor_updater Started
2022-10-23 18:16:37 [info] Elementor/Upgrades - _on_each_version Start
2022-10-23 18:16:38 [info] Elementor/Upgrades - _on_each_version Finished
2022-10-23 18:16:38 [info] Elementor/Upgrades - _v_3_8_0_fix_php8_image_custom_size Start
2022-10-23 18:16:38 [info] Elementor/Upgrades - _v_3_8_0_fix_php8_image_custom_size Finished
2022-10-23 18:16:38 [info] Elementor data updater process has been completed. [array (
'plugin' => 'Elementor',
'from' => '3.8.0-beta5',
'to' => '3.9.0-dev3',
)]
2022-10-23 18:17:06 [info] Elementor data updater process has been queued. [array (
'plugin' => 'Elementor Pro',
'from' => '3.7.7',
'to' => '3.9.0-dev4',
)]
2022-10-23 18:17:07 [info] elementor-pro::elementor_pro_updater Started
2022-10-23 18:17:07 [info] Elementor Pro/Upgrades - _on_each_version Start
2022-10-23 18:17:07 [info] Elementor Pro/Upgrades - _on_each_version Finished
2022-10-23 18:17:07 [info] Elementor data updater process has been completed. [array (
'plugin' => 'Elementor Pro',
'from' => '3.7.7',
'to' => '3.9.0-dev4',
)]
PHP: showing 7 of 7PHP: 2022-10-20 22:46:31 [warning X 5][/var/web/site/public_html/wp-content/plugins/elementor/core/common/modules/ajax/module.php::171] Undefined array key "data" [array (
'trace' => '
#0: Elementor\Core\Logger\Manager -> shutdown()
',
)]
PHP: 2022-10-21 02:42:03 [error X 30][/var/web/site/public_html/wp-content/plugins/elementor/includes/managers/elements.php::71] Uncaught TypeError: Elementor\Elements_Manager::create_element_instance(): Argument #1 ($element_data) must be of type array, bool given, called in /var/web/site/public_html/wp-content/plugins/elementor-pro/modules/payments/module.php on line 289 and defined in /var/web/site/public_html/wp-content/plugins/elementor/includes/managers/elements.php:71
Stack trace:
#0 /var/web/site/public_html/wp-content/plugins/elementor-pro/modules/payments/module.php(289): Elementor\Elements_Manager->create_element_instance(false)
#1 /var/web/site/public_html/wp-includes/class-wp-hook.php(307): ElementorPro\Modules\Payments\Module->submit_stripe_form('')
#2 /var/web/site/public_html/wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters('', Array)
#3 /var/web/site/public_html/wp-includes/plugin.php(476): WP_Hook->do_action(Array)
#4 /var/web/site/public_html/wp-admin/admin-ajax.php(202): do_action('wp_ajax_nopriv_...')
#5 {main}
thrown [array (
'trace' => '
#0: Elementor\Core\Logger\Manager -> shutdown()
',
)]
PHP: 2022-10-21 03:02:34 [error X 23][/var/web/site/public_html/wp-content/plugins/elementor/includes/managers/elements.php::71] Uncaught TypeError: Elementor\Elements_Manager::create_element_instance(): Argument #1 ($element_data) must be of type array, bool given, called in /var/web/site/public_html/wp-content/plugins/elementor-pro/modules/payments/module.php on line 289 and defined in /var/web/site/public_html/wp-content/plugins/elementor/includes/managers/elements.php:71
Stack trace:
#0 /var/web/site/public_html/wp-content/plugins/elementor-pro/modules/payments/module.php(289): Elementor\Elements_Manager->create_element_instance(false)
#1 /var/web/site/public_html/wp-includes/class-wp-hook.php(307): ElementorPro\Modules\Payments\Module->submit_stripe_form('')
#2 /var/web/site/public_html/wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters('', Array)
#3 /var/web/site/public_html/wp-includes/plugin.php(476): WP_Hook->do_action(Array)
#4 /var/web/site/public_html/wp-admin/admin-ajax.php(187): do_action('wp_ajax_submit_...')
#5 {main}
thrown [array (
'trace' => '
#0: Elementor\Core\Logger\Manager -> shutdown()
',
)]
PHP: 2022-10-23 17:50:01 [warning X 3][/var/web/site/public_html/wp-content/plugins/elementor-pro/modules/payments/module.php::300] Undefined array key "stripe_test_env_tax_rates_list" [array (
'trace' => '
#0: Elementor\Core\Logger\Manager -> shutdown()
',
)]
PHP: 2022-10-23 18:16:38 [notice X 383][/var/web/staging/public_html/wp-content/plugins/elementor/modules/performance-lab/module.php::31] Required parameter $url follows optional parameter $size [array (
'trace' => '
#0: Elementor\Core\Logger\Manager -> shutdown()
',
)]
PHP: 2022-10-23 18:23:10 [warning X 1][/var/web/staging/public_html/wp-content/plugins/elementor-pro/core/app/modules/site-editor/data/endpoints/templates.php::150] Undefined array key "condition_type" [array (
'trace' => '
#0: /var/web/staging/public_html/wp-content/plugins/elementor-pro/core/app/modules/site-editor/data/endpoints/templates.php(150): Elementor\Core\Logger\Manager -> rest_error_handler()
#1: ElementorPro\Core\App\Modules\SiteEditor\Data\Endpoints\Templates -> normalize_template_json_item()
#2: /var/web/staging/public_html/wp-content/plugins/elementor-pro/core/app/modules/site-editor/data/endpoints/templates.php(120): class type array_map()
#3: /var/web/staging/public_html/wp-content/plugins/elementor-pro/core/app/modules/site-editor/data/endpoints/templates.php(59): ElementorPro\Core\App\Modules\SiteEditor\Data\Endpoints\Templates -> normalize_templates_json()
#4: /var/web/staging/public_html/wp-content/plugins/elementor/data/base/endpoint.php(158): ElementorPro\Core\App\Modules\SiteEditor\Data\Endpoints\Templates -> get_items()
',
)]
PHP: 2022-10-23 18:27:05 [error X 4][/var/web/staging/public_html/wp-content/plugins/elementor/includes/managers/elements.php::71] Uncaught TypeError: Elementor\Elements_Manager::create_element_instance(): Argument #1 ($element_data) must be of type array, bool given, called in /var/web/staging/public_html/wp-content/plugins/elementor-pro/modules/payments/module.php on line 289 and defined in /var/web/staging/public_html/wp-content/plugins/elementor/includes/managers/elements.php:71
Stack trace:
#0 /var/web/staging/public_html/wp-content/plugins/elementor-pro/modules/payments/module.php(289): Elementor\Elements_Manager->create_element_instance(false)
#1 /var/web/staging/public_html/wp-includes/class-wp-hook.php(307): ElementorPro\Modules\Payments\Module->submit_stripe_form('')
#2 /var/web/staging/public_html/wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters('', Array)
#3 /var/web/staging/public_html/wp-includes/plugin.php(476): WP_Hook->do_action(Array)
#4 /var/web/staging/public_html/wp-admin/admin-ajax.php(187): do_action('wp_ajax_submit_...')
#5 {main}
thrown [array (
'trace' => '
#0: Elementor\Core\Logger\Manager -> shutdown()
',
)]
== Elementor - Compatibility Tag ==
Elementor Pro: Incompatible
== Elementor Pro - Compatibility Tag ==
|
1.0
|
🐞 Bug Report: Loop Builder - Stripe Button is throwing a 500 error when added to a loop template - ### Prerequisites
- [X] I have searched for similar issues in both open and closed tickets and cannot find a duplicate.
- [X] The issue still exists against the latest stable version of Elementor.
### Description
I added a stripe button into a loop and it doesn't work.
### Steps to reproduce
1) Create a loop with Stripe Button included.
2) Open dev tools
3) Test using Stripe Test Secret Key.
### Isolating the problem
- [X] This bug happens with only Elementor plugin active (and Elementor Pro).
- [ ] This bug happens with a Blank WordPress theme active ([Hello theme](https://wordpress.org/themes/hello-elementor/)).
- [X] I can reproduce this bug consistently using the steps above.
### System Info
== Server Environment ==
Operating System: Linux
Software: nginx/1.20.1
MySQL version: mariadb.org binary distribution v10.3.36
PHP Version: 8.0.24
PHP Memory Limit: 256M
PHP Max Input Vars: 3000
PHP Max Post Size: 128M
GD Installed: Yes
ZIP Installed: Yes
Write Permissions: All right
Elementor Library: Connected
== WordPress Environment ==
Version: 6.0.3
Site URL: https://utahmountainair.staging.tempurl.host
Home URL: https://utahmountainair.staging.tempurl.host
WP Multisite: No
Max Upload Size: 128 MB
Memory limit: 40M
Max Memory limit: 256M
Permalink Structure: /%postname%/
Language: en-US
Timezone: 0
Admin Email: daniel@finidev.tech
Debug Mode: Inactive
== Theme ==
Name: Hello Elementor
Version: 2.6.1
Author: Elementor Team
Child Theme: No
== User ==
Role: administrator
WP Profile lang: en_US
User Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/16.0 Safari/605.1.15
== Active Plugins ==
Defender Pro
Version: 3.3.3
Author: WPMU DEV
Duplicate Page
Version: 4.4.9
Author: mndpsingh287
Easy Media Replace
Version: 0.1.3
Author: Nabil Lemsieh
Elementor
Version: 3.9.0-dev3
Author: Elementor.com
Elementor Beta (Developer Edition)
Version: 1.1.1
Author: Elementor.com
Elementor Pro
Version: 3.9.0-dev4
Author: Elementor.com
Hummingbird Pro
Version: 3.3.5
Author: WPMU DEV
JetEngine
Version: 3.0.5
Author: Crocoblock
JetEngine - Custom visibility conditions
Version: 1.1.0
Author: Crocoblock
Smush Pro
Version: 3.12.2
Author: WPMU DEV
Ultimate Dashboard
Version: 3.7.1.5
Author: David Vongries
Ultimate Dashboard PRO
Version: 3.6.4
Author: David Vongries
User Profile Picture
Version: 2.6.0
Author: Cozmoslabs
WPMU DEV Dashboard
Version: 4.11.17
Author: WPMU DEV
== Must-Use Plugins ==
Hosting
Version: 0.1
Author:
== Elements Usage ==
container : 1
button : 1
heading : 1
image : 1
text-editor : 1
footer : 1
heading : 7
icon : 3
icon-list : 2
image : 2
social-icons : 1
spacer : 2
text-editor : 1
theme-site-logo : 1
header : 1
button : 3
heading : 2
icon-list : 2
nav-menu : 1
theme-site-logo : 1
loop : 0
container : 1
heading : 2
image : 1
spacer : 2
stripe-button : 1
text-editor : 1
page : 9
button : 41
counter : 16
divider : 13
form : 6
google_maps : 1
heading : 104
html : 1
icon : 3
icon-box : 17
icon-list : 10
image : 72
image-box : 14
progress : 7
social-icons : 1
spacer : 1
text-editor : 63
toggle : 5
video : 1
section : 1
button : 9
counter : 4
form : 1
heading : 17
icon : 1
icon-list : 3
image : 24
image-box : 1
text-editor : 7
video : 1
widget : 0
form : 1
wp-page : 7
button : 11
container : 18
counter : 9
form : 2
heading : 35
html : 2
icon : 5
icon-list : 6
image : 40
jet-listing-dynamic-field : 2
jet-listing-grid : 3
loop-grid : 1
spacer : 5
text-editor : 12
video : 1
loop-item : 1
container : 1
heading : 2
image : 1
spacer : 2
stripe-button : 1
text-editor : 1
jet-listing-items : 0
button : 2
container : 7
heading : 2
icon-box : 1
image : 3
jet-listing-dynamic-field : 5
spacer : 2
stripe-button : 1
text-editor : 1
== Elementor Experiments ==
Optimized DOM Output: Active
Improved Asset Loading: Active
Improved CSS Loading: Active
Inline Font Icons: Active
Accessibility Improvements: Active
Additional Custom Breakpoints: Active
Import Export Website Kit: Active
Hide native WordPress widgets from search results: Active
admin_menu_rearrangement: Inactive by default
Flexbox Container: Active
Default to New Theme Builder: Active
Hello Theme Header & Footer: Active
Landing Pages: Active
Color Sampler: Active
Favorite Widgets: Active
Admin Top Bar: Active
Page Transitions: Active
Notes: Active
Loop: Active
Form Submissions: Active
Scroll Snap: Active
== Log ==
Log: showing 20 of 352022-10-03 08:08:06 [info] Elementor data updater process has been completed. [array (
'plugin' => 'Elementor',
'from' => '3.7.7',
'to' => '3.7.8',
)]
2022-10-23 18:10:54 [info] Elementor data updater process has been queued. [array (
'plugin' => 'Elementor',
'from' => '3.7.8',
'to' => '3.8.0-beta5',
)]
2022-10-23 18:10:54 [info] elementor::elementor_updater Started
2022-10-23 18:10:55 [info] Elementor/Upgrades - _on_each_version Start
2022-10-23 18:10:56 [info] Elementor/Upgrades - _on_each_version Finished
2022-10-23 18:10:56 [info] Elementor/Upgrades - _v_3_8_0_fix_php8_image_custom_size Start
2022-10-23 18:10:56 [info] Elementor/Upgrades - _v_3_8_0_fix_php8_image_custom_size Finished
2022-10-23 18:10:56 [info] Elementor data updater process has been completed. [array (
'plugin' => 'Elementor',
'from' => '3.7.8',
'to' => '3.8.0-beta5',
)]
2022-10-23 18:16:35 [info] Elementor data updater process has been queued. [array (
'plugin' => 'Elementor',
'from' => '3.8.0-beta5',
'to' => '3.9.0-dev3',
)]
2022-10-23 18:16:37 [info] elementor::elementor_updater Started
2022-10-23 18:16:37 [info] Elementor/Upgrades - _on_each_version Start
2022-10-23 18:16:38 [info] Elementor/Upgrades - _on_each_version Finished
2022-10-23 18:16:38 [info] Elementor/Upgrades - _v_3_8_0_fix_php8_image_custom_size Start
2022-10-23 18:16:38 [info] Elementor/Upgrades - _v_3_8_0_fix_php8_image_custom_size Finished
2022-10-23 18:16:38 [info] Elementor data updater process has been completed. [array (
'plugin' => 'Elementor',
'from' => '3.8.0-beta5',
'to' => '3.9.0-dev3',
)]
2022-10-23 18:17:06 [info] Elementor data updater process has been queued. [array (
'plugin' => 'Elementor Pro',
'from' => '3.7.7',
'to' => '3.9.0-dev4',
)]
2022-10-23 18:17:07 [info] elementor-pro::elementor_pro_updater Started
2022-10-23 18:17:07 [info] Elementor Pro/Upgrades - _on_each_version Start
2022-10-23 18:17:07 [info] Elementor Pro/Upgrades - _on_each_version Finished
2022-10-23 18:17:07 [info] Elementor data updater process has been completed. [array (
'plugin' => 'Elementor Pro',
'from' => '3.7.7',
'to' => '3.9.0-dev4',
)]
PHP: showing 7 of 7PHP: 2022-10-20 22:46:31 [warning X 5][/var/web/site/public_html/wp-content/plugins/elementor/core/common/modules/ajax/module.php::171] Undefined array key "data" [array (
'trace' => '
#0: Elementor\Core\Logger\Manager -> shutdown()
',
)]
PHP: 2022-10-21 02:42:03 [error X 30][/var/web/site/public_html/wp-content/plugins/elementor/includes/managers/elements.php::71] Uncaught TypeError: Elementor\Elements_Manager::create_element_instance(): Argument #1 ($element_data) must be of type array, bool given, called in /var/web/site/public_html/wp-content/plugins/elementor-pro/modules/payments/module.php on line 289 and defined in /var/web/site/public_html/wp-content/plugins/elementor/includes/managers/elements.php:71
Stack trace:
#0 /var/web/site/public_html/wp-content/plugins/elementor-pro/modules/payments/module.php(289): Elementor\Elements_Manager->create_element_instance(false)
#1 /var/web/site/public_html/wp-includes/class-wp-hook.php(307): ElementorPro\Modules\Payments\Module->submit_stripe_form('')
#2 /var/web/site/public_html/wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters('', Array)
#3 /var/web/site/public_html/wp-includes/plugin.php(476): WP_Hook->do_action(Array)
#4 /var/web/site/public_html/wp-admin/admin-ajax.php(202): do_action('wp_ajax_nopriv_...')
#5 {main}
thrown [array (
'trace' => '
#0: Elementor\Core\Logger\Manager -> shutdown()
',
)]
PHP: 2022-10-21 03:02:34 [error X 23][/var/web/site/public_html/wp-content/plugins/elementor/includes/managers/elements.php::71] Uncaught TypeError: Elementor\Elements_Manager::create_element_instance(): Argument #1 ($element_data) must be of type array, bool given, called in /var/web/site/public_html/wp-content/plugins/elementor-pro/modules/payments/module.php on line 289 and defined in /var/web/site/public_html/wp-content/plugins/elementor/includes/managers/elements.php:71
Stack trace:
#0 /var/web/site/public_html/wp-content/plugins/elementor-pro/modules/payments/module.php(289): Elementor\Elements_Manager->create_element_instance(false)
#1 /var/web/site/public_html/wp-includes/class-wp-hook.php(307): ElementorPro\Modules\Payments\Module->submit_stripe_form('')
#2 /var/web/site/public_html/wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters('', Array)
#3 /var/web/site/public_html/wp-includes/plugin.php(476): WP_Hook->do_action(Array)
#4 /var/web/site/public_html/wp-admin/admin-ajax.php(187): do_action('wp_ajax_submit_...')
#5 {main}
thrown [array (
'trace' => '
#0: Elementor\Core\Logger\Manager -> shutdown()
',
)]
PHP: 2022-10-23 17:50:01 [warning X 3][/var/web/site/public_html/wp-content/plugins/elementor-pro/modules/payments/module.php::300] Undefined array key "stripe_test_env_tax_rates_list" [array (
'trace' => '
#0: Elementor\Core\Logger\Manager -> shutdown()
',
)]
PHP: 2022-10-23 18:16:38 [notice X 383][/var/web/staging/public_html/wp-content/plugins/elementor/modules/performance-lab/module.php::31] Required parameter $url follows optional parameter $size [array (
'trace' => '
#0: Elementor\Core\Logger\Manager -> shutdown()
',
)]
PHP: 2022-10-23 18:23:10 [warning X 1][/var/web/staging/public_html/wp-content/plugins/elementor-pro/core/app/modules/site-editor/data/endpoints/templates.php::150] Undefined array key "condition_type" [array (
'trace' => '
#0: /var/web/staging/public_html/wp-content/plugins/elementor-pro/core/app/modules/site-editor/data/endpoints/templates.php(150): Elementor\Core\Logger\Manager -> rest_error_handler()
#1: ElementorPro\Core\App\Modules\SiteEditor\Data\Endpoints\Templates -> normalize_template_json_item()
#2: /var/web/staging/public_html/wp-content/plugins/elementor-pro/core/app/modules/site-editor/data/endpoints/templates.php(120): class type array_map()
#3: /var/web/staging/public_html/wp-content/plugins/elementor-pro/core/app/modules/site-editor/data/endpoints/templates.php(59): ElementorPro\Core\App\Modules\SiteEditor\Data\Endpoints\Templates -> normalize_templates_json()
#4: /var/web/staging/public_html/wp-content/plugins/elementor/data/base/endpoint.php(158): ElementorPro\Core\App\Modules\SiteEditor\Data\Endpoints\Templates -> get_items()
',
)]
PHP: 2022-10-23 18:27:05 [error X 4][/var/web/staging/public_html/wp-content/plugins/elementor/includes/managers/elements.php::71] Uncaught TypeError: Elementor\Elements_Manager::create_element_instance(): Argument #1 ($element_data) must be of type array, bool given, called in /var/web/staging/public_html/wp-content/plugins/elementor-pro/modules/payments/module.php on line 289 and defined in /var/web/staging/public_html/wp-content/plugins/elementor/includes/managers/elements.php:71
Stack trace:
#0 /var/web/staging/public_html/wp-content/plugins/elementor-pro/modules/payments/module.php(289): Elementor\Elements_Manager->create_element_instance(false)
#1 /var/web/staging/public_html/wp-includes/class-wp-hook.php(307): ElementorPro\Modules\Payments\Module->submit_stripe_form('')
#2 /var/web/staging/public_html/wp-includes/class-wp-hook.php(331): WP_Hook->apply_filters('', Array)
#3 /var/web/staging/public_html/wp-includes/plugin.php(476): WP_Hook->do_action(Array)
#4 /var/web/staging/public_html/wp-admin/admin-ajax.php(187): do_action('wp_ajax_submit_...')
#5 {main}
thrown [array (
'trace' => '
#0: Elementor\Core\Logger\Manager -> shutdown()
',
)]
== Elementor - Compatibility Tag ==
Elementor Pro: Incompatible
== Elementor Pro - Compatibility Tag ==
|
non_process
|
🐞 bug report loop builder stripe button is throwing a error when added to a loop template prerequisites i have searched for similar issues in both open and closed tickets and cannot find a duplicate the issue still exists against the latest stable version of elementor description i added a stripe button into a loop and it doesn t work steps to reproduce create a loop with stripe button included open dev tools test using stripe test secret key isolating the problem this bug happens with only elementor plugin active and elementor pro this bug happens with a blank wordpress theme active i can reproduce this bug consistently using the steps above system info server environment operating system linux software nginx mysql version mariadb org binary distribution php version php memory limit php max input vars php max post size gd installed yes zip installed yes write permissions all right elementor library connected wordpress environment version site url home url wp multisite no max upload size mb memory limit max memory limit permalink structure postname language en us timezone admin email daniel finidev tech debug mode inactive theme name hello elementor version author elementor team child theme no user role administrator wp profile lang en us user agent mozilla macintosh intel mac os x applewebkit khtml like gecko version safari active plugins defender pro version author wpmu dev duplicate page version author easy media replace version author nabil lemsieh elementor version author elementor com elementor beta developer edition version author elementor com elementor pro version author elementor com hummingbird pro version author wpmu dev jetengine version author crocoblock jetengine custom visibility conditions version author crocoblock smush pro version author wpmu dev ultimate dashboard version author david vongries ultimate dashboard pro version author david vongries user profile picture version author cozmoslabs wpmu dev dashboard version author wpmu dev must use plugins hosting version author elements usage container button heading image text editor footer heading icon icon list image social icons spacer text editor theme site logo header button heading icon list nav menu theme site logo loop container heading image spacer stripe button text editor page button counter divider form google maps heading html icon icon box icon list image image box progress social icons spacer text editor toggle video section button counter form heading icon icon list image image box text editor video widget form wp page button container counter form heading html icon icon list image jet listing dynamic field jet listing grid loop grid spacer text editor video loop item container heading image spacer stripe button text editor jet listing items button container heading icon box image jet listing dynamic field spacer stripe button text editor elementor experiments optimized dom output active improved asset loading active improved css loading active inline font icons active accessibility improvements active additional custom breakpoints active import export website kit active hide native wordpress widgets from search results active admin menu rearrangement inactive by default flexbox container active default to new theme builder active hello theme header footer active landing pages active color sampler active favorite widgets active admin top bar active page transitions active notes active loop active form submissions active scroll snap active log log showing of elementor data updater process has been completed array plugin elementor from to elementor data updater process has been queued array plugin elementor from to elementor elementor updater started elementor upgrades on each version start elementor upgrades on each version finished elementor upgrades v fix image custom size start elementor upgrades v fix image custom size finished elementor data updater process has been completed array plugin elementor from to elementor data updater process has been queued array plugin elementor from to elementor elementor updater started elementor upgrades on each version start elementor upgrades on each version finished elementor upgrades v fix image custom size start elementor upgrades v fix image custom size finished elementor data updater process has been completed array plugin elementor from to elementor data updater process has been queued array plugin elementor pro from to elementor pro elementor pro updater started elementor pro upgrades on each version start elementor pro upgrades on each version finished elementor data updater process has been completed array plugin elementor pro from to php showing of undefined array key data array trace elementor core logger manager shutdown php uncaught typeerror elementor elements manager create element instance argument element data must be of type array bool given called in var web site public html wp content plugins elementor pro modules payments module php on line and defined in var web site public html wp content plugins elementor includes managers elements php stack trace var web site public html wp content plugins elementor pro modules payments module php elementor elements manager create element instance false var web site public html wp includes class wp hook php elementorpro modules payments module submit stripe form var web site public html wp includes class wp hook php wp hook apply filters array var web site public html wp includes plugin php wp hook do action array var web site public html wp admin admin ajax php do action wp ajax nopriv main thrown array trace elementor core logger manager shutdown php uncaught typeerror elementor elements manager create element instance argument element data must be of type array bool given called in var web site public html wp content plugins elementor pro modules payments module php on line and defined in var web site public html wp content plugins elementor includes managers elements php stack trace var web site public html wp content plugins elementor pro modules payments module php elementor elements manager create element instance false var web site public html wp includes class wp hook php elementorpro modules payments module submit stripe form var web site public html wp includes class wp hook php wp hook apply filters array var web site public html wp includes plugin php wp hook do action array var web site public html wp admin admin ajax php do action wp ajax submit main thrown array trace elementor core logger manager shutdown php undefined array key stripe test env tax rates list array trace elementor core logger manager shutdown php required parameter url follows optional parameter size array trace elementor core logger manager shutdown php undefined array key condition type array trace var web staging public html wp content plugins elementor pro core app modules site editor data endpoints templates php elementor core logger manager rest error handler elementorpro core app modules siteeditor data endpoints templates normalize template json item var web staging public html wp content plugins elementor pro core app modules site editor data endpoints templates php class type array map var web staging public html wp content plugins elementor pro core app modules site editor data endpoints templates php elementorpro core app modules siteeditor data endpoints templates normalize templates json var web staging public html wp content plugins elementor data base endpoint php elementorpro core app modules siteeditor data endpoints templates get items php uncaught typeerror elementor elements manager create element instance argument element data must be of type array bool given called in var web staging public html wp content plugins elementor pro modules payments module php on line and defined in var web staging public html wp content plugins elementor includes managers elements php stack trace var web staging public html wp content plugins elementor pro modules payments module php elementor elements manager create element instance false var web staging public html wp includes class wp hook php elementorpro modules payments module submit stripe form var web staging public html wp includes class wp hook php wp hook apply filters array var web staging public html wp includes plugin php wp hook do action array var web staging public html wp admin admin ajax php do action wp ajax submit main thrown array trace elementor core logger manager shutdown elementor compatibility tag elementor pro incompatible elementor pro compatibility tag
| 0
|
3,729
| 6,733,142,344
|
IssuesEvent
|
2017-10-18 13:58:37
|
york-region-tpss/stp
|
https://api.github.com/repos/york-region-tpss/stp
|
closed
|
Contract preparation single item view - process request
|
process workflow
|
Update the status of the item as request fired by buttons
|
1.0
|
Contract preparation single item view - process request - Update the status of the item as request fired by buttons
|
process
|
contract preparation single item view process request update the status of the item as request fired by buttons
| 1
|
352,483
| 32,072,612,898
|
IssuesEvent
|
2023-09-25 08:59:42
|
chesslablab/react-chess
|
https://api.github.com/repos/chesslablab/react-chess
|
opened
|
Transcribe "Play C20 and load the PGN movetext" to Jest
|
good first issue hacktoberfest testing
|
Transcribe the **Play C20 and load the PGN movetext** test to Jest.
Happy coding,
|
1.0
|
Transcribe "Play C20 and load the PGN movetext" to Jest - Transcribe the **Play C20 and load the PGN movetext** test to Jest.
Happy coding,
|
non_process
|
transcribe play and load the pgn movetext to jest transcribe the play and load the pgn movetext test to jest happy coding
| 0
|
190,803
| 6,822,681,207
|
IssuesEvent
|
2017-11-07 20:56:42
|
status-im/status-react
|
https://api.github.com/repos/status-im/status-react
|
closed
|
Long address text overlaps chat area on Chats tab [iOS only]
|
bounty bug medium-priority
|
### Description
*Type*: Bug
*Summary*: only on iOS: Long address overlaps chat area in the Chats tab if chat is the last one in the list. It's expected that only first x words will be shown to use 2 lines instead of showing the complete address.
Ok scenario (first 2 lines are shown in chats for address is in green box on the picture below)
Issue (more than 2 lines are shown for an address is in red box on the picture below)

#### Expected behavior
Only 2 lines of the address are shown in Chats screen, so it fits the area for the chat
#### Actual behavior
Long address (e.g. 5 lines) is fully shown and overlaps chat area in the Chats tab
### Reproduction
- Open Status
- Open Chats
- Create a new group chat "Test", add Jarrad as a member.
- in created group chat tap on command /location
- in the input field next to location command type "Museum of medical instruments". As a result, suggestions list will be shown with this name
- Tap on "Museum of medical instruments of the Hos..." item in the suggestions list. As a result, location message with an address and a map is shown in the chat
- now return back to Chats list (tap on icon "<" in top left corner)
- if new chat with location command is not shown at the bottom of Chats list then open each chat below our "Test" chat and send something to the chat, e.g. word "text" - as a result, the chat will go up in the list of chats. You need to repeat it until the "Test" chat is at the very bottom in the Chats
- Check how location text is shown (currently, it takes 5 lines instead of expected 2 lines)
### Additional Information
* Status version: 0.9.8-63-g06bff32f+ and up to 0.9.12
* Operating System:
Real device iPhone 6s, iOS 10.2.1
|
1.0
|
Long address text overlaps chat area on Chats tab [iOS only] - ### Description
*Type*: Bug
*Summary*: only on iOS: Long address overlaps chat area in the Chats tab if chat is the last one in the list. It's expected that only first x words will be shown to use 2 lines instead of showing the complete address.
Ok scenario (first 2 lines are shown in chats for address is in green box on the picture below)
Issue (more than 2 lines are shown for an address is in red box on the picture below)

#### Expected behavior
Only 2 lines of the address are shown in Chats screen, so it fits the area for the chat
#### Actual behavior
Long address (e.g. 5 lines) is fully shown and overlaps chat area in the Chats tab
### Reproduction
- Open Status
- Open Chats
- Create a new group chat "Test", add Jarrad as a member.
- in created group chat tap on command /location
- in the input field next to location command type "Museum of medical instruments". As a result, suggestions list will be shown with this name
- Tap on "Museum of medical instruments of the Hos..." item in the suggestions list. As a result, location message with an address and a map is shown in the chat
- now return back to Chats list (tap on icon "<" in top left corner)
- if new chat with location command is not shown at the bottom of Chats list then open each chat below our "Test" chat and send something to the chat, e.g. word "text" - as a result, the chat will go up in the list of chats. You need to repeat it until the "Test" chat is at the very bottom in the Chats
- Check how location text is shown (currently, it takes 5 lines instead of expected 2 lines)
### Additional Information
* Status version: 0.9.8-63-g06bff32f+ and up to 0.9.12
* Operating System:
Real device iPhone 6s, iOS 10.2.1
|
non_process
|
long address text overlaps chat area on chats tab description type bug summary only on ios long address overlaps chat area in the chats tab if chat is the last one in the list it s expected that only first x words will be shown to use lines instead of showing the complete address ok scenario first lines are shown in chats for address is in green box on the picture below issue more than lines are shown for an address is in red box on the picture below expected behavior only lines of the address are shown in chats screen so it fits the area for the chat actual behavior long address e g lines is fully shown and overlaps chat area in the chats tab reproduction open status open chats create a new group chat test add jarrad as a member in created group chat tap on command location in the input field next to location command type museum of medical instruments as a result suggestions list will be shown with this name tap on museum of medical instruments of the hos item in the suggestions list as a result location message with an address and a map is shown in the chat now return back to chats list tap on icon in top left corner if new chat with location command is not shown at the bottom of chats list then open each chat below our test chat and send something to the chat e g word text as a result the chat will go up in the list of chats you need to repeat it until the test chat is at the very bottom in the chats check how location text is shown currently it takes lines instead of expected lines additional information status version and up to operating system real device iphone ios
| 0
|
61,004
| 14,597,690,032
|
IssuesEvent
|
2020-12-20 21:16:17
|
turkdevops/sourcegraph
|
https://api.github.com/repos/turkdevops/sourcegraph
|
opened
|
CVE-2019-10747 (High) detected in set-value-2.0.0.tgz, set-value-0.4.3.tgz
|
security vulnerability
|
## CVE-2019-10747 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>set-value-2.0.0.tgz</b>, <b>set-value-0.4.3.tgz</b></p></summary>
<p>
<details><summary><b>set-value-2.0.0.tgz</b></p></summary>
<p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p>
<p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz</a></p>
<p>Path to dependency file: sourcegraph/yarn.lock</p>
<p>Path to vulnerable library: sourcegraph/yarn.lock</p>
<p>
Dependency Hierarchy:
- gulp-4.0.2.tgz (Root Library)
- gulp-cli-2.2.0.tgz
- matchdep-2.0.0.tgz
- findup-sync-2.0.0.tgz
- micromatch-3.1.10.tgz
- extglob-2.0.4.tgz
- expand-brackets-2.1.4.tgz
- snapdragon-0.8.2.tgz
- base-0.11.2.tgz
- cache-base-1.0.1.tgz
- :x: **set-value-2.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>set-value-0.4.3.tgz</b></p></summary>
<p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p>
<p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz">https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz</a></p>
<p>Path to dependency file: sourcegraph/yarn.lock</p>
<p>Path to vulnerable library: sourcegraph/yarn.lock</p>
<p>
Dependency Hierarchy:
- gulp-4.0.2.tgz (Root Library)
- gulp-cli-2.2.0.tgz
- matchdep-2.0.0.tgz
- findup-sync-2.0.0.tgz
- micromatch-3.1.10.tgz
- extglob-2.0.4.tgz
- expand-brackets-2.1.4.tgz
- snapdragon-0.8.2.tgz
- base-0.11.2.tgz
- cache-base-1.0.1.tgz
- union-value-1.0.0.tgz
- :x: **set-value-0.4.3.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/sourcegraph/commit/a1adef14ff38d04f6dc3d8a762577426decb340e">a1adef14ff38d04f6dc3d8a762577426decb340e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
set-value is vulnerable to Prototype Pollution in versions lower than 3.0.1. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using any of the constructor, prototype and _proto_ payloads.
<p>Publish Date: 2019-08-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10747>CVE-2019-10747</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f">https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f</a></p>
<p>Release Date: 2019-07-24</p>
<p>Fix Resolution: 2.0.1,3.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-10747 (High) detected in set-value-2.0.0.tgz, set-value-0.4.3.tgz - ## CVE-2019-10747 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>set-value-2.0.0.tgz</b>, <b>set-value-0.4.3.tgz</b></p></summary>
<p>
<details><summary><b>set-value-2.0.0.tgz</b></p></summary>
<p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p>
<p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz</a></p>
<p>Path to dependency file: sourcegraph/yarn.lock</p>
<p>Path to vulnerable library: sourcegraph/yarn.lock</p>
<p>
Dependency Hierarchy:
- gulp-4.0.2.tgz (Root Library)
- gulp-cli-2.2.0.tgz
- matchdep-2.0.0.tgz
- findup-sync-2.0.0.tgz
- micromatch-3.1.10.tgz
- extglob-2.0.4.tgz
- expand-brackets-2.1.4.tgz
- snapdragon-0.8.2.tgz
- base-0.11.2.tgz
- cache-base-1.0.1.tgz
- :x: **set-value-2.0.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>set-value-0.4.3.tgz</b></p></summary>
<p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p>
<p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz">https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz</a></p>
<p>Path to dependency file: sourcegraph/yarn.lock</p>
<p>Path to vulnerable library: sourcegraph/yarn.lock</p>
<p>
Dependency Hierarchy:
- gulp-4.0.2.tgz (Root Library)
- gulp-cli-2.2.0.tgz
- matchdep-2.0.0.tgz
- findup-sync-2.0.0.tgz
- micromatch-3.1.10.tgz
- extglob-2.0.4.tgz
- expand-brackets-2.1.4.tgz
- snapdragon-0.8.2.tgz
- base-0.11.2.tgz
- cache-base-1.0.1.tgz
- union-value-1.0.0.tgz
- :x: **set-value-0.4.3.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/sourcegraph/commit/a1adef14ff38d04f6dc3d8a762577426decb340e">a1adef14ff38d04f6dc3d8a762577426decb340e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
set-value is vulnerable to Prototype Pollution in versions lower than 3.0.1. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using any of the constructor, prototype and _proto_ payloads.
<p>Publish Date: 2019-08-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10747>CVE-2019-10747</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f">https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f</a></p>
<p>Release Date: 2019-07-24</p>
<p>Fix Resolution: 2.0.1,3.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in set value tgz set value tgz cve high severity vulnerability vulnerable libraries set value tgz set value tgz set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href path to dependency file sourcegraph yarn lock path to vulnerable library sourcegraph yarn lock dependency hierarchy gulp tgz root library gulp cli tgz matchdep tgz findup sync tgz micromatch tgz extglob tgz expand brackets tgz snapdragon tgz base tgz cache base tgz x set value tgz vulnerable library set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href path to dependency file sourcegraph yarn lock path to vulnerable library sourcegraph yarn lock dependency hierarchy gulp tgz root library gulp cli tgz matchdep tgz findup sync tgz micromatch tgz extglob tgz expand brackets tgz snapdragon tgz base tgz cache base tgz union value tgz x set value tgz vulnerable library found in head commit a href found in base branch main vulnerability details set value is vulnerable to prototype pollution in versions lower than the function mixin deep could be tricked into adding or modifying properties of object prototype using any of the constructor prototype and proto payloads publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
13,489
| 10,281,001,987
|
IssuesEvent
|
2019-08-26 07:22:14
|
ampproject/amp.dev
|
https://api.github.com/repos/ampproject/amp.dev
|
reopened
|
Search: empty search window if no result is found
|
Category: Infrastructure P1: High Type: Bug
|
Steps to reproduce:
1. enter `amp-scri` as search query
2. wait
=> no result is shown, only an empty screen
<img width="1193" alt="Screen Shot 2019-08-24 at 20 51 58" src="https://user-images.githubusercontent.com/380472/63641616-7c64c280-c6b1-11e9-9cec-0cf7310e6210.png">
//cc @sebil @tharders
|
1.0
|
Search: empty search window if no result is found - Steps to reproduce:
1. enter `amp-scri` as search query
2. wait
=> no result is shown, only an empty screen
<img width="1193" alt="Screen Shot 2019-08-24 at 20 51 58" src="https://user-images.githubusercontent.com/380472/63641616-7c64c280-c6b1-11e9-9cec-0cf7310e6210.png">
//cc @sebil @tharders
|
non_process
|
search empty search window if no result is found steps to reproduce enter amp scri as search query wait no result is shown only an empty screen img width alt screen shot at src cc sebil tharders
| 0
|
14,668
| 10,210,643,079
|
IssuesEvent
|
2019-08-14 15:11:31
|
unipept/unipept
|
https://api.github.com/repos/unipept/unipept
|
closed
|
calculate_lca should calculate lca of organisms
|
/services/cli migrated
|
Instead of a list of peptides the input will be a list of taxon ids.
output should be a single name
_[Original issue](https://github.ugent.be/unipept/unipept/issues/216) by ghost on Thu Feb 20 2014 at 14:42._
_Closed by an unknown user on Tue Mar 18 2014 at 10:48._
|
1.0
|
calculate_lca should calculate lca of organisms - Instead of a list of peptides the input will be a list of taxon ids.
output should be a single name
_[Original issue](https://github.ugent.be/unipept/unipept/issues/216) by ghost on Thu Feb 20 2014 at 14:42._
_Closed by an unknown user on Tue Mar 18 2014 at 10:48._
|
non_process
|
calculate lca should calculate lca of organisms instead of a list of peptides the input will be a list of taxon ids output should be a single name by ghost on thu feb at closed by an unknown user on tue mar at
| 0
|
11,175
| 13,957,695,090
|
IssuesEvent
|
2020-10-24 08:11:26
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
MT - MITA: Harvesting MT
|
Geoportal Harvesting process MT - Malta
|
Dear Angelo,
I trust this message finds you well, On behalf of the Maltese INSPIRE Team, I wish you and all the INSPIRE Team at JRC a Happy New Year. Angelo, kindly can you perform a harvest on the Maltese CSW at your convenience as we need to check the results of some changes we did on our metadata.
Thanks and Regards,
Rene
|
1.0
|
MT - MITA: Harvesting MT - Dear Angelo,
I trust this message finds you well, On behalf of the Maltese INSPIRE Team, I wish you and all the INSPIRE Team at JRC a Happy New Year. Angelo, kindly can you perform a harvest on the Maltese CSW at your convenience as we need to check the results of some changes we did on our metadata.
Thanks and Regards,
Rene
|
process
|
mt mita harvesting mt dear angelo i trust this message finds you well on behalf of the maltese inspire team i wish you and all the inspire team at jrc a happy new year angelo kindly can you perform a harvest on the maltese csw at your convenience as we need to check the results of some changes we did on our metadata thanks and regards rene
| 1
|
239
| 2,564,828,875
|
IssuesEvent
|
2015-02-06 22:42:25
|
sympy/csympy
|
https://api.github.com/repos/sympy/csympy
|
opened
|
Be as fast as FORM
|
performance
|
[FORM](https://github.com/vermaseren/form) is a great tool for manipulating large expressions, the main application is particle physics (so it has built-in capabilities for traces of gamma matrices, etc.), but it's a general tool. In a nutshell, it expands everything immediately, and all operations in FORM are term by term (so no factorization for example, where you need to analyze the whole expression). Read the tutorial and reference manual for lots of examples of what it can do:
http://www.nikhef.nl/~form/maindir/documentation/tutorial/online/online.html
http://www.nikhef.nl/~form/maindir/documentation/reference/online/online.html
We should try to be as fast with CSymPy (while allowing more general manipulations).
Create `expand2.frm`:
```
#: TermsInSmall 1000000
#: SmallSize 100000000
Symbols x,y,z,w;
Local n = 15;
Local r = (w + (w + z + y + x)^n)*(w + z + y + x)^n;
.end
```
and run it:
```
$ form expand2.frm
FORM 4.1 (Feb 6 2015) 64-bits Run: Fri Feb 6 15:36:10 2015
#: TermsInSmall 1000000
#: SmallSize 100000000
Symbols x,y,z,w;
Local n = 15;
Local r = (w + (w + z + y + x)^n)*(w + z + y + x)^n;
.end
Time = 0.00 sec Generated terms = 1
n Terms in output = 1
Bytes used = 20
Time = 0.87 sec Generated terms = 666672
r Terms in output = 6272
Bytes used = 246080
0.87 sec out of 0.88 sec
```
Compared to CSymPy:
```
$ ./expand2
Expanding: (w + (w + z + y + x)^15)*(w + z + y + x)^15
865ms
number of terms: 6272
```
So we are about as fast. But we need to try more benchmarks.
|
True
|
Be as fast as FORM - [FORM](https://github.com/vermaseren/form) is a great tool for manipulating large expressions, the main application is particle physics (so it has built-in capabilities for traces of gamma matrices, etc.), but it's a general tool. In a nutshell, it expands everything immediately, and all operations in FORM are term by term (so no factorization for example, where you need to analyze the whole expression). Read the tutorial and reference manual for lots of examples of what it can do:
http://www.nikhef.nl/~form/maindir/documentation/tutorial/online/online.html
http://www.nikhef.nl/~form/maindir/documentation/reference/online/online.html
We should try to be as fast with CSymPy (while allowing more general manipulations).
Create `expand2.frm`:
```
#: TermsInSmall 1000000
#: SmallSize 100000000
Symbols x,y,z,w;
Local n = 15;
Local r = (w + (w + z + y + x)^n)*(w + z + y + x)^n;
.end
```
and run it:
```
$ form expand2.frm
FORM 4.1 (Feb 6 2015) 64-bits Run: Fri Feb 6 15:36:10 2015
#: TermsInSmall 1000000
#: SmallSize 100000000
Symbols x,y,z,w;
Local n = 15;
Local r = (w + (w + z + y + x)^n)*(w + z + y + x)^n;
.end
Time = 0.00 sec Generated terms = 1
n Terms in output = 1
Bytes used = 20
Time = 0.87 sec Generated terms = 666672
r Terms in output = 6272
Bytes used = 246080
0.87 sec out of 0.88 sec
```
Compared to CSymPy:
```
$ ./expand2
Expanding: (w + (w + z + y + x)^15)*(w + z + y + x)^15
865ms
number of terms: 6272
```
So we are about as fast. But we need to try more benchmarks.
|
non_process
|
be as fast as form is a great tool for manipulating large expressions the main application is particle physics so it has built in capabilities for traces of gamma matrices etc but it s a general tool in a nutshell it expands everything immediately and all operations in form are term by term so no factorization for example where you need to analyze the whole expression read the tutorial and reference manual for lots of examples of what it can do we should try to be as fast with csympy while allowing more general manipulations create frm termsinsmall smallsize symbols x y z w local n local r w w z y x n w z y x n end and run it form frm form feb bits run fri feb termsinsmall smallsize symbols x y z w local n local r w w z y x n w z y x n end time sec generated terms n terms in output bytes used time sec generated terms r terms in output bytes used sec out of sec compared to csympy expanding w w z y x w z y x number of terms so we are about as fast but we need to try more benchmarks
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.