Unnamed: 0
int64
1
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
3
438
labels
stringlengths
4
308
body
stringlengths
7
254k
index
stringclasses
7 values
text_combine
stringlengths
96
254k
label
stringclasses
2 values
text
stringlengths
96
246k
binary_label
int64
0
1
139,704
12,878,229,310
IssuesEvent
2020-07-11 15:28:26
UC-Davis-molecular-computing/scadnano-python-package
https://api.github.com/repos/UC-Davis-molecular-computing/scadnano-python-package
opened
add CONTRIBUTING document
documentation high priority
This repo should have a CONTRIBUTING.md document, similar to that of the [web interface](https://github.com/UC-Davis-molecular-computing/scadnano/blob/master/CONTRIBUTING.md), explaining how to contribute. It should follow the same basic model, although it does not require the full explanation of the React/Redux architecture, since the Python package architecture is much simpler.
1.0
add CONTRIBUTING document - This repo should have a CONTRIBUTING.md document, similar to that of the [web interface](https://github.com/UC-Davis-molecular-computing/scadnano/blob/master/CONTRIBUTING.md), explaining how to contribute. It should follow the same basic model, although it does not require the full explanation of the React/Redux architecture, since the Python package architecture is much simpler.
non_main
add contributing document this repo should have a contributing md document similar to that of the explaining how to contribute it should follow the same basic model although it does not require the full explanation of the react redux architecture since the python package architecture is much simpler
0
16,836
9,536,669,789
IssuesEvent
2019-04-30 10:20:25
Garados007/Werwolf
https://api.github.com/repos/Garados007/Werwolf
closed
Optimiere Abrufe bei Spielrundenwechsel
difficult performance
Bei einem Wechsel der aktuellen Spielrunde werden ein Großteil der Daten (z.B. Chateinträge) verworfen und müssen neu abgerufen werden. Ein Teil davon ändert sich aber nicht in der nächsten Runde und soll nur ausgeblendet werden oder ungültige registrierte periodische Abrfragen existieren. Diese Abfragen lassen sich optimieren.
True
Optimiere Abrufe bei Spielrundenwechsel - Bei einem Wechsel der aktuellen Spielrunde werden ein Großteil der Daten (z.B. Chateinträge) verworfen und müssen neu abgerufen werden. Ein Teil davon ändert sich aber nicht in der nächsten Runde und soll nur ausgeblendet werden oder ungültige registrierte periodische Abrfragen existieren. Diese Abfragen lassen sich optimieren.
non_main
optimiere abrufe bei spielrundenwechsel bei einem wechsel der aktuellen spielrunde werden ein großteil der daten z b chateinträge verworfen und müssen neu abgerufen werden ein teil davon ändert sich aber nicht in der nächsten runde und soll nur ausgeblendet werden oder ungültige registrierte periodische abrfragen existieren diese abfragen lassen sich optimieren
0
5,212
26,464,339,927
IssuesEvent
2023-01-16 21:18:18
bazelbuild/intellij
https://api.github.com/repos/bazelbuild/intellij
closed
Flag --incompatible_disable_starlark_host_transitions will break IntelliJ Plugin Aspect Google in Bazel 7.0
type: bug product: IntelliJ topic: bazel awaiting-maintainer
Incompatible flag `--incompatible_disable_starlark_host_transitions` will be enabled by default in the next major release (Bazel 7.0), thus breaking IntelliJ Plugin Aspect Google. Please migrate to fix this and unblock the flip of this flag. The flag is documented here: [bazelbuild/bazel#17032](https://github.com/bazelbuild/bazel/issues/17032). Please check the following CI builds for build and test results: - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4af5-4f0f-b210-a8bc53dc2a30) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4af2-4446-bfab-24a8d74e2d5d) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4af7-465b-827b-23bc5b62642f) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4afc-4532-9b35-a42b7306f517) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4afa-4b25-9644-b488cef16618) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4aff-4819-ac6d-a9edbfe3305c) Never heard of incompatible flags before? We have [documentation](https://docs.bazel.build/versions/master/backward-compatibility.html) that explains everything. If you have any questions, please file an issue in https://github.com/bazelbuild/continuous-integration.
True
Flag --incompatible_disable_starlark_host_transitions will break IntelliJ Plugin Aspect Google in Bazel 7.0 - Incompatible flag `--incompatible_disable_starlark_host_transitions` will be enabled by default in the next major release (Bazel 7.0), thus breaking IntelliJ Plugin Aspect Google. Please migrate to fix this and unblock the flip of this flag. The flag is documented here: [bazelbuild/bazel#17032](https://github.com/bazelbuild/bazel/issues/17032). Please check the following CI builds for build and test results: - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4af5-4f0f-b210-a8bc53dc2a30) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4af2-4446-bfab-24a8d74e2d5d) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4af7-465b-827b-23bc5b62642f) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4afc-4532-9b35-a42b7306f517) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4afa-4b25-9644-b488cef16618) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154b-4aff-4819-ac6d-a9edbfe3305c) Never heard of incompatible flags before? We have [documentation](https://docs.bazel.build/versions/master/backward-compatibility.html) that explains everything. If you have any questions, please file an issue in https://github.com/bazelbuild/continuous-integration.
main
flag incompatible disable starlark host transitions will break intellij plugin aspect google in bazel incompatible flag incompatible disable starlark host transitions will be enabled by default in the next major release bazel thus breaking intellij plugin aspect google please migrate to fix this and unblock the flip of this flag the flag is documented here please check the following ci builds for build and test results never heard of incompatible flags before we have that explains everything if you have any questions please file an issue in
1
2,868
10,275,584,891
IssuesEvent
2019-08-24 09:14:44
arcticicestudio/arctic
https://api.github.com/repos/arcticicestudio/arctic
opened
ESLint
context-workflow scope-dx scope-maintainability scope-quality scope-stability type-feature
<p align="center"><img src="https://user-images.githubusercontent.com/7836623/63634799-a8555900-c65b-11e9-8988-dd0bc5d9c03b.png" /></p> Integrate [ESLint][], the _pluggable_ and de-facto standard linting utility for JavaScript. ### Configuration Preset The configuration presets that will be used are [@arcticicestudio/eslint-config][pr] that implements the [Arctic Ice Studio JavaScript Style][stg-js]. It comes with the following peer dependencies: - [eslint][esl-gh] It it built on top of [@arcticicestudio/eslint-config-base][pr-b] that includes various rules of the following plugins and rule presets that are therefore also required peer dependencies: - [eslint-config-prettier][esl-c-pr] - [eslint-plugin-import][esl-p-i] - [eslint-plugin-jsx-a11y][esl-p-a11y] - [eslint-plugin-prettier][esl-p-pr] - [eslint-plugin-react-hooks][esl-p-r-h] - [eslint-plugin-react][esl-p-r] Since _arctic_ will be built with [TypeScript][ts], the [@arcticicestudio/eslint-config-typescript][pr-ts] preset will be extended to add support for _TypeScript_ source file linting and compatibility with [Prettier][] through the [`@arcticicestudio/eslint-config-typescript/prettier` extension entry point][pr-ts-d#ep]. This preset requires the following peer dependencies: - [@typescript-eslint/eslint-plugin][esl-ts-p] - [@typescript-eslint/parser][esl-ts-pa] - [typescript][gh-ts] Since the custom presets are still in major version `0` note that the version range should be `>=0.x.x <1.0.0` to avoid the “SemVer Major Zero Caveat”. When defining package versions with the the carat `^` or tilde `~` range selector it won't affect packages with a major version of `0`. _yarn_ will resolve these packages to their exact version until the major version is greater or equal to `1`. To avoid this caveat the more detailed version range `>=0.x.x <1.0.0` should be used to resolve all versions greater or equal to `0.x.x` but less than `1.0.0`. This will always use the latest `0.x.x` version and removes the need to increment the version manually on each new release. To allow to lint TypeScript code the `@typescript-eslint/parser` parser will be used and [specified as main parser][esl-d-parser]. Also to make use of the latest experimental Babel features and proposals, [eslint-plugin-babel][esl-p-b] will be added with the following rule configurations: - `babel/camelcase` with level `error` - doesn't complain about optional chaining (`let foo = bar?.a_b;`). Note that the [core rule `camelcase`][esl-r-cc] must be disabled! - `babel/no-unused-expressions` with level `error` - doesn't fail when using `do` expressions or optional chaining (`a?.b()`). Note that the [core rule `no-unused-expressions`][esl-r-nue] must be disabled! See the [documentation of provided rules][esl-p-b#r]and required configurations to use them. The `.eslintrc.js` configuration file will be placed in the project root next to the `.eslintignore` file to define ignore pattern. ### Package script To allow to run the JavaScript linting separately a `lint:js` npm script/task will be added to be included in the main `lint` script flow. To use the great [auto-fixing][esl-d-cli#af] feature another `format:js` script/task will be added. ## Tasks - [ ] Install required packages to as development dependencies: - [@arcticicestudio/eslint-config-typescript][npm-esl-c-ais-ts] - [@arcticicestudio/eslint-config][npm-esl-c-ais] - [eslint-config-prettier][npm-esl-c-pr] - [eslint-plugin-babel][npm-esl-p-b] - [eslint-plugin-import][npm-esl-p-i] - [eslint-plugin-jsx-a11y][npm-esl-p-a11y] - [eslint-plugin-prettier][npm-esl-p-pr] - [eslint-plugin-react-hooks][npm-esl-p-r-h] - [eslint-plugin-react][npm-esl-p-r] - [eslint][npm-esl] - [typescript][npm-ts] - [ ] Implement `.eslintrc.js` configuration file. - [ ] Extend installed presets. - [ ] Integrate [eslint-plugin-babel][npm-esl-p-b] - [ ] Enable `babel/no-unused-expressions` and `babel/camelcase` rules including the deactivation of their associated core rules. - [ ] Add `babel` to the array of enabled plugins. - [ ] Implement `.eslintignore` ignore pattern file. - [ ] Implement npm `lint:js`, `lint:ts` and `format:js` scripts. - [ ] Lint current code base for the first time and fix possible JavaScript style guide violations. [esl-c-pr]: https://github.com/prettier/eslint-config-prettier [esl-d-cli#af]: https://eslint.org/docs/user-guide/command-line-interface#fixing-problems [esl-d-parser]: https://eslint.org/docs/user-guide/configuring#specifying-parser [esl-p-a11y]: https://github.com/evcohen/eslint-plugin-jsx-a11y [esl-p-b]: https://github.com/babel/eslint-plugin-babel [esl-p-b#r]: https://github.com/babel/eslint-plugin-babel#rules [esl-p-i]: https://github.com/benmosher/eslint-plugin-import [esl-p-pr]: https://github.com/prettier/eslint-plugin-prettier [esl-p-r-h]: https://github.com/facebook/react/tree/master/packages/eslint-plugin-react-hooks [esl-p-r]: https://github.com/yannickcr/eslint-plugin-react [esl-r-cc]: https://eslint.org/docs/rules/camelcase [esl-r-nue]: https://eslint.org/docs/rules/no-unused-expressions [esl-ts-p]: https://github.com/typescript-eslint/typescript-eslint/tree/master/packages/eslint-plugin [esl-ts-pa]: https://github.com/typescript-eslint/typescript-eslint/tree/master/packages/parser [eslint]: https://eslint.org [gh-esl]: https://github.com/eslint/eslint [gh-ts]: https://github.com/Microsoft/TypeScript [npm-esl-c-ais-ts]: https://www.npmjs.com/package/%40arcticicestudio/eslint-config-typescript [npm-esl-c-ais]: https://www.npmjs.com/package/%40arcticicestudio/eslint-config [npm-esl-c-pr]: https://www.npmjs.com/package/eslint-config-prettier [npm-esl-p-a11y]: https://www.npmjs.com/package/eslint-plugin-jsx-a11y [npm-esl-p-b]: https://www.npmjs.com/package/eslint-plugin-babel [npm-esl-p-i]: https://www.npmjs.com/package/eslint-plugin-import [npm-esl-p-pr]: https://www.npmjs.com/package/eslint-plugin-prettier [npm-esl-p-r-h]: https://www.npmjs.com/package/eslint-plugin-react-hooks [npm-esl-p-r]: https://www.npmjs.com/package/eslint-plugin-react [npm-esl]: https://www.npmjs.com/package/eslint [npm-ts]: https://www.npmjs.com/package/typescript [pr-b]: https://github.com/arcticicestudio/styleguide-javascript/tree/develop/packages/%40arcticicestudio/eslint-config-base [pr-ts-d#ep]: https://github.com/arcticicestudio/styleguide-javascript/blob/develop/packages/%40arcticicestudio/eslint-config-typescript/README.md#entry-points [pr-ts]: https://github.com/arcticicestudio/styleguide-javascript/tree/develop/packages/%40arcticicestudio/eslint-config-typescript [pr]: https://github.com/arcticicestudio/styleguide-javascript/tree/develop/packages/%40arcticicestudio/eslint-config [prettier]: https://prettier.io [stg-js]: https://arcticicestudio.github.io/styleguide-javascript [ts]: https://www.typescriptlang.org
True
ESLint - <p align="center"><img src="https://user-images.githubusercontent.com/7836623/63634799-a8555900-c65b-11e9-8988-dd0bc5d9c03b.png" /></p> Integrate [ESLint][], the _pluggable_ and de-facto standard linting utility for JavaScript. ### Configuration Preset The configuration presets that will be used are [@arcticicestudio/eslint-config][pr] that implements the [Arctic Ice Studio JavaScript Style][stg-js]. It comes with the following peer dependencies: - [eslint][esl-gh] It it built on top of [@arcticicestudio/eslint-config-base][pr-b] that includes various rules of the following plugins and rule presets that are therefore also required peer dependencies: - [eslint-config-prettier][esl-c-pr] - [eslint-plugin-import][esl-p-i] - [eslint-plugin-jsx-a11y][esl-p-a11y] - [eslint-plugin-prettier][esl-p-pr] - [eslint-plugin-react-hooks][esl-p-r-h] - [eslint-plugin-react][esl-p-r] Since _arctic_ will be built with [TypeScript][ts], the [@arcticicestudio/eslint-config-typescript][pr-ts] preset will be extended to add support for _TypeScript_ source file linting and compatibility with [Prettier][] through the [`@arcticicestudio/eslint-config-typescript/prettier` extension entry point][pr-ts-d#ep]. This preset requires the following peer dependencies: - [@typescript-eslint/eslint-plugin][esl-ts-p] - [@typescript-eslint/parser][esl-ts-pa] - [typescript][gh-ts] Since the custom presets are still in major version `0` note that the version range should be `>=0.x.x <1.0.0` to avoid the “SemVer Major Zero Caveat”. When defining package versions with the the carat `^` or tilde `~` range selector it won't affect packages with a major version of `0`. _yarn_ will resolve these packages to their exact version until the major version is greater or equal to `1`. To avoid this caveat the more detailed version range `>=0.x.x <1.0.0` should be used to resolve all versions greater or equal to `0.x.x` but less than `1.0.0`. This will always use the latest `0.x.x` version and removes the need to increment the version manually on each new release. To allow to lint TypeScript code the `@typescript-eslint/parser` parser will be used and [specified as main parser][esl-d-parser]. Also to make use of the latest experimental Babel features and proposals, [eslint-plugin-babel][esl-p-b] will be added with the following rule configurations: - `babel/camelcase` with level `error` - doesn't complain about optional chaining (`let foo = bar?.a_b;`). Note that the [core rule `camelcase`][esl-r-cc] must be disabled! - `babel/no-unused-expressions` with level `error` - doesn't fail when using `do` expressions or optional chaining (`a?.b()`). Note that the [core rule `no-unused-expressions`][esl-r-nue] must be disabled! See the [documentation of provided rules][esl-p-b#r]and required configurations to use them. The `.eslintrc.js` configuration file will be placed in the project root next to the `.eslintignore` file to define ignore pattern. ### Package script To allow to run the JavaScript linting separately a `lint:js` npm script/task will be added to be included in the main `lint` script flow. To use the great [auto-fixing][esl-d-cli#af] feature another `format:js` script/task will be added. ## Tasks - [ ] Install required packages to as development dependencies: - [@arcticicestudio/eslint-config-typescript][npm-esl-c-ais-ts] - [@arcticicestudio/eslint-config][npm-esl-c-ais] - [eslint-config-prettier][npm-esl-c-pr] - [eslint-plugin-babel][npm-esl-p-b] - [eslint-plugin-import][npm-esl-p-i] - [eslint-plugin-jsx-a11y][npm-esl-p-a11y] - [eslint-plugin-prettier][npm-esl-p-pr] - [eslint-plugin-react-hooks][npm-esl-p-r-h] - [eslint-plugin-react][npm-esl-p-r] - [eslint][npm-esl] - [typescript][npm-ts] - [ ] Implement `.eslintrc.js` configuration file. - [ ] Extend installed presets. - [ ] Integrate [eslint-plugin-babel][npm-esl-p-b] - [ ] Enable `babel/no-unused-expressions` and `babel/camelcase` rules including the deactivation of their associated core rules. - [ ] Add `babel` to the array of enabled plugins. - [ ] Implement `.eslintignore` ignore pattern file. - [ ] Implement npm `lint:js`, `lint:ts` and `format:js` scripts. - [ ] Lint current code base for the first time and fix possible JavaScript style guide violations. [esl-c-pr]: https://github.com/prettier/eslint-config-prettier [esl-d-cli#af]: https://eslint.org/docs/user-guide/command-line-interface#fixing-problems [esl-d-parser]: https://eslint.org/docs/user-guide/configuring#specifying-parser [esl-p-a11y]: https://github.com/evcohen/eslint-plugin-jsx-a11y [esl-p-b]: https://github.com/babel/eslint-plugin-babel [esl-p-b#r]: https://github.com/babel/eslint-plugin-babel#rules [esl-p-i]: https://github.com/benmosher/eslint-plugin-import [esl-p-pr]: https://github.com/prettier/eslint-plugin-prettier [esl-p-r-h]: https://github.com/facebook/react/tree/master/packages/eslint-plugin-react-hooks [esl-p-r]: https://github.com/yannickcr/eslint-plugin-react [esl-r-cc]: https://eslint.org/docs/rules/camelcase [esl-r-nue]: https://eslint.org/docs/rules/no-unused-expressions [esl-ts-p]: https://github.com/typescript-eslint/typescript-eslint/tree/master/packages/eslint-plugin [esl-ts-pa]: https://github.com/typescript-eslint/typescript-eslint/tree/master/packages/parser [eslint]: https://eslint.org [gh-esl]: https://github.com/eslint/eslint [gh-ts]: https://github.com/Microsoft/TypeScript [npm-esl-c-ais-ts]: https://www.npmjs.com/package/%40arcticicestudio/eslint-config-typescript [npm-esl-c-ais]: https://www.npmjs.com/package/%40arcticicestudio/eslint-config [npm-esl-c-pr]: https://www.npmjs.com/package/eslint-config-prettier [npm-esl-p-a11y]: https://www.npmjs.com/package/eslint-plugin-jsx-a11y [npm-esl-p-b]: https://www.npmjs.com/package/eslint-plugin-babel [npm-esl-p-i]: https://www.npmjs.com/package/eslint-plugin-import [npm-esl-p-pr]: https://www.npmjs.com/package/eslint-plugin-prettier [npm-esl-p-r-h]: https://www.npmjs.com/package/eslint-plugin-react-hooks [npm-esl-p-r]: https://www.npmjs.com/package/eslint-plugin-react [npm-esl]: https://www.npmjs.com/package/eslint [npm-ts]: https://www.npmjs.com/package/typescript [pr-b]: https://github.com/arcticicestudio/styleguide-javascript/tree/develop/packages/%40arcticicestudio/eslint-config-base [pr-ts-d#ep]: https://github.com/arcticicestudio/styleguide-javascript/blob/develop/packages/%40arcticicestudio/eslint-config-typescript/README.md#entry-points [pr-ts]: https://github.com/arcticicestudio/styleguide-javascript/tree/develop/packages/%40arcticicestudio/eslint-config-typescript [pr]: https://github.com/arcticicestudio/styleguide-javascript/tree/develop/packages/%40arcticicestudio/eslint-config [prettier]: https://prettier.io [stg-js]: https://arcticicestudio.github.io/styleguide-javascript [ts]: https://www.typescriptlang.org
main
eslint integrate the pluggable and de facto standard linting utility for javascript configuration preset the configuration presets that will be used are that implements the it comes with the following peer dependencies it it built on top of that includes various rules of the following plugins and rule presets that are therefore also required peer dependencies since arctic will be built with the preset will be extended to add support for typescript source file linting and compatibility with through the this preset requires the following peer dependencies since the custom presets are still in major version note that the version range should be x x to avoid the “semver major zero caveat” when defining package versions with the the carat or tilde range selector it won t affect packages with a major version of yarn will resolve these packages to their exact version until the major version is greater or equal to to avoid this caveat the more detailed version range x x should be used to resolve all versions greater or equal to x x but less than this will always use the latest x x version and removes the need to increment the version manually on each new release to allow to lint typescript code the typescript eslint parser parser will be used and also to make use of the latest experimental babel features and proposals will be added with the following rule configurations babel camelcase with level error doesn t complain about optional chaining let foo bar a b note that the must be disabled babel no unused expressions with level error doesn t fail when using do expressions or optional chaining a b note that the must be disabled see the and required configurations to use them the eslintrc js configuration file will be placed in the project root next to the eslintignore file to define ignore pattern package script to allow to run the javascript linting separately a lint js npm script task will be added to be included in the main lint script flow to use the great feature another format js script task will be added tasks install required packages to as development dependencies implement eslintrc js configuration file extend installed presets integrate enable babel no unused expressions and babel camelcase rules including the deactivation of their associated core rules add babel to the array of enabled plugins implement eslintignore ignore pattern file implement npm lint js lint ts and format js scripts lint current code base for the first time and fix possible javascript style guide violations
1
2,140
7,360,244,817
IssuesEvent
2018-03-10 16:37:21
MDAnalysis/mdanalysis
https://api.github.com/repos/MDAnalysis/mdanalysis
closed
MAINT: asv custom builds
maintainability performance
It seems `asv` now has [basic support for custom builds](https://github.com/airspeed-velocity/asv/pull/611) where `setup.py`is not in the project root. We should likely investigate if this will suffice for our purposes so that we no longer have to maintain a custom branch. If it suffices for our purposes we should almost certainly abandon maintaining our own `asv` branch & simplify the benchmarking docs / procedure, etc. I think we have some related issues I could link to, etc., but just bookmarking here for now.
True
MAINT: asv custom builds - It seems `asv` now has [basic support for custom builds](https://github.com/airspeed-velocity/asv/pull/611) where `setup.py`is not in the project root. We should likely investigate if this will suffice for our purposes so that we no longer have to maintain a custom branch. If it suffices for our purposes we should almost certainly abandon maintaining our own `asv` branch & simplify the benchmarking docs / procedure, etc. I think we have some related issues I could link to, etc., but just bookmarking here for now.
main
maint asv custom builds it seems asv now has where setup py is not in the project root we should likely investigate if this will suffice for our purposes so that we no longer have to maintain a custom branch if it suffices for our purposes we should almost certainly abandon maintaining our own asv branch simplify the benchmarking docs procedure etc i think we have some related issues i could link to etc but just bookmarking here for now
1
27,054
12,509,811,485
IssuesEvent
2020-06-02 17:33:50
cityofaustin/atd-data-tech
https://api.github.com/repos/cityofaustin/atd-data-tech
closed
Meet with RPP Folks to Discuss Online Payments
Product: Residential Parking Permit Digitization Service: Apps Service: Product Type: Meeting Workgroup: PE
Scheduled for 6/2/2020 at 11 am ### Objective Try to talk about RPP Online payments so that everyone is on the same page ### Participants - Jacob, Jason, Joseph, John, Stephanie & Diana ### Agenda - Debrief on status of Online Payments with comptroller office - Configuration of Online Payments - Whether this will be ready by July 1 or if needs to be pushed to Aug 3 ------ - [x] Schedule meeting - [ ] Optional: Schedule debrief - [ ] Meet and take notes - [ ] Optional: Debrief with DTS team members - [ ] Create resulting issues
2.0
Meet with RPP Folks to Discuss Online Payments - Scheduled for 6/2/2020 at 11 am ### Objective Try to talk about RPP Online payments so that everyone is on the same page ### Participants - Jacob, Jason, Joseph, John, Stephanie & Diana ### Agenda - Debrief on status of Online Payments with comptroller office - Configuration of Online Payments - Whether this will be ready by July 1 or if needs to be pushed to Aug 3 ------ - [x] Schedule meeting - [ ] Optional: Schedule debrief - [ ] Meet and take notes - [ ] Optional: Debrief with DTS team members - [ ] Create resulting issues
non_main
meet with rpp folks to discuss online payments scheduled for at am objective try to talk about rpp online payments so that everyone is on the same page participants jacob jason joseph john stephanie diana agenda debrief on status of online payments with comptroller office configuration of online payments whether this will be ready by july or if needs to be pushed to aug schedule meeting optional schedule debrief meet and take notes optional debrief with dts team members create resulting issues
0
3,123
11,959,670,993
IssuesEvent
2020-04-04 23:00:27
microsoft/DirectXMesh
https://api.github.com/repos/microsoft/DirectXMesh
opened
Remove use of DWORD in public interface
maintainence
I changed most of the functions to take C++ Standard Types, but never fixed up ``DWORD`` out of concern for changing link signatures. I should really change it to something standard.
True
Remove use of DWORD in public interface - I changed most of the functions to take C++ Standard Types, but never fixed up ``DWORD`` out of concern for changing link signatures. I should really change it to something standard.
main
remove use of dword in public interface i changed most of the functions to take c standard types but never fixed up dword out of concern for changing link signatures i should really change it to something standard
1
2,230
7,869,447,714
IssuesEvent
2018-06-24 14:07:04
arcticicestudio/nord-hyper
https://api.github.com/repos/arcticicestudio/nord-hyper
opened
Add Prettier
context-workflow scope-maintainability type-improvement
<p align="center"><img src="https://prettier.io/icon.png" width="128" height="128" /></p> [Prettier][] should be used in development with the _Prettier → ESLint_ formatting flow. [prettier]: https://prettier.io
True
Add Prettier - <p align="center"><img src="https://prettier.io/icon.png" width="128" height="128" /></p> [Prettier][] should be used in development with the _Prettier → ESLint_ formatting flow. [prettier]: https://prettier.io
main
add prettier should be used in development with the prettier → eslint formatting flow
1
80,554
15,586,295,162
IssuesEvent
2021-03-18 01:37:00
attesch/myretail
https://api.github.com/repos/attesch/myretail
opened
CVE-2020-10673 (High) detected in jackson-databind-2.9.4.jar
security vulnerability
## CVE-2020-10673 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /myretail/build.gradle</p> <p>Path to vulnerable library: myretail/build.gradle</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.0.0.RELEASE.jar (Root Library) - spring-boot-starter-json-2.0.0.RELEASE.jar - :x: **jackson-databind-2.9.4.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus). <p>Publish Date: 2020-03-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673>CVE-2020-10673</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2660">https://github.com/FasterXML/jackson-databind/issues/2660</a></p> <p>Release Date: 2020-03-18</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-10673 (High) detected in jackson-databind-2.9.4.jar - ## CVE-2020-10673 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /myretail/build.gradle</p> <p>Path to vulnerable library: myretail/build.gradle</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.0.0.RELEASE.jar (Root Library) - spring-boot-starter-json-2.0.0.RELEASE.jar - :x: **jackson-databind-2.9.4.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.caucho.config.types.ResourceRef (aka caucho-quercus). <p>Publish Date: 2020-03-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10673>CVE-2020-10673</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2660">https://github.com/FasterXML/jackson-databind/issues/2660</a></p> <p>Release Date: 2020-03-18</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file myretail build gradle path to vulnerable library myretail build gradle dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com caucho config types resourceref aka caucho quercus publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
7,674
9,924,791,215
IssuesEvent
2019-07-01 10:33:09
sisbell/tor-android-service
https://api.github.com/repos/sisbell/tor-android-service
closed
Settings Changes
Orbot Compatibility
Changes to - Socks Ports, DNS, IPv6Traffic, HTTP_PROXY - Auto resolving of more ports https://github.com/guardianproject/orbot/commit/56917567cd21a734a35f3bee0e56ba23793b6887
True
Settings Changes - Changes to - Socks Ports, DNS, IPv6Traffic, HTTP_PROXY - Auto resolving of more ports https://github.com/guardianproject/orbot/commit/56917567cd21a734a35f3bee0e56ba23793b6887
non_main
settings changes changes to socks ports dns http proxy auto resolving of more ports
0
12,227
3,059,488,500
IssuesEvent
2015-08-14 15:14:57
mysociety/fixmystreet-international
https://api.github.com/repos/mysociety/fixmystreet-international
closed
[Design] Define logo size we require for app and MakeMyIsland site
design
- [ ] Check if the logo will be updated - [ ] Get requirements from designers
1.0
[Design] Define logo size we require for app and MakeMyIsland site - - [ ] Check if the logo will be updated - [ ] Get requirements from designers
non_main
define logo size we require for app and makemyisland site check if the logo will be updated get requirements from designers
0
5,404
27,115,681,186
IssuesEvent
2023-02-15 18:22:31
VA-Explorer/va_explorer
https://api.github.com/repos/VA-Explorer/va_explorer
closed
Calculate and highlight outlier data within VA trends
Type: Maintainance good first issue Domain: Frontend Status: Inactive
**What is the expected state?** As a data analyst I expect to be able to tell at a glance if my trends data contains outliers so I can quickly identify them or further investigate them. I would like the trends tab chart to somehow highlight data that qualifies as an outlier. **What is the actual state?** The trends chart currently shows data without regard to outliers. It doesn't calculate them and it doesn't show them. **Relevant context** - `va_analytics/dash_apps/va_dashboard.py`
True
Calculate and highlight outlier data within VA trends - **What is the expected state?** As a data analyst I expect to be able to tell at a glance if my trends data contains outliers so I can quickly identify them or further investigate them. I would like the trends tab chart to somehow highlight data that qualifies as an outlier. **What is the actual state?** The trends chart currently shows data without regard to outliers. It doesn't calculate them and it doesn't show them. **Relevant context** - `va_analytics/dash_apps/va_dashboard.py`
main
calculate and highlight outlier data within va trends what is the expected state as a data analyst i expect to be able to tell at a glance if my trends data contains outliers so i can quickly identify them or further investigate them i would like the trends tab chart to somehow highlight data that qualifies as an outlier what is the actual state the trends chart currently shows data without regard to outliers it doesn t calculate them and it doesn t show them relevant context va analytics dash apps va dashboard py
1
2,759
2,642,559,175
IssuesEvent
2015-03-12 01:11:05
Annexa/Moki-Ecommerce
https://api.github.com/repos/Annexa/Moki-Ecommerce
opened
Check all taxes are calculated correctly based on international location
Page design
International distros: 0% tax
1.0
Check all taxes are calculated correctly based on international location - International distros: 0% tax
non_main
check all taxes are calculated correctly based on international location international distros tax
0
5,819
30,792,568,420
IssuesEvent
2023-07-31 17:16:33
jupyter-naas/awesome-notebooks
https://api.github.com/repos/jupyter-naas/awesome-notebooks
closed
JSON - Handle Nested Data
templates maintainer
This notebook will show how to handle nested data in JSON format using Python library. It is usefull for organizations to quickly parse and extract data from complex JSON files.
True
JSON - Handle Nested Data - This notebook will show how to handle nested data in JSON format using Python library. It is usefull for organizations to quickly parse and extract data from complex JSON files.
main
json handle nested data this notebook will show how to handle nested data in json format using python library it is usefull for organizations to quickly parse and extract data from complex json files
1
25,035
4,183,770,098
IssuesEvent
2016-06-23 02:20:24
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
closed
DboSource: Logging of query time and numRows in tests
datasource Defect
This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: 2.7.11 * Platform and Target: Apache, MySQL, PHPUnit tests ### What you did Running PHPUnit tests and debugging performance problems with MySQL/creating fixtures. ### Expected Behavior When setting debug=2, I want to see the correct sum of query execution time in the query log at the end of the page. ### Actual Behavior First: I know the logging is disabled for CREATE, DROP and TRUNCATE TABLE statements, so their execution times are not added. Changing this is not part of this issue, but it could be an idea to make this configurable. The problem is that $this->took is not reset before BEGIN/COMMIT/ROLLBACK and in insertMulti(), so the last execution time is always summed up. I fixed this in the attached patch: [DboSource.patch.txt](https://github.com/cakephp/cakephp/files/325906/DboSource.patch.txt)
1.0
DboSource: Logging of query time and numRows in tests - This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: 2.7.11 * Platform and Target: Apache, MySQL, PHPUnit tests ### What you did Running PHPUnit tests and debugging performance problems with MySQL/creating fixtures. ### Expected Behavior When setting debug=2, I want to see the correct sum of query execution time in the query log at the end of the page. ### Actual Behavior First: I know the logging is disabled for CREATE, DROP and TRUNCATE TABLE statements, so their execution times are not added. Changing this is not part of this issue, but it could be an idea to make this configurable. The problem is that $this->took is not reset before BEGIN/COMMIT/ROLLBACK and in insertMulti(), so the last execution time is always summed up. I fixed this in the attached patch: [DboSource.patch.txt](https://github.com/cakephp/cakephp/files/325906/DboSource.patch.txt)
non_main
dbosource logging of query time and numrows in tests this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target apache mysql phpunit tests what you did running phpunit tests and debugging performance problems with mysql creating fixtures expected behavior when setting debug i want to see the correct sum of query execution time in the query log at the end of the page actual behavior first i know the logging is disabled for create drop and truncate table statements so their execution times are not added changing this is not part of this issue but it could be an idea to make this configurable the problem is that this took is not reset before begin commit rollback and in insertmulti so the last execution time is always summed up i fixed this in the attached patch
0
392,899
11,597,409,779
IssuesEvent
2020-02-24 20:49:03
cds-snc/report-a-cybercrime
https://api.github.com/repos/cds-snc/report-a-cybercrime
closed
Should remove "Location" field in the database
bug medium priority
## Summary With a currently version, the location is removed from "Enter your location" page. Should the location be removed as well in the database? ![location](https://user-images.githubusercontent.com/60269104/74657180-af1dde00-515d-11ea-9c5c-4ca0ec87bfa5.jpg) ## Steps to reproduce > How exactly can the bug be reproduced? Be very specific. ## Unresolved questions > Are there any related issues you consider out of scope for this issue that could be addressed in the future?
1.0
Should remove "Location" field in the database - ## Summary With a currently version, the location is removed from "Enter your location" page. Should the location be removed as well in the database? ![location](https://user-images.githubusercontent.com/60269104/74657180-af1dde00-515d-11ea-9c5c-4ca0ec87bfa5.jpg) ## Steps to reproduce > How exactly can the bug be reproduced? Be very specific. ## Unresolved questions > Are there any related issues you consider out of scope for this issue that could be addressed in the future?
non_main
should remove location field in the database summary with a currently version the location is removed from enter your location page should the location be removed as well in the database steps to reproduce how exactly can the bug be reproduced be very specific unresolved questions are there any related issues you consider out of scope for this issue that could be addressed in the future
0
3,590
14,480,928,846
IssuesEvent
2020-12-10 11:53:14
utm-cssc/website
https://api.github.com/repos/utm-cssc/website
opened
🥅 Initiative: Winter 2021 Course
Domain: User Experience Role: Maintainer Role: Product Owner
### Motivation 🏁 <!-- A clear and concise motivation for this initiative? How will this help execute the vision of the org? --> The purpose of this issue is to define the goals for the Winter 2021 iteration of the CSSC Website development course. ### Initiative Overview 👁️‍🗨️ <!-- A clear and concise description of what the initiative is. --> I propose a team of 4 student developers + 1 student mentor to deliver the following: **Implementation Details 🛠️** <!--- Please share a plan to help realize this initiative --> - Migrate the Firebase services to a self-hosted Parse + MongoDB solution - Create a self-hosted analytics server using Plausible Analytics - Create an automation server - Complete the missing features of the course manager - Create a functioning calendar where authorized students can submit events using a form and - Create a vibrant mentorship portal similar to ADPList - At the end of the project, any university should be able to launch their own instance of a technology-enabled student organization that comes with a website, a backend server, an automation server, authentication API, and thorough documentation. ### Impact 💥 When I visited universities in Pakistan, I saw some of the same issues faced by students and their organizations that I did at UTM: - Students don't have access to a unified portal for guidance and resources - There isn't a strong mentorship community - Working with side project devtools are hard and there aren't effective templates ### Additional details ℹ️ <!-- Is there anything else you can add about the proposal? You might want to link to related issues here, if you haven't already. --> There is room for collaborating with universities in Pakistan. I can envision team from AJKU, UET Mardan, and UTM all collaborating to create an effective student
True
🥅 Initiative: Winter 2021 Course - ### Motivation 🏁 <!-- A clear and concise motivation for this initiative? How will this help execute the vision of the org? --> The purpose of this issue is to define the goals for the Winter 2021 iteration of the CSSC Website development course. ### Initiative Overview 👁️‍🗨️ <!-- A clear and concise description of what the initiative is. --> I propose a team of 4 student developers + 1 student mentor to deliver the following: **Implementation Details 🛠️** <!--- Please share a plan to help realize this initiative --> - Migrate the Firebase services to a self-hosted Parse + MongoDB solution - Create a self-hosted analytics server using Plausible Analytics - Create an automation server - Complete the missing features of the course manager - Create a functioning calendar where authorized students can submit events using a form and - Create a vibrant mentorship portal similar to ADPList - At the end of the project, any university should be able to launch their own instance of a technology-enabled student organization that comes with a website, a backend server, an automation server, authentication API, and thorough documentation. ### Impact 💥 When I visited universities in Pakistan, I saw some of the same issues faced by students and their organizations that I did at UTM: - Students don't have access to a unified portal for guidance and resources - There isn't a strong mentorship community - Working with side project devtools are hard and there aren't effective templates ### Additional details ℹ️ <!-- Is there anything else you can add about the proposal? You might want to link to related issues here, if you haven't already. --> There is room for collaborating with universities in Pakistan. I can envision team from AJKU, UET Mardan, and UTM all collaborating to create an effective student
main
🥅 initiative winter course motivation 🏁 a clear and concise motivation for this initiative how will this help execute the vision of the org the purpose of this issue is to define the goals for the winter iteration of the cssc website development course initiative overview 👁️‍🗨️ a clear and concise description of what the initiative is i propose a team of student developers student mentor to deliver the following implementation details 🛠️ migrate the firebase services to a self hosted parse mongodb solution create a self hosted analytics server using plausible analytics create an automation server complete the missing features of the course manager create a functioning calendar where authorized students can submit events using a form and create a vibrant mentorship portal similar to adplist at the end of the project any university should be able to launch their own instance of a technology enabled student organization that comes with a website a backend server an automation server authentication api and thorough documentation impact 💥 when i visited universities in pakistan i saw some of the same issues faced by students and their organizations that i did at utm students don t have access to a unified portal for guidance and resources there isn t a strong mentorship community working with side project devtools are hard and there aren t effective templates additional details ℹ️ is there anything else you can add about the proposal you might want to link to related issues here if you haven t already there is room for collaborating with universities in pakistan i can envision team from ajku uet mardan and utm all collaborating to create an effective student
1
62,309
17,023,894,364
IssuesEvent
2021-07-03 04:25:02
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Andrew S. Edwards Memorial Triangle
Component: admin Priority: minor Resolution: invalid Type: defect
**[Submitted to the original trac issue database at 12.10am, Saturday, 11th January 2014]** The Andrew S. Edwards Memorial Triangle, which is at the triangle of D Street, Louisiana Avenue, and N Capitol Street in Washington DC is in Waze maps and Google maps, but not on openstreetmap.
1.0
Andrew S. Edwards Memorial Triangle - **[Submitted to the original trac issue database at 12.10am, Saturday, 11th January 2014]** The Andrew S. Edwards Memorial Triangle, which is at the triangle of D Street, Louisiana Avenue, and N Capitol Street in Washington DC is in Waze maps and Google maps, but not on openstreetmap.
non_main
andrew s edwards memorial triangle the andrew s edwards memorial triangle which is at the triangle of d street louisiana avenue and n capitol street in washington dc is in waze maps and google maps but not on openstreetmap
0
8,653
6,608,849,381
IssuesEvent
2017-09-19 12:40:18
Openki/Openki
https://api.github.com/repos/Openki/Openki
opened
Load testing
Big Expertise needed Performance Tech
Some Entry-points: - https://www.mongodb.com/presentations/mongodb-europe-2016-debugging-mongodb-performance - https://github.com/rueckstiess/mtools/wiki/mplotqueries - Record DPP Calls --> https://chrome.google.com/webstore/detail/meteor-devtools/ippapidnnboiophakmmhkdlchoccbgje?hl=en - Reproduce Calls --> https://github.com/alanning/meteor-load-test
True
Load testing - Some Entry-points: - https://www.mongodb.com/presentations/mongodb-europe-2016-debugging-mongodb-performance - https://github.com/rueckstiess/mtools/wiki/mplotqueries - Record DPP Calls --> https://chrome.google.com/webstore/detail/meteor-devtools/ippapidnnboiophakmmhkdlchoccbgje?hl=en - Reproduce Calls --> https://github.com/alanning/meteor-load-test
non_main
load testing some entry points record dpp calls reproduce calls
0
803
4,423,286,535
IssuesEvent
2016-08-16 07:58:20
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
ansible 2.1.0, s3 module bug
aws bug_report cloud waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME s3 module ##### ANSIBLE VERSION ``` ansible 2.1.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION Only default settings ##### OS / ENVIRONMENT I think N/A. But we use Ubuntu: ``` NAME="Ubuntu" VERSION="14.04.4 LTS, Trusty Tahr" ID=ubuntu ID_LIKE=debian PRETTY_NAME="Ubuntu 14.04.4 LTS" VERSION_ID="14.04" HOME_URL="http://www.ubuntu.com/" SUPPORT_URL="http://help.ubuntu.com/" BUG_REPORT_URL="http://bugs.launchpad.net/ubuntu/" ``` ##### SUMMARY Error during using s3 module. For example, during get file from s3 bucket. ##### STEPS TO REPRODUCE Run playbook ``` --- - name: Test playbook hosts: localhost tasks: - name: Get file from S3 s3: bucket=<bucket_name> object=/path/to/file> dest=/tmp/file mode=get ``` ##### EXPECTED RESULTS File to /tmp directory ##### ACTUAL RESULTS Error during ansible task running. ``` TASK [Get file from S3] ************************************************ fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "msg": "Failed to connect to S3: Region does not seem to be available for aws module boto.s3. If the region definitely exists, you may need to upgrade boto or extend with endpoints_path"} ```
True
ansible 2.1.0, s3 module bug - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME s3 module ##### ANSIBLE VERSION ``` ansible 2.1.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION Only default settings ##### OS / ENVIRONMENT I think N/A. But we use Ubuntu: ``` NAME="Ubuntu" VERSION="14.04.4 LTS, Trusty Tahr" ID=ubuntu ID_LIKE=debian PRETTY_NAME="Ubuntu 14.04.4 LTS" VERSION_ID="14.04" HOME_URL="http://www.ubuntu.com/" SUPPORT_URL="http://help.ubuntu.com/" BUG_REPORT_URL="http://bugs.launchpad.net/ubuntu/" ``` ##### SUMMARY Error during using s3 module. For example, during get file from s3 bucket. ##### STEPS TO REPRODUCE Run playbook ``` --- - name: Test playbook hosts: localhost tasks: - name: Get file from S3 s3: bucket=<bucket_name> object=/path/to/file> dest=/tmp/file mode=get ``` ##### EXPECTED RESULTS File to /tmp directory ##### ACTUAL RESULTS Error during ansible task running. ``` TASK [Get file from S3] ************************************************ fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "msg": "Failed to connect to S3: Region does not seem to be available for aws module boto.s3. If the region definitely exists, you may need to upgrade boto or extend with endpoints_path"} ```
main
ansible module bug issue type bug report component name module ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration only default settings os environment i think n a but we use ubuntu name ubuntu version lts trusty tahr id ubuntu id like debian pretty name ubuntu lts version id home url support url bug report url summary error during using module for example during get file from bucket steps to reproduce run playbook name test playbook hosts localhost tasks name get file from bucket object path to file dest tmp file mode get expected results file to tmp directory actual results error during ansible task running task fatal failed changed false failed true msg failed to connect to region does not seem to be available for aws module boto if the region definitely exists you may need to upgrade boto or extend with endpoints path
1
27,741
8,031,831,597
IssuesEvent
2018-07-28 07:20:51
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
On Windows at least building with a -j > 1 scons parameter fails with a file not found even though the file exists
bug confirmed platform:windows topic:buildsystem
**Operating system or device - Godot version:** Windows 10 Pro 64-bit, Godot source 64-bit build (tested with master branch) Visual Studio 2015 Community, Windows 10 SDK Kit 10586. Python 2.7.11, Scons 2.5.0, and Pywin32 220 all these tools are 32-bit. **Issue description** (what happened, and what was expected): When I issue the command: scons platform=windows target=release_debug -jx Where "x" is a number > 1 then the build process will eventually give a file not found error. Going to where it says the file is not found it is present. Specifying -j1 on the build line works fine. **Steps to reproduce:** scons platform=windows target=release_debug -j2
1.0
On Windows at least building with a -j > 1 scons parameter fails with a file not found even though the file exists - **Operating system or device - Godot version:** Windows 10 Pro 64-bit, Godot source 64-bit build (tested with master branch) Visual Studio 2015 Community, Windows 10 SDK Kit 10586. Python 2.7.11, Scons 2.5.0, and Pywin32 220 all these tools are 32-bit. **Issue description** (what happened, and what was expected): When I issue the command: scons platform=windows target=release_debug -jx Where "x" is a number > 1 then the build process will eventually give a file not found error. Going to where it says the file is not found it is present. Specifying -j1 on the build line works fine. **Steps to reproduce:** scons platform=windows target=release_debug -j2
non_main
on windows at least building with a j scons parameter fails with a file not found even though the file exists operating system or device godot version windows pro bit godot source bit build tested with master branch visual studio community windows sdk kit python scons and all these tools are bit issue description what happened and what was expected when i issue the command scons platform windows target release debug jx where x is a number then the build process will eventually give a file not found error going to where it says the file is not found it is present specifying on the build line works fine steps to reproduce scons platform windows target release debug
0
46,449
11,842,775,455
IssuesEvent
2020-03-24 00:06:37
AObuchow/lsp4xml-extensions-maven
https://api.github.com/repos/AObuchow/lsp4xml-extensions-maven
opened
Build is timing out on CI
build testing
Recently the CI builds have been going on for the total allocated time on GitHub actions (6 hours). Odds are, some test is missing a timeout duration and going on until the entire build times out. I will have to ensure all tests have a timeout duration.
1.0
Build is timing out on CI - Recently the CI builds have been going on for the total allocated time on GitHub actions (6 hours). Odds are, some test is missing a timeout duration and going on until the entire build times out. I will have to ensure all tests have a timeout duration.
non_main
build is timing out on ci recently the ci builds have been going on for the total allocated time on github actions hours odds are some test is missing a timeout duration and going on until the entire build times out i will have to ensure all tests have a timeout duration
0
77,248
7,569,604,560
IssuesEvent
2018-04-23 05:39:36
apache/incubator-superset
https://api.github.com/repos/apache/incubator-superset
closed
[js-testing] write more tests for javascripts/explore/reducers/exploreReducer.js
js-testing
current coverage: 34%. bring to 70%.
1.0
[js-testing] write more tests for javascripts/explore/reducers/exploreReducer.js - current coverage: 34%. bring to 70%.
non_main
write more tests for javascripts explore reducers explorereducer js current coverage bring to
0
2,051
6,952,510,854
IssuesEvent
2017-12-06 17:41:37
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
closed
Weblate unable to push new translations
maintainability
Now that the master branch is protected, Weblate fails to push new translations. One possible solution would be to change the weblate user to administrator, but that's not really ideal… Any better ideas? Also, it's currently failing to merge translation files. I will look into solving that.
True
Weblate unable to push new translations - Now that the master branch is protected, Weblate fails to push new translations. One possible solution would be to change the weblate user to administrator, but that's not really ideal… Any better ideas? Also, it's currently failing to merge translation files. I will look into solving that.
main
weblate unable to push new translations now that the master branch is protected weblate fails to push new translations one possible solution would be to change the weblate user to administrator but that s not really ideal… any better ideas also it s currently failing to merge translation files i will look into solving that
1
1,464
6,363,118,697
IssuesEvent
2017-07-31 16:18:54
duckduckgo/zeroclickinfo-goodies
https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies
closed
Conversions: Trigger mph to kmh conversion
Category: Highest Impact Tasks Maintainer Active Topic: Conversions
These don't trigger unless you change it to "kmph": **55 mph in kmh** [https://duckduckgo.com/?q=55%20mph%20in%20kmh&kp=1&kad=wt_WT&kl=wt-wt](https://duckduckgo.com/?q=55%20mph%20in%20kmh&kp=1&kad=wt_WT&kl=wt-wt) **661 mph to km/h** [https://duckduckgo.com/?q=661%20mph%20to%20km%2Fh&kp=1&kad=wt_WT&kl=wt-wt](https://duckduckgo.com/?q=661%20mph%20to%20km%2Fh&kp=1&kad=wt_WT&kl=wt-wt) ------ IA Page: http://duck.co/ia/view/conversions [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @mintsoft
True
Conversions: Trigger mph to kmh conversion - These don't trigger unless you change it to "kmph": **55 mph in kmh** [https://duckduckgo.com/?q=55%20mph%20in%20kmh&kp=1&kad=wt_WT&kl=wt-wt](https://duckduckgo.com/?q=55%20mph%20in%20kmh&kp=1&kad=wt_WT&kl=wt-wt) **661 mph to km/h** [https://duckduckgo.com/?q=661%20mph%20to%20km%2Fh&kp=1&kad=wt_WT&kl=wt-wt](https://duckduckgo.com/?q=661%20mph%20to%20km%2Fh&kp=1&kad=wt_WT&kl=wt-wt) ------ IA Page: http://duck.co/ia/view/conversions [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @mintsoft
main
conversions trigger mph to kmh conversion these don t trigger unless you change it to kmph mph in kmh mph to km h ia page mintsoft
1
4,225
20,909,147,562
IssuesEvent
2022-03-24 07:27:12
pypiserver/pypiserver
https://api.github.com/repos/pypiserver/pypiserver
opened
Rename `master` branch to `main`
type.Maintainance
To be consistent with the guidelines. The proposal is to follow the suggestions here https://github.com/github/renaming. Personally, I believe this would be a good step, other ideas are welcome. Planning to do this _after_ the next release is published (#419) ✌️
True
Rename `master` branch to `main` - To be consistent with the guidelines. The proposal is to follow the suggestions here https://github.com/github/renaming. Personally, I believe this would be a good step, other ideas are welcome. Planning to do this _after_ the next release is published (#419) ✌️
main
rename master branch to main to be consistent with the guidelines the proposal is to follow the suggestions here personally i believe this would be a good step other ideas are welcome planning to do this after the next release is published ✌️
1
94,360
19,534,025,517
IssuesEvent
2021-12-31 00:18:02
GTNewHorizons/GT-New-Horizons-Modpack
https://api.github.com/repos/GTNewHorizons/GT-New-Horizons-Modpack
closed
Witchery++ baba yaga spawn rate
Status: CodeComplete Status: Need to be Tested
#### Which modpack version are you using? 2.0.4.6 # #### What do you suggest instead/what changes do you propose? Change the spawn rate of the bags yaga using the crystal ball. The times I tried, I must have put 8-9 hours into the dumb fortunes and got the spawn once. It's just dumb, it's not interesting to do the fortunes once. Let alone hundreds of times. Please
1.0
Witchery++ baba yaga spawn rate - #### Which modpack version are you using? 2.0.4.6 # #### What do you suggest instead/what changes do you propose? Change the spawn rate of the bags yaga using the crystal ball. The times I tried, I must have put 8-9 hours into the dumb fortunes and got the spawn once. It's just dumb, it's not interesting to do the fortunes once. Let alone hundreds of times. Please
non_main
witchery baba yaga spawn rate which modpack version are you using what do you suggest instead what changes do you propose change the spawn rate of the bags yaga using the crystal ball the times i tried i must have put hours into the dumb fortunes and got the spawn once it s just dumb it s not interesting to do the fortunes once let alone hundreds of times please
0
4,356
22,035,596,864
IssuesEvent
2022-05-28 14:23:28
BioArchLinux/Packages
https://api.github.com/repos/BioArchLinux/Packages
closed
[MAINTAIN] sukana
maintain
**Please check all your packages requirement** saddly, there is one package **r-metabma** when compile it, it always broke the server, please check a normal 16G RAM can compile it or not. I have rm it to the prepare **Log of the bug** <details> ``` -` root@bioarchlinux .o+` ----------------- `ooo/ OS: Arch Linux [x86_64] `+oooo: Host: X10SLL-F 0123456789 `+oooooo: Kernel: 5.15.43-1-lts -+oooooo+: Uptime: 28 mins `/:-:++oooo+: Packages: 309 (pacman) `/++++/+++++++: Shell: bash 5.1.16 `/++++++++++++++: Resolution: 1024x768 `/+++ooooooooooooo/` Terminal: /dev/pts/0 ./ooosssso++osssssso+` CPU: Intel Xeon E3-1230 v3 (8) @ 3.7Gz .oossssso-````/ossssss+` GPU: ASPEED Technology, Inc. ASPEED Gy -osssssso. :ssssssso. Memory: 1598MiB / 15962MiB (10%) :osssssss/ osssso+++. Disk (/): 52GB / 218GB (23%) /ossssssss/ +ssssooo/- Locale: en_US.UTF-8 `/ossssso+/:- -:/+osssso+- `+sso+:-` `.-/+oso: ████████████████████████ `++:. `-/+/ ████████████████████████ .` `/ ``` </details> **Description** Our sponsor give us a quite greater server, but even this new server can't pass compile, so please local test these packages, every package and then fix them as soon as possible. Because I will close the lilac timer unless it would be broken several times maybe in the future.
True
[MAINTAIN] sukana - **Please check all your packages requirement** saddly, there is one package **r-metabma** when compile it, it always broke the server, please check a normal 16G RAM can compile it or not. I have rm it to the prepare **Log of the bug** <details> ``` -` root@bioarchlinux .o+` ----------------- `ooo/ OS: Arch Linux [x86_64] `+oooo: Host: X10SLL-F 0123456789 `+oooooo: Kernel: 5.15.43-1-lts -+oooooo+: Uptime: 28 mins `/:-:++oooo+: Packages: 309 (pacman) `/++++/+++++++: Shell: bash 5.1.16 `/++++++++++++++: Resolution: 1024x768 `/+++ooooooooooooo/` Terminal: /dev/pts/0 ./ooosssso++osssssso+` CPU: Intel Xeon E3-1230 v3 (8) @ 3.7Gz .oossssso-````/ossssss+` GPU: ASPEED Technology, Inc. ASPEED Gy -osssssso. :ssssssso. Memory: 1598MiB / 15962MiB (10%) :osssssss/ osssso+++. Disk (/): 52GB / 218GB (23%) /ossssssss/ +ssssooo/- Locale: en_US.UTF-8 `/ossssso+/:- -:/+osssso+- `+sso+:-` `.-/+oso: ████████████████████████ `++:. `-/+/ ████████████████████████ .` `/ ``` </details> **Description** Our sponsor give us a quite greater server, but even this new server can't pass compile, so please local test these packages, every package and then fix them as soon as possible. Because I will close the lilac timer unless it would be broken several times maybe in the future.
main
sukana please check all your packages requirement saddly there is one package r metabma when compile it it always broke the server please check a normal ram can compile it or not i have rm it to the prepare log of the bug root bioarchlinux o ooo os arch linux oooo host f oooooo kernel lts oooooo uptime mins oooo packages pacman shell bash resolution ooooooooooooo terminal dev pts ooosssso osssssso cpu intel xeon oossssso ossssss gpu aspeed technology inc aspeed gy osssssso ssssssso memory osssssss osssso disk ossssssss ssssooo locale en us utf ossssso osssso sso oso ████████████████████████ ████████████████████████ description our sponsor give us a quite greater server but even this new server can t pass compile so please local test these packages every package and then fix them as soon as possible because i will close the lilac timer unless it would be broken several times maybe in the future
1
65,249
19,297,087,777
IssuesEvent
2021-12-12 19:15:26
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
closed
Installation order issue
Type: Documentation Type: Defect
I broke three fresh Fedora 34/35 installations after having installed ZFS as mentioned in your documentation here: https://openzfs.github.io/openzfs-docs/Getting%20Started/Fedora/index.html The command... ``` dnf install -y kernel-devel zfs ``` ...causes to the following error: ``` Loading new zfs-2.1.1 DKMS files... Building for 5.15.6-100.fc34.x86_64 **Module build for kernel 5.15.6-100.fc34.x86_64 was skipped since the kernel headers for this kernel does not seem to be installed.** Installieren : zfs-2.1.1-1.fc34.x86_64 203/204 Ausgeführtes Scriptlet: zfs-2.1.1-1.fc34.x86_64 203/204 Created symlink /etc/systemd/system/zfs-import.target.wants/zfs-import-cache.service → /usr/lib/systemd/system/zfs-import-cache.service. Created symlink /etc/systemd/system/zfs.target.wants/zfs-mount.service → /usr/lib/systemd/system/zfs-mount.service. Created symlink /etc/systemd/system/zfs.target.wants/zfs-share.service → /usr/lib/systemd/system/zfs-share.service. Created symlink /etc/systemd/system/zed.service → /usr/lib/systemd/system/zfs-zed.service. Created symlink /etc/systemd/system/zfs.target.wants/zfs-zed.service → /usr/lib/systemd/system/zfs-zed.service. Created symlink /etc/systemd/system/multi-user.target.wants/zfs.target → /usr/lib/systemd/system/zfs.target. Created symlink /etc/systemd/system/zfs.target.wants/zfs-import.target → /usr/lib/systemd/system/zfs-import.target. Created symlink /etc/systemd/system/zfs-volumes.target.wants/zfs-volume-wait.service → /usr/lib/systemd/system/zfs-volume-wait.service. Installieren : kernel-devel-5.15.6-100.fc34.x86_64 204/204 Ausgeführtes Scriptlet: kernel-devel-5.15.6-100.fc34.x86_64 204/204 Ausgeführtes Scriptlet: zfs-2.1.1-1.fc34.x86_64 204/204 Ausgeführtes Scriptlet: kernel-devel-5.15.6-100.fc34.x86_64 ``` But as you can see this will fail because the kernel-devel becomes installed after ZFS. Thus the ZFS installation fails. And WHY THE HELL is this important information that the installation failed NOT MARKED RED so everyone can see this within this 100drets of terminal lines. I installed Fedora now for the third time not having seen this message before and wondering about the following error when executing modprobe: ``` modprobe: FATAL: Module zfs not found in directory /lib/modules/5.15.6-100.fc34.x86_64 ``` I wasted a hole day on getting ZFS working again now and moved back to Fedora 34 finally because it is not working with Fedora 35 anymore. So PLEASE update the documentation or even better the install routine if possible. I am getting to old for this linux frickelzeug.
1.0
Installation order issue - I broke three fresh Fedora 34/35 installations after having installed ZFS as mentioned in your documentation here: https://openzfs.github.io/openzfs-docs/Getting%20Started/Fedora/index.html The command... ``` dnf install -y kernel-devel zfs ``` ...causes to the following error: ``` Loading new zfs-2.1.1 DKMS files... Building for 5.15.6-100.fc34.x86_64 **Module build for kernel 5.15.6-100.fc34.x86_64 was skipped since the kernel headers for this kernel does not seem to be installed.** Installieren : zfs-2.1.1-1.fc34.x86_64 203/204 Ausgeführtes Scriptlet: zfs-2.1.1-1.fc34.x86_64 203/204 Created symlink /etc/systemd/system/zfs-import.target.wants/zfs-import-cache.service → /usr/lib/systemd/system/zfs-import-cache.service. Created symlink /etc/systemd/system/zfs.target.wants/zfs-mount.service → /usr/lib/systemd/system/zfs-mount.service. Created symlink /etc/systemd/system/zfs.target.wants/zfs-share.service → /usr/lib/systemd/system/zfs-share.service. Created symlink /etc/systemd/system/zed.service → /usr/lib/systemd/system/zfs-zed.service. Created symlink /etc/systemd/system/zfs.target.wants/zfs-zed.service → /usr/lib/systemd/system/zfs-zed.service. Created symlink /etc/systemd/system/multi-user.target.wants/zfs.target → /usr/lib/systemd/system/zfs.target. Created symlink /etc/systemd/system/zfs.target.wants/zfs-import.target → /usr/lib/systemd/system/zfs-import.target. Created symlink /etc/systemd/system/zfs-volumes.target.wants/zfs-volume-wait.service → /usr/lib/systemd/system/zfs-volume-wait.service. Installieren : kernel-devel-5.15.6-100.fc34.x86_64 204/204 Ausgeführtes Scriptlet: kernel-devel-5.15.6-100.fc34.x86_64 204/204 Ausgeführtes Scriptlet: zfs-2.1.1-1.fc34.x86_64 204/204 Ausgeführtes Scriptlet: kernel-devel-5.15.6-100.fc34.x86_64 ``` But as you can see this will fail because the kernel-devel becomes installed after ZFS. Thus the ZFS installation fails. And WHY THE HELL is this important information that the installation failed NOT MARKED RED so everyone can see this within this 100drets of terminal lines. I installed Fedora now for the third time not having seen this message before and wondering about the following error when executing modprobe: ``` modprobe: FATAL: Module zfs not found in directory /lib/modules/5.15.6-100.fc34.x86_64 ``` I wasted a hole day on getting ZFS working again now and moved back to Fedora 34 finally because it is not working with Fedora 35 anymore. So PLEASE update the documentation or even better the install routine if possible. I am getting to old for this linux frickelzeug.
non_main
installation order issue i broke three fresh fedora installations after having installed zfs as mentioned in your documentation here the command dnf install y kernel devel zfs causes to the following error loading new zfs dkms files building for module build for kernel was skipped since the kernel headers for this kernel does not seem to be installed installieren zfs ausgeführtes scriptlet zfs created symlink etc systemd system zfs import target wants zfs import cache service → usr lib systemd system zfs import cache service created symlink etc systemd system zfs target wants zfs mount service → usr lib systemd system zfs mount service created symlink etc systemd system zfs target wants zfs share service → usr lib systemd system zfs share service created symlink etc systemd system zed service → usr lib systemd system zfs zed service created symlink etc systemd system zfs target wants zfs zed service → usr lib systemd system zfs zed service created symlink etc systemd system multi user target wants zfs target → usr lib systemd system zfs target created symlink etc systemd system zfs target wants zfs import target → usr lib systemd system zfs import target created symlink etc systemd system zfs volumes target wants zfs volume wait service → usr lib systemd system zfs volume wait service installieren kernel devel ausgeführtes scriptlet kernel devel ausgeführtes scriptlet zfs ausgeführtes scriptlet kernel devel but as you can see this will fail because the kernel devel becomes installed after zfs thus the zfs installation fails and why the hell is this important information that the installation failed not marked red so everyone can see this within this of terminal lines i installed fedora now for the third time not having seen this message before and wondering about the following error when executing modprobe modprobe fatal module zfs not found in directory lib modules i wasted a hole day on getting zfs working again now and moved back to fedora finally because it is not working with fedora anymore so please update the documentation or even better the install routine if possible i am getting to old for this linux frickelzeug
0
28,308
6,978,852,497
IssuesEvent
2017-12-12 19:03:45
flutter/flutter
https://api.github.com/repos/flutter/flutter
opened
Firebase codelab uses Android Studio as example, but doesn't show IntelliJ version.
dev: docs - codelab
For installing Google Repository, on [this page](https://codelabs.developers.google.com/codelabs/flutter-firebase/index.html#4), the codelab says to use "Android Studio > Tools > Android > SDK Manager > SDK Tools", but there is a similar method on IntelliJ that it doesn't show (or have screen shots for) which starts with File > Settings... > Appearance & Behavior > System Settings > Android SDK. We should either standardize on one true IDE, or show steps for both.
1.0
Firebase codelab uses Android Studio as example, but doesn't show IntelliJ version. - For installing Google Repository, on [this page](https://codelabs.developers.google.com/codelabs/flutter-firebase/index.html#4), the codelab says to use "Android Studio > Tools > Android > SDK Manager > SDK Tools", but there is a similar method on IntelliJ that it doesn't show (or have screen shots for) which starts with File > Settings... > Appearance & Behavior > System Settings > Android SDK. We should either standardize on one true IDE, or show steps for both.
non_main
firebase codelab uses android studio as example but doesn t show intellij version for installing google repository on the codelab says to use android studio tools android sdk manager sdk tools but there is a similar method on intellij that it doesn t show or have screen shots for which starts with file settings appearance behavior system settings android sdk we should either standardize on one true ide or show steps for both
0
557
4,006,257,118
IssuesEvent
2016-05-12 14:25:19
ESAPI/esapi-java-legacy
https://api.github.com/repos/ESAPI/esapi-java-legacy
opened
Create File Validator that checks Magic Bytes as Opposed to Extensions
Component-Validator enhancement Maintainability Validation
This Might make more sense as its own little mini-module, but we can discuss it here. Apache [Tika](https://tika.apache.org/) is a library that can be used to parse and inspect file headers to determine actual file types to mitigate the issue of say, renaming `netcat.exe` to `imsafe.docx` for the purposes of getting a file into a file store/system/whatever. The current esapi file validator literally just looks at a list of extensions in a properties file, and checks that the extensions are allowed. This won't catch the case where someone uploaded a file and renamed the extension, which we could catch and throw an intrusion exception. Tika is a heavy, heavy library however, and would bring in a slew of parsing dependencies. (it is painful in an airgap.) So should we do it here, or should OWASP create a depedent downstream project that relies on ESAPI interfaces, in order to manage the dependencies? @kwwall @planetlevel do you have any preferences?
True
Create File Validator that checks Magic Bytes as Opposed to Extensions - This Might make more sense as its own little mini-module, but we can discuss it here. Apache [Tika](https://tika.apache.org/) is a library that can be used to parse and inspect file headers to determine actual file types to mitigate the issue of say, renaming `netcat.exe` to `imsafe.docx` for the purposes of getting a file into a file store/system/whatever. The current esapi file validator literally just looks at a list of extensions in a properties file, and checks that the extensions are allowed. This won't catch the case where someone uploaded a file and renamed the extension, which we could catch and throw an intrusion exception. Tika is a heavy, heavy library however, and would bring in a slew of parsing dependencies. (it is painful in an airgap.) So should we do it here, or should OWASP create a depedent downstream project that relies on ESAPI interfaces, in order to manage the dependencies? @kwwall @planetlevel do you have any preferences?
main
create file validator that checks magic bytes as opposed to extensions this might make more sense as its own little mini module but we can discuss it here apache is a library that can be used to parse and inspect file headers to determine actual file types to mitigate the issue of say renaming netcat exe to imsafe docx for the purposes of getting a file into a file store system whatever the current esapi file validator literally just looks at a list of extensions in a properties file and checks that the extensions are allowed this won t catch the case where someone uploaded a file and renamed the extension which we could catch and throw an intrusion exception tika is a heavy heavy library however and would bring in a slew of parsing dependencies it is painful in an airgap so should we do it here or should owasp create a depedent downstream project that relies on esapi interfaces in order to manage the dependencies kwwall planetlevel do you have any preferences
1
5,322
26,882,654,281
IssuesEvent
2023-02-05 20:29:37
IzK-ArcOS/ArcOS-Environment
https://api.github.com/repos/IzK-ArcOS/ArcOS-Environment
closed
Refactoring ArcOS & implementing build systems
enhancement maintainability
As the codebase grows, it might be time to rethink the project structure, code and build systems. Right now, ArcOS doesn't use build systems aside from `electron-packager`, has no project-wide linter, is written in plain JavaScript and CSS without preprocessors, does not follow directory structure conventions and the codebase is – overall – very messy. It's also tightly coupled to the Electron runtime which makes it harder to run it using, say, the Tauri framework. The following list proposes a refactoring roadmap in chronological order: - [x] Implement ESLint & Prettier - [ ] Rethink the project structure - [ ] Implement ESBuild - [ ] Use ES Modules - [ ] Switch from CSS to SCSS - [ ] Switch from JavaScript to TypeScript - [ ] Improve classes to follow SOLID principles - [ ] Decouple ArcOS from Electron - [ ] Refactor the codebase This list will be updated from time to time. Feel free to add on to or to correct this list.
True
Refactoring ArcOS & implementing build systems - As the codebase grows, it might be time to rethink the project structure, code and build systems. Right now, ArcOS doesn't use build systems aside from `electron-packager`, has no project-wide linter, is written in plain JavaScript and CSS without preprocessors, does not follow directory structure conventions and the codebase is – overall – very messy. It's also tightly coupled to the Electron runtime which makes it harder to run it using, say, the Tauri framework. The following list proposes a refactoring roadmap in chronological order: - [x] Implement ESLint & Prettier - [ ] Rethink the project structure - [ ] Implement ESBuild - [ ] Use ES Modules - [ ] Switch from CSS to SCSS - [ ] Switch from JavaScript to TypeScript - [ ] Improve classes to follow SOLID principles - [ ] Decouple ArcOS from Electron - [ ] Refactor the codebase This list will be updated from time to time. Feel free to add on to or to correct this list.
main
refactoring arcos implementing build systems as the codebase grows it might be time to rethink the project structure code and build systems right now arcos doesn t use build systems aside from electron packager has no project wide linter is written in plain javascript and css without preprocessors does not follow directory structure conventions and the codebase is – overall – very messy it s also tightly coupled to the electron runtime which makes it harder to run it using say the tauri framework the following list proposes a refactoring roadmap in chronological order implement eslint prettier rethink the project structure implement esbuild use es modules switch from css to scss switch from javascript to typescript improve classes to follow solid principles decouple arcos from electron refactor the codebase this list will be updated from time to time feel free to add on to or to correct this list
1
2,706
9,531,849,402
IssuesEvent
2019-04-29 17:01:47
codestation/qcma
https://api.github.com/repos/codestation/qcma
closed
Request - Release with latest changes
unmaintained
No changes have been applied past months. Is there some possibility to make a release with all changes including removal of deprecated ffmpeg flags? Thanks.
True
Request - Release with latest changes - No changes have been applied past months. Is there some possibility to make a release with all changes including removal of deprecated ffmpeg flags? Thanks.
main
request release with latest changes no changes have been applied past months is there some possibility to make a release with all changes including removal of deprecated ffmpeg flags thanks
1
321,491
9,799,309,435
IssuesEvent
2019-06-11 14:10:42
python/mypy
https://api.github.com/repos/python/mypy
closed
None is allowed in generic instantiation regardless of bound
bug new-semantic-analyzer priority-1-normal topic-strict-optional topic-type-variables
This is a little (inconsequential) oddity I've noticed. It seems like `None` bypasses any generic bounds even in strict mode. * Are you reporting a bug, or opening a feature request? Bug. * Please insert below the code you are checking with mypy. ```python from typing import TypeVar, Generic T = TypeVar('T', bound=str) # str or any other type except None class G(Generic[T]): pass # error: Type argument "builtins.int" of "C" must be a subtype of "builtins.tuple[Any]" class CheckItNotNone(G[int]): pass # No error. class CheckItNone(G[None]): pass ``` * What is the actual behavior/output? Annotated in the snippet. * What is the behavior/output you expect? `CheckItNone` should emit an error `Type argument "builtins.NoneType" of "C" must be a subtype of "builtins.str"`, similar to the not-None case. * What are the versions of mypy and Python you are using? mypy ~~0.641~~ 0.650, Python 3.5.6 * What are the mypy flags you are using? (For example --strict-optional) `--strict --strict-optional`
1.0
None is allowed in generic instantiation regardless of bound - This is a little (inconsequential) oddity I've noticed. It seems like `None` bypasses any generic bounds even in strict mode. * Are you reporting a bug, or opening a feature request? Bug. * Please insert below the code you are checking with mypy. ```python from typing import TypeVar, Generic T = TypeVar('T', bound=str) # str or any other type except None class G(Generic[T]): pass # error: Type argument "builtins.int" of "C" must be a subtype of "builtins.tuple[Any]" class CheckItNotNone(G[int]): pass # No error. class CheckItNone(G[None]): pass ``` * What is the actual behavior/output? Annotated in the snippet. * What is the behavior/output you expect? `CheckItNone` should emit an error `Type argument "builtins.NoneType" of "C" must be a subtype of "builtins.str"`, similar to the not-None case. * What are the versions of mypy and Python you are using? mypy ~~0.641~~ 0.650, Python 3.5.6 * What are the mypy flags you are using? (For example --strict-optional) `--strict --strict-optional`
non_main
none is allowed in generic instantiation regardless of bound this is a little inconsequential oddity i ve noticed it seems like none bypasses any generic bounds even in strict mode are you reporting a bug or opening a feature request bug please insert below the code you are checking with mypy python from typing import typevar generic t typevar t bound str str or any other type except none class g generic pass error type argument builtins int of c must be a subtype of builtins tuple class checkitnotnone g pass no error class checkitnone g pass what is the actual behavior output annotated in the snippet what is the behavior output you expect checkitnone should emit an error type argument builtins nonetype of c must be a subtype of builtins str similar to the not none case what are the versions of mypy and python you are using mypy python what are the mypy flags you are using for example strict optional strict strict optional
0
176,014
28,013,578,577
IssuesEvent
2023-03-27 20:34:15
flutter/flutter
https://api.github.com/repos/flutter/flutter
closed
Button splash radius does not update when pressed
framework f: material design has reproducible steps found in release: 3.7 found in release: 3.8
## Steps to Reproduce 1. Update buttons border radius when pressed. **Expected results:** Buttons splash border radius updates when pressed. ![image](https://user-images.githubusercontent.com/5238952/221961096-6de03290-9f49-4acf-b31f-7b2fd9eba0ba.png) **Actual results:** Buttons splash border radius does not update when pressed. ![image](https://user-images.githubusercontent.com/5238952/221960957-8b82d1be-ac15-401f-b973-bd6b4e6121fa.png) <details> <summary>Code sample</summary> ```dart import 'package:flutter/material.dart'; void main() { runApp(const MyApp()); } class MyApp extends StatelessWidget { const MyApp({super.key}); @override Widget build(BuildContext context) { return MaterialApp( theme: ThemeData.dark( useMaterial3: true, ), home: Scaffold( body: Center( child: FilledButton( style: ButtonStyle( fixedSize: MaterialStateProperty.all(const Size(80, 80)), shape: MaterialStateProperty.resolveWith( (states) => RoundedRectangleBorder( borderRadius: BorderRadius.circular( states.contains(MaterialState.pressed) ? 20 : 50, ), ), ), ), child: const Text('Button'), onPressed: () {}, ), ), ), ); } } ``` </details> <details> <summary>Logs</summary> ``` [√] Flutter (Channel stable, 3.7.3, on Microsoft Windows [Version 10.0.22623.1325], locale en-US) • Flutter version 3.7.3 on channel stable at C:\flutter • Upstream repository https://github.com/flutter/flutter.git • Framework revision 9944297138 (3 weeks ago), 2023-02-08 15:46:04 -0800 • Engine revision 248290d6d5 • Dart version 2.19.2 • DevTools version 2.20.1 ``` </details> This circular button that changes border radius when pressed is found in many of googles apps. ![image](https://user-images.githubusercontent.com/5238952/221962373-1b9b8023-536d-4b8a-a8b8-296a3d00aa7e.png)
1.0
Button splash radius does not update when pressed - ## Steps to Reproduce 1. Update buttons border radius when pressed. **Expected results:** Buttons splash border radius updates when pressed. ![image](https://user-images.githubusercontent.com/5238952/221961096-6de03290-9f49-4acf-b31f-7b2fd9eba0ba.png) **Actual results:** Buttons splash border radius does not update when pressed. ![image](https://user-images.githubusercontent.com/5238952/221960957-8b82d1be-ac15-401f-b973-bd6b4e6121fa.png) <details> <summary>Code sample</summary> ```dart import 'package:flutter/material.dart'; void main() { runApp(const MyApp()); } class MyApp extends StatelessWidget { const MyApp({super.key}); @override Widget build(BuildContext context) { return MaterialApp( theme: ThemeData.dark( useMaterial3: true, ), home: Scaffold( body: Center( child: FilledButton( style: ButtonStyle( fixedSize: MaterialStateProperty.all(const Size(80, 80)), shape: MaterialStateProperty.resolveWith( (states) => RoundedRectangleBorder( borderRadius: BorderRadius.circular( states.contains(MaterialState.pressed) ? 20 : 50, ), ), ), ), child: const Text('Button'), onPressed: () {}, ), ), ), ); } } ``` </details> <details> <summary>Logs</summary> ``` [√] Flutter (Channel stable, 3.7.3, on Microsoft Windows [Version 10.0.22623.1325], locale en-US) • Flutter version 3.7.3 on channel stable at C:\flutter • Upstream repository https://github.com/flutter/flutter.git • Framework revision 9944297138 (3 weeks ago), 2023-02-08 15:46:04 -0800 • Engine revision 248290d6d5 • Dart version 2.19.2 • DevTools version 2.20.1 ``` </details> This circular button that changes border radius when pressed is found in many of googles apps. ![image](https://user-images.githubusercontent.com/5238952/221962373-1b9b8023-536d-4b8a-a8b8-296a3d00aa7e.png)
non_main
button splash radius does not update when pressed steps to reproduce update buttons border radius when pressed expected results buttons splash border radius updates when pressed actual results buttons splash border radius does not update when pressed code sample dart import package flutter material dart void main runapp const myapp class myapp extends statelesswidget const myapp super key override widget build buildcontext context return materialapp theme themedata dark true home scaffold body center child filledbutton style buttonstyle fixedsize materialstateproperty all const size shape materialstateproperty resolvewith states roundedrectangleborder borderradius borderradius circular states contains materialstate pressed child const text button onpressed logs flutter channel stable on microsoft windows locale en us • flutter version on channel stable at c flutter • upstream repository • framework revision weeks ago • engine revision • dart version • devtools version this circular button that changes border radius when pressed is found in many of googles apps
0
1,590
6,572,372,993
IssuesEvent
2017-09-11 01:48:35
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
ecs_service - Application Load Balancer register
affects_2.1 aws cloud feature_idea waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME ecs_service ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### OS / ENVIRONMENT Distributor ID: Ubuntu Description: Ubuntu 14.04.5 LTS Release: 14.04 Codename: trusty ##### SUMMARY Hey guys. Does the ecs_service load balancer infos support the register of the new type of Load Balancer (Application Load Balancer) from Amazon? If yes, can I have a brief explanation of how to do this, once this new type of load balancer have some infos that doesn't exist on ecs_service documentation? If no, there is any plans of do this? Thank you.
True
ecs_service - Application Load Balancer register - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME ecs_service ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### OS / ENVIRONMENT Distributor ID: Ubuntu Description: Ubuntu 14.04.5 LTS Release: 14.04 Codename: trusty ##### SUMMARY Hey guys. Does the ecs_service load balancer infos support the register of the new type of Load Balancer (Application Load Balancer) from Amazon? If yes, can I have a brief explanation of how to do this, once this new type of load balancer have some infos that doesn't exist on ecs_service documentation? If no, there is any plans of do this? Thank you.
main
ecs service application load balancer register issue type feature idea component name ecs service ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides os environment distributor id ubuntu description ubuntu lts release codename trusty summary hey guys does the ecs service load balancer infos support the register of the new type of load balancer application load balancer from amazon if yes can i have a brief explanation of how to do this once this new type of load balancer have some infos that doesn t exist on ecs service documentation if no there is any plans of do this thank you
1
240,660
20,067,723,023
IssuesEvent
2022-02-04 00:06:39
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
[test-failed]: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/monitoring/beats/cluster·js - Monitoring app beats cluster "before all" hook for "shows beats panel with data"
failed-test test-cloud
**Version: 8.1.0** **Class: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/monitoring/beats/cluster·js** **Stack Trace:** ``` Error: retry.try timeout: Error: expected testSubject(superDatePickerQuickMenu) to exist at TestSubjects.existOrFail (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/common/test_subjects.ts:44:13) at /var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:43:11 at runAttempt (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:29:15) at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:68:21) at RetryService.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry.ts:31:12) at setState (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:31:7) at Object.open (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:53:9) at TimePickerPageObject.getRefreshConfig (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/page_objects/time_picker.ts:183:5) at TimePickerPageObject.pauseAutoRefresh (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/page_objects/time_picker.ts:283:27) at setup (test/functional/apps/monitoring/_get_lifecycle_methods.js:47:7) at onFailure (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:17:9) at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:59:13) at RetryService.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry.ts:31:12) at setState (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:31:7) at Object.open (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:53:9) at TimePickerPageObject.getRefreshConfig (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/page_objects/time_picker.ts:183:5) at TimePickerPageObject.pauseAutoRefresh (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/page_objects/time_picker.ts:283:27) at setup (test/functional/apps/monitoring/_get_lifecycle_methods.js:47:7) at Context.<anonymous> (test/functional/apps/monitoring/beats/cluster.js:18:7) at Object.apply (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` **Other test failures:** _Test Report: https://internal-ci.elastic.co/view/Stack%20Tests/job/elastic+estf-cloud-kibana-tests/2876/testReport/_
2.0
[test-failed]: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/monitoring/beats/cluster·js - Monitoring app beats cluster "before all" hook for "shows beats panel with data" - **Version: 8.1.0** **Class: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/monitoring/beats/cluster·js** **Stack Trace:** ``` Error: retry.try timeout: Error: expected testSubject(superDatePickerQuickMenu) to exist at TestSubjects.existOrFail (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/common/test_subjects.ts:44:13) at /var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:43:11 at runAttempt (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:29:15) at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:68:21) at RetryService.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry.ts:31:12) at setState (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:31:7) at Object.open (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:53:9) at TimePickerPageObject.getRefreshConfig (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/page_objects/time_picker.ts:183:5) at TimePickerPageObject.pauseAutoRefresh (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/page_objects/time_picker.ts:283:27) at setup (test/functional/apps/monitoring/_get_lifecycle_methods.js:47:7) at onFailure (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:17:9) at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:59:13) at RetryService.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry.ts:31:12) at setState (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:31:7) at Object.open (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/menu_toggle.ts:53:9) at TimePickerPageObject.getRefreshConfig (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/page_objects/time_picker.ts:183:5) at TimePickerPageObject.pauseAutoRefresh (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/page_objects/time_picker.ts:283:27) at setup (test/functional/apps/monitoring/_get_lifecycle_methods.js:47:7) at Context.<anonymous> (test/functional/apps/monitoring/beats/cluster.js:18:7) at Object.apply (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` **Other test failures:** _Test Report: https://internal-ci.elastic.co/view/Stack%20Tests/job/elastic+estf-cloud-kibana-tests/2876/testReport/_
non_main
chrome x pack ui functional x pack test functional apps monitoring beats cluster·js monitoring app beats cluster before all hook for shows beats panel with data version class chrome x pack ui functional x pack test functional apps monitoring beats cluster·js stack trace error retry try timeout error expected testsubject superdatepickerquickmenu to exist at testsubjects existorfail var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services common test subjects ts at var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services menu toggle ts at runattempt var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry for success ts at retryforsuccess var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry for success ts at retryservice try var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry ts at setstate var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services menu toggle ts at object open var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services menu toggle ts at timepickerpageobject getrefreshconfig var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional page objects time picker ts at timepickerpageobject pauseautorefresh var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional page objects time picker ts at setup test functional apps monitoring get lifecycle methods js at onfailure var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry for success ts at retryforsuccess var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry for success ts at retryservice try var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry ts at setstate var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services menu toggle ts at object open var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services menu toggle ts at timepickerpageobject getrefreshconfig var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional page objects time picker ts at timepickerpageobject pauseautorefresh var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional page objects time picker ts at setup test functional apps monitoring get lifecycle methods js at context test functional apps monitoring beats cluster js at object apply var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana node modules kbn test target node functional test runner lib mocha wrap function js other test failures test report
0
1,365
5,889,631,907
IssuesEvent
2017-05-17 13:21:32
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
maven_artifact does not use classifier when constructing destination name of artifact
affects_2.2 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME maven_artifact ##### ANSIBLE VERSION ``` $ ansible --version ansible 2.2.0 (devel 1fc44e4103) last updated 2016/05/02 13:44:52 (GMT +1050) lib/ansible/modules/core: (detached HEAD b6ad3b6773) last updated 2016/05/02 13:45:21 (GMT +1050) lib/ansible/modules/extras: (devel 1846de2809) last updated 2016/05/02 13:28:20 (GMT +1050) config file = xxx/ansible-playbooks/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION ``` [defaults] host_key_checking = False [ssh_connection] control_path = %(directory)s/%%h-%%r ``` ##### OS / ENVIRONMENT OS X El Capitan v10.11.4 ##### SUMMARY When specifying a classifier for maven_artifact and a directory as the dest parameter, the destination filename is constructed assuming no classifier. I.e. `artifact_id + "-" + version + "." + extension` vs `artifact_id + "-" + version + "-" + classifier + "." + extension` This causes problems when say downloading multiple artifacts with same same groupId:artifactId:version coordinates. Depending of the version of lib/ansible/modules/extras used, either subsequent artifacts are skipped because the destination exists, or is overwritten as the MD5 hash is different. ##### STEPS TO REPRODUCE ``` --- - hosts: test-hosts tasks: - name: Create temporary directory file: path: /tmp/ansible-test state: directory - name: Download Stanford Core NLP maven_artifact: group_id: edu.stanford.nlp artifact_id: stanford-corenlp version: 3.6.0 dest: /tmp/ansible-test/ state: present - name: Download Stanford Core NLP Models maven_artifact: group_id: edu.stanford.nlp artifact_id: stanford-corenlp version: 3.6.0 classifier: models dest: /tmp/ansible-test/ state: present ``` ##### EXPECTED RESULTS Should download two artifacts to ``` /tmp/ansible-test/stanford-corenlp-3.6.0.jar, and /tmp/ansible-test/stanford-corenlp-3.6.0-models.jar ``` ##### ACTUAL RESULTS One file `/tmp/ansible-test/stanford-corenlp-3.6.0.jar`
True
maven_artifact does not use classifier when constructing destination name of artifact - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME maven_artifact ##### ANSIBLE VERSION ``` $ ansible --version ansible 2.2.0 (devel 1fc44e4103) last updated 2016/05/02 13:44:52 (GMT +1050) lib/ansible/modules/core: (detached HEAD b6ad3b6773) last updated 2016/05/02 13:45:21 (GMT +1050) lib/ansible/modules/extras: (devel 1846de2809) last updated 2016/05/02 13:28:20 (GMT +1050) config file = xxx/ansible-playbooks/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION ``` [defaults] host_key_checking = False [ssh_connection] control_path = %(directory)s/%%h-%%r ``` ##### OS / ENVIRONMENT OS X El Capitan v10.11.4 ##### SUMMARY When specifying a classifier for maven_artifact and a directory as the dest parameter, the destination filename is constructed assuming no classifier. I.e. `artifact_id + "-" + version + "." + extension` vs `artifact_id + "-" + version + "-" + classifier + "." + extension` This causes problems when say downloading multiple artifacts with same same groupId:artifactId:version coordinates. Depending of the version of lib/ansible/modules/extras used, either subsequent artifacts are skipped because the destination exists, or is overwritten as the MD5 hash is different. ##### STEPS TO REPRODUCE ``` --- - hosts: test-hosts tasks: - name: Create temporary directory file: path: /tmp/ansible-test state: directory - name: Download Stanford Core NLP maven_artifact: group_id: edu.stanford.nlp artifact_id: stanford-corenlp version: 3.6.0 dest: /tmp/ansible-test/ state: present - name: Download Stanford Core NLP Models maven_artifact: group_id: edu.stanford.nlp artifact_id: stanford-corenlp version: 3.6.0 classifier: models dest: /tmp/ansible-test/ state: present ``` ##### EXPECTED RESULTS Should download two artifacts to ``` /tmp/ansible-test/stanford-corenlp-3.6.0.jar, and /tmp/ansible-test/stanford-corenlp-3.6.0-models.jar ``` ##### ACTUAL RESULTS One file `/tmp/ansible-test/stanford-corenlp-3.6.0.jar`
main
maven artifact does not use classifier when constructing destination name of artifact issue type bug report component name maven artifact ansible version ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras devel last updated gmt config file xxx ansible playbooks ansible cfg configured module search path default w o overrides configuration host key checking false control path directory s h r os environment os x el capitan summary when specifying a classifier for maven artifact and a directory as the dest parameter the destination filename is constructed assuming no classifier i e artifact id version extension vs artifact id version classifier extension this causes problems when say downloading multiple artifacts with same same groupid artifactid version coordinates depending of the version of lib ansible modules extras used either subsequent artifacts are skipped because the destination exists or is overwritten as the hash is different steps to reproduce hosts test hosts tasks name create temporary directory file path tmp ansible test state directory name download stanford core nlp maven artifact group id edu stanford nlp artifact id stanford corenlp version dest tmp ansible test state present name download stanford core nlp models maven artifact group id edu stanford nlp artifact id stanford corenlp version classifier models dest tmp ansible test state present expected results should download two artifacts to tmp ansible test stanford corenlp jar and tmp ansible test stanford corenlp models jar actual results one file tmp ansible test stanford corenlp jar
1
72,152
31,167,983,340
IssuesEvent
2023-08-16 21:27:30
BCDevOps/developer-experience
https://api.github.com/repos/BCDevOps/developer-experience
opened
Shrink disk on 6 Silver app nodes
*team/ DXC* *team/ ops and shared services*
**Describe the issue** In an effort to save money, test out using 500G disks on 6 VM workers instead of 800G. **What is the Value/Impact?** Reduced costs **What is the plan? How will this get completed?** Create CHG and send notifications Work with VMware team to rebuild the nodes with smaller disks **Identify any dependencies** VMware **Definition of done** Nodes rebuilt with smaller disks New ticket created to check performance impacts after the next upgrade
1.0
Shrink disk on 6 Silver app nodes - **Describe the issue** In an effort to save money, test out using 500G disks on 6 VM workers instead of 800G. **What is the Value/Impact?** Reduced costs **What is the plan? How will this get completed?** Create CHG and send notifications Work with VMware team to rebuild the nodes with smaller disks **Identify any dependencies** VMware **Definition of done** Nodes rebuilt with smaller disks New ticket created to check performance impacts after the next upgrade
non_main
shrink disk on silver app nodes describe the issue in an effort to save money test out using disks on vm workers instead of what is the value impact reduced costs what is the plan how will this get completed create chg and send notifications work with vmware team to rebuild the nodes with smaller disks identify any dependencies vmware definition of done nodes rebuilt with smaller disks new ticket created to check performance impacts after the next upgrade
0
3,232
12,368,706,361
IssuesEvent
2020-05-18 14:13:29
Kashdeya/Tiny-Progressions
https://api.github.com/repos/Kashdeya/Tiny-Progressions
closed
[Suggestion] Toggle Birthday Pickaxe Cake
Version not Maintainted
I think the Birthday Pickaxe is a great tool with an adjusted recipe. I can already do that with craft-tweaker, but what I can't change is whether or not the cake is placed and takes a massive hit of durability from it. I get that this pickaxe came from a joke, but it would be great for that to be toggleable.
True
[Suggestion] Toggle Birthday Pickaxe Cake - I think the Birthday Pickaxe is a great tool with an adjusted recipe. I can already do that with craft-tweaker, but what I can't change is whether or not the cake is placed and takes a massive hit of durability from it. I get that this pickaxe came from a joke, but it would be great for that to be toggleable.
main
toggle birthday pickaxe cake i think the birthday pickaxe is a great tool with an adjusted recipe i can already do that with craft tweaker but what i can t change is whether or not the cake is placed and takes a massive hit of durability from it i get that this pickaxe came from a joke but it would be great for that to be toggleable
1
22,714
32,038,498,240
IssuesEvent
2023-09-22 17:14:44
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
120850Release X.Y.Z - $MONTH $YEAR
P1 type: process release team-OSS
# Status of Bazel X.Y.Z - Expected first release candidate date: [date] - Expected release date: [date] - [List of release blockers](link-to-milestone) To report a release-blocking bug, please add a comment with the text `@bazel-io flag` to the issue. A release manager will triage it and add it to the milestone. To cherry-pick a mainline commit into X.Y.Z, simply send a PR against the `release-X.Y.Z` branch. **Task list:** <!-- The first item is only needed for major releases (X.0.0) --> - [ ] Pick release baseline: [link to base commit] - [ ] Create release candidate: X.Y.Zrc1 - [ ] Check downstream projects - [ ] Create [draft release announcement](https://docs.google.com/document/d/1pu2ARPweOCTxPsRR8snoDtkC9R51XWRyBXeiC6Ql5so/edit) <!-- Note that there should be a new Bazel Release Announcement document for every major release. For minor and patch releases, use the latest open doc. --> - [ ] Send the release announcement PR for review: [link to bazel-blog PR] <!-- Only for major releases. --> - [ ] Push the release and notify package maintainers: [link to comment notifying package maintainers] - [ ] Update the documentation - [ ] Push the blog post: [link to blog post] <!-- Only for major releases. --> - [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
1.0
120850Release X.Y.Z - $MONTH $YEAR - # Status of Bazel X.Y.Z - Expected first release candidate date: [date] - Expected release date: [date] - [List of release blockers](link-to-milestone) To report a release-blocking bug, please add a comment with the text `@bazel-io flag` to the issue. A release manager will triage it and add it to the milestone. To cherry-pick a mainline commit into X.Y.Z, simply send a PR against the `release-X.Y.Z` branch. **Task list:** <!-- The first item is only needed for major releases (X.0.0) --> - [ ] Pick release baseline: [link to base commit] - [ ] Create release candidate: X.Y.Zrc1 - [ ] Check downstream projects - [ ] Create [draft release announcement](https://docs.google.com/document/d/1pu2ARPweOCTxPsRR8snoDtkC9R51XWRyBXeiC6Ql5so/edit) <!-- Note that there should be a new Bazel Release Announcement document for every major release. For minor and patch releases, use the latest open doc. --> - [ ] Send the release announcement PR for review: [link to bazel-blog PR] <!-- Only for major releases. --> - [ ] Push the release and notify package maintainers: [link to comment notifying package maintainers] - [ ] Update the documentation - [ ] Push the blog post: [link to blog post] <!-- Only for major releases. --> - [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
non_main
x y z month year status of bazel x y z expected first release candidate date expected release date link to milestone to report a release blocking bug please add a comment with the text bazel io flag to the issue a release manager will triage it and add it to the milestone to cherry pick a mainline commit into x y z simply send a pr against the release x y z branch task list pick release baseline create release candidate x y check downstream projects create send the release announcement pr for review push the release and notify package maintainers update the documentation push the blog post update the
0
4,954
25,455,594,266
IssuesEvent
2022-11-24 13:57:15
pace/bricks
https://api.github.com/repos/pace/bricks
closed
service generation: extend Makefile
T::Maintainance
When generating a new service, also generate an extended Makefile. Something like this: ```Makefile .PHONY: all lint test integration GOPATH?=~/go all: bin/foobarctl bin/foobard bin/foobarctl: cmd/foobarctl go build -mod=vendor -o $@ ./$< bin/foobard: cmd/foobard go build -mod=vendor -o $@ ./$< lint: $(GOPATH)/bin/golangci-lint $(GOPATH)/bin/golangci-lint run $(GOPATH)/bin/golangci-lint: curl -sfL https://install.goreleaser.com/github.com/golangci/golangci-lint.sh | sh -s -- -b $(GOPATH)/bin v1.15.0 test: go test -mod=vendor -count=1 -v -cover -race -short ./... integration: go test -mod=vendor -p 1 -count=1 -v -cover -race -run TestIntegration ./... ```
True
service generation: extend Makefile - When generating a new service, also generate an extended Makefile. Something like this: ```Makefile .PHONY: all lint test integration GOPATH?=~/go all: bin/foobarctl bin/foobard bin/foobarctl: cmd/foobarctl go build -mod=vendor -o $@ ./$< bin/foobard: cmd/foobard go build -mod=vendor -o $@ ./$< lint: $(GOPATH)/bin/golangci-lint $(GOPATH)/bin/golangci-lint run $(GOPATH)/bin/golangci-lint: curl -sfL https://install.goreleaser.com/github.com/golangci/golangci-lint.sh | sh -s -- -b $(GOPATH)/bin v1.15.0 test: go test -mod=vendor -count=1 -v -cover -race -short ./... integration: go test -mod=vendor -p 1 -count=1 -v -cover -race -run TestIntegration ./... ```
main
service generation extend makefile when generating a new service also generate an extended makefile something like this makefile phony all lint test integration gopath go all bin foobarctl bin foobard bin foobarctl cmd foobarctl go build mod vendor o bin foobard cmd foobard go build mod vendor o lint gopath bin golangci lint gopath bin golangci lint run gopath bin golangci lint curl sfl sh s b gopath bin test go test mod vendor count v cover race short integration go test mod vendor p count v cover race run testintegration
1
90,314
18,107,037,987
IssuesEvent
2021-09-22 20:21:34
StanfordBioinformatics/pulsar_lims
https://api.github.com/repos/StanfordBioinformatics/pulsar_lims
closed
ENCODE data submission: ENCSR079SHZ Western, marker missing
Encode IP submission
https://www.encodeproject.org/experiments/ENCSR079SHZ/ Hi Tao and Cory, can you check this experiment. I believe the biosample characterization is actually okay but it needs labels on the marker. I don’t think it’s in Pulsar either. Thanks, Annika
1.0
ENCODE data submission: ENCSR079SHZ Western, marker missing - https://www.encodeproject.org/experiments/ENCSR079SHZ/ Hi Tao and Cory, can you check this experiment. I believe the biosample characterization is actually okay but it needs labels on the marker. I don’t think it’s in Pulsar either. Thanks, Annika
non_main
encode data submission western marker missing hi tao and cory can you check this experiment i believe the biosample characterization is actually okay but it needs labels on the marker i don’t think it’s in pulsar either thanks annika
0
4,013
18,736,464,756
IssuesEvent
2021-11-04 08:20:25
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
Process custom global transform macros
type/feature maintainer/need-response
<!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed). --> ### Describe your idea/feature/enhancement As far as I can tell, `sam local start-api` ignores the `Transforms` section of the template. This is problematic for anybody trying to do template transforms before the serverless transform because sam wont understand custom resources or template syntax supported by the transform. In my example, I have a transform that replaces `:latest` in a layer arn with the latest version of that layer. Sam doesn't run the transform, so it gives me an invalid layer error. ### Proposal I would like if sam local could run the transforms before trying to parse resources from the template. _This may be out of scope of sam-cli, but I couldn't find an issue for it, and wanted to at least have a discussion about it._ Things to consider: 1. Supporting transforms that happen after serverless transform seems impossible because as far as i can tell sam local doesnt actually run the serverless transform, it just picks up on the serverless resources.
True
Process custom global transform macros - <!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed). --> ### Describe your idea/feature/enhancement As far as I can tell, `sam local start-api` ignores the `Transforms` section of the template. This is problematic for anybody trying to do template transforms before the serverless transform because sam wont understand custom resources or template syntax supported by the transform. In my example, I have a transform that replaces `:latest` in a layer arn with the latest version of that layer. Sam doesn't run the transform, so it gives me an invalid layer error. ### Proposal I would like if sam local could run the transforms before trying to parse resources from the template. _This may be out of scope of sam-cli, but I couldn't find an issue for it, and wanted to at least have a discussion about it._ Things to consider: 1. Supporting transforms that happen after serverless transform seems impossible because as far as i can tell sam local doesnt actually run the serverless transform, it just picks up on the serverless resources.
main
process custom global transform macros describe your idea feature enhancement as far as i can tell sam local start api ignores the transforms section of the template this is problematic for anybody trying to do template transforms before the serverless transform because sam wont understand custom resources or template syntax supported by the transform in my example i have a transform that replaces latest in a layer arn with the latest version of that layer sam doesn t run the transform so it gives me an invalid layer error proposal i would like if sam local could run the transforms before trying to parse resources from the template this may be out of scope of sam cli but i couldn t find an issue for it and wanted to at least have a discussion about it things to consider supporting transforms that happen after serverless transform seems impossible because as far as i can tell sam local doesnt actually run the serverless transform it just picks up on the serverless resources
1
240,625
20,051,989,962
IssuesEvent
2022-02-03 07:54:34
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/transform/creation_runtime_mappings·ts - transform creation with runtime mappings batch transform with unique rt_airline_lower and sort by time and runtime mappings navigates through the wizard and sets all needed fields
failed-test
A test failed on a tracked branch ``` Error: Transform id input text should be 'fq_2_1643874372355' (got '') at Assertion.assert (/opt/local-ssd/buildkite/builds/kb-n2-4-b847c70509228dfd/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:100:11) at Assertion.eql (/opt/local-ssd/buildkite/builds/kb-n2-4-b847c70509228dfd/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:244:8) at Object.assertTransformIdValue (test/functional/services/transform/wizard.ts:676:36) at runMicrotasks (<anonymous>) at processTicksAndRejections (node:internal/process/task_queues:96:5) at Object.setTransformId (test/functional/services/transform/wizard.ts:686:7) at Context.<anonymous> (test/functional/apps/transform/creation_runtime_mappings.ts:391:11) at Object.apply (/opt/local-ssd/buildkite/builds/kb-n2-4-b847c70509228dfd/elastic/kibana-hourly/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) { actual: '', expected: 'fq_2_1643874372355', showDiff: true } ``` First failure: [CI Build - 8.1](https://buildkite.com/elastic/kibana-hourly/builds/9139#7ef2bfb6-0908-4145-9a68-a9d43fd78082) <!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/transform/creation_runtime_mappings·ts","test.name":"transform creation with runtime mappings batch transform with unique rt_airline_lower and sort by time and runtime mappings navigates through the wizard and sets all needed fields","test.failCount":1}} -->
1.0
Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/transform/creation_runtime_mappings·ts - transform creation with runtime mappings batch transform with unique rt_airline_lower and sort by time and runtime mappings navigates through the wizard and sets all needed fields - A test failed on a tracked branch ``` Error: Transform id input text should be 'fq_2_1643874372355' (got '') at Assertion.assert (/opt/local-ssd/buildkite/builds/kb-n2-4-b847c70509228dfd/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:100:11) at Assertion.eql (/opt/local-ssd/buildkite/builds/kb-n2-4-b847c70509228dfd/elastic/kibana-hourly/kibana/node_modules/@kbn/expect/expect.js:244:8) at Object.assertTransformIdValue (test/functional/services/transform/wizard.ts:676:36) at runMicrotasks (<anonymous>) at processTicksAndRejections (node:internal/process/task_queues:96:5) at Object.setTransformId (test/functional/services/transform/wizard.ts:686:7) at Context.<anonymous> (test/functional/apps/transform/creation_runtime_mappings.ts:391:11) at Object.apply (/opt/local-ssd/buildkite/builds/kb-n2-4-b847c70509228dfd/elastic/kibana-hourly/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) { actual: '', expected: 'fq_2_1643874372355', showDiff: true } ``` First failure: [CI Build - 8.1](https://buildkite.com/elastic/kibana-hourly/builds/9139#7ef2bfb6-0908-4145-9a68-a9d43fd78082) <!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/transform/creation_runtime_mappings·ts","test.name":"transform creation with runtime mappings batch transform with unique rt_airline_lower and sort by time and runtime mappings navigates through the wizard and sets all needed fields","test.failCount":1}} -->
non_main
failing test chrome x pack ui functional tests x pack test functional apps transform creation runtime mappings·ts transform creation with runtime mappings batch transform with unique rt airline lower and sort by time and runtime mappings navigates through the wizard and sets all needed fields a test failed on a tracked branch error transform id input text should be fq got at assertion assert opt local ssd buildkite builds kb elastic kibana hourly kibana node modules kbn expect expect js at assertion eql opt local ssd buildkite builds kb elastic kibana hourly kibana node modules kbn expect expect js at object asserttransformidvalue test functional services transform wizard ts at runmicrotasks at processticksandrejections node internal process task queues at object settransformid test functional services transform wizard ts at context test functional apps transform creation runtime mappings ts at object apply opt local ssd buildkite builds kb elastic kibana hourly kibana node modules kbn test target node functional test runner lib mocha wrap function js actual expected fq showdiff true first failure
0
679
4,226,587,416
IssuesEvent
2016-07-02 15:21:23
duckduckgo/zeroclickinfo-spice
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
closed
Amazon: Add shipping information
Maintainer Input Requested
The amazon product price shown does not include shipping. Add a separate shipping field? ie. https://www.amazon.com/dp/0545302021 9.49 + 5.49 shipping ------ IA Page: http://duck.co/ia/view/products [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @bsstoner
True
Amazon: Add shipping information - The amazon product price shown does not include shipping. Add a separate shipping field? ie. https://www.amazon.com/dp/0545302021 9.49 + 5.49 shipping ------ IA Page: http://duck.co/ia/view/products [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @bsstoner
main
amazon add shipping information the amazon product price shown does not include shipping add a separate shipping field ie shipping ia page bsstoner
1
33,480
7,720,467,561
IssuesEvent
2018-05-23 23:20:15
wallabyjs/public
https://api.github.com/repos/wallabyjs/public
closed
VSCode - Show code coverage at the bottom right corner
UX VS Code feature request
Hi! One feature that would be really nice to have until the webapp support multiple projects is to have code coverage displayed in the bottom right corner of the VS Code extension. **Proposition:** Show code coverage at the bottom right by default Add a setting that can disable code coverage. If multi-root workspace is supported at the same time (see my other issue #1353) then we would need a way to see the % per projects. There is 2 ways I have in mind for showing that: - Average of all projects with hover that show the drilldown (fastest one) - The project % in the opened file, if nothing is selected then fallback to average of all projects with hover (nice to have) **Nice to have:** Thresholds per wallaby.js file to show the coverage indicator in VS Code in red (or an icon) if under the threshold. In multiple projects it would react the same as in single project where the indicator shows red (or icon) if one of the project is under the threshold.
1.0
VSCode - Show code coverage at the bottom right corner - Hi! One feature that would be really nice to have until the webapp support multiple projects is to have code coverage displayed in the bottom right corner of the VS Code extension. **Proposition:** Show code coverage at the bottom right by default Add a setting that can disable code coverage. If multi-root workspace is supported at the same time (see my other issue #1353) then we would need a way to see the % per projects. There is 2 ways I have in mind for showing that: - Average of all projects with hover that show the drilldown (fastest one) - The project % in the opened file, if nothing is selected then fallback to average of all projects with hover (nice to have) **Nice to have:** Thresholds per wallaby.js file to show the coverage indicator in VS Code in red (or an icon) if under the threshold. In multiple projects it would react the same as in single project where the indicator shows red (or icon) if one of the project is under the threshold.
non_main
vscode show code coverage at the bottom right corner hi one feature that would be really nice to have until the webapp support multiple projects is to have code coverage displayed in the bottom right corner of the vs code extension proposition show code coverage at the bottom right by default add a setting that can disable code coverage if multi root workspace is supported at the same time see my other issue then we would need a way to see the per projects there is ways i have in mind for showing that average of all projects with hover that show the drilldown fastest one the project in the opened file if nothing is selected then fallback to average of all projects with hover nice to have nice to have thresholds per wallaby js file to show the coverage indicator in vs code in red or an icon if under the threshold in multiple projects it would react the same as in single project where the indicator shows red or icon if one of the project is under the threshold
0
1,686
6,574,166,156
IssuesEvent
2017-09-11 11:47:18
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
ec2_elb_lb misconfigures minimum size, desired size, and maximum size
affects_2.2 aws bug_report cloud waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ec2_elb_lb ##### ANSIBLE VERSION ``` ansible 2.2.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION Default ##### OS / ENVIRONMENT Operating on: ubuntu 14.04 Managing: ubuntu 14.04 ##### SUMMARY <!--- Explain the problem briefly --> When updating an existing ELB with min size 2 and max size 10, inevitably sometimes the resulting configuration has a min size of 3 and max size of 11 (as well as a desired size of 3). This doesn't show up anywhere in my playbook (see below). ##### STEPS TO REPRODUCE ``` --- # roles/auto-scaling/tasks/main.yml - name: Retrieve current Auto Scaling Group properties command: "aws --region {{ region }} autoscaling describe-auto-scaling-groups --auto-scaling-group-names webapp-{{ e }}" register: asg_properties_result - name: Set asg_properties variable from JSON output if the Auto Scaling Group already exists set_fact: asg_properties: "{{ (asg_properties_result.stdout | from_json).AutoScalingGroups[0] }}" when: (asg_properties_result.stdout | from_json).AutoScalingGroups | count - name: Configure Auto Scaling Group and perform rolling deploy ec2_asg: region: "{{ region }}" name: "webapp-{{ e }}" launch_config_name: "{{ ec2_lc.name }}" availability_zones: "{{ zone }}" health_check_type: ELB health_check_period: 300 tags: - Name: "webapp-{{ e }}" environment: "{{ e }}" desired_capacity: 2 # desired_capacity: "{{ asg_properties.DesiredCapacity | default(2) }}" # broken due to bug in ansible 2.0.1.0 replace_all_instances: yes replace_batch_size: 1 # replace_batch_size: "{{ (asg_properties.DesiredCapacity | default(2) / 4) | round(0, 'ceil') | int }}" # broken due to bug in ansible 2.0.1.0 min_size: 2 max_size: 10 load_balancers: - "webapp-{{ e }}" state: present wait_timeout: 420 register: asg_result - name: Configure Scaling Policies ec2_scaling_policy: region: "{{ region }}" name: "{{ item.name }}" asg_name: "webapp-{{ e }}" state: present adjustment_type: "{{ item.adjustment_type }}" min_adjustment_step: "{{ item.min_adjustment_step }}" scaling_adjustment: "{{ item.scaling_adjustment }}" cooldown: "{{ item.cooldown }}" with_items: - name: "Increase Group Size" adjustment_type: "ChangeInCapacity" scaling_adjustment: +1 min_adjustment_step: 1 cooldown: 180 - name: "Decrease Group Size" adjustment_type: "ChangeInCapacity" scaling_adjustment: -1 min_adjustment_step: 1 cooldown: 300 register: sp_result - name: Determine Metric Alarm configuration set_fact: metric_alarms: - name: "{{ asg_name }}-ScaleUp" comparison: ">=" threshold: 50.0 alarm_actions: - "{{ sp_result.results[0].arn }}" - name: "{{ asg_name }}-ScaleDown" comparison: "<=" threshold: 20.0 alarm_actions: - "{{ sp_result.results[1].arn }}" - name: Configure Metric Alarms and link to Scaling Policies ec2_metric_alarm: region: "{{ region }}" name: "{{ item.name }}" state: present metric: "CPUUtilization" namespace: "AWS/EC2" statistic: "Average" comparison: "{{ item.comparison }}" threshold: "{{ item.threshold }}" period: 60 evaluation_periods: 5 unit: "Percent" dimensions: AutoScalingGroupName: "webapp-{{ e }}" alarm_actions: "{{ item.alarm_actions }}" with_items: "{{metric_alarms}}" # when: max_size > 1 register: ma_result ``` ##### EXPECTED RESULTS I expect results that look like the bottom line in the table, instead I get results like the top line. ![alt text](https://i.gyazo.com/031adc4de5340c5361f2821169716fc2.png "Results") ##### ACTUAL RESULTS I expect results that look like the bottom line in the table, instead I get results like the top line. ![alt text](https://i.gyazo.com/031adc4de5340c5361f2821169716fc2.png "Results")
True
ec2_elb_lb misconfigures minimum size, desired size, and maximum size - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ec2_elb_lb ##### ANSIBLE VERSION ``` ansible 2.2.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION Default ##### OS / ENVIRONMENT Operating on: ubuntu 14.04 Managing: ubuntu 14.04 ##### SUMMARY <!--- Explain the problem briefly --> When updating an existing ELB with min size 2 and max size 10, inevitably sometimes the resulting configuration has a min size of 3 and max size of 11 (as well as a desired size of 3). This doesn't show up anywhere in my playbook (see below). ##### STEPS TO REPRODUCE ``` --- # roles/auto-scaling/tasks/main.yml - name: Retrieve current Auto Scaling Group properties command: "aws --region {{ region }} autoscaling describe-auto-scaling-groups --auto-scaling-group-names webapp-{{ e }}" register: asg_properties_result - name: Set asg_properties variable from JSON output if the Auto Scaling Group already exists set_fact: asg_properties: "{{ (asg_properties_result.stdout | from_json).AutoScalingGroups[0] }}" when: (asg_properties_result.stdout | from_json).AutoScalingGroups | count - name: Configure Auto Scaling Group and perform rolling deploy ec2_asg: region: "{{ region }}" name: "webapp-{{ e }}" launch_config_name: "{{ ec2_lc.name }}" availability_zones: "{{ zone }}" health_check_type: ELB health_check_period: 300 tags: - Name: "webapp-{{ e }}" environment: "{{ e }}" desired_capacity: 2 # desired_capacity: "{{ asg_properties.DesiredCapacity | default(2) }}" # broken due to bug in ansible 2.0.1.0 replace_all_instances: yes replace_batch_size: 1 # replace_batch_size: "{{ (asg_properties.DesiredCapacity | default(2) / 4) | round(0, 'ceil') | int }}" # broken due to bug in ansible 2.0.1.0 min_size: 2 max_size: 10 load_balancers: - "webapp-{{ e }}" state: present wait_timeout: 420 register: asg_result - name: Configure Scaling Policies ec2_scaling_policy: region: "{{ region }}" name: "{{ item.name }}" asg_name: "webapp-{{ e }}" state: present adjustment_type: "{{ item.adjustment_type }}" min_adjustment_step: "{{ item.min_adjustment_step }}" scaling_adjustment: "{{ item.scaling_adjustment }}" cooldown: "{{ item.cooldown }}" with_items: - name: "Increase Group Size" adjustment_type: "ChangeInCapacity" scaling_adjustment: +1 min_adjustment_step: 1 cooldown: 180 - name: "Decrease Group Size" adjustment_type: "ChangeInCapacity" scaling_adjustment: -1 min_adjustment_step: 1 cooldown: 300 register: sp_result - name: Determine Metric Alarm configuration set_fact: metric_alarms: - name: "{{ asg_name }}-ScaleUp" comparison: ">=" threshold: 50.0 alarm_actions: - "{{ sp_result.results[0].arn }}" - name: "{{ asg_name }}-ScaleDown" comparison: "<=" threshold: 20.0 alarm_actions: - "{{ sp_result.results[1].arn }}" - name: Configure Metric Alarms and link to Scaling Policies ec2_metric_alarm: region: "{{ region }}" name: "{{ item.name }}" state: present metric: "CPUUtilization" namespace: "AWS/EC2" statistic: "Average" comparison: "{{ item.comparison }}" threshold: "{{ item.threshold }}" period: 60 evaluation_periods: 5 unit: "Percent" dimensions: AutoScalingGroupName: "webapp-{{ e }}" alarm_actions: "{{ item.alarm_actions }}" with_items: "{{metric_alarms}}" # when: max_size > 1 register: ma_result ``` ##### EXPECTED RESULTS I expect results that look like the bottom line in the table, instead I get results like the top line. ![alt text](https://i.gyazo.com/031adc4de5340c5361f2821169716fc2.png "Results") ##### ACTUAL RESULTS I expect results that look like the bottom line in the table, instead I get results like the top line. ![alt text](https://i.gyazo.com/031adc4de5340c5361f2821169716fc2.png "Results")
main
elb lb misconfigures minimum size desired size and maximum size issue type bug report component name elb lb ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration default os environment operating on ubuntu managing ubuntu summary when updating an existing elb with min size and max size inevitably sometimes the resulting configuration has a min size of and max size of as well as a desired size of this doesn t show up anywhere in my playbook see below steps to reproduce roles auto scaling tasks main yml name retrieve current auto scaling group properties command aws region region autoscaling describe auto scaling groups auto scaling group names webapp e register asg properties result name set asg properties variable from json output if the auto scaling group already exists set fact asg properties asg properties result stdout from json autoscalinggroups when asg properties result stdout from json autoscalinggroups count name configure auto scaling group and perform rolling deploy asg region region name webapp e launch config name lc name availability zones zone health check type elb health check period tags name webapp e environment e desired capacity desired capacity asg properties desiredcapacity default broken due to bug in ansible replace all instances yes replace batch size replace batch size asg properties desiredcapacity default round ceil int broken due to bug in ansible min size max size load balancers webapp e state present wait timeout register asg result name configure scaling policies scaling policy region region name item name asg name webapp e state present adjustment type item adjustment type min adjustment step item min adjustment step scaling adjustment item scaling adjustment cooldown item cooldown with items name increase group size adjustment type changeincapacity scaling adjustment min adjustment step cooldown name decrease group size adjustment type changeincapacity scaling adjustment min adjustment step cooldown register sp result name determine metric alarm configuration set fact metric alarms name asg name scaleup comparison threshold alarm actions sp result results arn name asg name scaledown comparison threshold alarm actions sp result results arn name configure metric alarms and link to scaling policies metric alarm region region name item name state present metric cpuutilization namespace aws statistic average comparison item comparison threshold item threshold period evaluation periods unit percent dimensions autoscalinggroupname webapp e alarm actions item alarm actions with items metric alarms when max size register ma result expected results i expect results that look like the bottom line in the table instead i get results like the top line results actual results i expect results that look like the bottom line in the table instead i get results like the top line results
1
3,680
15,037,148,754
IssuesEvent
2021-02-02 16:01:34
IITIDIDX597/sp_2021_team1
https://api.github.com/repos/IITIDIDX597/sp_2021_team1
opened
Most frequently annotated content (administrator side)
Epic: 4 Personal control of information Epic: 5 Maintaining the system Story Week 3
**Project Goal:** S Lab is a tailored integrative learning and collaboration platform for clinicians that combines the latest research and tacit knowledge gained from experience in a practical way, while at the same time foster deeper learning experiences in order to deliver better AbilityLab Patient care. **Hill Statement:** Individual Clinicians can reference relevant, continuously evolving information for their patient's therapy needs to self-manage their approach & patient care plan development in a single platform. **Sub-Hill Statements:** 1. Clinicians can choose what information and knowledge is more relevant to their patient's needs, and have the ability to highlight, annotate, and save that specific information to their personal folder for reference in the future ### **Story Details:** As a: administrator I want: to see what is most frequently highlighted / annotated by clinicians So that: look at them and published to other clinicians on the platform
True
Most frequently annotated content (administrator side) - **Project Goal:** S Lab is a tailored integrative learning and collaboration platform for clinicians that combines the latest research and tacit knowledge gained from experience in a practical way, while at the same time foster deeper learning experiences in order to deliver better AbilityLab Patient care. **Hill Statement:** Individual Clinicians can reference relevant, continuously evolving information for their patient's therapy needs to self-manage their approach & patient care plan development in a single platform. **Sub-Hill Statements:** 1. Clinicians can choose what information and knowledge is more relevant to their patient's needs, and have the ability to highlight, annotate, and save that specific information to their personal folder for reference in the future ### **Story Details:** As a: administrator I want: to see what is most frequently highlighted / annotated by clinicians So that: look at them and published to other clinicians on the platform
main
most frequently annotated content administrator side project goal s lab is a tailored integrative learning and collaboration platform for clinicians that combines the latest research and tacit knowledge gained from experience in a practical way while at the same time foster deeper learning experiences in order to deliver better abilitylab patient care hill statement individual clinicians can reference relevant continuously evolving information for their patient s therapy needs to self manage their approach patient care plan development in a single platform sub hill statements clinicians can choose what information and knowledge is more relevant to their patient s needs and have the ability to highlight annotate and save that specific information to their personal folder for reference in the future story details as a administrator i want to see what is most frequently highlighted annotated by clinicians so that look at them and published to other clinicians on the platform
1
4,255
21,096,763,045
IssuesEvent
2022-04-04 11:03:51
Lissy93/dashy
https://api.github.com/repos/Lissy93/dashy
closed
[FEATURE_REQUEST] Email notification when a service is down
🦄 Feature Request 🛑 No Response 👤 Awaiting Maintainer Response
### Is your feature request related to a problem? If so, please describe. Hi, First thank you for this wonderfull piece of software, it is very handy and easy to use. Regarding the http status check, it would be great if you could implement an email notification when a service is down. I do not want to deploy heavier monitoring tool, as they will be way too much for my use case. ### Describe the solution you'd like A simple email notification with the service name when a service is seen as down. ### Priority Medium (Would be very useful) ### Is this something you would be keen to implement _No response_
True
[FEATURE_REQUEST] Email notification when a service is down - ### Is your feature request related to a problem? If so, please describe. Hi, First thank you for this wonderfull piece of software, it is very handy and easy to use. Regarding the http status check, it would be great if you could implement an email notification when a service is down. I do not want to deploy heavier monitoring tool, as they will be way too much for my use case. ### Describe the solution you'd like A simple email notification with the service name when a service is seen as down. ### Priority Medium (Would be very useful) ### Is this something you would be keen to implement _No response_
main
email notification when a service is down is your feature request related to a problem if so please describe hi first thank you for this wonderfull piece of software it is very handy and easy to use regarding the http status check it would be great if you could implement an email notification when a service is down i do not want to deploy heavier monitoring tool as they will be way too much for my use case describe the solution you d like a simple email notification with the service name when a service is seen as down priority medium would be very useful is this something you would be keen to implement no response
1
3,014
11,140,135,438
IssuesEvent
2019-12-21 11:50:33
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
fix terraform module support for no workspace
affects_2.9 bug cloud has_pr module needs_maintainer needs_triage python3 support:community
##### SUMMARY Some terraform backends don't support workspaces. The module should not fail when using them. The module fails with messages like: - `Failed to list Terraform workspaces:\r\nworkspaces not supported\n` - `Failed to list Terraform workspaces:\r\nnamed states not supported\n` Related to: - https://github.com/ansible/ansible/issues/43134#issuecomment-480873362 - https://github.com/ansible/ansible/issues/59089 - https://github.com/ansible/ansible/pull/57402 ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME terraform module ##### ANSIBLE VERSION <!--- Paste verbatim output from "ansible --version" between quotes --> ```paste below ansible 2.9.0 config file = /home/$USER/.ansible.cfg configured module search path = ['/home/$USER/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.7/site-packages/ansible executable location = /usr/bin/ansible python version = 3.7.4 (default, Oct 4 2019, 06:57:26) [GCC 9.2.0] ``` ##### CONFIGURATION <!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes --> ```paste below ANSIBLE_SSH_ARGS(/home/$USER/.ansible.cfg) = -o ControlMaster=auto -o ControlPersist=60s DEFAULT_HOST_LIST(/home/$USER/.ansible.cfg) = ['/home/$USER/.ansible/hosts'] ``` ##### OS / ENVIRONMENT Archlinux up to date. ##### STEPS TO REPRODUCE <!--- Describe exactly how to reproduce the problem, using a minimal test-case --> - Have an artifactory account accessible - configure terraform with the artifactory backend ```hcl terraform { backend "artifactory" { url = "https://artifactory.example.com/artifactory" repo = "terraform" subpath = "test" } } ``` <!--- Paste example playbooks or commands between quotes below --> ```yaml - name: test terraform no workspace hosts: localhost tasks: - name: apply configuration terraform: project_path: "{{ playbook_dir }}/terraform/" force_init: true backend_config: username: "{{ artifactory_user }}" password: "{{ artifactory_password }}" ``` <!--- HINT: You can paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- Describe what you expected to happen when running the steps above --> The module should work without workspaces support from terraform. No error message. ##### ACTUAL RESULTS <!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes --> ```paste below <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: $USER <127.0.0.1> EXEC /bin/sh -c 'echo ~$USER && sleep 0' <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/$USER/.ansible/tmp/ansible-tmp-1573660522.1152759-71895853408161 `" && echo ansible-tmp-1573660522.1152759-71895853408161="` echo /home/$USER/.ansible/tmp/ansible-tmp-1573660522.1152759-71895853408161 `" ) && sleep 0' Using module file /usr/lib/python3.7/site-packages/ansible/modules/cloud/misc/terraform.py <127.0.0.1> PUT /home/$USER/.ansible/tmp/ansible-local-100380yrb34ja/tmpcdahl6p2 TO /home/$USER/.ansible/tmp/ansible-tmp-1573660522.1152759-71895853408161/AnsiballZ_terraform.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/$USER/.ansible/tmp/ansible-tmp-1573660522.1152759-71895853408161/ /home/$USER/.ansible/tmp/ansible-tmp-1573660522.1152759-71895853408161/AnsiballZ_terraform.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c '/usr/bin/python /home/$USER/.ansible/tmp/ansible-tmp-1573660522.1152759-71895853408161/AnsiballZ_terraform.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'rm -f -r /home/$USER/.ansible/tmp/ansible-tmp-1573660522.1152759-71895853408161/ > /dev/null 2>&1 && sleep 0' fatal: [localhost]: FAILED! => { "changed": false, "invocation": { "module_args": { "backend_config": { "password": "XXXXX", "username": "XXXXX" }, "binary_path": null, "force_init": true, "lock": true, "lock_timeout": null, "plan_file": null, "project_path": "XXX", "purge_workspace": false, "state": "present", "state_file": null, "targets": [], "variables": {}, "variables_file": null, "workspace": "default" } }, "msg": "Failed to list Terraform workspaces:\r\nworkspaces not supported\n" ``` ##### FIX A fix was proposed: https://github.com/ansible/ansible/pull/57402 I tried another approach: defaulting to the default workspace when the workspace listing fails: ```diff diff --git a/lib/ansible/modules/cloud/misc/terraform.py b/lib/ansible/modules/cloud/misc/terraform.py index 1d54be2f4b..5c9832c8b0 100644 --- a/lib/ansible/modules/cloud/misc/terraform.py +++ b/lib/ansible/modules/cloud/misc/terraform.py @@ -208,7 +208,7 @@ def get_workspace_context(bin_path, project_path): workspace_ctx = {"current": "default", "all": []} command = [bin_path, 'workspace', 'list', '-no-color'] rc, out, err = module.run_command(command, cwd=project_path) - if rc != 0: + if rc != 0 and err.strip() != "workspaces not supported": module.fail_json(msg="Failed to list Terraform workspaces:\r\n{0}".format(err)) for item in out.split('\n'): stripped_item = item.strip() ``` The test works for the artifactory backend, but I suspect it won't on other backends. The error message `workspaces not supported` seems backend specific.
True
fix terraform module support for no workspace - ##### SUMMARY Some terraform backends don't support workspaces. The module should not fail when using them. The module fails with messages like: - `Failed to list Terraform workspaces:\r\nworkspaces not supported\n` - `Failed to list Terraform workspaces:\r\nnamed states not supported\n` Related to: - https://github.com/ansible/ansible/issues/43134#issuecomment-480873362 - https://github.com/ansible/ansible/issues/59089 - https://github.com/ansible/ansible/pull/57402 ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME terraform module ##### ANSIBLE VERSION <!--- Paste verbatim output from "ansible --version" between quotes --> ```paste below ansible 2.9.0 config file = /home/$USER/.ansible.cfg configured module search path = ['/home/$USER/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.7/site-packages/ansible executable location = /usr/bin/ansible python version = 3.7.4 (default, Oct 4 2019, 06:57:26) [GCC 9.2.0] ``` ##### CONFIGURATION <!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes --> ```paste below ANSIBLE_SSH_ARGS(/home/$USER/.ansible.cfg) = -o ControlMaster=auto -o ControlPersist=60s DEFAULT_HOST_LIST(/home/$USER/.ansible.cfg) = ['/home/$USER/.ansible/hosts'] ``` ##### OS / ENVIRONMENT Archlinux up to date. ##### STEPS TO REPRODUCE <!--- Describe exactly how to reproduce the problem, using a minimal test-case --> - Have an artifactory account accessible - configure terraform with the artifactory backend ```hcl terraform { backend "artifactory" { url = "https://artifactory.example.com/artifactory" repo = "terraform" subpath = "test" } } ``` <!--- Paste example playbooks or commands between quotes below --> ```yaml - name: test terraform no workspace hosts: localhost tasks: - name: apply configuration terraform: project_path: "{{ playbook_dir }}/terraform/" force_init: true backend_config: username: "{{ artifactory_user }}" password: "{{ artifactory_password }}" ``` <!--- HINT: You can paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- Describe what you expected to happen when running the steps above --> The module should work without workspaces support from terraform. No error message. ##### ACTUAL RESULTS <!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes --> ```paste below <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: $USER <127.0.0.1> EXEC /bin/sh -c 'echo ~$USER && sleep 0' <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /home/$USER/.ansible/tmp/ansible-tmp-1573660522.1152759-71895853408161 `" && echo ansible-tmp-1573660522.1152759-71895853408161="` echo /home/$USER/.ansible/tmp/ansible-tmp-1573660522.1152759-71895853408161 `" ) && sleep 0' Using module file /usr/lib/python3.7/site-packages/ansible/modules/cloud/misc/terraform.py <127.0.0.1> PUT /home/$USER/.ansible/tmp/ansible-local-100380yrb34ja/tmpcdahl6p2 TO /home/$USER/.ansible/tmp/ansible-tmp-1573660522.1152759-71895853408161/AnsiballZ_terraform.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/$USER/.ansible/tmp/ansible-tmp-1573660522.1152759-71895853408161/ /home/$USER/.ansible/tmp/ansible-tmp-1573660522.1152759-71895853408161/AnsiballZ_terraform.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c '/usr/bin/python /home/$USER/.ansible/tmp/ansible-tmp-1573660522.1152759-71895853408161/AnsiballZ_terraform.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'rm -f -r /home/$USER/.ansible/tmp/ansible-tmp-1573660522.1152759-71895853408161/ > /dev/null 2>&1 && sleep 0' fatal: [localhost]: FAILED! => { "changed": false, "invocation": { "module_args": { "backend_config": { "password": "XXXXX", "username": "XXXXX" }, "binary_path": null, "force_init": true, "lock": true, "lock_timeout": null, "plan_file": null, "project_path": "XXX", "purge_workspace": false, "state": "present", "state_file": null, "targets": [], "variables": {}, "variables_file": null, "workspace": "default" } }, "msg": "Failed to list Terraform workspaces:\r\nworkspaces not supported\n" ``` ##### FIX A fix was proposed: https://github.com/ansible/ansible/pull/57402 I tried another approach: defaulting to the default workspace when the workspace listing fails: ```diff diff --git a/lib/ansible/modules/cloud/misc/terraform.py b/lib/ansible/modules/cloud/misc/terraform.py index 1d54be2f4b..5c9832c8b0 100644 --- a/lib/ansible/modules/cloud/misc/terraform.py +++ b/lib/ansible/modules/cloud/misc/terraform.py @@ -208,7 +208,7 @@ def get_workspace_context(bin_path, project_path): workspace_ctx = {"current": "default", "all": []} command = [bin_path, 'workspace', 'list', '-no-color'] rc, out, err = module.run_command(command, cwd=project_path) - if rc != 0: + if rc != 0 and err.strip() != "workspaces not supported": module.fail_json(msg="Failed to list Terraform workspaces:\r\n{0}".format(err)) for item in out.split('\n'): stripped_item = item.strip() ``` The test works for the artifactory backend, but I suspect it won't on other backends. The error message `workspaces not supported` seems backend specific.
main
fix terraform module support for no workspace summary some terraform backends don t support workspaces the module should not fail when using them the module fails with messages like failed to list terraform workspaces r nworkspaces not supported n failed to list terraform workspaces r nnamed states not supported n related to issue type bug report component name terraform module ansible version paste below ansible config file home user ansible cfg configured module search path ansible python module location usr lib site packages ansible executable location usr bin ansible python version default oct configuration paste below ansible ssh args home user ansible cfg o controlmaster auto o controlpersist default host list home user ansible cfg os environment archlinux up to date steps to reproduce have an artifactory account accessible configure terraform with the artifactory backend hcl terraform backend artifactory url repo terraform subpath test yaml name test terraform no workspace hosts localhost tasks name apply configuration terraform project path playbook dir terraform force init true backend config username artifactory user password artifactory password expected results the module should work without workspaces support from terraform no error message actual results paste below establish local connection for user user exec bin sh c echo user sleep exec bin sh c umask mkdir p echo home user ansible tmp ansible tmp echo ansible tmp echo home user ansible tmp ansible tmp sleep using module file usr lib site packages ansible modules cloud misc terraform py put home user ansible tmp ansible local to home user ansible tmp ansible tmp ansiballz terraform py exec bin sh c chmod u x home user ansible tmp ansible tmp home user ansible tmp ansible tmp ansiballz terraform py sleep exec bin sh c usr bin python home user ansible tmp ansible tmp ansiballz terraform py sleep exec bin sh c rm f r home user ansible tmp ansible tmp dev null sleep fatal failed changed false invocation module args backend config password xxxxx username xxxxx binary path null force init true lock true lock timeout null plan file null project path xxx purge workspace false state present state file null targets variables variables file null workspace default msg failed to list terraform workspaces r nworkspaces not supported n fix a fix was proposed i tried another approach defaulting to the default workspace when the workspace listing fails diff diff git a lib ansible modules cloud misc terraform py b lib ansible modules cloud misc terraform py index a lib ansible modules cloud misc terraform py b lib ansible modules cloud misc terraform py def get workspace context bin path project path workspace ctx current default all command rc out err module run command command cwd project path if rc if rc and err strip workspaces not supported module fail json msg failed to list terraform workspaces r n format err for item in out split n stripped item item strip the test works for the artifactory backend but i suspect it won t on other backends the error message workspaces not supported seems backend specific
1
25,644
12,266,741,674
IssuesEvent
2020-05-07 09:28:09
PrestaShop/PrestaShop
https://api.github.com/repos/PrestaShop/PrestaShop
closed
Accented & special characters misinterpreted in customer service messages
1.7.6.4 1.7.6.5 BO Customer service NMI
#### Describe the bug Accented & special characters misinterpreted in customer service messages (SAV in french). #### Expected behavior A well interpretation of accented & special characters as customer wrote it in original message. #### Steps to Reproduce Steps to reproduce the behavior: 1. In FO, write a message with accented & special characters as customer 2. In BO, in SAV > SAV page (in FR), message is wrong interpreted **Screenshots** ![sav-accents](https://user-images.githubusercontent.com/7419525/80588058-e808e780-8a17-11ea-8871-90b84275b059.jpg) Some messages are unreadable! ![accents](https://user-images.githubusercontent.com/7419525/80588094-f8b95d80-8a17-11ea-86c8-a572163bf1a2.jpg) #### Additional information * PrestaShop version: 1.7.6.4 (fresh install, classic theme, no core/override change) * PHP version: 7.1
1.0
Accented & special characters misinterpreted in customer service messages - #### Describe the bug Accented & special characters misinterpreted in customer service messages (SAV in french). #### Expected behavior A well interpretation of accented & special characters as customer wrote it in original message. #### Steps to Reproduce Steps to reproduce the behavior: 1. In FO, write a message with accented & special characters as customer 2. In BO, in SAV > SAV page (in FR), message is wrong interpreted **Screenshots** ![sav-accents](https://user-images.githubusercontent.com/7419525/80588058-e808e780-8a17-11ea-8871-90b84275b059.jpg) Some messages are unreadable! ![accents](https://user-images.githubusercontent.com/7419525/80588094-f8b95d80-8a17-11ea-86c8-a572163bf1a2.jpg) #### Additional information * PrestaShop version: 1.7.6.4 (fresh install, classic theme, no core/override change) * PHP version: 7.1
non_main
accented special characters misinterpreted in customer service messages describe the bug accented special characters misinterpreted in customer service messages sav in french expected behavior a well interpretation of accented special characters as customer wrote it in original message steps to reproduce steps to reproduce the behavior in fo write a message with accented special characters as customer in bo in sav sav page in fr message is wrong interpreted screenshots some messages are unreadable additional information prestashop version fresh install classic theme no core override change php version
0
3,273
12,501,076,921
IssuesEvent
2020-06-02 00:08:45
unoplatform/uno
https://api.github.com/repos/unoplatform/uno
closed
Building Uno.UI.sln from the command line does not download %TEMP%\mono-wasm-e894d683f9f
area/vswin kind/bug kind/contributor-experience kind/maintainer-experience
## Current behavior ``` empty %TEMP%\*.* git clone master cd src msbuild /m /t:restore Uno.UI.sln msbuild /m Uno.UI.sln ``` - Compilation fails with `%TEMP%\mono-wasm-e894d683f9f` does not exist. - Opening Visual Studio and doing right click -> build SamplesApp.Wasm will create `%TEMP%\mono-wasm-e894d683f9f`. - Afterwards msbuild from the command line will function. ## Expected behavior ``` git clone master cd src msbuild just works ``` ## How to reproduce it (as minimally and precisely as possible) Forgot to grab a binlog. Opening up early on the off-chance others come across it and as a self-reminder. ## Environment Nuget Package: Package Version(s): Affected platform(s): - [ ] iOS - [ ] Android - [x] WebAssembly - [ ] Windows - [ ] Build tasks Visual Studio - [ ] 2017 (version: ) - [x] 2019 (version: ) - [ ] for Mac (version: ) Relevant plugins - [ ] Resharper (version: ) ## Anything else we need to know? <details> <summary>Microsoft Visual Studio Enterprise 2019 - Version 16.1.6 Details</summary> Microsoft Visual Studio Enterprise 2019 Version 16.1.6 VisualStudio.16.Release/16.1.6+29102.190 Microsoft .NET Framework Version 4.8.03752 Installed Version: Enterprise Visual C++ 2019 00435-60000-00000-AA184 Microsoft Visual C++ 2019 Application Insights Tools for Visual Studio Package 9.1.00429.1 Application Insights Tools for Visual Studio ASP.NET and Web Tools 2019 16.1.429.50124 ASP.NET and Web Tools 2019 ASP.NET Web Frameworks and Tools 2019 16.1.429.50124 For additional information, visit https://www.asp.net/ Azure App Service Tools v3.0.0 16.1.429.50124 Azure App Service Tools v3.0.0 Azure Functions and Web Jobs Tools 16.1.429.50124 Azure Functions and Web Jobs Tools C# Tools 3.1.1-beta4-19281-06+58a4b1e79aea28115e66b06f850c83a3f1fcb6d3 C# components used in the IDE. Depending on your project type and settings, a different version of the compiler may be used. Child Process Debugging Power Tool 1.0 Power tool to add child process debugging to Visual Studio. Common Azure Tools 1.10 Provides common services for use by Azure Mobile Services and Microsoft Azure Tools. Extensibility Message Bus 1.1.77 (master@24013d5) Provides common messaging-based MEF services for loosely coupled Visual Studio extension components communication and integration. FormatDocumentOnSave 1.0 Enables auto formatting of the code when you save a file. Visual Studio supports auto formatting of the code with the CTRL+E,D or CTRL+E,F key shortcuts but with this extension the command 'Format Document' is executed on Save. You can find the source here: https://github.com/Elders/VSE-FormatDocumentOnSave IntelliCode Extension 1.0 IntelliCode Visual Studio Extension Detailed Info Microsoft Azure Tools 2.9 Microsoft Azure Tools for Microsoft Visual Studio 0x10 - v2.9.20419.2 Microsoft Continuous Delivery Tools for Visual Studio 0.4 Simplifying the configuration of Azure DevOps pipelines from within the Visual Studio IDE. Microsoft JVM Debugger 1.0 Provides support for connecting the Visual Studio debugger to JDWP compatible Java Virtual Machines Microsoft Library Manager 1.0 Install client-side libraries easily to any web project Microsoft MI-Based Debugger 1.0 Provides support for connecting Visual Studio to MI compatible debuggers Microsoft Visual C++ Wizards 1.0 Microsoft Visual C++ Wizards Microsoft Visual Studio Tools for Containers 1.1 Develop, run, validate your ASP.NET Core applications in the target environment. F5 your application directly into a container with debugging, or CTRL + F5 to edit & refresh your app without having to rebuild the container. Microsoft Visual Studio VC Package 1.0 Microsoft Visual Studio VC Package Mono Debugging for Visual Studio 16.1.1 (2473f22) Support for debugging Mono processes with Visual Studio. Node.js Tools 1.5.10424.1 Commit Hash:c3ce0ae0b29c0b3a755ffc12f8a685fe7ddd3600 Adds support for developing and debugging Node.js apps in Visual Studio NuGet Package Manager 5.1.0 NuGet Package Manager in Visual Studio. For more information about NuGet, visit https://docs.nuget.org/ OzCodePackage Extension 1.0 OzCodePackage Visual Studio Extension Detailed Info ProjectServicesPackage Extension 1.0 ProjectServicesPackage Visual Studio Extension Detailed Info ResourcePackage Extension 1.0 ResourcePackage Visual Studio Extension Detailed Info ResourcePackage Extension 1.0 ResourcePackage Visual Studio Extension Detailed Info Snapshot Debugging Extension 1.0 Snapshot Debugging Visual Studio Extension Detailed Info SQL Server Data Tools 16.0.61904.23160 Microsoft SQL Server Data Tools Syntax Visualizer 1.0 An extension for visualizing Roslyn SyntaxTrees. Test Adapter for Boost.Test 1.0 Enables Visual Studio's testing tools with unit tests written for Boost.Test. The use terms and Third Party Notices are available in the extension installation directory. Test Adapter for Google Test 1.0 Enables Visual Studio's testing tools with unit tests written for Google Test. The use terms and Third Party Notices are available in the extension installation directory. TypeScript Tools 16.0.10506.2004 TypeScript Tools for Microsoft Visual Studio Visual Basic Tools 3.1.1-beta4-19281-06+58a4b1e79aea28115e66b06f850c83a3f1fcb6d3 Visual Basic components used in the IDE. Depending on your project type and settings, a different version of the compiler may be used. Visual F# Tools 10.4 for F# 4.6 16.1.0-beta.19253.3+42526fe359672a05fd562dc16a91a43d0fe047a7 Microsoft Visual F# Tools 10.4 for F# 4.6 Visual Studio Code Debug Adapter Host Package 1.0 Interop layer for hosting Visual Studio Code debug adapters in Visual Studio Visual Studio Tools for CMake 1.0 Visual Studio Tools for CMake Visual Studio Tools for CMake 1.0 Visual Studio Tools for CMake Visual Studio Tools for Containers 1.0 Visual Studio Tools for Containers VisualStudio.Mac 1.0 Mac Extension for Visual Studio Xamarin 16.1.0.545 (d16-1@db7c858e8) Visual Studio extension to enable development for Xamarin.iOS and Xamarin.Android. Xamarin Designer 16.1.0.418 (remotes/origin/d16-1@5b958bb10) Visual Studio extension to enable Xamarin Designer tools in Visual Studio. Xamarin Templates 16.2.112 (4db4af4) Templates for building iOS, Android, and Windows apps with Xamarin and Xamarin.Forms. Xamarin.Android SDK 9.3.0.23 (HEAD/d0b48056f) Xamarin.Android Reference Assemblies and MSBuild support. Mono: mono/mono/2018-08@3a07bd426d3 Java.Interop: xamarin/java.interop/d16-1@5ddc3e3 LibZipSharp: grendello/LibZipSharp/d16-1@44de300 LibZip: nih-at/libzip/rel-1-5-1@b95cf3f ProGuard: xamarin/proguard/master@905836d SQLite: xamarin/sqlite/3.27.1@8212a2d Xamarin.Android Tools: xamarin/xamarin-android-tools/d16-1@acabd26 Xamarin.iOS and Xamarin.Mac SDK 12.10.0.157 (6bd9475) Xamarin.iOS and Xamarin.Mac Reference Assemblies and MSBuild support. </details>
True
Building Uno.UI.sln from the command line does not download %TEMP%\mono-wasm-e894d683f9f - ## Current behavior ``` empty %TEMP%\*.* git clone master cd src msbuild /m /t:restore Uno.UI.sln msbuild /m Uno.UI.sln ``` - Compilation fails with `%TEMP%\mono-wasm-e894d683f9f` does not exist. - Opening Visual Studio and doing right click -> build SamplesApp.Wasm will create `%TEMP%\mono-wasm-e894d683f9f`. - Afterwards msbuild from the command line will function. ## Expected behavior ``` git clone master cd src msbuild just works ``` ## How to reproduce it (as minimally and precisely as possible) Forgot to grab a binlog. Opening up early on the off-chance others come across it and as a self-reminder. ## Environment Nuget Package: Package Version(s): Affected platform(s): - [ ] iOS - [ ] Android - [x] WebAssembly - [ ] Windows - [ ] Build tasks Visual Studio - [ ] 2017 (version: ) - [x] 2019 (version: ) - [ ] for Mac (version: ) Relevant plugins - [ ] Resharper (version: ) ## Anything else we need to know? <details> <summary>Microsoft Visual Studio Enterprise 2019 - Version 16.1.6 Details</summary> Microsoft Visual Studio Enterprise 2019 Version 16.1.6 VisualStudio.16.Release/16.1.6+29102.190 Microsoft .NET Framework Version 4.8.03752 Installed Version: Enterprise Visual C++ 2019 00435-60000-00000-AA184 Microsoft Visual C++ 2019 Application Insights Tools for Visual Studio Package 9.1.00429.1 Application Insights Tools for Visual Studio ASP.NET and Web Tools 2019 16.1.429.50124 ASP.NET and Web Tools 2019 ASP.NET Web Frameworks and Tools 2019 16.1.429.50124 For additional information, visit https://www.asp.net/ Azure App Service Tools v3.0.0 16.1.429.50124 Azure App Service Tools v3.0.0 Azure Functions and Web Jobs Tools 16.1.429.50124 Azure Functions and Web Jobs Tools C# Tools 3.1.1-beta4-19281-06+58a4b1e79aea28115e66b06f850c83a3f1fcb6d3 C# components used in the IDE. Depending on your project type and settings, a different version of the compiler may be used. Child Process Debugging Power Tool 1.0 Power tool to add child process debugging to Visual Studio. Common Azure Tools 1.10 Provides common services for use by Azure Mobile Services and Microsoft Azure Tools. Extensibility Message Bus 1.1.77 (master@24013d5) Provides common messaging-based MEF services for loosely coupled Visual Studio extension components communication and integration. FormatDocumentOnSave 1.0 Enables auto formatting of the code when you save a file. Visual Studio supports auto formatting of the code with the CTRL+E,D or CTRL+E,F key shortcuts but with this extension the command 'Format Document' is executed on Save. You can find the source here: https://github.com/Elders/VSE-FormatDocumentOnSave IntelliCode Extension 1.0 IntelliCode Visual Studio Extension Detailed Info Microsoft Azure Tools 2.9 Microsoft Azure Tools for Microsoft Visual Studio 0x10 - v2.9.20419.2 Microsoft Continuous Delivery Tools for Visual Studio 0.4 Simplifying the configuration of Azure DevOps pipelines from within the Visual Studio IDE. Microsoft JVM Debugger 1.0 Provides support for connecting the Visual Studio debugger to JDWP compatible Java Virtual Machines Microsoft Library Manager 1.0 Install client-side libraries easily to any web project Microsoft MI-Based Debugger 1.0 Provides support for connecting Visual Studio to MI compatible debuggers Microsoft Visual C++ Wizards 1.0 Microsoft Visual C++ Wizards Microsoft Visual Studio Tools for Containers 1.1 Develop, run, validate your ASP.NET Core applications in the target environment. F5 your application directly into a container with debugging, or CTRL + F5 to edit & refresh your app without having to rebuild the container. Microsoft Visual Studio VC Package 1.0 Microsoft Visual Studio VC Package Mono Debugging for Visual Studio 16.1.1 (2473f22) Support for debugging Mono processes with Visual Studio. Node.js Tools 1.5.10424.1 Commit Hash:c3ce0ae0b29c0b3a755ffc12f8a685fe7ddd3600 Adds support for developing and debugging Node.js apps in Visual Studio NuGet Package Manager 5.1.0 NuGet Package Manager in Visual Studio. For more information about NuGet, visit https://docs.nuget.org/ OzCodePackage Extension 1.0 OzCodePackage Visual Studio Extension Detailed Info ProjectServicesPackage Extension 1.0 ProjectServicesPackage Visual Studio Extension Detailed Info ResourcePackage Extension 1.0 ResourcePackage Visual Studio Extension Detailed Info ResourcePackage Extension 1.0 ResourcePackage Visual Studio Extension Detailed Info Snapshot Debugging Extension 1.0 Snapshot Debugging Visual Studio Extension Detailed Info SQL Server Data Tools 16.0.61904.23160 Microsoft SQL Server Data Tools Syntax Visualizer 1.0 An extension for visualizing Roslyn SyntaxTrees. Test Adapter for Boost.Test 1.0 Enables Visual Studio's testing tools with unit tests written for Boost.Test. The use terms and Third Party Notices are available in the extension installation directory. Test Adapter for Google Test 1.0 Enables Visual Studio's testing tools with unit tests written for Google Test. The use terms and Third Party Notices are available in the extension installation directory. TypeScript Tools 16.0.10506.2004 TypeScript Tools for Microsoft Visual Studio Visual Basic Tools 3.1.1-beta4-19281-06+58a4b1e79aea28115e66b06f850c83a3f1fcb6d3 Visual Basic components used in the IDE. Depending on your project type and settings, a different version of the compiler may be used. Visual F# Tools 10.4 for F# 4.6 16.1.0-beta.19253.3+42526fe359672a05fd562dc16a91a43d0fe047a7 Microsoft Visual F# Tools 10.4 for F# 4.6 Visual Studio Code Debug Adapter Host Package 1.0 Interop layer for hosting Visual Studio Code debug adapters in Visual Studio Visual Studio Tools for CMake 1.0 Visual Studio Tools for CMake Visual Studio Tools for CMake 1.0 Visual Studio Tools for CMake Visual Studio Tools for Containers 1.0 Visual Studio Tools for Containers VisualStudio.Mac 1.0 Mac Extension for Visual Studio Xamarin 16.1.0.545 (d16-1@db7c858e8) Visual Studio extension to enable development for Xamarin.iOS and Xamarin.Android. Xamarin Designer 16.1.0.418 (remotes/origin/d16-1@5b958bb10) Visual Studio extension to enable Xamarin Designer tools in Visual Studio. Xamarin Templates 16.2.112 (4db4af4) Templates for building iOS, Android, and Windows apps with Xamarin and Xamarin.Forms. Xamarin.Android SDK 9.3.0.23 (HEAD/d0b48056f) Xamarin.Android Reference Assemblies and MSBuild support. Mono: mono/mono/2018-08@3a07bd426d3 Java.Interop: xamarin/java.interop/d16-1@5ddc3e3 LibZipSharp: grendello/LibZipSharp/d16-1@44de300 LibZip: nih-at/libzip/rel-1-5-1@b95cf3f ProGuard: xamarin/proguard/master@905836d SQLite: xamarin/sqlite/3.27.1@8212a2d Xamarin.Android Tools: xamarin/xamarin-android-tools/d16-1@acabd26 Xamarin.iOS and Xamarin.Mac SDK 12.10.0.157 (6bd9475) Xamarin.iOS and Xamarin.Mac Reference Assemblies and MSBuild support. </details>
main
building uno ui sln from the command line does not download temp mono wasm current behavior empty temp git clone master cd src msbuild m t restore uno ui sln msbuild m uno ui sln compilation fails with temp mono wasm  does not exist opening visual studio and doing right click build samplesapp wasm will create temp mono wasm afterwards msbuild from the command line will function expected behavior git clone master cd src msbuild just works how to reproduce it as minimally and precisely as possible forgot to grab a binlog opening up early on the off chance others come across it and as a self reminder environment nuget package package version s affected platform s ios android webassembly windows build tasks visual studio version version for mac version relevant plugins resharper version anything else we need to know microsoft visual studio enterprise version details microsoft visual studio enterprise version visualstudio release microsoft net framework version installed version enterprise visual c microsoft visual c application insights tools for visual studio package application insights tools for visual studio asp net and web tools asp net and web tools asp net web frameworks and tools for additional information visit azure app service tools azure app service tools azure functions and web jobs tools azure functions and web jobs tools c tools c components used in the ide depending on your project type and settings a different version of the compiler may be used child process debugging power tool power tool to add child process debugging to visual studio common azure tools provides common services for use by azure mobile services and microsoft azure tools extensibility message bus master provides common messaging based mef services for loosely coupled visual studio extension components communication and integration formatdocumentonsave enables auto formatting of the code when you save a file visual studio supports auto formatting of the code with the ctrl e d or ctrl e f key shortcuts but with this extension the command format document is executed on save you can find the source here intellicode extension intellicode visual studio extension detailed info microsoft azure tools microsoft azure tools for microsoft visual studio microsoft continuous delivery tools for visual studio simplifying the configuration of azure devops pipelines from within the visual studio ide microsoft jvm debugger provides support for connecting the visual studio debugger to jdwp compatible java virtual machines microsoft library manager install client side libraries easily to any web project microsoft mi based debugger provides support for connecting visual studio to mi compatible debuggers microsoft visual c wizards microsoft visual c wizards microsoft visual studio tools for containers develop run validate your asp net core applications in the target environment your application directly into a container with debugging or ctrl to edit refresh your app without having to rebuild the container microsoft visual studio vc package microsoft visual studio vc package mono debugging for visual studio support for debugging mono processes with visual studio node js tools commit hash adds support for developing and debugging node js apps in visual studio nuget package manager nuget package manager in visual studio for more information about nuget visit ozcodepackage extension ozcodepackage visual studio extension detailed info projectservicespackage extension projectservicespackage visual studio extension detailed info resourcepackage extension resourcepackage visual studio extension detailed info resourcepackage extension resourcepackage visual studio extension detailed info snapshot debugging extension snapshot debugging visual studio extension detailed info sql server data tools microsoft sql server data tools syntax visualizer an extension for visualizing roslyn syntaxtrees test adapter for boost test enables visual studio s testing tools with unit tests written for boost test the use terms and third party notices are available in the extension installation directory test adapter for google test enables visual studio s testing tools with unit tests written for google test the use terms and third party notices are available in the extension installation directory typescript tools typescript tools for microsoft visual studio visual basic tools visual basic components used in the ide depending on your project type and settings a different version of the compiler may be used visual f tools for f beta microsoft visual f tools for f visual studio code debug adapter host package interop layer for hosting visual studio code debug adapters in visual studio visual studio tools for cmake visual studio tools for cmake visual studio tools for cmake visual studio tools for cmake visual studio tools for containers visual studio tools for containers visualstudio mac mac extension for visual studio xamarin visual studio extension to enable development for xamarin ios and xamarin android xamarin designer remotes origin visual studio extension to enable xamarin designer tools in visual studio xamarin templates templates for building ios android and windows apps with xamarin and xamarin forms xamarin android sdk head xamarin android reference assemblies and msbuild support mono mono mono java interop xamarin java interop libzipsharp grendello libzipsharp libzip nih at libzip rel proguard xamarin proguard master sqlite xamarin sqlite xamarin android tools xamarin xamarin android tools xamarin ios and xamarin mac sdk xamarin ios and xamarin mac reference assemblies and msbuild support
1
3,630
14,679,371,940
IssuesEvent
2020-12-31 06:49:45
backdrop-ops/contrib
https://api.github.com/repos/backdrop-ops/contrib
closed
Maintainer change request
Maintainer change request
Thank you for adopting an abandoned Backdrop project. Please note the procedure to add new maintainers: 1. If you haven't already, please join the Backdrop Contrib group by submitting an application.~ 2. File an issue in the current project's issue queue requesting to help maintain the project. 3. If written permission is granted by a current maintainer, create a PR that adds your name to the README.md file in the list of maintainers. Either you, or the current maintainer may merge the PR. 4. If the project does not have a listed maintainer, or if a current maintainer does not respond within 2 weeks, create *this issue* to take over the project. If you have reached step 4, above, you are in the correct place. **Please include a link to the issue you filed for the abandoned project.** https://github.com/backdrop-contrib/reference/issues/52 After confirming the project has been abandoned, a Backdrop Contrib administrator will add your name to the list of maintainers in that project's README.md file. Once your name is listed in the README.md, you may act as the current maintainer.
True
Maintainer change request - Thank you for adopting an abandoned Backdrop project. Please note the procedure to add new maintainers: 1. If you haven't already, please join the Backdrop Contrib group by submitting an application.~ 2. File an issue in the current project's issue queue requesting to help maintain the project. 3. If written permission is granted by a current maintainer, create a PR that adds your name to the README.md file in the list of maintainers. Either you, or the current maintainer may merge the PR. 4. If the project does not have a listed maintainer, or if a current maintainer does not respond within 2 weeks, create *this issue* to take over the project. If you have reached step 4, above, you are in the correct place. **Please include a link to the issue you filed for the abandoned project.** https://github.com/backdrop-contrib/reference/issues/52 After confirming the project has been abandoned, a Backdrop Contrib administrator will add your name to the list of maintainers in that project's README.md file. Once your name is listed in the README.md, you may act as the current maintainer.
main
maintainer change request thank you for adopting an abandoned backdrop project please note the procedure to add new maintainers if you haven t already please join the backdrop contrib group by submitting an application file an issue in the current project s issue queue requesting to help maintain the project if written permission is granted by a current maintainer create a pr that adds your name to the readme md file in the list of maintainers either you or the current maintainer may merge the pr if the project does not have a listed maintainer or if a current maintainer does not respond within weeks create this issue to take over the project if you have reached step above you are in the correct place please include a link to the issue you filed for the abandoned project after confirming the project has been abandoned a backdrop contrib administrator will add your name to the list of maintainers in that project s readme md file once your name is listed in the readme md you may act as the current maintainer
1
1,780
6,575,830,198
IssuesEvent
2017-09-11 17:29:34
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
hg: updating to empty changeset (a.k.a. revision -1 or null) doesn't work
affects_2.1 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME hg ##### ANSIBLE VERSION ``` ansible 2.1.1.0 config file = /home/argh/.ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION ##### OS / ENVIRONMENT Running from Debian GNU/Linux testing, managing Ubuntu 14.04.5 LTS. ##### SUMMARY Pulling all changes and updating a Mercurial repo to empty changeset (for backup/mirroring purposes, for example) used to work(?) with `revision: null`, now it doesn't. It also doesn't work with `revision: -1`, but works with `revision: 'null'` (i.e. with quotes). ##### STEPS TO REPRODUCE ``` yaml - name: Pull an example repo hg: repo: https://www.mercurial-scm.org/repo/hello/ dest: /tmp/hello revision: null # or -1 ``` ##### EXPECTED RESULTS The repo is cloned/pulled, and only contains .hg/ directory with file history, but no files in working directory. At the initial clone the report with -vv should say `{... "before": "", "after": "000000000000 default" ...}`. ##### ACTUAL RESULTS Cloning/pulling works fine, but there are `Makefile` and `hello.c`, updated to the latest revision. Also, with -vv the report says `{... "before": "", "after": "82e55d328c8c default tip" ...}`.
True
hg: updating to empty changeset (a.k.a. revision -1 or null) doesn't work - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME hg ##### ANSIBLE VERSION ``` ansible 2.1.1.0 config file = /home/argh/.ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION ##### OS / ENVIRONMENT Running from Debian GNU/Linux testing, managing Ubuntu 14.04.5 LTS. ##### SUMMARY Pulling all changes and updating a Mercurial repo to empty changeset (for backup/mirroring purposes, for example) used to work(?) with `revision: null`, now it doesn't. It also doesn't work with `revision: -1`, but works with `revision: 'null'` (i.e. with quotes). ##### STEPS TO REPRODUCE ``` yaml - name: Pull an example repo hg: repo: https://www.mercurial-scm.org/repo/hello/ dest: /tmp/hello revision: null # or -1 ``` ##### EXPECTED RESULTS The repo is cloned/pulled, and only contains .hg/ directory with file history, but no files in working directory. At the initial clone the report with -vv should say `{... "before": "", "after": "000000000000 default" ...}`. ##### ACTUAL RESULTS Cloning/pulling works fine, but there are `Makefile` and `hello.c`, updated to the latest revision. Also, with -vv the report says `{... "before": "", "after": "82e55d328c8c default tip" ...}`.
main
hg updating to empty changeset a k a revision or null doesn t work issue type bug report component name hg ansible version ansible config file home argh ansible cfg configured module search path default w o overrides configuration os environment running from debian gnu linux testing managing ubuntu lts summary pulling all changes and updating a mercurial repo to empty changeset for backup mirroring purposes for example used to work with revision null now it doesn t it also doesn t work with revision but works with revision null i e with quotes steps to reproduce yaml name pull an example repo hg repo dest tmp hello revision null or expected results the repo is cloned pulled and only contains hg directory with file history but no files in working directory at the initial clone the report with vv should say before after default actual results cloning pulling works fine but there are makefile and hello c updated to the latest revision also with vv the report says before after default tip
1
3,709
15,188,224,725
IssuesEvent
2021-02-15 14:50:15
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
Tag component with css variables
status: needs triage 🕵️‍♀️ status: waiting for maintainer response 💬 type: enhancement 💡
Hi! As I see, this example not working: https://www.carbondesignsystem.com/components/tag/code If I change the theme, the colors don't react, because the css custom variable is undefined and always the fallback is active. We are facing the same problem, because if we add this to our root scss, it's don't contains the vars for tag: ```scss @include carbon--theme($carbon--theme--g100, true); ``` Is this intentional? Or is there a mixin/function to get this variables? For the quick fix, I made a small mixin: ```scss @mixin get-variables($globalTheme, $tokens) { @each $key, $options in $tokens { $values: map-get($options, 'values'); @each $valueObject in $values { $theme: map-get($valueObject, 'theme'); $value: map-get($valueObject, 'value'); @if $theme == $globalTheme { @include custom-property($key, $value); } } } } ``` Before this we include the tokens of tag (https://github.com/carbon-design-system/carbon/blob/master/packages/components/src/components/tag/_tokens.scss), and we include our mixin like this: ```scss .root-element { @include get-variables($carbon--theme--g100, $tag-colors); } ``` And it's will generate: ```css .root-element { --cds-tag-background-red: #ffd7d9; --cds-tag-color-red: #750e13; --cds-tag-hover-red: #ffb3b8; ... } ``` What do you think about it?
True
Tag component with css variables - Hi! As I see, this example not working: https://www.carbondesignsystem.com/components/tag/code If I change the theme, the colors don't react, because the css custom variable is undefined and always the fallback is active. We are facing the same problem, because if we add this to our root scss, it's don't contains the vars for tag: ```scss @include carbon--theme($carbon--theme--g100, true); ``` Is this intentional? Or is there a mixin/function to get this variables? For the quick fix, I made a small mixin: ```scss @mixin get-variables($globalTheme, $tokens) { @each $key, $options in $tokens { $values: map-get($options, 'values'); @each $valueObject in $values { $theme: map-get($valueObject, 'theme'); $value: map-get($valueObject, 'value'); @if $theme == $globalTheme { @include custom-property($key, $value); } } } } ``` Before this we include the tokens of tag (https://github.com/carbon-design-system/carbon/blob/master/packages/components/src/components/tag/_tokens.scss), and we include our mixin like this: ```scss .root-element { @include get-variables($carbon--theme--g100, $tag-colors); } ``` And it's will generate: ```css .root-element { --cds-tag-background-red: #ffd7d9; --cds-tag-color-red: #750e13; --cds-tag-hover-red: #ffb3b8; ... } ``` What do you think about it?
main
tag component with css variables hi as i see this example not working if i change the theme the colors don t react because the css custom variable is undefined and always the fallback is active we are facing the same problem because if we add this to our root scss it s don t contains the vars for tag scss include carbon theme carbon theme true is this intentional or is there a mixin function to get this variables for the quick fix i made a small mixin scss mixin get variables globaltheme tokens each key options in tokens values map get options values each valueobject in values theme map get valueobject theme value map get valueobject value if theme globaltheme include custom property key value before this we include the tokens of tag and we include our mixin like this scss root element include get variables carbon theme tag colors and it s will generate css root element cds tag background red cds tag color red cds tag hover red what do you think about it
1
4,455
23,184,394,518
IssuesEvent
2022-08-01 07:02:35
beefproject/beef
https://api.github.com/repos/beefproject/beef
closed
Webcam HTML5 TypeError
Module Maintainability Medium
#### Environment What version/revision of BeEF are you using? 0.4.7.0-alpha On what version of Ruby? ruby 2.3.0p0 On what browser? Chrome Version 71.0.3578.98 (Official Build) (64-bit) On what operating system? ubuntu 16.04 LTS #### Configuration Are you using a non-default configuration? yes (to host on HTTPS) Have you enabled or disabled any BeEF extensions? #### Summary Unable to get webcam_html5 to function properly. The hooked victim's domain will have to be served from HTTPS origin per the requirements for GetUserMedia imposed in Chrome. Testing this locally with a HTTPS and allowing access to the webcam when prompted, I verified that GetUserMedia is being called; however I'm now receiving this error in the victim's browser console: Uncaught TypeError: Failed to execute 'createObjectURL' on 'URL': No function was found that matched the signature provided. at eval (eval at socket.onmessage (hook.js?_=1547657577331:1731), <anonymous>:4:756) #### Expected Behaviour What was the expected result? Captures a webcam picture via html5 when the user accepts camera access. #### Actual Behaviour What was the actual result? Uncaught TypeError #### Steps to Reproduce 1) Hook a browser that's using a HTTPS origin 2) Execute the webcam_html5 against the hooked victim 3) Victim clicks webcam access when prompted 4) Observe error in developer console for uncaught typeerror
True
Webcam HTML5 TypeError - #### Environment What version/revision of BeEF are you using? 0.4.7.0-alpha On what version of Ruby? ruby 2.3.0p0 On what browser? Chrome Version 71.0.3578.98 (Official Build) (64-bit) On what operating system? ubuntu 16.04 LTS #### Configuration Are you using a non-default configuration? yes (to host on HTTPS) Have you enabled or disabled any BeEF extensions? #### Summary Unable to get webcam_html5 to function properly. The hooked victim's domain will have to be served from HTTPS origin per the requirements for GetUserMedia imposed in Chrome. Testing this locally with a HTTPS and allowing access to the webcam when prompted, I verified that GetUserMedia is being called; however I'm now receiving this error in the victim's browser console: Uncaught TypeError: Failed to execute 'createObjectURL' on 'URL': No function was found that matched the signature provided. at eval (eval at socket.onmessage (hook.js?_=1547657577331:1731), <anonymous>:4:756) #### Expected Behaviour What was the expected result? Captures a webcam picture via html5 when the user accepts camera access. #### Actual Behaviour What was the actual result? Uncaught TypeError #### Steps to Reproduce 1) Hook a browser that's using a HTTPS origin 2) Execute the webcam_html5 against the hooked victim 3) Victim clicks webcam access when prompted 4) Observe error in developer console for uncaught typeerror
main
webcam typeerror environment what version revision of beef are you using alpha on what version of ruby ruby on what browser chrome version official build bit on what operating system ubuntu lts configuration are you using a non default configuration yes to host on https have you enabled or disabled any beef extensions summary unable to get webcam to function properly the hooked victim s domain will have to be served from https origin per the requirements for getusermedia imposed in chrome testing this locally with a https and allowing access to the webcam when prompted i verified that getusermedia is being called however i m now receiving this error in the victim s browser console uncaught typeerror failed to execute createobjecturl on url no function was found that matched the signature provided at eval eval at socket onmessage hook js expected behaviour what was the expected result captures a webcam picture via when the user accepts camera access actual behaviour what was the actual result uncaught typeerror steps to reproduce hook a browser that s using a https origin execute the webcam against the hooked victim victim clicks webcam access when prompted observe error in developer console for uncaught typeerror
1
5,445
27,263,875,306
IssuesEvent
2023-02-22 16:36:37
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
opened
Refactor Icon component to decouple "notification dot" functionality
type: enhancement work: frontend status: draft restricted: maintainers
In #2514 we modified the `Icon` component to support a red notification dot. We'd like to move that functionality out of `Icon` and into some other mechanism so that `Icon` is kept simple. We need to give some thought to the best way to design the code for the notification dot. Should it be a Svelte action so that we can add a dot to any DOM element? A component?
True
Refactor Icon component to decouple "notification dot" functionality - In #2514 we modified the `Icon` component to support a red notification dot. We'd like to move that functionality out of `Icon` and into some other mechanism so that `Icon` is kept simple. We need to give some thought to the best way to design the code for the notification dot. Should it be a Svelte action so that we can add a dot to any DOM element? A component?
main
refactor icon component to decouple notification dot functionality in we modified the icon component to support a red notification dot we d like to move that functionality out of icon and into some other mechanism so that icon is kept simple we need to give some thought to the best way to design the code for the notification dot should it be a svelte action so that we can add a dot to any dom element a component
1
840
4,480,013,828
IssuesEvent
2016-08-28 00:31:08
coniks-sys/coniks-java
https://api.github.com/repos/coniks-sys/coniks-java
closed
Use single logger in server and client
maintainability util
Change how logging works: - Have a single logger class - Use the following logging convention: "[class] error/exception message"
True
Use single logger in server and client - Change how logging works: - Have a single logger class - Use the following logging convention: "[class] error/exception message"
main
use single logger in server and client change how logging works have a single logger class use the following logging convention error exception message
1
50,861
7,641,457,893
IssuesEvent
2018-05-08 05:05:43
ConsenSys/mythril
https://api.github.com/repos/ConsenSys/mythril
closed
Installation instructions problems
need documentation
Hi, I have two issues with the installation instructions available at the README: 1. It's not mentioned there, but in macos you need to install `leveldb` before installing `mythril` from `pip`. `brew install leveldb` worked for me. If you don't install it first, `plyvel`'s compilation will fail. 2. It's not unclear if this message "If you plan to analyze Solidity code you'll also need the native version of solc. Solcjs is not supported." also applies when running `myth --truffle`, as you have to compile everything first.
1.0
Installation instructions problems - Hi, I have two issues with the installation instructions available at the README: 1. It's not mentioned there, but in macos you need to install `leveldb` before installing `mythril` from `pip`. `brew install leveldb` worked for me. If you don't install it first, `plyvel`'s compilation will fail. 2. It's not unclear if this message "If you plan to analyze Solidity code you'll also need the native version of solc. Solcjs is not supported." also applies when running `myth --truffle`, as you have to compile everything first.
non_main
installation instructions problems hi i have two issues with the installation instructions available at the readme it s not mentioned there but in macos you need to install leveldb before installing mythril from pip brew install leveldb worked for me if you don t install it first plyvel s compilation will fail it s not unclear if this message if you plan to analyze solidity code you ll also need the native version of solc solcjs is not supported also applies when running myth truffle as you have to compile everything first
0
3,912
17,469,305,010
IssuesEvent
2021-08-06 22:42:08
SNDST00M/material-dynmap
https://api.github.com/repos/SNDST00M/material-dynmap
opened
Cross-browser testing for endpoints
context-workflow scope-maintainability type-improvement status-evaluation
## 🏗 Feature Request #### Is your feature request related to a problem? Currently the CI workflow does not test whether UI elements are injected into the DOM after running the script. #### Describe the solution you'd like By adopting cross-browser testing in the stack, we can check that elements load as expected in all entrypoints. This will require some kind of event emitter for when Material Dynmap has activated and the script logic has completed.
True
Cross-browser testing for endpoints - ## 🏗 Feature Request #### Is your feature request related to a problem? Currently the CI workflow does not test whether UI elements are injected into the DOM after running the script. #### Describe the solution you'd like By adopting cross-browser testing in the stack, we can check that elements load as expected in all entrypoints. This will require some kind of event emitter for when Material Dynmap has activated and the script logic has completed.
main
cross browser testing for endpoints 🏗 feature request is your feature request related to a problem currently the ci workflow does not test whether ui elements are injected into the dom after running the script describe the solution you d like by adopting cross browser testing in the stack we can check that elements load as expected in all entrypoints this will require some kind of event emitter for when material dynmap has activated and the script logic has completed
1
14
2,515,188,394
IssuesEvent
2015-01-15 16:58:43
simplesamlphp/simplesamlphp
https://api.github.com/repos/simplesamlphp/simplesamlphp
opened
Extract the consentAdmin or consentSimpleAdmin out of the repository
enhancement low maintainability
Pick one of them, the other should get its own repository and allow installation through composer.
True
Extract the consentAdmin or consentSimpleAdmin out of the repository - Pick one of them, the other should get its own repository and allow installation through composer.
main
extract the consentadmin or consentsimpleadmin out of the repository pick one of them the other should get its own repository and allow installation through composer
1
306
3,078,425,253
IssuesEvent
2015-08-21 10:06:31
pavel-pimenov/flylinkdc-r5xx
https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx
closed
r500. Нелогичное поведение автовыбора настроек соединения.
bug imported Maintainability Priority-Medium
_From [bobrikov](https://code.google.com/u/bobrikov/) on May 06, 2011 12:43:04_ У меня стоит роутер с фключенным UPnP , в настройках соединения ставлю Фаервлл с UPnP. Соединяюсь - он пишеь в логе, что не удалось пробросить и сам включает активный режим. (может роутер глюкнул, не знаю, в стронге было всё ок) Зачем мне активный режим на роутере, если кроме пассива, больше ничего не даст найти ни файла. Да и хабы после поиска выкидывают за подмену ип. _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=449_
True
r500. Нелогичное поведение автовыбора настроек соединения. - _From [bobrikov](https://code.google.com/u/bobrikov/) on May 06, 2011 12:43:04_ У меня стоит роутер с фключенным UPnP , в настройках соединения ставлю Фаервлл с UPnP. Соединяюсь - он пишеь в логе, что не удалось пробросить и сам включает активный режим. (может роутер глюкнул, не знаю, в стронге было всё ок) Зачем мне активный режим на роутере, если кроме пассива, больше ничего не даст найти ни файла. Да и хабы после поиска выкидывают за подмену ип. _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=449_
main
нелогичное поведение автовыбора настроек соединения from on may у меня стоит роутер с фключенным upnp в настройках соединения ставлю фаервлл с upnp соединяюсь он пишеь в логе что не удалось пробросить и сам включает активный режим может роутер глюкнул не знаю в стронге было всё ок зачем мне активный режим на роутере если кроме пассива больше ничего не даст найти ни файла да и хабы после поиска выкидывают за подмену ип original issue
1
143,301
21,995,887,603
IssuesEvent
2022-05-26 06:16:02
stores-cedcommerce/Anthony-Store-Design
https://api.github.com/repos/stores-cedcommerce/Anthony-Store-Design
opened
The spacing from the left and right side is not equal.
Header section Mobile Design / UI / UX
**Actual result:** The spacing from the left and right side is not equal. ![image](https://user-images.githubusercontent.com/102131636/170428111-39d50a8a-be6f-4f24-9ecb-1ea07de72cb6.png) **Expected result:** The spacing can be equal from the left and right side.
1.0
The spacing from the left and right side is not equal. - **Actual result:** The spacing from the left and right side is not equal. ![image](https://user-images.githubusercontent.com/102131636/170428111-39d50a8a-be6f-4f24-9ecb-1ea07de72cb6.png) **Expected result:** The spacing can be equal from the left and right side.
non_main
the spacing from the left and right side is not equal actual result the spacing from the left and right side is not equal expected result the spacing can be equal from the left and right side
0
35,899
2,793,819,833
IssuesEvent
2015-05-11 13:37:16
elecoest/allevents-3-2
https://api.github.com/repos/elecoest/allevents-3-2
closed
Frontend - affichage événements passés ?
auto-migrated Priority-Medium Type-Enhancement
``` Est-il possible de faire afficher en FrontEnd à la fois les événements à venir et passés, comme AllEvent pour Joomla 1.x? ``` Original issue reported on code.google.com by `antoine....@gmail.com` on 10 Mar 2015 at 8:32
1.0
Frontend - affichage événements passés ? - ``` Est-il possible de faire afficher en FrontEnd à la fois les événements à venir et passés, comme AllEvent pour Joomla 1.x? ``` Original issue reported on code.google.com by `antoine....@gmail.com` on 10 Mar 2015 at 8:32
non_main
frontend affichage événements passés est il possible de faire afficher en frontend à la fois les événements à venir et passés comme allevent pour joomla x original issue reported on code google com by antoine gmail com on mar at
0
188,646
14,449,161,933
IssuesEvent
2020-12-08 07:37:42
facebook/react-native
https://api.github.com/repos/facebook/react-native
reopened
App sometimes reloads the js bundle when switching from background to foreground.
Needs: Attention Needs: Repro Needs: Verify on Latest Version
## Description App sometimes reloads the js bundle when switching from background to foreground. ## React Native version: 0.61 ## Steps To Reproduce Provide a detailed list of steps that reproduce the issue. Create a new react-native application. Install react-navigation. Create a stack navigator. Open app, go trough several screens. Switch app to another one in the background. Open as many apps as you can so you have a lot of them in the memory. Switch back and the react-native app reloads the bundle. ## Expected Results I should be able to manually disable this reaload. Since I don't kill the app I should go back to where I was. This is very annoying when you try to fill a form and you go to another application to copy and come back to paste.
1.0
App sometimes reloads the js bundle when switching from background to foreground. - ## Description App sometimes reloads the js bundle when switching from background to foreground. ## React Native version: 0.61 ## Steps To Reproduce Provide a detailed list of steps that reproduce the issue. Create a new react-native application. Install react-navigation. Create a stack navigator. Open app, go trough several screens. Switch app to another one in the background. Open as many apps as you can so you have a lot of them in the memory. Switch back and the react-native app reloads the bundle. ## Expected Results I should be able to manually disable this reaload. Since I don't kill the app I should go back to where I was. This is very annoying when you try to fill a form and you go to another application to copy and come back to paste.
non_main
app sometimes reloads the js bundle when switching from background to foreground description app sometimes reloads the js bundle when switching from background to foreground react native version steps to reproduce provide a detailed list of steps that reproduce the issue create a new react native application install react navigation create a stack navigator open app go trough several screens switch app to another one in the background open as many apps as you can so you have a lot of them in the memory switch back and the react native app reloads the bundle expected results i should be able to manually disable this reaload since i don t kill the app i should go back to where i was this is very annoying when you try to fill a form and you go to another application to copy and come back to paste
0
331,499
28,965,085,742
IssuesEvent
2023-05-10 07:18:37
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix backend_handler.test_set_backend
Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4712126583/jobs/8356932101" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4712126583/jobs/8356932101" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4712126583/jobs/8356932101" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4712126583/jobs/8356932101" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix backend_handler.test_set_backend - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4712126583/jobs/8356932101" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4712126583/jobs/8356932101" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4712126583/jobs/8356932101" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4712126583/jobs/8356932101" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
non_main
fix backend handler test set backend tensorflow img src torch img src numpy img src jax img src
0
4,375
22,274,625,853
IssuesEvent
2022-06-10 15:23:46
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
[Bug]: ParserError in index.scss
type: bug 🐛 status: needs triage 🕵️‍♀️ status: waiting for maintainer response 💬
### Package @carbon/react ### Browser Chrome ### Package version 1.3.0 ### React version 16.10 ### Description Building a using a scss including @use 'carbon/react' project fails # yarn build Creating an optimized production build... Browserslist: caniuse-lite is outdated. Please run: npx browserslist@latest --update-db Why you should do it regularly: https://github.com/browserslist/browserslist#browsers-data-updating https://github.com/carbon-design-system/carbon-tutorial/tree/v11-react-step-2Failed to compile. ./src/index.scss ParserError: Syntax Error at line: 1, column 31 ### Reproduction/example https://github.com/carbon-design-system/carbon-tutorial/tree/v11-react-step-2 ### Steps to reproduce git clone -b v11-react-step-2 https://github.com/carbon-design-system/carbon-tutorial.git cd carbon-tutorial yarn install yarn build ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md) - [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems
True
[Bug]: ParserError in index.scss - ### Package @carbon/react ### Browser Chrome ### Package version 1.3.0 ### React version 16.10 ### Description Building a using a scss including @use 'carbon/react' project fails # yarn build Creating an optimized production build... Browserslist: caniuse-lite is outdated. Please run: npx browserslist@latest --update-db Why you should do it regularly: https://github.com/browserslist/browserslist#browsers-data-updating https://github.com/carbon-design-system/carbon-tutorial/tree/v11-react-step-2Failed to compile. ./src/index.scss ParserError: Syntax Error at line: 1, column 31 ### Reproduction/example https://github.com/carbon-design-system/carbon-tutorial/tree/v11-react-step-2 ### Steps to reproduce git clone -b v11-react-step-2 https://github.com/carbon-design-system/carbon-tutorial.git cd carbon-tutorial yarn install yarn build ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md) - [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems
main
parsererror in index scss package carbon react browser chrome package version react version description building a using a scss including use carbon react project fails yarn build creating an optimized production build browserslist caniuse lite is outdated please run npx browserslist latest update db why you should do it regularly to compile src index scss parsererror syntax error at line column reproduction example steps to reproduce git clone b react step cd carbon tutorial yarn install yarn build code of conduct i agree to follow this project s i checked the for duplicate problems
1
369,517
10,914,202,241
IssuesEvent
2019-11-21 08:40:24
devinit/D-Portal
https://api.github.com/repos/devinit/D-Portal
closed
Filter by humanitarian flag ?
feature priority
I can see on an activity page whether the ``humanitaria`` flag is declared : eg: http://d-portal.org/q.html?aid=GB-CHC-1065972-DF168 ![humanitarian](https://user-images.githubusercontent.com/1209944/47169888-8d5bdf00-d2fc-11e8-98f2-264ade42cb84.png) At the d-portal search , is it possible to have a filter in order to only see/search humanitarian activities: - those with ``humanitarian="1" (or ="true") at the ``iati-activity`` element, - those with any ``transaction`` that contain a similar declaration ?
1.0
Filter by humanitarian flag ? - I can see on an activity page whether the ``humanitaria`` flag is declared : eg: http://d-portal.org/q.html?aid=GB-CHC-1065972-DF168 ![humanitarian](https://user-images.githubusercontent.com/1209944/47169888-8d5bdf00-d2fc-11e8-98f2-264ade42cb84.png) At the d-portal search , is it possible to have a filter in order to only see/search humanitarian activities: - those with ``humanitarian="1" (or ="true") at the ``iati-activity`` element, - those with any ``transaction`` that contain a similar declaration ?
non_main
filter by humanitarian flag i can see on an activity page whether the humanitaria flag is declared eg at the d portal search is it possible to have a filter in order to only see search humanitarian activities those with humanitarian or true at the iati activity element those with any transaction that contain a similar declaration
0
2,116
7,198,243,774
IssuesEvent
2018-02-05 12:04:31
Chromeroni/Hera-Chatbot
https://api.github.com/repos/Chromeroni/Hera-Chatbot
opened
Implement a logging framework
Maintainability to be reviewed
Implement a logging framework, such as log4J. Create daily logs, which will be stored on the bots current execution environment,
True
Implement a logging framework - Implement a logging framework, such as log4J. Create daily logs, which will be stored on the bots current execution environment,
main
implement a logging framework implement a logging framework such as create daily logs which will be stored on the bots current execution environment
1
91,740
3,862,440,666
IssuesEvent
2016-04-08 02:52:21
TranslationWMcs435/TranslationWMcs435
https://api.github.com/repos/TranslationWMcs435/TranslationWMcs435
closed
Make a more robust README and user instructions
Medium Priority
We know how to use the code, but Carlos (or another group) might not. Make the instructions clear, so that we can be sure people use it right.
1.0
Make a more robust README and user instructions - We know how to use the code, but Carlos (or another group) might not. Make the instructions clear, so that we can be sure people use it right.
non_main
make a more robust readme and user instructions we know how to use the code but carlos or another group might not make the instructions clear so that we can be sure people use it right
0
45,731
5,729,645,304
IssuesEvent
2017-04-21 07:00:01
nuxsmin/sysPass
https://api.github.com/repos/nuxsmin/sysPass
closed
'The user has no associated groups' when using LDAP
3rdParty NeedTests
I'm trying to authenticate against a Samba4 AD using the 2.0 Release of sysPass. Whenever I login using an AD Account, it fails with the message 'The user has no associated groups'. **My LDAP config in sysPass:** * Enable LDAP: true * Active Directory: false (because Samba apparently doesn't support inherited group membership) * Server: `ldaps://dc.example.com:636` * Bind User: `cn=syspass,cn=Users,dc=example,dc=com` * Bind Password: `secretpass` * Search base: `cn=Users,dc=example,dc=com` * Group: `cn=Domain Users,cn=Users,dc=example,dc=com` * Default Group: `Users` * Default Profile: `User` When I try logging in using the user `adm.max.mustermann`(which is present and working in other applications) I receive the aforementioned error message. The event log shows the following: > The user doesn't belong to the group > User : CN=Admin Max Mustermann,CN=Users,DC=example,DC=com > Group : Domain Users and > The user has no associated groups > Type : authBrowser > Type : authLdap > User : adm.max.mustermann I believe this behavior might have been introduced in one of the release candidates of 2.0, as I got this part to work in an earlier version (but ran into other issues which have since been fixed).
1.0
'The user has no associated groups' when using LDAP - I'm trying to authenticate against a Samba4 AD using the 2.0 Release of sysPass. Whenever I login using an AD Account, it fails with the message 'The user has no associated groups'. **My LDAP config in sysPass:** * Enable LDAP: true * Active Directory: false (because Samba apparently doesn't support inherited group membership) * Server: `ldaps://dc.example.com:636` * Bind User: `cn=syspass,cn=Users,dc=example,dc=com` * Bind Password: `secretpass` * Search base: `cn=Users,dc=example,dc=com` * Group: `cn=Domain Users,cn=Users,dc=example,dc=com` * Default Group: `Users` * Default Profile: `User` When I try logging in using the user `adm.max.mustermann`(which is present and working in other applications) I receive the aforementioned error message. The event log shows the following: > The user doesn't belong to the group > User : CN=Admin Max Mustermann,CN=Users,DC=example,DC=com > Group : Domain Users and > The user has no associated groups > Type : authBrowser > Type : authLdap > User : adm.max.mustermann I believe this behavior might have been introduced in one of the release candidates of 2.0, as I got this part to work in an earlier version (but ran into other issues which have since been fixed).
non_main
the user has no associated groups when using ldap i m trying to authenticate against a ad using the release of syspass whenever i login using an ad account it fails with the message the user has no associated groups my ldap config in syspass enable ldap true active directory false because samba apparently doesn t support inherited group membership server ldaps dc example com bind user cn syspass cn users dc example dc com bind password secretpass search base cn users dc example dc com group cn domain users cn users dc example dc com default group users default profile user when i try logging in using the user adm max mustermann which is present and working in other applications i receive the aforementioned error message the event log shows the following the user doesn t belong to the group user cn admin max mustermann cn users dc example dc com group domain users and the user has no associated groups type authbrowser type authldap user adm max mustermann i believe this behavior might have been introduced in one of the release candidates of as i got this part to work in an earlier version but ran into other issues which have since been fixed
0
30,258
4,577,017,815
IssuesEvent
2016-09-17 00:32:00
Becksteinlab/GromacsWrapper
https://api.github.com/repos/Becksteinlab/GromacsWrapper
closed
use tempdir in tests
tests
The current tests litter gromacs files and trajectories into the current directory. We should use [tempdir](https://pypi.python.org/pypi/tempdir/) to make sure that all temp files are properly contained and cleaned up afterwards.
1.0
use tempdir in tests - The current tests litter gromacs files and trajectories into the current directory. We should use [tempdir](https://pypi.python.org/pypi/tempdir/) to make sure that all temp files are properly contained and cleaned up afterwards.
non_main
use tempdir in tests the current tests litter gromacs files and trajectories into the current directory we should use to make sure that all temp files are properly contained and cleaned up afterwards
0
77,080
9,978,871,354
IssuesEvent
2019-07-09 21:02:07
directus/api
https://api.github.com/repos/directus/api
opened
API Functional Tests
documentation enhancement help wanted tests
<!-- 1. Do not delete this template or the issue will be closed 2. Ensure you're using the latest version of Directus 3. Post to the correct repo: App: https://github.com/directus/app/issues API: https://github.com/directus/api/issues (YOU ARE HERE) Docs: https://github.com/directus/docs/issues --> # Feature Request Create API Functional Tests ## What problem does this feature solve? Knowing that the API is working as expected; more thoroughly testing the API after updates/fixes... and before releasing. ## How do you think this should be implemented? TBD. @shealavington has been working on this for a while, but we need to align on exactly _how_ these tests will be structured. Here is what Shea wrote: > I've got [some tests for the Items endpoint running](https://github.com/directus/api-tests/tree/poc/wolfulus/tests). There's a few changes I'd like to make _(Such as outputting successful tests to the console)_, but I guess the first iteration is almost ready to be master-merged. Though on that note, I'm currently testing the 'Demo' directly, therefore some fail _(deletions)_ as it doesn't spawn a new instance every time haha. * *'Items' Tests:* * *Create*: Create a single item & Create multiple items in one payload. * *Read*: Get item, Get multiple items, Get all items _(by default should max out at 200 items)_ * *Update*: Update item, Update multiple items with the same data, Update multiple items with different data. * *Delete*: Delete item, delete multiple items. * *Server:* Ping should Pong * *Project:* Expected keys should exist when authed, and shouldn't when not authed. # Status ## Authentication * [ ] Tokens * [ ] Get Auth Token * [ ] Refresh Auth Token * [ ] Password Reset Request * [ ] Password Reset * [ ] SSO ## Query Parameters * [ ] Metadata * [ ] Fields * [ ] Limit * [ ] Offset * [ ] Single * [ ] Sorting * [ ] Status * [ ] Filtering * [ ] Language * [ ] Search Query * [ ] Groups * [ ] Skip Activity Logging * [ ] Activity Comment ## Items * [X] Get Item * [X] Get Multiple Items * [X] Get All Items * [X] Create Item * [X] Create Items * [X] Update Item * [X] Update Items * [X] Delete Item * [X] Delete Items * [ ] Get Item Revision * [ ] Get Item Revisions * [ ] Revert Item * [ ] Revert Items? ## System * [ ] Activity * [ ] Collections * [ ] Collection Presets * [ ] Fields * [ ] Files * [ ] Folders * [ ] Permissions * [ ] Relations * [ ] Revisions * [ ] Roles * [ ] Settings * [ ] Users ## Utilities * [ ] Hash String * [ ] Match Hashed String * [ ] Generate Random String ## Mail * [ ] Send Email ## Extensions * [ ] Get Extensions * [ ] Get Interface * [ ] Get Page * [ ] Get Custom Endpoint ## Server * [ ] Information * [ ] Ping ## Projects * [ ] Information * [ ] Update * [ ] Create Project * [ ] Projects Auth Config * [ ] Projects CORS Config * [ ] Projects Cache Config * [ ] Projects Storage Config * [ ] Projects Mail Config * [ ] Projects Rate Limit Config ## Field Types * [ ] Webhooks ## Directus Objects * [ ] Activity Object * [ ] Activity Seen Object * [ ] Collection Object * [ ] Collection Preset Object * [ ] Field Object * [ ] File Object * [ ] Folder Object * [ ] Permission Object * [ ] Relation Object * [ ] Revision Object * [ ] Role Object * [ ] Setting Object * [ ] User Object * [ ] User Role Object ## SCIM * [ ] Overview * [ ] Create SCIM User * [ ] Get SCIM Users * [ ] Get SCIM User * [ ] Update SCIM User * [ ] Create SCIM Group * [ ] Get SCIM Groups * [ ] Get SCIM Group * [ ] Update SCIM Group * [ ] Delete SCIM Group
1.0
API Functional Tests - <!-- 1. Do not delete this template or the issue will be closed 2. Ensure you're using the latest version of Directus 3. Post to the correct repo: App: https://github.com/directus/app/issues API: https://github.com/directus/api/issues (YOU ARE HERE) Docs: https://github.com/directus/docs/issues --> # Feature Request Create API Functional Tests ## What problem does this feature solve? Knowing that the API is working as expected; more thoroughly testing the API after updates/fixes... and before releasing. ## How do you think this should be implemented? TBD. @shealavington has been working on this for a while, but we need to align on exactly _how_ these tests will be structured. Here is what Shea wrote: > I've got [some tests for the Items endpoint running](https://github.com/directus/api-tests/tree/poc/wolfulus/tests). There's a few changes I'd like to make _(Such as outputting successful tests to the console)_, but I guess the first iteration is almost ready to be master-merged. Though on that note, I'm currently testing the 'Demo' directly, therefore some fail _(deletions)_ as it doesn't spawn a new instance every time haha. * *'Items' Tests:* * *Create*: Create a single item & Create multiple items in one payload. * *Read*: Get item, Get multiple items, Get all items _(by default should max out at 200 items)_ * *Update*: Update item, Update multiple items with the same data, Update multiple items with different data. * *Delete*: Delete item, delete multiple items. * *Server:* Ping should Pong * *Project:* Expected keys should exist when authed, and shouldn't when not authed. # Status ## Authentication * [ ] Tokens * [ ] Get Auth Token * [ ] Refresh Auth Token * [ ] Password Reset Request * [ ] Password Reset * [ ] SSO ## Query Parameters * [ ] Metadata * [ ] Fields * [ ] Limit * [ ] Offset * [ ] Single * [ ] Sorting * [ ] Status * [ ] Filtering * [ ] Language * [ ] Search Query * [ ] Groups * [ ] Skip Activity Logging * [ ] Activity Comment ## Items * [X] Get Item * [X] Get Multiple Items * [X] Get All Items * [X] Create Item * [X] Create Items * [X] Update Item * [X] Update Items * [X] Delete Item * [X] Delete Items * [ ] Get Item Revision * [ ] Get Item Revisions * [ ] Revert Item * [ ] Revert Items? ## System * [ ] Activity * [ ] Collections * [ ] Collection Presets * [ ] Fields * [ ] Files * [ ] Folders * [ ] Permissions * [ ] Relations * [ ] Revisions * [ ] Roles * [ ] Settings * [ ] Users ## Utilities * [ ] Hash String * [ ] Match Hashed String * [ ] Generate Random String ## Mail * [ ] Send Email ## Extensions * [ ] Get Extensions * [ ] Get Interface * [ ] Get Page * [ ] Get Custom Endpoint ## Server * [ ] Information * [ ] Ping ## Projects * [ ] Information * [ ] Update * [ ] Create Project * [ ] Projects Auth Config * [ ] Projects CORS Config * [ ] Projects Cache Config * [ ] Projects Storage Config * [ ] Projects Mail Config * [ ] Projects Rate Limit Config ## Field Types * [ ] Webhooks ## Directus Objects * [ ] Activity Object * [ ] Activity Seen Object * [ ] Collection Object * [ ] Collection Preset Object * [ ] Field Object * [ ] File Object * [ ] Folder Object * [ ] Permission Object * [ ] Relation Object * [ ] Revision Object * [ ] Role Object * [ ] Setting Object * [ ] User Object * [ ] User Role Object ## SCIM * [ ] Overview * [ ] Create SCIM User * [ ] Get SCIM Users * [ ] Get SCIM User * [ ] Update SCIM User * [ ] Create SCIM Group * [ ] Get SCIM Groups * [ ] Get SCIM Group * [ ] Update SCIM Group * [ ] Delete SCIM Group
non_main
api functional tests do not delete this template or the issue will be closed ensure you re using the latest version of directus post to the correct repo app api you are here docs feature request create api functional tests what problem does this feature solve knowing that the api is working as expected more thoroughly testing the api after updates fixes and before releasing how do you think this should be implemented tbd shealavington has been working on this for a while but we need to align on exactly how these tests will be structured here is what shea wrote i ve got there s a few changes i d like to make such as outputting successful tests to the console but i guess the first iteration is almost ready to be master merged though on that note i m currently testing the demo directly therefore some fail deletions as it doesn t spawn a new instance every time haha items tests create create a single item create multiple items in one payload read get item get multiple items get all items by default should max out at items update update item update multiple items with the same data update multiple items with different data delete delete item delete multiple items server ping should pong project expected keys should exist when authed and shouldn t when not authed status authentication tokens get auth token refresh auth token password reset request password reset sso query parameters metadata fields limit offset single sorting status filtering language search query groups skip activity logging activity comment items get item get multiple items get all items create item create items update item update items delete item delete items get item revision get item revisions revert item revert items system activity collections collection presets fields files folders permissions relations revisions roles settings users utilities hash string match hashed string generate random string mail send email extensions get extensions get interface get page get custom endpoint server information ping projects information update create project projects auth config projects cors config projects cache config projects storage config projects mail config projects rate limit config field types webhooks directus objects activity object activity seen object collection object collection preset object field object file object folder object permission object relation object revision object role object setting object user object user role object scim overview create scim user get scim users get scim user update scim user create scim group get scim groups get scim group update scim group delete scim group
0
30,533
2,724,004,158
IssuesEvent
2015-04-14 15:33:47
CruxFramework/crux-widgets
https://api.github.com/repos/CruxFramework/crux-widgets
closed
I need to restart devMode when I add a declared i18n message on crux page
bug CruxCore imported Milestone-4.0.0 Priority-Medium
_From [thi...@cruxframework.org](https://code.google.com/u/114650528804514463329/) on January 07, 2011 15:22:06_ I need to restart devMode when I add a declared i18n message on crux page _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=232_
1.0
I need to restart devMode when I add a declared i18n message on crux page - _From [thi...@cruxframework.org](https://code.google.com/u/114650528804514463329/) on January 07, 2011 15:22:06_ I need to restart devMode when I add a declared i18n message on crux page _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=232_
non_main
i need to restart devmode when i add a declared message on crux page from on january i need to restart devmode when i add a declared message on crux page original issue
0
2,602
8,838,195,623
IssuesEvent
2019-01-05 14:40:47
arcticicestudio/arctic-landscape
https://api.github.com/repos/arcticicestudio/arctic-landscape
closed
Initial Project Setup
context-workflow scope-configurability scope-maintainability type-feature
<p align="center"><img src="https://user-images.githubusercontent.com/7836623/50725336-7c9d5700-10fc-11e9-8a15-6df32f266976.png" /></p> Implement the initial project setup based on Arctic Ice Studio's standards and conventions.
True
Initial Project Setup - <p align="center"><img src="https://user-images.githubusercontent.com/7836623/50725336-7c9d5700-10fc-11e9-8a15-6df32f266976.png" /></p> Implement the initial project setup based on Arctic Ice Studio's standards and conventions.
main
initial project setup implement the initial project setup based on arctic ice studio s standards and conventions
1
2,021
6,757,629,242
IssuesEvent
2017-10-24 11:33:28
Kristinita/Erics-Green-Room
https://api.github.com/repos/Kristinita/Erics-Green-Room
closed
[Feature request] Открытие комнаты сразу после введения команды
need-maintainer web
### 1. Запрос Неплохо было бы, если б комнаты открывались сразу после команд введения комнат. ### 2. Аргументация 1. Экономия времени игрокам. Не нужно каждый раз делать лишнее движение мышью, чтобы открыть комнату. 1. Упрощение работы сервиса для новичков. Так, у lollipop возникли проблемы из-за того, что она не поняла, куда кликать. ### 3. Шаги для воспроизведения Зашёл на Альфа-хаб → ввёл команду `!newroom GREEN`: ![newroom GREEN](https://i.imgur.com/uqVAdez.png) ### 4. Желаемое поведение Активной становится вкладка с комнатой `GREEN`. ### 5. Актуальное поведение Активной остаётся вкладка с основным чатом, но не комнатой. Спасибо.
True
[Feature request] Открытие комнаты сразу после введения команды - ### 1. Запрос Неплохо было бы, если б комнаты открывались сразу после команд введения комнат. ### 2. Аргументация 1. Экономия времени игрокам. Не нужно каждый раз делать лишнее движение мышью, чтобы открыть комнату. 1. Упрощение работы сервиса для новичков. Так, у lollipop возникли проблемы из-за того, что она не поняла, куда кликать. ### 3. Шаги для воспроизведения Зашёл на Альфа-хаб → ввёл команду `!newroom GREEN`: ![newroom GREEN](https://i.imgur.com/uqVAdez.png) ### 4. Желаемое поведение Активной становится вкладка с комнатой `GREEN`. ### 5. Актуальное поведение Активной остаётся вкладка с основным чатом, но не комнатой. Спасибо.
main
открытие комнаты сразу после введения команды запрос неплохо было бы если б комнаты открывались сразу после команд введения комнат аргументация экономия времени игрокам не нужно каждый раз делать лишнее движение мышью чтобы открыть комнату упрощение работы сервиса для новичков так у lollipop возникли проблемы из за того что она не поняла куда кликать шаги для воспроизведения зашёл на альфа хаб → ввёл команду newroom green желаемое поведение активной становится вкладка с комнатой green актуальное поведение активной остаётся вкладка с основным чатом но не комнатой спасибо
1
2,697
9,417,455,573
IssuesEvent
2019-04-10 16:46:41
victorcombalweiss/datapuppy
https://api.github.com/repos/victorcombalweiss/datapuppy
opened
Move to Typescript
maintainability
As of now web application's front-end is written in raw Javascript. In order to have better maintainability it would be good to switch to Typescript.
True
Move to Typescript - As of now web application's front-end is written in raw Javascript. In order to have better maintainability it would be good to switch to Typescript.
main
move to typescript as of now web application s front end is written in raw javascript in order to have better maintainability it would be good to switch to typescript
1
280,048
24,277,229,904
IssuesEvent
2022-09-28 14:39:23
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: unoptimized-query-oracle/disable-rules=all failed
C-test-failure O-robot O-roachtest branch-master release-blocker
roachtest.unoptimized-query-oracle/disable-rules=all [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6661603?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6661603?buildTab=artifacts#/unoptimized-query-oracle/disable-rules=all) on master @ [4982322e91c37c6d0249dc5597b9e02eec19848d](https://github.com/cockroachdb/cockroach/commits/4982322e91c37c6d0249dc5597b9e02eec19848d): ``` test artifacts and logs in: /artifacts/unoptimized-query-oracle/disable-rules=all/run_1 query_comparison_util.go:234,query_comparison_util.go:66,unoptimized_query_oracle.go:51,test_runner.go:928: . 23546 statements run: expected unoptimized and optimized results to be equal (1) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.(*queryComparisonHelper).makeError | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/query_comparison_util.go:342 | [...repeated from below...] Wraps: (2) . 23546 statements run Wraps: (3) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runUnoptimizedQueryOracleImpl | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/unoptimized_query_oracle.go:165 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerUnoptimizedQueryOracle.func1.1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/unoptimized_query_oracle.go:54 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runOneRoundQueryComparison | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/query_comparison_util.go:233 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runQueryComparison | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/query_comparison_util.go:66 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerUnoptimizedQueryOracle.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/unoptimized_query_oracle.go:51 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:928 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1594 Wraps: (4) expected unoptimized and optimized results to be equal |   []string{ |    strings.Join({ |    "-10.1,0,-106", | +  ".00000000000000000", |    ",0,52475396.30853651200", |    }, ""), |   } | sql: SELECT | (-10.1):::DECIMAL AS col_79099, | 0:::OID AS col_79100, | ((-106):::INT8::INT8 / 1:::INT8::INT8)::DECIMAL AS col_79101, | 0:::OID AS col_79102, | 52475396.30853651200:::DECIMAL AS col_79103 | FROM | defaultdb.public.table5@[0] AS tab_28857 | ORDER BY | tab_28857.col5_1 DESC Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.leafError ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #88558 roachtest: unoptimized-query-oracle/disable-rules=all failed [C-test-failure O-roachtest O-robot T-sql-queries branch-release-22.2] </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*unoptimized-query-oracle/disable-rules=all.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: unoptimized-query-oracle/disable-rules=all failed - roachtest.unoptimized-query-oracle/disable-rules=all [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6661603?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6661603?buildTab=artifacts#/unoptimized-query-oracle/disable-rules=all) on master @ [4982322e91c37c6d0249dc5597b9e02eec19848d](https://github.com/cockroachdb/cockroach/commits/4982322e91c37c6d0249dc5597b9e02eec19848d): ``` test artifacts and logs in: /artifacts/unoptimized-query-oracle/disable-rules=all/run_1 query_comparison_util.go:234,query_comparison_util.go:66,unoptimized_query_oracle.go:51,test_runner.go:928: . 23546 statements run: expected unoptimized and optimized results to be equal (1) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.(*queryComparisonHelper).makeError | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/query_comparison_util.go:342 | [...repeated from below...] Wraps: (2) . 23546 statements run Wraps: (3) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runUnoptimizedQueryOracleImpl | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/unoptimized_query_oracle.go:165 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerUnoptimizedQueryOracle.func1.1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/unoptimized_query_oracle.go:54 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runOneRoundQueryComparison | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/query_comparison_util.go:233 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runQueryComparison | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/query_comparison_util.go:66 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerUnoptimizedQueryOracle.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/unoptimized_query_oracle.go:51 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:928 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1594 Wraps: (4) expected unoptimized and optimized results to be equal |   []string{ |    strings.Join({ |    "-10.1,0,-106", | +  ".00000000000000000", |    ",0,52475396.30853651200", |    }, ""), |   } | sql: SELECT | (-10.1):::DECIMAL AS col_79099, | 0:::OID AS col_79100, | ((-106):::INT8::INT8 / 1:::INT8::INT8)::DECIMAL AS col_79101, | 0:::OID AS col_79102, | 52475396.30853651200:::DECIMAL AS col_79103 | FROM | defaultdb.public.table5@[0] AS tab_28857 | ORDER BY | tab_28857.col5_1 DESC Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.leafError ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #88558 roachtest: unoptimized-query-oracle/disable-rules=all failed [C-test-failure O-roachtest O-robot T-sql-queries branch-release-22.2] </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*unoptimized-query-oracle/disable-rules=all.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
non_main
roachtest unoptimized query oracle disable rules all failed roachtest unoptimized query oracle disable rules all with on master test artifacts and logs in artifacts unoptimized query oracle disable rules all run query comparison util go query comparison util go unoptimized query oracle go test runner go statements run expected unoptimized and optimized results to be equal attached stack trace stack trace github com cockroachdb cockroach pkg cmd roachtest tests querycomparisonhelper makeerror github com cockroachdb cockroach pkg cmd roachtest tests query comparison util go wraps statements run wraps attached stack trace stack trace github com cockroachdb cockroach pkg cmd roachtest tests rununoptimizedqueryoracleimpl github com cockroachdb cockroach pkg cmd roachtest tests unoptimized query oracle go github com cockroachdb cockroach pkg cmd roachtest tests registerunoptimizedqueryoracle github com cockroachdb cockroach pkg cmd roachtest tests unoptimized query oracle go github com cockroachdb cockroach pkg cmd roachtest tests runoneroundquerycomparison github com cockroachdb cockroach pkg cmd roachtest tests query comparison util go github com cockroachdb cockroach pkg cmd roachtest tests runquerycomparison github com cockroachdb cockroach pkg cmd roachtest tests query comparison util go github com cockroachdb cockroach pkg cmd roachtest tests registerunoptimizedqueryoracle github com cockroachdb cockroach pkg cmd roachtest tests unoptimized query oracle go main testrunner runtest main pkg cmd roachtest test runner go runtime goexit goroot src runtime asm s wraps expected unoptimized and optimized results to be equal    string    strings join               sql select decimal as col oid as col decimal as col oid as col decimal as col from defaultdb public as tab order by tab desc error types withstack withstack errutil withprefix withstack withstack errutil leaferror parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see same failure on other branches roachtest unoptimized query oracle disable rules all failed cc cockroachdb sql queries
0
141,220
11,404,482,006
IssuesEvent
2020-01-31 09:54:14
raiden-network/raiden
https://api.github.com/repos/raiden-network/raiden
closed
Remove **mode: managed** from all the scenarios
Component / Scenario Player Flag / Testing
## Abstract Copying from https://github.com/raiden-network/scenario-player/issues/360 Issue 360 of the scenario player each scenario has a `mode: managed` config in most or all of the scenarios. ``` nodes: mode: managed count: 4 raiden_version: local ``` ## Motivation Though this is pretty harmless since we do not use this feature any longer it is advised to remove it so that it doesnt create any confusion ## Specification Removal of `mode: managed` from all the scenarios ## Backwards Compatibility Since this is a harmless removal it does not need to be backward compatible.
1.0
Remove **mode: managed** from all the scenarios - ## Abstract Copying from https://github.com/raiden-network/scenario-player/issues/360 Issue 360 of the scenario player each scenario has a `mode: managed` config in most or all of the scenarios. ``` nodes: mode: managed count: 4 raiden_version: local ``` ## Motivation Though this is pretty harmless since we do not use this feature any longer it is advised to remove it so that it doesnt create any confusion ## Specification Removal of `mode: managed` from all the scenarios ## Backwards Compatibility Since this is a harmless removal it does not need to be backward compatible.
non_main
remove mode managed from all the scenarios abstract copying from issue of the scenario player each scenario has a mode managed config in most or all of the scenarios nodes mode managed count raiden version local motivation though this is pretty harmless since we do not use this feature any longer it is advised to remove it so that it doesnt create any confusion specification removal of mode managed from all the scenarios backwards compatibility since this is a harmless removal it does not need to be backward compatible
0
2,024
6,757,640,733
IssuesEvent
2017-10-24 11:36:14
Kristinita/Erics-Green-Room
https://api.github.com/repos/Kristinita/Erics-Green-Room
closed
[Feature request] Новые метаданные: npt и kns
enhancement need-maintainer
### 1. Запрос Неплохо было бы, если б добавили 2 новых типа метаданных: `*-npt` и `*-kns-`. ### 2. Аргументация #### 2.1. npt **Не путать**. Дополнительное указание, чтобы игрок не путался (см. [**тао**](http://kristinita.ru/Gingerinas/%D0%A2%D0%B0%D0%BE)) в понятиях. Особенно, если понятия, между которыми игрок путается, находятся не в одном и том же пакете → их не получится включить в 1 тао-блок. Примеры: ```markdown https://i.imgur.com/hmiOLxx.png Сестра Фрикса, которую спас вместе с братом летающий златорунный овен. Упала и погибла.@Гелла*-info-«Море», где погибла Гелла, стало называться Геллеспонтом*-npt-С Геллой — вампиршей из свиты Воланда, которая ходила в одном переднике ``` Одна Гелла — мифологический персонаж, другая — литературный. ```markdown Алчевск, 1931—61*Ворошиловск*-info-Алчевск — по фамилии мецената Алексея Алчевского, основателя Донецко-Юрьевского металлургического сообщества*-npt-С городом Ворошилов — бывшим названием Уссурийска и Ворошиловском — бывшим названием Ставрополя*-proof-27 ``` Один Ворошиловск в пакете, где переименования украинских городов; другой, а также город Ворошилов — в пакете с российскими городами. я не могу сделать межпакетные тао-блоки. #### 2.2. kns **Как ни странно**. Дополнительное указание, что факт, который может смущать игрока, на самом деле верный. Пример: ```markdown Истра, 1781—930*Воскресенск*-info-Город на месте слободы близ Воскресенского монастыря в Московской области. Переименован по названию реки.*-npt-С другим городом Воскресенск Московской области, где базируется клуб «Химик»*-kns-Воскресенск, где клуб «Химик», назван так в 1938 ``` *То есть*: не путать этот Воскресенск с другим Воскресенском, который был назван в 1938. Как ни странно, город получил название «Воскресенск» в 1938, при Сталине. ### 3. Примеры отображения ```markdown [10:05:33 PM] <RED> Вопрос №1 из 110: -------------------------------------------------------- Сестра Фрикса, которую спас вместе с братом летающий златорунный овен. Упала и погибла. -------------------------------------------------------- [10:05:39 PM] <RED> Скрыто: <орнитоптера_Королевы_Александры> гелла [10:05:42 PM] <орнитоптера_Королевы_Александры> гелла [10:05:42 PM] <RED> орнитоптера_Королевы_Александры - даёт правильный ответ [10:05:42 PM] <RED> Правильный ответ: "Гелла" [10:05:42 PM] <RED> Комментарий: «Море», где погибла Гелла, стало называться Геллеспонтом [10:05:42 PM] <RED> Не путать: С Геллой — вампиршей из свиты Воланда, которая ходила в одном переднике [10:05:42 PM] <RED> Счёт: орнитоптера_Королевы_Александры (1) ``` ```markdown [10:05:33 PM] <RED> Вопрос №2 из 110: -------------------------------------------------------- Алчевск, 1931—61 -------------------------------------------------------- [10:05:42 PM] <орнитоптера_Королевы_Александры> Ворошиловск [10:05:42 PM] <RED> орнитоптера_Королевы_Александры - даёт правильный ответ [10:05:42 PM] <RED> Правильный ответ: "Ворошиловск" [10:05:42 PM] <RED> Комментарий: Алчевск — по фамилии мецената Алексея Алчевского, основателя Донецко-Юрьевского металлургического сообщества [10:05:42 PM] <RED> Не путать: С городом Ворошилов — бывшим названием Уссурийска и Ворошиловском — бывшим названием Ставрополя [10:05:42 PM] <RED> Счёт: орнитоптера_Королевы_Александры (2) ``` ```markdown [10:05:33 PM] <RED> Вопрос №3 из 110: -------------------------------------------------------- Истра, 1781—930 -------------------------------------------------------- [10:05:42 PM] <орнитоптера_Королевы_Александры> Воскресенск [10:05:42 PM] <RED> орнитоптера_Королевы_Александры - даёт правильный ответ [10:05:42 PM] <RED> Правильный ответ: "Воскресенск" [10:05:42 PM] <RED> Комментарий: Город на месте слободы близ Воскресенского монастыря в Московской области. Переименован по названию реки. [10:05:42 PM] <RED> Не путать: С другим городом Воскресенск Московской области, где базируется клуб «Химик» [10:05:42 PM] <RED> Как ни странно: Воскресенск, где клуб «Химик», назван так в 1938 [10:05:42 PM] <RED> Счёт: орнитоптера_Королевы_Александры (3) ``` Спасибо.
True
[Feature request] Новые метаданные: npt и kns - ### 1. Запрос Неплохо было бы, если б добавили 2 новых типа метаданных: `*-npt` и `*-kns-`. ### 2. Аргументация #### 2.1. npt **Не путать**. Дополнительное указание, чтобы игрок не путался (см. [**тао**](http://kristinita.ru/Gingerinas/%D0%A2%D0%B0%D0%BE)) в понятиях. Особенно, если понятия, между которыми игрок путается, находятся не в одном и том же пакете → их не получится включить в 1 тао-блок. Примеры: ```markdown https://i.imgur.com/hmiOLxx.png Сестра Фрикса, которую спас вместе с братом летающий златорунный овен. Упала и погибла.@Гелла*-info-«Море», где погибла Гелла, стало называться Геллеспонтом*-npt-С Геллой — вампиршей из свиты Воланда, которая ходила в одном переднике ``` Одна Гелла — мифологический персонаж, другая — литературный. ```markdown Алчевск, 1931—61*Ворошиловск*-info-Алчевск — по фамилии мецената Алексея Алчевского, основателя Донецко-Юрьевского металлургического сообщества*-npt-С городом Ворошилов — бывшим названием Уссурийска и Ворошиловском — бывшим названием Ставрополя*-proof-27 ``` Один Ворошиловск в пакете, где переименования украинских городов; другой, а также город Ворошилов — в пакете с российскими городами. я не могу сделать межпакетные тао-блоки. #### 2.2. kns **Как ни странно**. Дополнительное указание, что факт, который может смущать игрока, на самом деле верный. Пример: ```markdown Истра, 1781—930*Воскресенск*-info-Город на месте слободы близ Воскресенского монастыря в Московской области. Переименован по названию реки.*-npt-С другим городом Воскресенск Московской области, где базируется клуб «Химик»*-kns-Воскресенск, где клуб «Химик», назван так в 1938 ``` *То есть*: не путать этот Воскресенск с другим Воскресенском, который был назван в 1938. Как ни странно, город получил название «Воскресенск» в 1938, при Сталине. ### 3. Примеры отображения ```markdown [10:05:33 PM] <RED> Вопрос №1 из 110: -------------------------------------------------------- Сестра Фрикса, которую спас вместе с братом летающий златорунный овен. Упала и погибла. -------------------------------------------------------- [10:05:39 PM] <RED> Скрыто: <орнитоптера_Королевы_Александры> гелла [10:05:42 PM] <орнитоптера_Королевы_Александры> гелла [10:05:42 PM] <RED> орнитоптера_Королевы_Александры - даёт правильный ответ [10:05:42 PM] <RED> Правильный ответ: "Гелла" [10:05:42 PM] <RED> Комментарий: «Море», где погибла Гелла, стало называться Геллеспонтом [10:05:42 PM] <RED> Не путать: С Геллой — вампиршей из свиты Воланда, которая ходила в одном переднике [10:05:42 PM] <RED> Счёт: орнитоптера_Королевы_Александры (1) ``` ```markdown [10:05:33 PM] <RED> Вопрос №2 из 110: -------------------------------------------------------- Алчевск, 1931—61 -------------------------------------------------------- [10:05:42 PM] <орнитоптера_Королевы_Александры> Ворошиловск [10:05:42 PM] <RED> орнитоптера_Королевы_Александры - даёт правильный ответ [10:05:42 PM] <RED> Правильный ответ: "Ворошиловск" [10:05:42 PM] <RED> Комментарий: Алчевск — по фамилии мецената Алексея Алчевского, основателя Донецко-Юрьевского металлургического сообщества [10:05:42 PM] <RED> Не путать: С городом Ворошилов — бывшим названием Уссурийска и Ворошиловском — бывшим названием Ставрополя [10:05:42 PM] <RED> Счёт: орнитоптера_Королевы_Александры (2) ``` ```markdown [10:05:33 PM] <RED> Вопрос №3 из 110: -------------------------------------------------------- Истра, 1781—930 -------------------------------------------------------- [10:05:42 PM] <орнитоптера_Королевы_Александры> Воскресенск [10:05:42 PM] <RED> орнитоптера_Королевы_Александры - даёт правильный ответ [10:05:42 PM] <RED> Правильный ответ: "Воскресенск" [10:05:42 PM] <RED> Комментарий: Город на месте слободы близ Воскресенского монастыря в Московской области. Переименован по названию реки. [10:05:42 PM] <RED> Не путать: С другим городом Воскресенск Московской области, где базируется клуб «Химик» [10:05:42 PM] <RED> Как ни странно: Воскресенск, где клуб «Химик», назван так в 1938 [10:05:42 PM] <RED> Счёт: орнитоптера_Королевы_Александры (3) ``` Спасибо.
main
новые метаданные npt и kns запрос неплохо было бы если б добавили новых типа метаданных npt и kns аргументация npt не путать дополнительное указание чтобы игрок не путался см в понятиях особенно если понятия между которыми игрок путается находятся не в одном и том же пакете → их не получится включить в тао блок примеры markdown сестра фрикса которую спас вместе с братом летающий златорунный овен упала и погибла гелла info «море» где погибла гелла стало называться геллеспонтом npt с геллой — вампиршей из свиты воланда которая ходила в одном переднике одна гелла — мифологический персонаж другая — литературный markdown алчевск — ворошиловск info алчевск — по фамилии мецената алексея алчевского основателя донецко юрьевского металлургического сообщества npt с городом ворошилов — бывшим названием уссурийска и ворошиловском — бывшим названием ставрополя proof один ворошиловск в пакете где переименования украинских городов другой а также город ворошилов — в пакете с российскими городами я не могу сделать межпакетные тао блоки kns как ни странно дополнительное указание что факт который может смущать игрока на самом деле верный пример markdown истра — воскресенск info город на месте слободы близ воскресенского монастыря в московской области переименован по названию реки npt с другим городом воскресенск московской области где базируется клуб «химик» kns воскресенск где клуб «химик» назван так в то есть не путать этот воскресенск с другим воскресенском который был назван в как ни странно город получил название «воскресенск» в при сталине примеры отображения markdown вопрос № из сестра фрикса которую спас вместе с братом летающий златорунный овен упала и погибла скрыто гелла гелла орнитоптера королевы александры даёт правильный ответ правильный ответ гелла комментарий «море» где погибла гелла стало называться геллеспонтом не путать с геллой — вампиршей из свиты воланда которая ходила в одном переднике счёт орнитоптера королевы александры markdown вопрос № из алчевск — ворошиловск орнитоптера королевы александры даёт правильный ответ правильный ответ ворошиловск комментарий алчевск — по фамилии мецената алексея алчевского основателя донецко юрьевского металлургического сообщества не путать с городом ворошилов — бывшим названием уссурийска и ворошиловском — бывшим названием ставрополя счёт орнитоптера королевы александры markdown вопрос № из истра — воскресенск орнитоптера королевы александры даёт правильный ответ правильный ответ воскресенск комментарий город на месте слободы близ воскресенского монастыря в московской области переименован по названию реки не путать с другим городом воскресенск московской области где базируется клуб «химик» как ни странно воскресенск где клуб «химик» назван так в счёт орнитоптера королевы александры спасибо
1
99,624
16,449,124,312
IssuesEvent
2021-05-21 01:08:09
SmartBear/readyapi-swagger-assertion-plugin
https://api.github.com/repos/SmartBear/readyapi-swagger-assertion-plugin
opened
CVE-2021-29505 (Medium) detected in xstream-1.3.1.jar
security vulnerability
## CVE-2021-29505 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.3.1.jar</b></p></summary> <p></p> <p>Path to dependency file: readyapi-swagger-assertion-plugin/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/thoughtworks/xstream/1.3.1/xstream-1.3.1.jar</p> <p> Dependency Hierarchy: - ready-api-soapui-pro-1.7.0.jar (Root Library) - ready-api-soapui-1.7.0.jar - :x: **xstream-1.3.1.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> XStream is vulnerable to a Remote Command Execution attack before version 1.4.17. The vulnerability may allow a remote attacker has sufficient rights to execute commands of the host only by manipulating the processed input stream. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. <p>Publish Date: 2021-03-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29505>CVE-2021-29505</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-7chv-rrw6-w6fc">https://github.com/advisories/GHSA-7chv-rrw6-w6fc</a></p> <p>Release Date: 2021-03-31</p> <p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.17</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.3.1","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:1.7.0;com.smartbear:ready-api-soapui:1.7.0;com.thoughtworks.xstream:xstream:1.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.17"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-29505","vulnerabilityDetails":"XStream is vulnerable to a Remote Command Execution attack before version 1.4.17. The vulnerability may allow a remote attacker has sufficient rights to execute commands of the host only by manipulating the processed input stream. No user is affected, who followed the recommendation to setup XStream\u0027s security framework with a whitelist limited to the minimal required types.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29505","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-29505 (Medium) detected in xstream-1.3.1.jar - ## CVE-2021-29505 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.3.1.jar</b></p></summary> <p></p> <p>Path to dependency file: readyapi-swagger-assertion-plugin/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/thoughtworks/xstream/1.3.1/xstream-1.3.1.jar</p> <p> Dependency Hierarchy: - ready-api-soapui-pro-1.7.0.jar (Root Library) - ready-api-soapui-1.7.0.jar - :x: **xstream-1.3.1.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> XStream is vulnerable to a Remote Command Execution attack before version 1.4.17. The vulnerability may allow a remote attacker has sufficient rights to execute commands of the host only by manipulating the processed input stream. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. <p>Publish Date: 2021-03-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29505>CVE-2021-29505</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-7chv-rrw6-w6fc">https://github.com/advisories/GHSA-7chv-rrw6-w6fc</a></p> <p>Release Date: 2021-03-31</p> <p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.17</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.thoughtworks.xstream","packageName":"xstream","packageVersion":"1.3.1","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:1.7.0;com.smartbear:ready-api-soapui:1.7.0;com.thoughtworks.xstream:xstream:1.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.thoughtworks.xstream:xstream:1.4.17"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-29505","vulnerabilityDetails":"XStream is vulnerable to a Remote Command Execution attack before version 1.4.17. The vulnerability may allow a remote attacker has sufficient rights to execute commands of the host only by manipulating the processed input stream. No user is affected, who followed the recommendation to setup XStream\u0027s security framework with a whitelist limited to the minimal required types.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29505","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_main
cve medium detected in xstream jar cve medium severity vulnerability vulnerable library xstream jar path to dependency file readyapi swagger assertion plugin pom xml path to vulnerable library home wss scanner repository thoughtworks xstream xstream jar dependency hierarchy ready api soapui pro jar root library ready api soapui jar x xstream jar vulnerable library found in base branch master vulnerability details xstream is vulnerable to a remote command execution attack before version the vulnerability may allow a remote attacker has sufficient rights to execute commands of the host only by manipulating the processed input stream no user is affected who followed the recommendation to setup xstream s security framework with a whitelist limited to the minimal required types publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com thoughtworks xstream xstream isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree com smartbear ready api soapui pro com smartbear ready api soapui com thoughtworks xstream xstream isminimumfixversionavailable true minimumfixversion com thoughtworks xstream xstream basebranches vulnerabilityidentifier cve vulnerabilitydetails xstream is vulnerable to a remote command execution attack before version the vulnerability may allow a remote attacker has sufficient rights to execute commands of the host only by manipulating the processed input stream no user is affected who followed the recommendation to setup xstream security framework with a whitelist limited to the minimal required types vulnerabilityurl
0
489,747
14,111,909,253
IssuesEvent
2020-11-07 02:18:01
RogueLogix/BiggerSeries
https://api.github.com/repos/RogueLogix/BiggerSeries
closed
Missing side configuration for Cyanite Reprocessor
Priority: Low UI
Currently, there is no ability to reassign sides for the Cyanite Reprocessor (as in, set one side as an input, and another as an output).
1.0
Missing side configuration for Cyanite Reprocessor - Currently, there is no ability to reassign sides for the Cyanite Reprocessor (as in, set one side as an input, and another as an output).
non_main
missing side configuration for cyanite reprocessor currently there is no ability to reassign sides for the cyanite reprocessor as in set one side as an input and another as an output
0
305,495
26,390,750,610
IssuesEvent
2023-01-12 15:30:49
CSOIreland/PxStat
https://api.github.com/repos/CSOIreland/PxStat
closed
[BUG] Spurious data point being picked up if the semi colon is on the next line after the DATA in a px file
bug released tested fixed
Additional blank data point being picked up if the semi colon is on the next line after the DATA in a px file, e.g. DATA= 77.8 91.7 93.6 104.0 114.4 117.3 108.8 106.7 113.5 107.7 112.1 107.7 147.6 143.1 164.3 159.8 151.1 157.7 ;
1.0
[BUG] Spurious data point being picked up if the semi colon is on the next line after the DATA in a px file - Additional blank data point being picked up if the semi colon is on the next line after the DATA in a px file, e.g. DATA= 77.8 91.7 93.6 104.0 114.4 117.3 108.8 106.7 113.5 107.7 112.1 107.7 147.6 143.1 164.3 159.8 151.1 157.7 ;
non_main
spurious data point being picked up if the semi colon is on the next line after the data in a px file additional blank data point being picked up if the semi colon is on the next line after the data in a px file e g data
0
148,245
23,331,363,191
IssuesEvent
2022-08-09 05:43:08
TheSuperHackers/GeneralsGamePatch
https://api.github.com/repos/TheSuperHackers/GeneralsGamePatch
opened
China Satellite Hack One is not useful
Design Controversial Minor
China Satellite Hack One is not useful. It spies permanently on Command Centers only. This discovery does not provide any significant intel on opponent. China Satellite Hack II spies on all enemy structures and units for a certain duration. Related * #708 * #837 ### Proposal 1 Make Satellite Hack One see Supply Depots too, so it combines better with Cash Hack Generals Power. ### Proposal 2 Turn Satellite Hack One into Spy Power like USA Strategy Scan and have it reveal all structures only for a certain duration. ### Proposal 3 Keep all permanent Satellite Hack One benefits after Satellite Hack Two was researched. Can combine with Proposal 1.
1.0
China Satellite Hack One is not useful - China Satellite Hack One is not useful. It spies permanently on Command Centers only. This discovery does not provide any significant intel on opponent. China Satellite Hack II spies on all enemy structures and units for a certain duration. Related * #708 * #837 ### Proposal 1 Make Satellite Hack One see Supply Depots too, so it combines better with Cash Hack Generals Power. ### Proposal 2 Turn Satellite Hack One into Spy Power like USA Strategy Scan and have it reveal all structures only for a certain duration. ### Proposal 3 Keep all permanent Satellite Hack One benefits after Satellite Hack Two was researched. Can combine with Proposal 1.
non_main
china satellite hack one is not useful china satellite hack one is not useful it spies permanently on command centers only this discovery does not provide any significant intel on opponent china satellite hack ii spies on all enemy structures and units for a certain duration related proposal make satellite hack one see supply depots too so it combines better with cash hack generals power proposal turn satellite hack one into spy power like usa strategy scan and have it reveal all structures only for a certain duration proposal keep all permanent satellite hack one benefits after satellite hack two was researched can combine with proposal
0
746,537
26,034,092,402
IssuesEvent
2022-12-22 01:50:10
battlecode/galaxy
https://api.github.com/repos/battlecode/galaxy
closed
Easy deploys for client (via shell script or GH CI)
type: feature priority: p1 critical module: devops
calling this "frontend" since the deployment process is more in the push-a-website thing Also to make sure that we don't lose sight of it til we find a better place critical cuz bad deploys make life much harder during competition
1.0
Easy deploys for client (via shell script or GH CI) - calling this "frontend" since the deployment process is more in the push-a-website thing Also to make sure that we don't lose sight of it til we find a better place critical cuz bad deploys make life much harder during competition
non_main
easy deploys for client via shell script or gh ci calling this frontend since the deployment process is more in the push a website thing also to make sure that we don t lose sight of it til we find a better place critical cuz bad deploys make life much harder during competition
0
4,269
21,386,952,621
IssuesEvent
2022-04-21 00:21:46
Lissy93/dashy
https://api.github.com/repos/Lissy93/dashy
closed
[QUESTION] How to use a named docker volume
🤷‍♂️ Question 👤 Awaiting Maintainer Response
### Question Hi, I'm trying to deploy Dashy on my docker cluster, however for that I need to be able to use a named volume for persistent storage. I saw in your documentation an exemple of docker compose but with a direct file mapping for persistence. I tried this ``` dashy: image: lissy93/dashy container_name: BattleStation_Dashy # Pass in your config file below, by specifying the path on your host machine volumes: - persistent-Dashy:/public/ # Set any environmental variables environment: - NODE_ENV=production # Specify your user ID and group ID. You can find this by running `id -u` and `id -g` - UID=500 - GID=500 # Specify restart policy restart: unless-stopped # Configure healthchecks healthcheck: test: ['CMD', 'node', '/app/services/healthcheck'] interval: 1m30s timeout: 10s retries: 3 start_period: 40s labels: - "traefik.enable=true" - "traefik.http.routers.battleStation_dashy_web.entrypoints=web" - "traefik.http.routers.battleStation_dashy_web.rule=Host(`XXXXXXXXX`)" - "traefik.http.routers.battleStation_dashy_websecure.entrypoints=websecure" - "traefik.http.routers.battleStation_dashy_websecure.tls=true" - "traefik.http.routers.battleStation_dashy_websecure.rule=Host(`XXXXXXXX`)" ``` Is there a way to specify the path of the config storage ? maybe with and ENV variable ? PS : I reposted the question as I did't see a way to reopen the previous one auto closed by your bot :-) ### Category Setup and Deployment ### Please tick the boxes - [X] You are using a [supported](https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions) version of Dashy (check the first two digits of the version number) - [X] You've checked that this [question hasn't already been raised](https://github.com/Lissy93/dashy/issues?q=is%3Aissue) - [X] You've checked the [docs](https://github.com/Lissy93/dashy/tree/master/docs#readme) and [troubleshooting](https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting) guide - [X] You agree to the [code of conduct](https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct)
True
[QUESTION] How to use a named docker volume - ### Question Hi, I'm trying to deploy Dashy on my docker cluster, however for that I need to be able to use a named volume for persistent storage. I saw in your documentation an exemple of docker compose but with a direct file mapping for persistence. I tried this ``` dashy: image: lissy93/dashy container_name: BattleStation_Dashy # Pass in your config file below, by specifying the path on your host machine volumes: - persistent-Dashy:/public/ # Set any environmental variables environment: - NODE_ENV=production # Specify your user ID and group ID. You can find this by running `id -u` and `id -g` - UID=500 - GID=500 # Specify restart policy restart: unless-stopped # Configure healthchecks healthcheck: test: ['CMD', 'node', '/app/services/healthcheck'] interval: 1m30s timeout: 10s retries: 3 start_period: 40s labels: - "traefik.enable=true" - "traefik.http.routers.battleStation_dashy_web.entrypoints=web" - "traefik.http.routers.battleStation_dashy_web.rule=Host(`XXXXXXXXX`)" - "traefik.http.routers.battleStation_dashy_websecure.entrypoints=websecure" - "traefik.http.routers.battleStation_dashy_websecure.tls=true" - "traefik.http.routers.battleStation_dashy_websecure.rule=Host(`XXXXXXXX`)" ``` Is there a way to specify the path of the config storage ? maybe with and ENV variable ? PS : I reposted the question as I did't see a way to reopen the previous one auto closed by your bot :-) ### Category Setup and Deployment ### Please tick the boxes - [X] You are using a [supported](https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions) version of Dashy (check the first two digits of the version number) - [X] You've checked that this [question hasn't already been raised](https://github.com/Lissy93/dashy/issues?q=is%3Aissue) - [X] You've checked the [docs](https://github.com/Lissy93/dashy/tree/master/docs#readme) and [troubleshooting](https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting) guide - [X] You agree to the [code of conduct](https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct)
main
how to use a named docker volume question hi i m trying to deploy dashy on my docker cluster however for that i need to be able to use a named volume for persistent storage i saw in your documentation an exemple of docker compose but with a direct file mapping for persistence i tried this dashy image dashy container name battlestation dashy pass in your config file below by specifying the path on your host machine volumes persistent dashy public set any environmental variables environment node env production specify your user id and group id you can find this by running id u and id g uid gid specify restart policy restart unless stopped configure healthchecks healthcheck test interval timeout retries start period labels traefik enable true traefik http routers battlestation dashy web entrypoints web traefik http routers battlestation dashy web rule host xxxxxxxxx traefik http routers battlestation dashy websecure entrypoints websecure traefik http routers battlestation dashy websecure tls true traefik http routers battlestation dashy websecure rule host xxxxxxxx is there a way to specify the path of the config storage maybe with and env variable ps i reposted the question as i did t see a way to reopen the previous one auto closed by your bot category setup and deployment please tick the boxes you are using a version of dashy check the first two digits of the version number you ve checked that this you ve checked the and guide you agree to the
1
3,136
12,041,259,054
IssuesEvent
2020-04-14 08:31:57
travisbrown/dhallj
https://api.github.com/repos/travisbrown/dhallj
opened
Clean up escaping
maintainability
There are currently many different string encodings with different escapings, with lots of transition points between them. These transitions are currently handled in a fairly ad-hoc way, with escaping done at each of them as needed to get tests to pass. I don't know of any specific bugs in this respect at the moment, but I'm sure there are some in there, and anyway the current situation won't be fun to maintain in the longer term. I need to work through exactly what needs to be escaped where, and to clean up the transitions. This should be relatively easy given the tests we have now.
True
Clean up escaping - There are currently many different string encodings with different escapings, with lots of transition points between them. These transitions are currently handled in a fairly ad-hoc way, with escaping done at each of them as needed to get tests to pass. I don't know of any specific bugs in this respect at the moment, but I'm sure there are some in there, and anyway the current situation won't be fun to maintain in the longer term. I need to work through exactly what needs to be escaped where, and to clean up the transitions. This should be relatively easy given the tests we have now.
main
clean up escaping there are currently many different string encodings with different escapings with lots of transition points between them these transitions are currently handled in a fairly ad hoc way with escaping done at each of them as needed to get tests to pass i don t know of any specific bugs in this respect at the moment but i m sure there are some in there and anyway the current situation won t be fun to maintain in the longer term i need to work through exactly what needs to be escaped where and to clean up the transitions this should be relatively easy given the tests we have now
1
1,642
6,572,667,365
IssuesEvent
2017-09-11 04:14:21
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
route53_zone fails if the private zone configured in the play already exists.
affects_2.2 aws bug_report cloud waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME route53_zone ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT $ cat /etc/redhat-release CentOS Linux release 7.2.1511 (Core) $ rpm -q ansible ansible-2.2.0.0-3.el7.noarch $ rpm -qa|grep boto python2-botocore-1.4.58-1.el7.noarch python2-boto3-1.4.0-1.el7.noarch python2-boto-2.43.0-1.el7.noarch ##### SUMMARY route53_zone fails if the private zone configured in the play already exists. If the zone does not exist, then it creates the zone successfully. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` $ cat route53_zone.yml - hosts: localhost tasks: - name: Create private DNS zone route53_zone: zone: example.com vpc_id: vpc-NNNNNN vpc_region: us-east-1 state: present register: zone - debug: var=zone $ ansible-playbook route53_zone.yml [WARNING]: provided hosts list is empty, only localhost is available PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* ok: [localhost] TASK [Create private DNS zone] ************************************************* changed: [localhost] TASK [debug] ******************************************************************* ok: [localhost] => { "zone": { "changed": true, "set": { "comment": "", "name": "example.com.", "private_zone": true, "vpc_id": "vpc-NNNNNN", "vpc_region": "us-east-1", "zone_id": "XYZXYZXYZXYZXYZ" } } } PLAY RECAP ********************************************************************* localhost : ok=3 changed=1 unreachable=0 failed=0 $ ansible-playbook route53_zone.yml [WARNING]: provided hosts list is empty, only localhost is available PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* ok: [localhost] TASK [Create private DNS zone] ************************************************* An exception occurred during task execution. To see the full traceback, use -vvv. The error was: TypeError: list indices must be integers, not str fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_qr2D8d/ansible_module_route53_zone.py\", line 223, in <module>\n main()\n File \"/tmp/ansible_qr2D8d/ansible_module_route53_zone.py\", line 190, in main\n vpc_details = details['GetHostedZoneResponse']['VPCs']['VPC']\nTypeError: list indices must be integers, not str\n", "module_stdout": "", "msg": "MODULE FAILURE"} to retry, use: --limit @/home/acharrett/route53_zone.retry PLAY RECAP ********************************************************************* localhost : ok=1 changed=0 unreachable=0 failed=1 ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS On the second run, I would expect Ansible to return changed=0 as the zone already existed. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> ``` File "/tmp/ansible_ShBpJS/ansible_module_route53_zone.py", line 190, in main vpc_details = details['GetHostedZoneResponse']['VPCs']['VPC'] TypeError: list indices must be integers, not str ``` <!--- Paste verbatim command output between quotes below --> ``` $ ansible-playbook route53_zone.yml -vvvv Using /etc/ansible/ansible.cfg as config file [WARNING]: provided hosts list is empty, only localhost is available Loading callback plugin default of type stdout, v2.0 from /usr/lib/python2.7/site-packages/ansible/plugins/callback/__init__.pyc PLAYBOOK: route53_zone.yml ********************************************************** 1 plays in route53_zone.yml PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* Using module file /usr/lib/python2.7/site-packages/ansible/modules/core/system/setup.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: acharrett <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479389391.9-221283713905247 `" && echo ansible-tmp-1479389391.9-221283713905247="` echo $HOME/.ansible/tmp/ansible-tmp-1479389391.9-221283713905247 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpIxIEz5 TO /home/acharrett/.ansible/tmp/ansible-tmp-1479389391.9-221283713905247/setup.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/acharrett/.ansible/tmp/ansible-tmp-1479389391.9-221283713905247/ /home/acharrett/.ansible/tmp/ansible-tmp-1479389391.9-221283713905247/setup.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c '/usr/bin/python2 /home/acharrett/.ansible/tmp/ansible-tmp-1479389391.9-221283713905247/setup.py; rm -rf "/home/acharrett/.ansible/tmp/ansible-tmp-1479389391.9-221283713905247/" > /dev/null 2>&1 && sleep 0' ok: [localhost] TASK [Create private DNS zone] ************************************************* task path: /home/acharrett/route53_zone.yml:4 Using module file /usr/lib/python2.7/site-packages/ansible/modules/extras/cloud/amazon/route53_zone.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: acharrett <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479389392.16-161591944519754 `" && echo ansible-tmp-1479389392.16-161591944519754="` echo $HOME/.ansible/tmp/ansible-tmp-1479389392.16-161591944519754 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpvxNMWn TO /home/acharrett/.ansible/tmp/ansible-tmp-1479389392.16-161591944519754/route53_zone.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/acharrett/.ansible/tmp/ansible-tmp-1479389392.16-161591944519754/ /home/acharrett/.ansible/tmp/ansible-tmp-1479389392.16-161591944519754/route53_zone.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c '/usr/bin/python2 /home/acharrett/.ansible/tmp/ansible-tmp-1479389392.16-161591944519754/route53_zone.py; rm -rf "/home/acharrett/.ansible/tmp/ansible-tmp-1479389392.16-161591944519754/" > /dev/null 2>&1 && sleep 0' An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/tmp/ansible_ShBpJS/ansible_module_route53_zone.py", line 223, in <module> main() File "/tmp/ansible_ShBpJS/ansible_module_route53_zone.py", line 190, in main vpc_details = details['GetHostedZoneResponse']['VPCs']['VPC'] TypeError: list indices must be integers, not str fatal: [localhost]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_name": "route53_zone" }, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_ShBpJS/ansible_module_route53_zone.py\", line 223, in <module>\n main()\n File \"/tmp/ansible_ShBpJS/ansible_module_route53_zone.py\", line 190, in main\n vpc_details = details['GetHostedZoneResponse']['VPCs']['VPC']\nTypeError: list indices must be integers, not str\n", "module_stdout": "", "msg": "MODULE FAILURE" } to retry, use: --limit @/home/acharrett/route53.retry PLAY RECAP ********************************************************************* localhost : ok=1 changed=0 unreachable=0 failed=1 ``` ##### FURTHER INFORMATION If I replace these lines (starting line 190 of ansible/modules/extras/cloud/amazon/route53_zone.py): ``` vpc_details = details['GetHostedZoneResponse']['VPCs']['VPC'] current_vpc_id = vpc_details['VPCId'] current_vpc_region = vpc_details['VPCRegion'] ``` with these lines ``` vpc_details = details.GetHostedZoneResponse.VPCs[0] current_vpc_id = vpc_details.VPCId current_vpc_region = vpc_details.VPCRegion ``` Then the module works as expected
True
route53_zone fails if the private zone configured in the play already exists. - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME route53_zone ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT $ cat /etc/redhat-release CentOS Linux release 7.2.1511 (Core) $ rpm -q ansible ansible-2.2.0.0-3.el7.noarch $ rpm -qa|grep boto python2-botocore-1.4.58-1.el7.noarch python2-boto3-1.4.0-1.el7.noarch python2-boto-2.43.0-1.el7.noarch ##### SUMMARY route53_zone fails if the private zone configured in the play already exists. If the zone does not exist, then it creates the zone successfully. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` $ cat route53_zone.yml - hosts: localhost tasks: - name: Create private DNS zone route53_zone: zone: example.com vpc_id: vpc-NNNNNN vpc_region: us-east-1 state: present register: zone - debug: var=zone $ ansible-playbook route53_zone.yml [WARNING]: provided hosts list is empty, only localhost is available PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* ok: [localhost] TASK [Create private DNS zone] ************************************************* changed: [localhost] TASK [debug] ******************************************************************* ok: [localhost] => { "zone": { "changed": true, "set": { "comment": "", "name": "example.com.", "private_zone": true, "vpc_id": "vpc-NNNNNN", "vpc_region": "us-east-1", "zone_id": "XYZXYZXYZXYZXYZ" } } } PLAY RECAP ********************************************************************* localhost : ok=3 changed=1 unreachable=0 failed=0 $ ansible-playbook route53_zone.yml [WARNING]: provided hosts list is empty, only localhost is available PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* ok: [localhost] TASK [Create private DNS zone] ************************************************* An exception occurred during task execution. To see the full traceback, use -vvv. The error was: TypeError: list indices must be integers, not str fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_qr2D8d/ansible_module_route53_zone.py\", line 223, in <module>\n main()\n File \"/tmp/ansible_qr2D8d/ansible_module_route53_zone.py\", line 190, in main\n vpc_details = details['GetHostedZoneResponse']['VPCs']['VPC']\nTypeError: list indices must be integers, not str\n", "module_stdout": "", "msg": "MODULE FAILURE"} to retry, use: --limit @/home/acharrett/route53_zone.retry PLAY RECAP ********************************************************************* localhost : ok=1 changed=0 unreachable=0 failed=1 ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS On the second run, I would expect Ansible to return changed=0 as the zone already existed. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> ``` File "/tmp/ansible_ShBpJS/ansible_module_route53_zone.py", line 190, in main vpc_details = details['GetHostedZoneResponse']['VPCs']['VPC'] TypeError: list indices must be integers, not str ``` <!--- Paste verbatim command output between quotes below --> ``` $ ansible-playbook route53_zone.yml -vvvv Using /etc/ansible/ansible.cfg as config file [WARNING]: provided hosts list is empty, only localhost is available Loading callback plugin default of type stdout, v2.0 from /usr/lib/python2.7/site-packages/ansible/plugins/callback/__init__.pyc PLAYBOOK: route53_zone.yml ********************************************************** 1 plays in route53_zone.yml PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* Using module file /usr/lib/python2.7/site-packages/ansible/modules/core/system/setup.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: acharrett <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479389391.9-221283713905247 `" && echo ansible-tmp-1479389391.9-221283713905247="` echo $HOME/.ansible/tmp/ansible-tmp-1479389391.9-221283713905247 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpIxIEz5 TO /home/acharrett/.ansible/tmp/ansible-tmp-1479389391.9-221283713905247/setup.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/acharrett/.ansible/tmp/ansible-tmp-1479389391.9-221283713905247/ /home/acharrett/.ansible/tmp/ansible-tmp-1479389391.9-221283713905247/setup.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c '/usr/bin/python2 /home/acharrett/.ansible/tmp/ansible-tmp-1479389391.9-221283713905247/setup.py; rm -rf "/home/acharrett/.ansible/tmp/ansible-tmp-1479389391.9-221283713905247/" > /dev/null 2>&1 && sleep 0' ok: [localhost] TASK [Create private DNS zone] ************************************************* task path: /home/acharrett/route53_zone.yml:4 Using module file /usr/lib/python2.7/site-packages/ansible/modules/extras/cloud/amazon/route53_zone.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: acharrett <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479389392.16-161591944519754 `" && echo ansible-tmp-1479389392.16-161591944519754="` echo $HOME/.ansible/tmp/ansible-tmp-1479389392.16-161591944519754 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpvxNMWn TO /home/acharrett/.ansible/tmp/ansible-tmp-1479389392.16-161591944519754/route53_zone.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/acharrett/.ansible/tmp/ansible-tmp-1479389392.16-161591944519754/ /home/acharrett/.ansible/tmp/ansible-tmp-1479389392.16-161591944519754/route53_zone.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c '/usr/bin/python2 /home/acharrett/.ansible/tmp/ansible-tmp-1479389392.16-161591944519754/route53_zone.py; rm -rf "/home/acharrett/.ansible/tmp/ansible-tmp-1479389392.16-161591944519754/" > /dev/null 2>&1 && sleep 0' An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/tmp/ansible_ShBpJS/ansible_module_route53_zone.py", line 223, in <module> main() File "/tmp/ansible_ShBpJS/ansible_module_route53_zone.py", line 190, in main vpc_details = details['GetHostedZoneResponse']['VPCs']['VPC'] TypeError: list indices must be integers, not str fatal: [localhost]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_name": "route53_zone" }, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_ShBpJS/ansible_module_route53_zone.py\", line 223, in <module>\n main()\n File \"/tmp/ansible_ShBpJS/ansible_module_route53_zone.py\", line 190, in main\n vpc_details = details['GetHostedZoneResponse']['VPCs']['VPC']\nTypeError: list indices must be integers, not str\n", "module_stdout": "", "msg": "MODULE FAILURE" } to retry, use: --limit @/home/acharrett/route53.retry PLAY RECAP ********************************************************************* localhost : ok=1 changed=0 unreachable=0 failed=1 ``` ##### FURTHER INFORMATION If I replace these lines (starting line 190 of ansible/modules/extras/cloud/amazon/route53_zone.py): ``` vpc_details = details['GetHostedZoneResponse']['VPCs']['VPC'] current_vpc_id = vpc_details['VPCId'] current_vpc_region = vpc_details['VPCRegion'] ``` with these lines ``` vpc_details = details.GetHostedZoneResponse.VPCs[0] current_vpc_id = vpc_details.VPCId current_vpc_region = vpc_details.VPCRegion ``` Then the module works as expected
main
zone fails if the private zone configured in the play already exists issue type bug report component name zone ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment cat etc redhat release centos linux release core rpm q ansible ansible noarch rpm qa grep boto botocore noarch noarch boto noarch summary zone fails if the private zone configured in the play already exists if the zone does not exist then it creates the zone successfully steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used cat zone yml hosts localhost tasks name create private dns zone zone zone example com vpc id vpc nnnnnn vpc region us east state present register zone debug var zone ansible playbook zone yml provided hosts list is empty only localhost is available play task ok task changed task ok zone changed true set comment name example com private zone true vpc id vpc nnnnnn vpc region us east zone id xyzxyzxyzxyzxyz play recap localhost ok changed unreachable failed ansible playbook zone yml provided hosts list is empty only localhost is available play task ok task an exception occurred during task execution to see the full traceback use vvv the error was typeerror list indices must be integers not str fatal failed changed false failed true module stderr traceback most recent call last n file tmp ansible ansible module zone py line in n main n file tmp ansible ansible module zone py line in main n vpc details details ntypeerror list indices must be integers not str n module stdout msg module failure to retry use limit home acharrett zone retry play recap localhost ok changed unreachable failed expected results on the second run i would expect ansible to return changed as the zone already existed actual results file tmp ansible shbpjs ansible module zone py line in main vpc details details typeerror list indices must be integers not str ansible playbook zone yml vvvv using etc ansible ansible cfg as config file provided hosts list is empty only localhost is available loading callback plugin default of type stdout from usr lib site packages ansible plugins callback init pyc playbook zone yml plays in zone yml play task using module file usr lib site packages ansible modules core system setup py establish local connection for user acharrett exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp to home acharrett ansible tmp ansible tmp setup py exec bin sh c chmod u x home acharrett ansible tmp ansible tmp home acharrett ansible tmp ansible tmp setup py sleep exec bin sh c usr bin home acharrett ansible tmp ansible tmp setup py rm rf home acharrett ansible tmp ansible tmp dev null sleep ok task task path home acharrett zone yml using module file usr lib site packages ansible modules extras cloud amazon zone py establish local connection for user acharrett exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmpvxnmwn to home acharrett ansible tmp ansible tmp zone py exec bin sh c chmod u x home acharrett ansible tmp ansible tmp home acharrett ansible tmp ansible tmp zone py sleep exec bin sh c usr bin home acharrett ansible tmp ansible tmp zone py rm rf home acharrett ansible tmp ansible tmp dev null sleep an exception occurred during task execution the full traceback is traceback most recent call last file tmp ansible shbpjs ansible module zone py line in main file tmp ansible shbpjs ansible module zone py line in main vpc details details typeerror list indices must be integers not str fatal failed changed false failed true invocation module name zone module stderr traceback most recent call last n file tmp ansible shbpjs ansible module zone py line in n main n file tmp ansible shbpjs ansible module zone py line in main n vpc details details ntypeerror list indices must be integers not str n module stdout msg module failure to retry use limit home acharrett retry play recap localhost ok changed unreachable failed further information if i replace these lines starting line of ansible modules extras cloud amazon zone py vpc details details current vpc id vpc details current vpc region vpc details with these lines vpc details details gethostedzoneresponse vpcs current vpc id vpc details vpcid current vpc region vpc details vpcregion then the module works as expected
1
3,409
13,181,917,852
IssuesEvent
2020-08-12 15:00:22
duo-labs/cloudmapper
https://api.github.com/repos/duo-labs/cloudmapper
closed
Color edges of network map based on protocol
map unmaintained_functionality
Right now edges only show what can talk to what and are all gray. I could provide more context by coloring the edges based on an assumption of the protocol used. For example: - Green: 443/tcp HTTPS - Blue: 80/tcp HTTP - Purple: 22/tcp SSH - Red: Everything - Yellow: Concerning protocols such as databases, rabbitmq, etc. - Gray: Anything else This could get ugly if I have 443, 80, 22, 3306 (MySQL), and say 8080. This would result in 5 colored edges. Maybe I want that though?
True
Color edges of network map based on protocol - Right now edges only show what can talk to what and are all gray. I could provide more context by coloring the edges based on an assumption of the protocol used. For example: - Green: 443/tcp HTTPS - Blue: 80/tcp HTTP - Purple: 22/tcp SSH - Red: Everything - Yellow: Concerning protocols such as databases, rabbitmq, etc. - Gray: Anything else This could get ugly if I have 443, 80, 22, 3306 (MySQL), and say 8080. This would result in 5 colored edges. Maybe I want that though?
main
color edges of network map based on protocol right now edges only show what can talk to what and are all gray i could provide more context by coloring the edges based on an assumption of the protocol used for example green tcp https blue tcp http purple tcp ssh red everything yellow concerning protocols such as databases rabbitmq etc gray anything else this could get ugly if i have mysql and say this would result in colored edges maybe i want that though
1
1,789
6,575,881,306
IssuesEvent
2017-09-11 17:41:33
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
ec2 badly handles non ec2 instance related limits
affects_2.1 aws bug_report cloud waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> cloud/amazon/ec2.py ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` Using: ansible 2.1.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides But, I have checked devel branch in this repo and the issue seems still not handled ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> GNU/Linux (Ubuntu 14.04.x and 16.04.x) 64-bit. ##### SUMMARY <!--- Explain the problem briefly --> `ec2.py` module does not report well failures coming from **non instance** related **limits** ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> Just try creating `N` instances with `M` EBS volumes each, making sure to exceed volume related limit. <!--- Paste example playbooks or commands between quotes below --> ``` - name: create nodes local_action: module: ec2 params, args count: <positive int> ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> At the least I need the invocation to fail with meaningful error: 1. fail the module invocation, and print error message explaining which limit was actually exceeded. In a perfect world, I would also expect playbook to gracefully fail with "rollback": 1. destroy already created nodes 2. print failure message as explained before 3. fail the task ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> ``` 20:56:39 TASK [my-aws-bootstrap : create cluster nodes] ******************************* 20:56:39 task path: /var/lib/jenkins/jobs/lab-start/workspace/mypipeline/playbooks/roles/aws-bootstrap/tasks/main.yml:32 20:56:39 <localhost> ESTABLISH LOCAL CONNECTION FOR USER: jenkins 20:56:39 <localhost> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 python && sleep 0' 21:16:42 fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"assign_public_ip": false, "aws_access_key": null, "aws_secret_key": null, "count": 16, "count_tag": null, "ebs_optimized": false, "ec2_url": null, "exact_count": null, "group": ["group1"], "group_id": null, "id": null, "image": "ami-xxxxxxxx", "instance_ids": null, "instance_profile_name": null, "instance_tags": {"Environment": "pipeline-lab", "Name": "node", "lab_id": "lab-100", "user": "user1"}, "instance_type": "i2.8xlarge", "kernel": null, "key_name": "userkey", "monitoring": true, "network_interfaces": null, "placement_group": null, "private_ip": null, "profile": null, "ramdisk": null, "region": "us-east-1", "security_token": null, "source_dest_check": true, "spot_launch_group": null, "spot_price": null, "spot_type": "one-time", "spot_wait_timeout": "600", "state": "present", "tenancy": "default", "termination_protection": false, "user_data": null, "validate_certs": true, "volumes": [{"delete_on_termination": true, "device_name": "/dev/xvda", "volume_size": 1000}, {"delete_on_termination": true, "device_name": "/dev/xvdl", "device_type": "gp2", "volume_size": "100"}, {"delete_on_termination": true, "device_name": "/dev/xvdm", "device_type": "gp2", "volume_size": "100"}, {"delete_on_termination": true, "device_name": "/dev/xvdn", "device_type": "gp2", "volume_size": "100"}, {"device_name": "/dev/xvdd", "ephemeral": "ephemeral0"}, {"device_name": "/dev/xvde", "ephemeral": "ephemeral1"}, {"device_name": "/dev/xvdf", "ephemeral": "ephemeral2"}, {"device_name": "/dev/xvdg", "ephemeral": "ephemeral3"}, {"device_name": "/dev/xvdh", "ephemeral": "ephemeral4"}, {"device_name": "/dev/xvdi", "ephemeral": "ephemeral5"}, {"device_name": "/dev/xvdj", "ephemeral": "ephemeral6"}, {"device_name": "/dev/xvdk", "ephemeral": "ephemeral7"}], "vpc_subnet_id": "subnet-xxxxxx", "wait": true, "wait_timeout": "1200", "zone": "us-east-1a"}, "module_name": "ec2"}, "msg": "wait for instances running timeout on Mon Sep 26 21:16:42 2016"} ``` <!--- Paste verbatim command output between quotes below --> ``` ```
True
ec2 badly handles non ec2 instance related limits - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> cloud/amazon/ec2.py ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` Using: ansible 2.1.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides But, I have checked devel branch in this repo and the issue seems still not handled ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> GNU/Linux (Ubuntu 14.04.x and 16.04.x) 64-bit. ##### SUMMARY <!--- Explain the problem briefly --> `ec2.py` module does not report well failures coming from **non instance** related **limits** ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> Just try creating `N` instances with `M` EBS volumes each, making sure to exceed volume related limit. <!--- Paste example playbooks or commands between quotes below --> ``` - name: create nodes local_action: module: ec2 params, args count: <positive int> ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> At the least I need the invocation to fail with meaningful error: 1. fail the module invocation, and print error message explaining which limit was actually exceeded. In a perfect world, I would also expect playbook to gracefully fail with "rollback": 1. destroy already created nodes 2. print failure message as explained before 3. fail the task ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> ``` 20:56:39 TASK [my-aws-bootstrap : create cluster nodes] ******************************* 20:56:39 task path: /var/lib/jenkins/jobs/lab-start/workspace/mypipeline/playbooks/roles/aws-bootstrap/tasks/main.yml:32 20:56:39 <localhost> ESTABLISH LOCAL CONNECTION FOR USER: jenkins 20:56:39 <localhost> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 python && sleep 0' 21:16:42 fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"assign_public_ip": false, "aws_access_key": null, "aws_secret_key": null, "count": 16, "count_tag": null, "ebs_optimized": false, "ec2_url": null, "exact_count": null, "group": ["group1"], "group_id": null, "id": null, "image": "ami-xxxxxxxx", "instance_ids": null, "instance_profile_name": null, "instance_tags": {"Environment": "pipeline-lab", "Name": "node", "lab_id": "lab-100", "user": "user1"}, "instance_type": "i2.8xlarge", "kernel": null, "key_name": "userkey", "monitoring": true, "network_interfaces": null, "placement_group": null, "private_ip": null, "profile": null, "ramdisk": null, "region": "us-east-1", "security_token": null, "source_dest_check": true, "spot_launch_group": null, "spot_price": null, "spot_type": "one-time", "spot_wait_timeout": "600", "state": "present", "tenancy": "default", "termination_protection": false, "user_data": null, "validate_certs": true, "volumes": [{"delete_on_termination": true, "device_name": "/dev/xvda", "volume_size": 1000}, {"delete_on_termination": true, "device_name": "/dev/xvdl", "device_type": "gp2", "volume_size": "100"}, {"delete_on_termination": true, "device_name": "/dev/xvdm", "device_type": "gp2", "volume_size": "100"}, {"delete_on_termination": true, "device_name": "/dev/xvdn", "device_type": "gp2", "volume_size": "100"}, {"device_name": "/dev/xvdd", "ephemeral": "ephemeral0"}, {"device_name": "/dev/xvde", "ephemeral": "ephemeral1"}, {"device_name": "/dev/xvdf", "ephemeral": "ephemeral2"}, {"device_name": "/dev/xvdg", "ephemeral": "ephemeral3"}, {"device_name": "/dev/xvdh", "ephemeral": "ephemeral4"}, {"device_name": "/dev/xvdi", "ephemeral": "ephemeral5"}, {"device_name": "/dev/xvdj", "ephemeral": "ephemeral6"}, {"device_name": "/dev/xvdk", "ephemeral": "ephemeral7"}], "vpc_subnet_id": "subnet-xxxxxx", "wait": true, "wait_timeout": "1200", "zone": "us-east-1a"}, "module_name": "ec2"}, "msg": "wait for instances running timeout on Mon Sep 26 21:16:42 2016"} ``` <!--- Paste verbatim command output between quotes below --> ``` ```
main
badly handles non instance related limits issue type bug report component name cloud amazon py ansible version using ansible config file etc ansible ansible cfg configured module search path default w o overrides but i have checked devel branch in this repo and the issue seems still not handled configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific gnu linux ubuntu x and x bit summary py module does not report well failures coming from non instance related limits steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used just try creating n instances with m ebs volumes each making sure to exceed volume related limit name create nodes local action module params args count expected results at the least i need the invocation to fail with meaningful error fail the module invocation and print error message explaining which limit was actually exceeded in a perfect world i would also expect playbook to gracefully fail with rollback destroy already created nodes print failure message as explained before fail the task actual results task task path var lib jenkins jobs lab start workspace mypipeline playbooks roles aws bootstrap tasks main yml establish local connection for user jenkins exec bin sh c lang en us utf lc all en us utf lc messages en us utf python sleep fatal failed changed false failed true invocation module args assign public ip false aws access key null aws secret key null count count tag null ebs optimized false url null exact count null group group id null id null image ami xxxxxxxx instance ids null instance profile name null instance tags environment pipeline lab name node lab id lab user instance type kernel null key name userkey monitoring true network interfaces null placement group null private ip null profile null ramdisk null region us east security token null source dest check true spot launch group null spot price null spot type one time spot wait timeout state present tenancy default termination protection false user data null validate certs true volumes vpc subnet id subnet xxxxxx wait true wait timeout zone us east module name msg wait for instances running timeout on mon sep
1
1,762
6,575,000,000
IssuesEvent
2017-09-11 14:44:21
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Azure NSG Documenation missing some fields
affects_2.3 azure cloud docs_report waiting_on_maintainer
##### ISSUE TYPE - Documentation Report ##### COMPONENT NAME http://docs.ansible.com/ansible/azure_rm_securitygroup_module.html ##### SUMMARY The documention does not indicate some fields related to rules - source_address_prefix - Does this support tags as in the PORTAL? - direction - Assumed set is Inbound/Outbound - access - Assumed set isDeny/Allow
True
Azure NSG Documenation missing some fields - ##### ISSUE TYPE - Documentation Report ##### COMPONENT NAME http://docs.ansible.com/ansible/azure_rm_securitygroup_module.html ##### SUMMARY The documention does not indicate some fields related to rules - source_address_prefix - Does this support tags as in the PORTAL? - direction - Assumed set is Inbound/Outbound - access - Assumed set isDeny/Allow
main
azure nsg documenation missing some fields issue type documentation report component name summary the documention does not indicate some fields related to rules source address prefix does this support tags as in the portal direction assumed set is inbound outbound access assumed set isdeny allow
1
220,052
16,886,075,213
IssuesEvent
2021-06-23 00:13:54
opendatacube/datacube-k8s-eks
https://api.github.com/repos/opendatacube/datacube-k8s-eks
closed
Add Contributing guidelines
documentation enhancement
We should add a contributing guideline to define our goals and the "rules" for developing components of the stack. Things like: - Kubernetes tools must be able to be disabled via a flag - static helm variables should be defined as yaml files etc.
1.0
Add Contributing guidelines - We should add a contributing guideline to define our goals and the "rules" for developing components of the stack. Things like: - Kubernetes tools must be able to be disabled via a flag - static helm variables should be defined as yaml files etc.
non_main
add contributing guidelines we should add a contributing guideline to define our goals and the rules for developing components of the stack things like kubernetes tools must be able to be disabled via a flag static helm variables should be defined as yaml files etc
0
35,886
2,793,819,791
IssuesEvent
2015-05-11 13:37:15
elecoest/allevents-3-2
https://api.github.com/repos/elecoest/allevents-3-2
closed
Frontend\page liste des événements &page événement\proportions photo
auto-migrated Priority-Medium Type-Enhancement
``` joomla 3.3.6 AllEvents version 3.2.11.9 Serait-il possible de conserver les proportions des vignettes sur la page liste des événements et sur la page événement? Comme dans le module "AllEvents-Module List" ``` Original issue reported on code.google.com by `jeanluc...@gmail.com` on 25 Feb 2015 at 6:18
1.0
Frontend\page liste des événements &page événement\proportions photo - ``` joomla 3.3.6 AllEvents version 3.2.11.9 Serait-il possible de conserver les proportions des vignettes sur la page liste des événements et sur la page événement? Comme dans le module "AllEvents-Module List" ``` Original issue reported on code.google.com by `jeanluc...@gmail.com` on 25 Feb 2015 at 6:18
non_main
frontend page liste des événements page événement proportions photo joomla allevents version serait il possible de conserver les proportions des vignettes sur la page liste des événements et sur la page événement comme dans le module allevents module list original issue reported on code google com by jeanluc gmail com on feb at
0
158,532
12,418,130,646
IssuesEvent
2020-05-22 22:54:49
saltstack/salt
https://api.github.com/repos/saltstack/salt
closed
Py3 Fedora 29 test failures - Failed to synchronize cache for repo 'updates'
Confirmed Test Failure WarRoom
### Description of Issue https://jenkinsci.saltstack.com/job/2019.2.1/view/Python3/job/salt-fedora-29-py3/192/ A bunch of test cases failed with following Error Message False is not True. Salt Comment: Error occurred installing package(s). Additional info follows: errors: - Running scope as unit: run-rddbe1290b5904a3fbfd55365bb31d148.scope Fedora 29 - x86_64 - Updates 2.7 kB/s | 400 B 00:00 Failed to synchronize cache for repo 'updates' Error: Failed to synchronize cache for repo 'updates' --------------- We haven't switched our system level repositories to artifactories. Gareth, did a quick analysis and found Salt isn't at fault here, but we still need to handle this error gracefully.
1.0
Py3 Fedora 29 test failures - Failed to synchronize cache for repo 'updates' - ### Description of Issue https://jenkinsci.saltstack.com/job/2019.2.1/view/Python3/job/salt-fedora-29-py3/192/ A bunch of test cases failed with following Error Message False is not True. Salt Comment: Error occurred installing package(s). Additional info follows: errors: - Running scope as unit: run-rddbe1290b5904a3fbfd55365bb31d148.scope Fedora 29 - x86_64 - Updates 2.7 kB/s | 400 B 00:00 Failed to synchronize cache for repo 'updates' Error: Failed to synchronize cache for repo 'updates' --------------- We haven't switched our system level repositories to artifactories. Gareth, did a quick analysis and found Salt isn't at fault here, but we still need to handle this error gracefully.
non_main
fedora test failures failed to synchronize cache for repo updates description of issue a bunch of test cases failed with following error message false is not true salt comment error occurred installing package s additional info follows errors running scope as unit run scope fedora updates kb s b failed to synchronize cache for repo updates error failed to synchronize cache for repo updates we haven t switched our system level repositories to artifactories gareth did a quick analysis and found salt isn t at fault here but we still need to handle this error gracefully
0
275,320
23,906,686,660
IssuesEvent
2022-09-09 02:04:31
NVIDIA/spark-rapids
https://api.github.com/repos/NVIDIA/spark-rapids
opened
[TEST] make avro_test enable method
test task
**Describe the bug** https://github.com/NVIDIA/spark-rapids/pull/6505/files#r965443759 avro_test enable method is inconsistent w/ other tests that require extra libs, let's make it follow the pattern in the future
1.0
[TEST] make avro_test enable method - **Describe the bug** https://github.com/NVIDIA/spark-rapids/pull/6505/files#r965443759 avro_test enable method is inconsistent w/ other tests that require extra libs, let's make it follow the pattern in the future
non_main
make avro test enable method describe the bug avro test enable method is inconsistent w other tests that require extra libs let s make it follow the pattern in the future
0
316,431
9,647,599,551
IssuesEvent
2019-05-17 14:18:34
meetalva/alva
https://api.github.com/repos/meetalva/alva
closed
Dark mode
help wanted priority: low type: feature
We should have a optional dark mode for the Alva application, so users can work better in Dark environments or have an better contrast if they are working on a light website.
1.0
Dark mode - We should have a optional dark mode for the Alva application, so users can work better in Dark environments or have an better contrast if they are working on a light website.
non_main
dark mode we should have a optional dark mode for the alva application so users can work better in dark environments or have an better contrast if they are working on a light website
0
606,170
18,756,451,983
IssuesEvent
2021-11-05 11:24:52
stackabletech/operator-rs
https://api.github.com/repos/stackabletech/operator-rs
closed
Remove labelselector type=krustlet when moving to docker images
priority/critical
We add a "type=krustlet" labelselector to only match "our" agent nodes. This is not required anymore when moving to full kubernetes and docker support https://github.com/stackabletech/operator-rs/blob/34639c05b842392539036bc2a37f096c9e27d9c8/src/krustlet.rs#L49
1.0
Remove labelselector type=krustlet when moving to docker images - We add a "type=krustlet" labelselector to only match "our" agent nodes. This is not required anymore when moving to full kubernetes and docker support https://github.com/stackabletech/operator-rs/blob/34639c05b842392539036bc2a37f096c9e27d9c8/src/krustlet.rs#L49
non_main
remove labelselector type krustlet when moving to docker images we add a type krustlet labelselector to only match our agent nodes this is not required anymore when moving to full kubernetes and docker support
0
5,425
27,215,745,643
IssuesEvent
2023-02-20 21:32:47
NIAEFEUP/website-niaefeup-backend
https://api.github.com/repos/NIAEFEUP/website-niaefeup-backend
opened
generations: Create type for school year
maintainability
Currently, a generation is identified by a string named schoolYear that is validated by the `SchoolYear` validator. Its format is `<xx-yy>` where yy=xx+1. This works fine but there are some places where we need to manage the years as integers and we might want to use school years for other models in the future. So, it'd be nice to create a `SchoolYear` type that makes it easier to get the school years and is always validated. We still want to save it as a string in the database, so a `Converter` might be useful, the same way that it's used for `Permissions`. The current places where this would be useful are: `SchoolYear` validator `inferSchoolYearIfNotSpecified` in `GenerationService`. Please check if there are more cases when you implement this.
True
generations: Create type for school year - Currently, a generation is identified by a string named schoolYear that is validated by the `SchoolYear` validator. Its format is `<xx-yy>` where yy=xx+1. This works fine but there are some places where we need to manage the years as integers and we might want to use school years for other models in the future. So, it'd be nice to create a `SchoolYear` type that makes it easier to get the school years and is always validated. We still want to save it as a string in the database, so a `Converter` might be useful, the same way that it's used for `Permissions`. The current places where this would be useful are: `SchoolYear` validator `inferSchoolYearIfNotSpecified` in `GenerationService`. Please check if there are more cases when you implement this.
main
generations create type for school year currently a generation is identified by a string named schoolyear that is validated by the schoolyear validator its format is where yy xx this works fine but there are some places where we need to manage the years as integers and we might want to use school years for other models in the future so it d be nice to create a schoolyear type that makes it easier to get the school years and is always validated we still want to save it as a string in the database so a converter might be useful the same way that it s used for permissions the current places where this would be useful are schoolyear validator inferschoolyearifnotspecified in generationservice please check if there are more cases when you implement this
1
5,794
30,698,826,833
IssuesEvent
2023-07-26 20:58:31
bazelbuild/intellij
https://api.github.com/repos/bazelbuild/intellij
closed
Intellij builds are broken with Bazel latest release 6.3.0
type: bug product: Android Studio product: CLion product: IntelliJ awaiting-maintainer
Android Studio Plugin Google: https://buildkite.com/bazel/android-studio-plugin-google/builds/4811#01898f88-8e27-46b0-8f8e-b9e7fcb2aba4 CLion Plugin Google: https://buildkite.com/bazel/clion-plugin-google/builds/4807#01898f88-bafc-49a7-99f9-58b49ae7b40e IntelliJ Plugin Google : https://buildkite.com/bazel/intellij-plugin-google/builds/4794#01898f8c-45fc-4774-8701-f74449b4ae4d IntelliJ UE Plugin Google : https://buildkite.com/bazel/intellij-ue-plugin-google/builds/4803#01898f8c-b0ee-4893-9267-5f15c4d835c7 IntelliJ Plugin Aspect Google:https://buildkite.com/bazel/intellij-plugin-aspect-google/builds/4811#01898f8c-ae31-4aab-8c66-d654731ea787 Logs : ``` querysync/java/com/google/idea/blaze/qsync/project/BuildGraphData.java:217: warning: as of release 10, 'var' is a restricted type name and cannot be used for type declarations or as the element type of an array var unused = result.sourceOwners(); ^ querysync/java/com/google/idea/blaze/qsync/project/BuildGraphData.java:217: error: cannot find symbol var unused = result.sourceOwners(); ^ symbol: class var location: class Builder ``` Steps: ``` git clone https://github.com/bazelbuild/intellij.git export USE_BAZEL_VERSION=6.3.0 # reproducing for Android-Studio-internal-stable bazel build --define=ij_product=android-studio-latest //aswb/... ``` CC GreenTeam @fweikert
True
Intellij builds are broken with Bazel latest release 6.3.0 - Android Studio Plugin Google: https://buildkite.com/bazel/android-studio-plugin-google/builds/4811#01898f88-8e27-46b0-8f8e-b9e7fcb2aba4 CLion Plugin Google: https://buildkite.com/bazel/clion-plugin-google/builds/4807#01898f88-bafc-49a7-99f9-58b49ae7b40e IntelliJ Plugin Google : https://buildkite.com/bazel/intellij-plugin-google/builds/4794#01898f8c-45fc-4774-8701-f74449b4ae4d IntelliJ UE Plugin Google : https://buildkite.com/bazel/intellij-ue-plugin-google/builds/4803#01898f8c-b0ee-4893-9267-5f15c4d835c7 IntelliJ Plugin Aspect Google:https://buildkite.com/bazel/intellij-plugin-aspect-google/builds/4811#01898f8c-ae31-4aab-8c66-d654731ea787 Logs : ``` querysync/java/com/google/idea/blaze/qsync/project/BuildGraphData.java:217: warning: as of release 10, 'var' is a restricted type name and cannot be used for type declarations or as the element type of an array var unused = result.sourceOwners(); ^ querysync/java/com/google/idea/blaze/qsync/project/BuildGraphData.java:217: error: cannot find symbol var unused = result.sourceOwners(); ^ symbol: class var location: class Builder ``` Steps: ``` git clone https://github.com/bazelbuild/intellij.git export USE_BAZEL_VERSION=6.3.0 # reproducing for Android-Studio-internal-stable bazel build --define=ij_product=android-studio-latest //aswb/... ``` CC GreenTeam @fweikert
main
intellij builds are broken with bazel latest release android studio plugin google clion plugin google intellij plugin google intellij ue plugin google intellij plugin aspect google logs querysync java com google idea blaze qsync project buildgraphdata java warning as of release var is a restricted type name and cannot be used for type declarations or as the element type of an array var unused result sourceowners querysync java com google idea blaze qsync project buildgraphdata java error cannot find symbol var unused result sourceowners symbol class var location class builder steps git clone export use bazel version reproducing for android studio internal stable bazel build define ij product android studio latest aswb cc greenteam fweikert
1