Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
21,616
30,022,518,007
IssuesEvent
2023-06-27 01:34:13
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Include error messages in incompatible change process
P2 type: process team-OSS stale
### Description of the problem / feature request: The [incompatible change process](https://bazel.build/breaking-changes-guide.html) is centered on github labels/issues for a proactive testing workflow. Good error messages are needed to complement this, to address anyone doing _reactive_ detection of bazel breakages (aka trying a new release and seeing what breaks; the most common pattern I see in the wild). ### Feature requests: what underlying problem are you trying to solve with this feature? Some breaking changes are going out with poor error messages, and it's not clear that that's being considered in the change approval process. I'd ideally like to see these changes have good error messages, and where not, have the error messages that users will see at least documented (for searchability), factored in as a cost to account for when these breaking changes are approved, and possibly highlighted centrally in the release notes as important to be aware of. A couple current examples: - [incompatible_py3_is_default](https://github.com/bazelbuild/bazel/issues/7359) does not indicate in the issue what error messages will look like if users are broken by this change, and it looks like it'll be opaque python errors. - [incompatible_remap_main_repo](https://github.com/bazelbuild/bazel/issues/7130) appears to cause opaque unrelated-looking failures. I should note that most rule changes like deprecating attributes _do_ have good error messages; it's only a small subset that don't. Though the format is not standard - some errors link github issues, some link the --incompatible_... flag to override with, and some link both. Including examples in the incompatible change GH issues should help reviewers ensure both flag and issue URL are output. ### What's the output of `bazel info release`? Bazel 0.26 is most recent as of when I'm filing this.
1.0
Include error messages in incompatible change process - ### Description of the problem / feature request: The [incompatible change process](https://bazel.build/breaking-changes-guide.html) is centered on github labels/issues for a proactive testing workflow. Good error messages are needed to complement this, to address anyone doing _reactive_ detection of bazel breakages (aka trying a new release and seeing what breaks; the most common pattern I see in the wild). ### Feature requests: what underlying problem are you trying to solve with this feature? Some breaking changes are going out with poor error messages, and it's not clear that that's being considered in the change approval process. I'd ideally like to see these changes have good error messages, and where not, have the error messages that users will see at least documented (for searchability), factored in as a cost to account for when these breaking changes are approved, and possibly highlighted centrally in the release notes as important to be aware of. A couple current examples: - [incompatible_py3_is_default](https://github.com/bazelbuild/bazel/issues/7359) does not indicate in the issue what error messages will look like if users are broken by this change, and it looks like it'll be opaque python errors. - [incompatible_remap_main_repo](https://github.com/bazelbuild/bazel/issues/7130) appears to cause opaque unrelated-looking failures. I should note that most rule changes like deprecating attributes _do_ have good error messages; it's only a small subset that don't. Though the format is not standard - some errors link github issues, some link the --incompatible_... flag to override with, and some link both. Including examples in the incompatible change GH issues should help reviewers ensure both flag and issue URL are output. ### What's the output of `bazel info release`? Bazel 0.26 is most recent as of when I'm filing this.
process
include error messages in incompatible change process description of the problem feature request the is centered on github labels issues for a proactive testing workflow good error messages are needed to complement this to address anyone doing reactive detection of bazel breakages aka trying a new release and seeing what breaks the most common pattern i see in the wild feature requests what underlying problem are you trying to solve with this feature some breaking changes are going out with poor error messages and it s not clear that that s being considered in the change approval process i d ideally like to see these changes have good error messages and where not have the error messages that users will see at least documented for searchability factored in as a cost to account for when these breaking changes are approved and possibly highlighted centrally in the release notes as important to be aware of a couple current examples does not indicate in the issue what error messages will look like if users are broken by this change and it looks like it ll be opaque python errors appears to cause opaque unrelated looking failures i should note that most rule changes like deprecating attributes do have good error messages it s only a small subset that don t though the format is not standard some errors link github issues some link the incompatible flag to override with and some link both including examples in the incompatible change gh issues should help reviewers ensure both flag and issue url are output what s the output of bazel info release bazel is most recent as of when i m filing this
1
52,294
6,608,339,607
IssuesEvent
2017-09-19 10:34:35
SleipnerTechnologies/saftens-bekaempelse
https://api.github.com/repos/SleipnerTechnologies/saftens-bekaempelse
closed
Forside Design
design feature
Lav forsiden til projektet. Skal have en store CTA i toppen, men tilmelding som indsamler, og derefter lidt generel info om blærekræft..... ikke for meget... det er det vi har info siden til #5 #2
1.0
Forside Design - Lav forsiden til projektet. Skal have en store CTA i toppen, men tilmelding som indsamler, og derefter lidt generel info om blærekræft..... ikke for meget... det er det vi har info siden til #5 #2
non_process
forside design lav forsiden til projektet skal have en store cta i toppen men tilmelding som indsamler og derefter lidt generel info om blærekræft ikke for meget det er det vi har info siden til
0
167,029
12,980,598,152
IssuesEvent
2020-07-22 05:45:00
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: gossip/chaos/nodes=9 failed
C-test-failure O-roachtest O-robot branch-provisional_202007220233_v20.2.0-alpha.2 release-blocker
[(roachtest).gossip/chaos/nodes=9 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2107811&tab=buildLog) on [provisional_202007220233_v20.2.0-alpha.2@d3119926d33d808c6384cf3e99a7f7435f395489](https://github.com/cockroachdb/cockroach/commits/d3119926d33d808c6384cf3e99a7f7435f395489): ``` The test failed on branch=provisional_202007220233_v20.2.0-alpha.2, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/gossip/chaos/nodes=9/run_1 gossip.go:64,gossip.go:102,gossip.go:114,gossip.go:124,test_runner.go:757: gossip did not stabilize in 20.0s cluster.go:1571,context.go:135,cluster.go:1560,test_runner.go:826: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-2107811-1595392378-29-n9cpu4 --oneshot --ignore-empty-nodes: exit status 1 4: 5293 7: 5342 5: 5343 9: 5321 2: 5279 6: 5267 8: 5375 1: 6007 3: dead Error: UNCLASSIFIED_PROBLEM: 3: dead (1) UNCLASSIFIED_PROBLEM Wraps: (2) attached stack trace | main.glob..func13 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1115 | main.wrap.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:266 | github.com/spf13/cobra.(*Command).execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:830 | github.com/spf13/cobra.(*Command).ExecuteC | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:914 | github.com/spf13/cobra.(*Command).Execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:864 | main.main | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1808 | runtime.main | /usr/local/go/src/runtime/proc.go:203 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1373 Wraps: (3) 3 safe details enclosed Wraps: (4) 3: dead Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *safedetails.withSafeDetails (4) *errors.errorString ``` <details><summary>More</summary><p> Artifacts: [/gossip/chaos/nodes=9](https://teamcity.cockroachdb.com/viewLog.html?buildId=2107811&tab=artifacts#/gossip/chaos/nodes=9) Related: - #49493 roachtest: gossip/chaos/nodes=9 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Agossip%2Fchaos%2Fnodes%3D9.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
2.0
roachtest: gossip/chaos/nodes=9 failed - [(roachtest).gossip/chaos/nodes=9 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2107811&tab=buildLog) on [provisional_202007220233_v20.2.0-alpha.2@d3119926d33d808c6384cf3e99a7f7435f395489](https://github.com/cockroachdb/cockroach/commits/d3119926d33d808c6384cf3e99a7f7435f395489): ``` The test failed on branch=provisional_202007220233_v20.2.0-alpha.2, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/gossip/chaos/nodes=9/run_1 gossip.go:64,gossip.go:102,gossip.go:114,gossip.go:124,test_runner.go:757: gossip did not stabilize in 20.0s cluster.go:1571,context.go:135,cluster.go:1560,test_runner.go:826: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-2107811-1595392378-29-n9cpu4 --oneshot --ignore-empty-nodes: exit status 1 4: 5293 7: 5342 5: 5343 9: 5321 2: 5279 6: 5267 8: 5375 1: 6007 3: dead Error: UNCLASSIFIED_PROBLEM: 3: dead (1) UNCLASSIFIED_PROBLEM Wraps: (2) attached stack trace | main.glob..func13 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1115 | main.wrap.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:266 | github.com/spf13/cobra.(*Command).execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:830 | github.com/spf13/cobra.(*Command).ExecuteC | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:914 | github.com/spf13/cobra.(*Command).Execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:864 | main.main | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1808 | runtime.main | /usr/local/go/src/runtime/proc.go:203 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1373 Wraps: (3) 3 safe details enclosed Wraps: (4) 3: dead Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *safedetails.withSafeDetails (4) *errors.errorString ``` <details><summary>More</summary><p> Artifacts: [/gossip/chaos/nodes=9](https://teamcity.cockroachdb.com/viewLog.html?buildId=2107811&tab=artifacts#/gossip/chaos/nodes=9) Related: - #49493 roachtest: gossip/chaos/nodes=9 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Agossip%2Fchaos%2Fnodes%3D9.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
non_process
roachtest gossip chaos nodes failed on the test failed on branch provisional alpha cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts gossip chaos nodes run gossip go gossip go gossip go gossip go test runner go gossip did not stabilize in cluster go context go cluster go test runner go dead node detection home agent work go src github com cockroachdb cockroach bin roachprod monitor teamcity oneshot ignore empty nodes exit status dead error unclassified problem dead unclassified problem wraps attached stack trace main glob home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go main wrap home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command executec home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go main main home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps safe details enclosed wraps dead error types errors unclassified withstack withstack safedetails withsafedetails errors errorstring more artifacts related roachtest gossip chaos nodes failed powered by
0
22,623
31,847,110,905
IssuesEvent
2023-09-14 20:52:46
KASHO7/project2
https://api.github.com/repos/KASHO7/project2
opened
Handling template-processor.js (problem 2)
Issue | Template Processor
Hayden will be working on creating the template processor file.
1.0
Handling template-processor.js (problem 2) - Hayden will be working on creating the template processor file.
process
handling template processor js problem hayden will be working on creating the template processor file
1
795,066
28,059,788,263
IssuesEvent
2023-03-29 11:53:27
KDT-Final-Team4/backend
https://api.github.com/repos/KDT-Final-Team4/backend
closed
[Feat] 공지사항 공개 여부 기능 구현
For : API Priority : Low Status : In progress Type : Feature
## 🔨개발 할 기능 - 공지사항 공개 여부 기능 구현 ## 🧩 세부 기능 해당 기능에 대한 세부 계획 작성 (ex. -[ ] 로그인 시 아이디 비번 입력 받기) - [x] 공지사항 상태 변경 기능 (공개/비공개) ## 📖 참고 사항
1.0
[Feat] 공지사항 공개 여부 기능 구현 - ## 🔨개발 할 기능 - 공지사항 공개 여부 기능 구현 ## 🧩 세부 기능 해당 기능에 대한 세부 계획 작성 (ex. -[ ] 로그인 시 아이디 비번 입력 받기) - [x] 공지사항 상태 변경 기능 (공개/비공개) ## 📖 참고 사항
non_process
공지사항 공개 여부 기능 구현 🔨개발 할 기능 공지사항 공개 여부 기능 구현 🧩 세부 기능 해당 기능에 대한 세부 계획 작성 ex 로그인 시 아이디 비번 입력 받기 공지사항 상태 변경 기능 공개 비공개 📖 참고 사항
0
4,893
5,376,069,190
IssuesEvent
2017-02-23 07:51:29
symfony/symfony
https://api.github.com/repos/symfony/symfony
closed
circular reference on security.access.decision_manager
Bug Security Status: Needs Review Unconfirmed
| Q | A | ---------------- | ----- | Bug report? | yes (possibly) | Feature request? | no | BC Break report? | unsure | RFC? | no | Symfony version | 2.8.15 While creating a custom voter as per [these instructions](http://symfony.com/doc/2.8/security/voters.html) I'm getting the following error when trying to inject the decision manager service: **with public in the service declaration: true** > Circular reference detected for service "security.access.decision_manager", path: "security.access.decision_manager". **with public in the service declaration: false** > Circular reference detected for service "security.access.decision_manager", path: "cache_warmer -> twig -> security.authorization_checker -> security.access.decision_manager". according to the docs, this should only **not** work < 2.8? However, when I then tried injecting the container, and pull out the authorization.checker service inside the voter, and I get a similar circular reference error, even though the docs say this is the workaround. So, I'm not sure if the docs are wrong, I've missed something, or this is a bug? **service declaration** ``` app.form_config.voter: class: AppBundle\Services\voters\ConfigAccessVoter arguments: ["@security.access.decision_manager"] tags: - { name: security.voter } public: false ```
True
circular reference on security.access.decision_manager - | Q | A | ---------------- | ----- | Bug report? | yes (possibly) | Feature request? | no | BC Break report? | unsure | RFC? | no | Symfony version | 2.8.15 While creating a custom voter as per [these instructions](http://symfony.com/doc/2.8/security/voters.html) I'm getting the following error when trying to inject the decision manager service: **with public in the service declaration: true** > Circular reference detected for service "security.access.decision_manager", path: "security.access.decision_manager". **with public in the service declaration: false** > Circular reference detected for service "security.access.decision_manager", path: "cache_warmer -> twig -> security.authorization_checker -> security.access.decision_manager". according to the docs, this should only **not** work < 2.8? However, when I then tried injecting the container, and pull out the authorization.checker service inside the voter, and I get a similar circular reference error, even though the docs say this is the workaround. So, I'm not sure if the docs are wrong, I've missed something, or this is a bug? **service declaration** ``` app.form_config.voter: class: AppBundle\Services\voters\ConfigAccessVoter arguments: ["@security.access.decision_manager"] tags: - { name: security.voter } public: false ```
non_process
circular reference on security access decision manager q a bug report yes possibly feature request no bc break report unsure rfc no symfony version while creating a custom voter as per i m getting the following error when trying to inject the decision manager service with public in the service declaration true circular reference detected for service security access decision manager path security access decision manager with public in the service declaration false circular reference detected for service security access decision manager path cache warmer twig security authorization checker security access decision manager according to the docs this should only not work however when i then tried injecting the container and pull out the authorization checker service inside the voter and i get a similar circular reference error even though the docs say this is the workaround so i m not sure if the docs are wrong i ve missed something or this is a bug service declaration app form config voter class appbundle services voters configaccessvoter arguments tags name security voter public false
0
32,354
8,838,953,533
IssuesEvent
2019-01-05 23:13:07
inexorgame/inexor-core
https://api.github.com/repos/inexorgame/inexor-core
closed
Create new release first as a draft, publish when packages uploaded
cat:BUG feat:build environment org:needs validation/discussion/decision
Right now we have dead download links for ca. 20 - 40 minutes after every merge into master. This is because our CI first creates a new GitHub release and instantly publish it. However, the building process for Windows and Linux is taking much longer than this and it takes time till these packages are successfully uploaded. The CI should first create the release as a draft and only publish it when the packages are sucessfully uploaded. Switching to a release draft should be a small change in the API call and publishing it can be done with an additional Travis job as an additional Travis build stage. We need to make sure that both Travis AND AppVeyor are finished before publishing.
1.0
Create new release first as a draft, publish when packages uploaded - Right now we have dead download links for ca. 20 - 40 minutes after every merge into master. This is because our CI first creates a new GitHub release and instantly publish it. However, the building process for Windows and Linux is taking much longer than this and it takes time till these packages are successfully uploaded. The CI should first create the release as a draft and only publish it when the packages are sucessfully uploaded. Switching to a release draft should be a small change in the API call and publishing it can be done with an additional Travis job as an additional Travis build stage. We need to make sure that both Travis AND AppVeyor are finished before publishing.
non_process
create new release first as a draft publish when packages uploaded right now we have dead download links for ca minutes after every merge into master this is because our ci first creates a new github release and instantly publish it however the building process for windows and linux is taking much longer than this and it takes time till these packages are successfully uploaded the ci should first create the release as a draft and only publish it when the packages are sucessfully uploaded switching to a release draft should be a small change in the api call and publishing it can be done with an additional travis job as an additional travis build stage we need to make sure that both travis and appveyor are finished before publishing
0
2,861
5,824,385,008
IssuesEvent
2017-05-07 12:33:07
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[FEATURE][processing] New algorithm for merging connected lines
HackFest needs-backport Processing Text User Manual
Original commit: https://github.com/qgis/QGIS/commit/299037a0bfcaf0251b2d4b763f5a8b9c1b8176b1 by nyalldawson This algorithm joins all connected parts of MultiLineString geometries into single LineString geometries. If any parts of the input MultiLineString geometries are not connected, the resultant geometry will be a MultiLineString containing any lines which could be merged and any non-connected line parts. (cherry-picked from 30fcaed634aa43d01528a031fd8f2d255f97f837)
1.0
[FEATURE][processing] New algorithm for merging connected lines - Original commit: https://github.com/qgis/QGIS/commit/299037a0bfcaf0251b2d4b763f5a8b9c1b8176b1 by nyalldawson This algorithm joins all connected parts of MultiLineString geometries into single LineString geometries. If any parts of the input MultiLineString geometries are not connected, the resultant geometry will be a MultiLineString containing any lines which could be merged and any non-connected line parts. (cherry-picked from 30fcaed634aa43d01528a031fd8f2d255f97f837)
process
new algorithm for merging connected lines original commit by nyalldawson this algorithm joins all connected parts of multilinestring geometries into single linestring geometries if any parts of the input multilinestring geometries are not connected the resultant geometry will be a multilinestring containing any lines which could be merged and any non connected line parts cherry picked from
1
7,261
10,420,651,110
IssuesEvent
2019-09-16 01:55:37
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Not possible to toggle use selected features in the Processing layer combobox
Bug Processing Regression
<!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). --> **Describe the bug** I am trying to import a selected features into a PostgreSQL database but I am unable to toggle the selected features button. If I reselect the layer in the processing dialog then the selected features become active **How to Reproduce** 1. Load a vector layer (I used a spatialite layer) 2. Select a random feature or features 3. Activate processing and select the alg Export to PostgreSQL 4. Notice that the `Selected Features Only` is greyed out. 5. Click on the layer again and it becomes active and I get the following error ``` 2019-07-09T11:06:20 WARNING Traceback (most recent call last): File "/usr/share/qgis/python/plugins/processing/gui/wrappers.py", line 1544, in parentValueChanged if value in self.parent_file_based_layers: TypeError: unhashable type: 'QgsProcessingFeatureSourceDefinition' ``` ![save](https://user-images.githubusercontent.com/2510900/60874651-a4e74900-a238-11e9-818d-c93573f2bd1e.gif) **QGIS and OS versions** ![update](https://user-images.githubusercontent.com/2510900/60874954-335bca80-a239-11e9-83c8-4299dbd16cdd.png) **Additional context** <!-- Add any other context about the problem here. -->
1.0
Not possible to toggle use selected features in the Processing layer combobox - <!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). --> **Describe the bug** I am trying to import a selected features into a PostgreSQL database but I am unable to toggle the selected features button. If I reselect the layer in the processing dialog then the selected features become active **How to Reproduce** 1. Load a vector layer (I used a spatialite layer) 2. Select a random feature or features 3. Activate processing and select the alg Export to PostgreSQL 4. Notice that the `Selected Features Only` is greyed out. 5. Click on the layer again and it becomes active and I get the following error ``` 2019-07-09T11:06:20 WARNING Traceback (most recent call last): File "/usr/share/qgis/python/plugins/processing/gui/wrappers.py", line 1544, in parentValueChanged if value in self.parent_file_based_layers: TypeError: unhashable type: 'QgsProcessingFeatureSourceDefinition' ``` ![save](https://user-images.githubusercontent.com/2510900/60874651-a4e74900-a238-11e9-818d-c93573f2bd1e.gif) **QGIS and OS versions** ![update](https://user-images.githubusercontent.com/2510900/60874954-335bca80-a239-11e9-83c8-4299dbd16cdd.png) **Additional context** <!-- Add any other context about the problem here. -->
process
not possible to toggle use selected features in the processing layer combobox bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue if the issue concerns a third party plugin then it cannot be fixed by the qgis team please raise your issue in the dedicated bug tracker for that specific plugin as listed in the plugin s description describe the bug i am trying to import a selected features into a postgresql database but i am unable to toggle the selected features button if i reselect the layer in the processing dialog then the selected features become active how to reproduce load a vector layer i used a spatialite layer select a random feature or features activate processing and select the alg export to postgresql notice that the selected features only is greyed out click on the layer again and it becomes active and i get the following error warning traceback most recent call last file usr share qgis python plugins processing gui wrappers py line in parentvaluechanged if value in self parent file based layers typeerror unhashable type qgsprocessingfeaturesourcedefinition qgis and os versions additional context
1
53,023
13,781,975,310
IssuesEvent
2020-10-08 16:56:54
ioana-nicolae/first
https://api.github.com/repos/ioana-nicolae/first
opened
CVE-2016-10735 (Medium) detected in bootstrap-3.3.7.tgz, bootstrap-3.1.1.tgz
security vulnerability
## CVE-2016-10735 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>bootstrap-3.3.7.tgz</b>, <b>bootstrap-3.1.1.tgz</b></p></summary> <p> <details><summary><b>bootstrap-3.3.7.tgz</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://registry.npmjs.org/bootstrap/-/bootstrap-3.3.7.tgz">https://registry.npmjs.org/bootstrap/-/bootstrap-3.3.7.tgz</a></p> <p>Path to dependency file: first/angular.js-master/angular.js-master/yarn.lock</p> <p>Path to vulnerable library: first/angular.js-master/angular.js-master/yarn.lock</p> <p> Dependency Hierarchy: - angular-benchpress-0.2.2.tgz (Root Library) - :x: **bootstrap-3.3.7.tgz** (Vulnerable Library) </details> <details><summary><b>bootstrap-3.1.1.tgz</b></p></summary> <p>Sleek, intuitive, and powerful front-end framework for faster and easier web development.</p> <p>Library home page: <a href="https://registry.npmjs.org/bootstrap/-/bootstrap-3.1.1.tgz">https://registry.npmjs.org/bootstrap/-/bootstrap-3.1.1.tgz</a></p> <p>Path to dependency file: first/angular.js-master/angular.js-master/yarn.lock</p> <p>Path to vulnerable library: first/angular.js-master/angular.js-master/yarn.lock</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.1.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/ioana-nicolae/first/commit/6e040516f98074b37f69996b76c1c99015c1d685">6e040516f98074b37f69996b76c1c99015c1d685</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/issues/20184">https://github.com/twbs/bootstrap/issues/20184</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: 3.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"bootstrap","packageVersion":"3.3.7","isTransitiveDependency":true,"dependencyTree":"angular-benchpress:0.2.2;bootstrap:3.3.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"javascript/Node.js","packageName":"bootstrap","packageVersion":"3.1.1","isTransitiveDependency":false,"dependencyTree":"bootstrap:3.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"}],"vulnerabilityIdentifier":"CVE-2016-10735","vulnerabilityDetails":"In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2016-10735 (Medium) detected in bootstrap-3.3.7.tgz, bootstrap-3.1.1.tgz - ## CVE-2016-10735 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>bootstrap-3.3.7.tgz</b>, <b>bootstrap-3.1.1.tgz</b></p></summary> <p> <details><summary><b>bootstrap-3.3.7.tgz</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://registry.npmjs.org/bootstrap/-/bootstrap-3.3.7.tgz">https://registry.npmjs.org/bootstrap/-/bootstrap-3.3.7.tgz</a></p> <p>Path to dependency file: first/angular.js-master/angular.js-master/yarn.lock</p> <p>Path to vulnerable library: first/angular.js-master/angular.js-master/yarn.lock</p> <p> Dependency Hierarchy: - angular-benchpress-0.2.2.tgz (Root Library) - :x: **bootstrap-3.3.7.tgz** (Vulnerable Library) </details> <details><summary><b>bootstrap-3.1.1.tgz</b></p></summary> <p>Sleek, intuitive, and powerful front-end framework for faster and easier web development.</p> <p>Library home page: <a href="https://registry.npmjs.org/bootstrap/-/bootstrap-3.1.1.tgz">https://registry.npmjs.org/bootstrap/-/bootstrap-3.1.1.tgz</a></p> <p>Path to dependency file: first/angular.js-master/angular.js-master/yarn.lock</p> <p>Path to vulnerable library: first/angular.js-master/angular.js-master/yarn.lock</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.1.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/ioana-nicolae/first/commit/6e040516f98074b37f69996b76c1c99015c1d685">6e040516f98074b37f69996b76c1c99015c1d685</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/issues/20184">https://github.com/twbs/bootstrap/issues/20184</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: 3.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"bootstrap","packageVersion":"3.3.7","isTransitiveDependency":true,"dependencyTree":"angular-benchpress:0.2.2;bootstrap:3.3.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"javascript/Node.js","packageName":"bootstrap","packageVersion":"3.1.1","isTransitiveDependency":false,"dependencyTree":"bootstrap:3.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"}],"vulnerabilityIdentifier":"CVE-2016-10735","vulnerabilityDetails":"In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in bootstrap tgz bootstrap tgz cve medium severity vulnerability vulnerable libraries bootstrap tgz bootstrap tgz bootstrap tgz the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file first angular js master angular js master yarn lock path to vulnerable library first angular js master angular js master yarn lock dependency hierarchy angular benchpress tgz root library x bootstrap tgz vulnerable library bootstrap tgz sleek intuitive and powerful front end framework for faster and easier web development library home page a href path to dependency file first angular js master angular js master yarn lock path to vulnerable library first angular js master angular js master yarn lock dependency hierarchy x bootstrap tgz vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap x before and x beta before beta xss is possible in the data target attribute a different vulnerability than cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails in bootstrap x before and x beta before beta xss is possible in the data target attribute a different vulnerability than cve vulnerabilityurl
0
11,063
13,894,765,695
IssuesEvent
2020-10-19 15:03:40
cncf/cnf-conformance
https://api.github.com/repos/cncf/cnf-conformance
opened
Implementation: Move build process to CircleCI from Travis CI
process sprint18
### Implementation: Move build process to CircleCI from Travis CI Short Description: - PoC was completed in #428 - this issue will cover the remaining implementation steps to start using CircleCI w/ the cnf-conformance GitHub repo ### Documentation Tasks: - [ ] Add comment suggesting updates as needed for: - [ ] the [test categories markdown](https://github.com/cncf/cnf-conformance/blob/master/TEST-CATEGORIES.md) - [ ] the [pseudo code markdown](https://github.com/cncf/cnf-conformance/blob/master/PSEUDO-CODE.md) - [ ] slide content updates, LINK_TO_UPDATES - [ ] the [README](https://github.com/cncf/cnf-conformance/blob/master/README.md) ### GitHub Repo Tasks: - [ ] TBD - [ ] Tag 1 or more people to peer review ### QA tasks Dev Review: - [ ] walk through A/C - [ ] do you get the expected result? - [ ] if yes, - [ ] move to `Needs Peer Review` column - [ ] create Pull Request and follow check list - [ ] Assign 1 or more people for peer review - [ ] if no, document what additional tasks will be needed Peer review: - [ ] walk through A/C - [ ] do you get the expected result? - [ ] if yes, - [ ] move to `Reviewer Approved` column - [ ] Approve pull request - [ ] if no, - [ ] document what did not go as expected, including error messages and screenshots (if possible) - [ ] Add comment to pull request - [ ] request changes to pull request
1.0
Implementation: Move build process to CircleCI from Travis CI - ### Implementation: Move build process to CircleCI from Travis CI Short Description: - PoC was completed in #428 - this issue will cover the remaining implementation steps to start using CircleCI w/ the cnf-conformance GitHub repo ### Documentation Tasks: - [ ] Add comment suggesting updates as needed for: - [ ] the [test categories markdown](https://github.com/cncf/cnf-conformance/blob/master/TEST-CATEGORIES.md) - [ ] the [pseudo code markdown](https://github.com/cncf/cnf-conformance/blob/master/PSEUDO-CODE.md) - [ ] slide content updates, LINK_TO_UPDATES - [ ] the [README](https://github.com/cncf/cnf-conformance/blob/master/README.md) ### GitHub Repo Tasks: - [ ] TBD - [ ] Tag 1 or more people to peer review ### QA tasks Dev Review: - [ ] walk through A/C - [ ] do you get the expected result? - [ ] if yes, - [ ] move to `Needs Peer Review` column - [ ] create Pull Request and follow check list - [ ] Assign 1 or more people for peer review - [ ] if no, document what additional tasks will be needed Peer review: - [ ] walk through A/C - [ ] do you get the expected result? - [ ] if yes, - [ ] move to `Reviewer Approved` column - [ ] Approve pull request - [ ] if no, - [ ] document what did not go as expected, including error messages and screenshots (if possible) - [ ] Add comment to pull request - [ ] request changes to pull request
process
implementation move build process to circleci from travis ci implementation move build process to circleci from travis ci short description poc was completed in this issue will cover the remaining implementation steps to start using circleci w the cnf conformance github repo documentation tasks add comment suggesting updates as needed for the the slide content updates link to updates the github repo tasks tbd tag or more people to peer review qa tasks dev review walk through a c do you get the expected result if yes move to needs peer review column create pull request and follow check list assign or more people for peer review if no document what additional tasks will be needed peer review walk through a c do you get the expected result if yes move to reviewer approved column approve pull request if no document what did not go as expected including error messages and screenshots if possible add comment to pull request request changes to pull request
1
306,393
9,392,635,104
IssuesEvent
2019-04-07 03:04:06
tra38/Paranoia_Super_Mission_Generator
https://api.github.com/repos/tra38/Paranoia_Super_Mission_Generator
opened
Create a machine learning algorithm to generate missions
low-priority
**User Story (MVP)** - Use should be able to run a neural network trained on "synthetic data" (missions generated by PARANOIA Super Mission Generator) to generate human-readable missions. **Post-MVP** - Neural network is to be trained on published PARANOIA missions instead of "synthetic data". Doing this will require getting permissions from Mongoose Publishing first. **Notes** - The reason why this issue is "low-priority" is that we have already gotten a good enough system for generating PARANOIA missions (aside from a few minor issues like 'complications'). Put it frankly, there is very little reason to add neural networks into the equation. Templates work fine. The reason why this issue even exists though is the realization that neural networks will wind up being the future of text generation. OpenAI recently wrote [a blog post in February 14th 2019](https://openai.com/blog/better-language-models/) about a neural network that can generate text, and the generated text (though hand-selected) are miles above what I would expect a neural network to generate. The public text generator that they released was also pretty decent as well. Even though there are still subtle flaws in the neural network they have right now, those subtle flaws can be fixed given enough time and resources. OpenAI have demonstrated what is possible, and what is possible will wind up being *inevitable*. gwern wrote [a blog post](https://www.gwern.net/GPT-2) explaining how he was able to use OpenAI's public text generator (nicknamed GPT-2-small) to create 19th-century poetry generators (GPT-2-poetry and GPT-2-poetry-prefix). This same sort of process could also be used to generate PARANOIA missions. Thus, if one wants to stay on the cutting-edge of natural language generation, I need to eventually master gwern's process. Of course, I don't need to do so immediately - our current generator works fine for now. But I do need to start preparing for the future. **Priority** - Low <hr> "The bitter lesson is based on the historical observations that 1) AI researchers have often tried to build knowledge into their agents, 2) this always helps in the short term, and is personally satisfying to the researcher, but 3) in the long run it plateaus and even inhibits further progress, and 4) breakthrough progress eventually arrives by an opposing approach based on scaling computation by search and learning. The eventual success is tinged with bitterness, and often incompletely digested, because it is success over a favored, human-centric approach."---["The Bitter Lesson"](http://www.incompleteideas.net/IncIdeas/BitterLesson.html) Note though that "The Bitter Lesson" assumes that (a) we do have access to increased computation (which is doubtful, considering the end of Moore's Law), and (b) we're willing to spend tons of money to use that increased computation (which is also doubtful). So even if "search and learning" winds up being superior to knowledge-based approaches, knowledge-based approaches may be more cost-effective and worthwhile in the short-term.
1.0
Create a machine learning algorithm to generate missions - **User Story (MVP)** - Use should be able to run a neural network trained on "synthetic data" (missions generated by PARANOIA Super Mission Generator) to generate human-readable missions. **Post-MVP** - Neural network is to be trained on published PARANOIA missions instead of "synthetic data". Doing this will require getting permissions from Mongoose Publishing first. **Notes** - The reason why this issue is "low-priority" is that we have already gotten a good enough system for generating PARANOIA missions (aside from a few minor issues like 'complications'). Put it frankly, there is very little reason to add neural networks into the equation. Templates work fine. The reason why this issue even exists though is the realization that neural networks will wind up being the future of text generation. OpenAI recently wrote [a blog post in February 14th 2019](https://openai.com/blog/better-language-models/) about a neural network that can generate text, and the generated text (though hand-selected) are miles above what I would expect a neural network to generate. The public text generator that they released was also pretty decent as well. Even though there are still subtle flaws in the neural network they have right now, those subtle flaws can be fixed given enough time and resources. OpenAI have demonstrated what is possible, and what is possible will wind up being *inevitable*. gwern wrote [a blog post](https://www.gwern.net/GPT-2) explaining how he was able to use OpenAI's public text generator (nicknamed GPT-2-small) to create 19th-century poetry generators (GPT-2-poetry and GPT-2-poetry-prefix). This same sort of process could also be used to generate PARANOIA missions. Thus, if one wants to stay on the cutting-edge of natural language generation, I need to eventually master gwern's process. Of course, I don't need to do so immediately - our current generator works fine for now. But I do need to start preparing for the future. **Priority** - Low <hr> "The bitter lesson is based on the historical observations that 1) AI researchers have often tried to build knowledge into their agents, 2) this always helps in the short term, and is personally satisfying to the researcher, but 3) in the long run it plateaus and even inhibits further progress, and 4) breakthrough progress eventually arrives by an opposing approach based on scaling computation by search and learning. The eventual success is tinged with bitterness, and often incompletely digested, because it is success over a favored, human-centric approach."---["The Bitter Lesson"](http://www.incompleteideas.net/IncIdeas/BitterLesson.html) Note though that "The Bitter Lesson" assumes that (a) we do have access to increased computation (which is doubtful, considering the end of Moore's Law), and (b) we're willing to spend tons of money to use that increased computation (which is also doubtful). So even if "search and learning" winds up being superior to knowledge-based approaches, knowledge-based approaches may be more cost-effective and worthwhile in the short-term.
non_process
create a machine learning algorithm to generate missions user story mvp use should be able to run a neural network trained on synthetic data missions generated by paranoia super mission generator to generate human readable missions post mvp neural network is to be trained on published paranoia missions instead of synthetic data doing this will require getting permissions from mongoose publishing first notes the reason why this issue is low priority is that we have already gotten a good enough system for generating paranoia missions aside from a few minor issues like complications put it frankly there is very little reason to add neural networks into the equation templates work fine the reason why this issue even exists though is the realization that neural networks will wind up being the future of text generation openai recently wrote about a neural network that can generate text and the generated text though hand selected are miles above what i would expect a neural network to generate the public text generator that they released was also pretty decent as well even though there are still subtle flaws in the neural network they have right now those subtle flaws can be fixed given enough time and resources openai have demonstrated what is possible and what is possible will wind up being inevitable gwern wrote explaining how he was able to use openai s public text generator nicknamed gpt small to create century poetry generators gpt poetry and gpt poetry prefix this same sort of process could also be used to generate paranoia missions thus if one wants to stay on the cutting edge of natural language generation i need to eventually master gwern s process of course i don t need to do so immediately our current generator works fine for now but i do need to start preparing for the future priority low the bitter lesson is based on the historical observations that ai researchers have often tried to build knowledge into their agents this always helps in the short term and is personally satisfying to the researcher but in the long run it plateaus and even inhibits further progress and breakthrough progress eventually arrives by an opposing approach based on scaling computation by search and learning the eventual success is tinged with bitterness and often incompletely digested because it is success over a favored human centric approach note though that the bitter lesson assumes that a we do have access to increased computation which is doubtful considering the end of moore s law and b we re willing to spend tons of money to use that increased computation which is also doubtful so even if search and learning winds up being superior to knowledge based approaches knowledge based approaches may be more cost effective and worthwhile in the short term
0
7,036
10,196,724,175
IssuesEvent
2019-08-12 21:33:30
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Critical Turkish translation SQL failure
Database/Postgres Priority:P2 Query Processor Type:Bug
- Chrome Version 71.0.3578.98 (Official Build) (64-bit) - Ubuntu 18.04 up-to-date - Postgres 10 - Metabase v0.31.2 - Docker - Postgres 10 for Internal Database (Separate server from other postgres) - All system locales, containers locales are default to en_us, even the docker host Reproduce steps; 1. Do initial setup of Metabase 2. Change Language from Admin panel to Turkish 3. Logout and Login Description of the problem; After changing the language to Turkish and re-login; main page has **only** `ERROR: syntax error at or near "İNNER" Position: 63` error message. Therefore it's impossible to reach any homepage elements. Loses all functionality. Turkish language has 2 kind of I; one is I and other is İ. Be aware second one has dot on it. As the error dumps at the end provide, the query Metabase trying to execute has at least one Turkish capital i with dots(which is illegal character in the SQL queries as you might know). I've also search in the Turkish translation files to detect if somehow contains that `İNNER` word but no success. External database has almost no data, but just schemas/tables (A migrated Django db). Inner DB also is just created with the metabase (under single docker-compose) **Note that** With default English language, my setup works flawless. This problem only occurs with Turkish Language. I also confirm, after changing back to English, issue resolves. Metabase container error log; ``` 12-27 22:05:52 ERROR metabase.middleware :: GET /api/database 500 (5 ms) (3 DB calls). {:message "ERROR: syntax error at or near \"İNNER\"\n Position: 63", :type org.postgresql.util.PSQLException, :stacktrace ("org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2433)" "org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2178)" "org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:306)" "org.postgresql.jdbc.PgStatement.executeInternal(PgStatement.java:441)" "org.postgresql.jdbc.PgStatement.execute(PgStatement.java:365)" "org.postgresql.jdbc.PgPreparedStatement.executeWithFlags(PgPreparedStatement.java:155)" "org.postgresql.jdbc.PgPreparedStatement.executeQuery(PgPreparedStatement.java:118)" "com.mchange.v2.c3p0.impl.NewProxyPreparedStatement.executeQuery(NewProxyPreparedStatement.java:353)" "clojure.java.jdbc$execute_query_with_params.invokeStatic(jdbc.clj:1002)" "clojure.java.jdbc$execute_query_with_params.invoke(jdbc.clj:996)" "clojure.java.jdbc$db_query_with_resultset_STAR_.invokeStatic(jdbc.clj:1025)" "clojure.java.jdbc$db_query_with_resultset_STAR_.invoke(jdbc.clj:1005)" "clojure.java.jdbc$query.invokeStatic(jdbc.clj:1099)" "clojure.java.jdbc$query.invoke(jdbc.clj:1056)" "toucan.db$query.invokeStatic(db.clj:275)" "toucan.db$query.doInvoke(db.clj:271)" "clojure.lang.RestFn.invoke(RestFn.java:410)" "--> models.user$permissions_set.invokeStatic(user.clj:237)" "models.user$permissions_set.invoke(user.clj:229)" "middleware$bind_current_user$fn__56179$fn__56182.invoke(middleware.clj:136)" "models.interface$current_user_permissions_set.invokeStatic(interface.clj:219)" "models.interface$current_user_permissions_set.invoke(interface.clj:219)" "models.interface$make_perms_check_fn$_has_perms_QMARK___23020.invoke(interface.clj:234)" "models.interface$make_perms_check_fn$_has_perms_QMARK___23020.invoke(interface.clj:232)" "models.interface$fn__22970$G__22963__22979.invoke(interface.clj:181)" "api.database$dbs_list.invokeStatic(database.clj:138)" "api.database$dbs_list.invoke(database.clj:137)" "api.database$fn__47920.invokeStatic(database.clj:150)" "api.database$fn__47920.invoke(database.clj:143)" "middleware$enforce_authentication$fn__56174.invoke(middleware.clj:113)" "api.routes$fn__56326.invokeStatic(routes.clj:65)" "api.routes$fn__56326.invoke(routes.clj:65)" "routes$fn__56409$fn__56410.doInvoke(routes.clj:108)" "routes$fn__56409.invokeStatic(routes.clj:103)" "routes$fn__56409.invoke(routes.clj:103)" "middleware$catch_api_exceptions$fn__56309.invoke(middleware.clj:436)" "middleware$log_api_call$fn__56287$fn__56289.invoke(middleware.clj:364)" "middleware$log_api_call$fn__56287.invoke(middleware.clj:363)" "middleware$add_security_headers$fn__56229.invoke(middleware.clj:252)" "core$wrap_streamed_json_response$fn__62673.invoke(core.clj:67)" "middleware$bind_current_user$fn__56179.invoke(middleware.clj:137)" "middleware$maybe_set_site_url$fn__56239.invoke(middleware.clj:290)" "middleware$add_content_type$fn__56232.invoke(middleware.clj:262)"), :sql-exception-chain ["PSQLException:" "Message: ERROR: syntax error at or near \"İNNER\"" "Position: 63" "SQLState: 42601" "Error Code: 0"]} ``` Metabase inner database container log; ``` 2018-12-27 19:24:00.876 UTC [32] ERROR: syntax error at or near "İNNER" at character 63 2018-12-27 19:24:00.876 UTC [32] STATEMENT: SELECT "p"."object" FROM "permissions_group_membership" "pgm" İNNER JOIN "permissions_group" "pg" ON "pgm"."group_id" = "pg"."id" İNNER JOIN "permissions" "p" ON "p"."group_id" = "pg"."id" WHERE "pgm"."user_id" = 1 ```
1.0
Critical Turkish translation SQL failure - - Chrome Version 71.0.3578.98 (Official Build) (64-bit) - Ubuntu 18.04 up-to-date - Postgres 10 - Metabase v0.31.2 - Docker - Postgres 10 for Internal Database (Separate server from other postgres) - All system locales, containers locales are default to en_us, even the docker host Reproduce steps; 1. Do initial setup of Metabase 2. Change Language from Admin panel to Turkish 3. Logout and Login Description of the problem; After changing the language to Turkish and re-login; main page has **only** `ERROR: syntax error at or near "İNNER" Position: 63` error message. Therefore it's impossible to reach any homepage elements. Loses all functionality. Turkish language has 2 kind of I; one is I and other is İ. Be aware second one has dot on it. As the error dumps at the end provide, the query Metabase trying to execute has at least one Turkish capital i with dots(which is illegal character in the SQL queries as you might know). I've also search in the Turkish translation files to detect if somehow contains that `İNNER` word but no success. External database has almost no data, but just schemas/tables (A migrated Django db). Inner DB also is just created with the metabase (under single docker-compose) **Note that** With default English language, my setup works flawless. This problem only occurs with Turkish Language. I also confirm, after changing back to English, issue resolves. Metabase container error log; ``` 12-27 22:05:52 ERROR metabase.middleware :: GET /api/database 500 (5 ms) (3 DB calls). {:message "ERROR: syntax error at or near \"İNNER\"\n Position: 63", :type org.postgresql.util.PSQLException, :stacktrace ("org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2433)" "org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2178)" "org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:306)" "org.postgresql.jdbc.PgStatement.executeInternal(PgStatement.java:441)" "org.postgresql.jdbc.PgStatement.execute(PgStatement.java:365)" "org.postgresql.jdbc.PgPreparedStatement.executeWithFlags(PgPreparedStatement.java:155)" "org.postgresql.jdbc.PgPreparedStatement.executeQuery(PgPreparedStatement.java:118)" "com.mchange.v2.c3p0.impl.NewProxyPreparedStatement.executeQuery(NewProxyPreparedStatement.java:353)" "clojure.java.jdbc$execute_query_with_params.invokeStatic(jdbc.clj:1002)" "clojure.java.jdbc$execute_query_with_params.invoke(jdbc.clj:996)" "clojure.java.jdbc$db_query_with_resultset_STAR_.invokeStatic(jdbc.clj:1025)" "clojure.java.jdbc$db_query_with_resultset_STAR_.invoke(jdbc.clj:1005)" "clojure.java.jdbc$query.invokeStatic(jdbc.clj:1099)" "clojure.java.jdbc$query.invoke(jdbc.clj:1056)" "toucan.db$query.invokeStatic(db.clj:275)" "toucan.db$query.doInvoke(db.clj:271)" "clojure.lang.RestFn.invoke(RestFn.java:410)" "--> models.user$permissions_set.invokeStatic(user.clj:237)" "models.user$permissions_set.invoke(user.clj:229)" "middleware$bind_current_user$fn__56179$fn__56182.invoke(middleware.clj:136)" "models.interface$current_user_permissions_set.invokeStatic(interface.clj:219)" "models.interface$current_user_permissions_set.invoke(interface.clj:219)" "models.interface$make_perms_check_fn$_has_perms_QMARK___23020.invoke(interface.clj:234)" "models.interface$make_perms_check_fn$_has_perms_QMARK___23020.invoke(interface.clj:232)" "models.interface$fn__22970$G__22963__22979.invoke(interface.clj:181)" "api.database$dbs_list.invokeStatic(database.clj:138)" "api.database$dbs_list.invoke(database.clj:137)" "api.database$fn__47920.invokeStatic(database.clj:150)" "api.database$fn__47920.invoke(database.clj:143)" "middleware$enforce_authentication$fn__56174.invoke(middleware.clj:113)" "api.routes$fn__56326.invokeStatic(routes.clj:65)" "api.routes$fn__56326.invoke(routes.clj:65)" "routes$fn__56409$fn__56410.doInvoke(routes.clj:108)" "routes$fn__56409.invokeStatic(routes.clj:103)" "routes$fn__56409.invoke(routes.clj:103)" "middleware$catch_api_exceptions$fn__56309.invoke(middleware.clj:436)" "middleware$log_api_call$fn__56287$fn__56289.invoke(middleware.clj:364)" "middleware$log_api_call$fn__56287.invoke(middleware.clj:363)" "middleware$add_security_headers$fn__56229.invoke(middleware.clj:252)" "core$wrap_streamed_json_response$fn__62673.invoke(core.clj:67)" "middleware$bind_current_user$fn__56179.invoke(middleware.clj:137)" "middleware$maybe_set_site_url$fn__56239.invoke(middleware.clj:290)" "middleware$add_content_type$fn__56232.invoke(middleware.clj:262)"), :sql-exception-chain ["PSQLException:" "Message: ERROR: syntax error at or near \"İNNER\"" "Position: 63" "SQLState: 42601" "Error Code: 0"]} ``` Metabase inner database container log; ``` 2018-12-27 19:24:00.876 UTC [32] ERROR: syntax error at or near "İNNER" at character 63 2018-12-27 19:24:00.876 UTC [32] STATEMENT: SELECT "p"."object" FROM "permissions_group_membership" "pgm" İNNER JOIN "permissions_group" "pg" ON "pgm"."group_id" = "pg"."id" İNNER JOIN "permissions" "p" ON "p"."group_id" = "pg"."id" WHERE "pgm"."user_id" = 1 ```
process
critical turkish translation sql failure chrome version official build bit ubuntu up to date postgres metabase docker postgres for internal database separate server from other postgres all system locales containers locales are default to en us even the docker host reproduce steps do initial setup of metabase change language from admin panel to turkish logout and login description of the problem after changing the language to turkish and re login main page has only error syntax error at or near i̇nner position error message therefore it s impossible to reach any homepage elements loses all functionality turkish language has kind of i one is i and other is i̇ be aware second one has dot on it as the error dumps at the end provide the query metabase trying to execute has at least one turkish capital i with dots which is illegal character in the sql queries as you might know i ve also search in the turkish translation files to detect if somehow contains that i̇nner word but no success external database has almost no data but just schemas tables a migrated django db inner db also is just created with the metabase under single docker compose note that with default english language my setup works flawless this problem only occurs with turkish language i also confirm after changing back to english issue resolves metabase container error log error metabase middleware get api database ms db calls message error syntax error at or near i̇nner n position type org postgresql util psqlexception stacktrace org postgresql core queryexecutorimpl receiveerrorresponse queryexecutorimpl java org postgresql core queryexecutorimpl processresults queryexecutorimpl java org postgresql core queryexecutorimpl execute queryexecutorimpl java org postgresql jdbc pgstatement executeinternal pgstatement java org postgresql jdbc pgstatement execute pgstatement java org postgresql jdbc pgpreparedstatement executewithflags pgpreparedstatement java org postgresql jdbc pgpreparedstatement executequery pgpreparedstatement java com mchange impl newproxypreparedstatement executequery newproxypreparedstatement java clojure java jdbc execute query with params invokestatic jdbc clj clojure java jdbc execute query with params invoke jdbc clj clojure java jdbc db query with resultset star invokestatic jdbc clj clojure java jdbc db query with resultset star invoke jdbc clj clojure java jdbc query invokestatic jdbc clj clojure java jdbc query invoke jdbc clj toucan db query invokestatic db clj toucan db query doinvoke db clj clojure lang restfn invoke restfn java models user permissions set invokestatic user clj models user permissions set invoke user clj middleware bind current user fn fn invoke middleware clj models interface current user permissions set invokestatic interface clj models interface current user permissions set invoke interface clj models interface make perms check fn has perms qmark invoke interface clj models interface make perms check fn has perms qmark invoke interface clj models interface fn g invoke interface clj api database dbs list invokestatic database clj api database dbs list invoke database clj api database fn invokestatic database clj api database fn invoke database clj middleware enforce authentication fn invoke middleware clj api routes fn invokestatic routes clj api routes fn invoke routes clj routes fn fn doinvoke routes clj routes fn invokestatic routes clj routes fn invoke routes clj middleware catch api exceptions fn invoke middleware clj middleware log api call fn fn invoke middleware clj middleware log api call fn invoke middleware clj middleware add security headers fn invoke middleware clj core wrap streamed json response fn invoke core clj middleware bind current user fn invoke middleware clj middleware maybe set site url fn invoke middleware clj middleware add content type fn invoke middleware clj sql exception chain metabase inner database container log utc error syntax error at or near i̇nner at character utc statement select p object from permissions group membership pgm i̇nner join permissions group pg on pgm group id pg id i̇nner join permissions p on p group id pg id where pgm user id
1
6,434
9,534,549,726
IssuesEvent
2019-04-30 02:09:16
de-ai/designengine.ai
https://api.github.com/repos/de-ai/designengine.ai
closed
Missing Sketch Fonts
Good First Issue New feature Processing Render
Determine from Sketch plugin if any fonts used aren't installed. Notify on render page for author to upload font files. - [x] Figure out how to identify missing fonts in plugin - [x] Send list of missing font names to db - [x] Notify user during processing that x fonts are missing - [ ] Upload drop + API call to record font w/ project & user - [ ] Download & install fonts on plugin machine, restart Sketch & re-open previous file - [ ] Reprocess if fonts found, otherwise send updated message to db/frontend - [ ] Notify uploader when new processing finished
1.0
Missing Sketch Fonts - Determine from Sketch plugin if any fonts used aren't installed. Notify on render page for author to upload font files. - [x] Figure out how to identify missing fonts in plugin - [x] Send list of missing font names to db - [x] Notify user during processing that x fonts are missing - [ ] Upload drop + API call to record font w/ project & user - [ ] Download & install fonts on plugin machine, restart Sketch & re-open previous file - [ ] Reprocess if fonts found, otherwise send updated message to db/frontend - [ ] Notify uploader when new processing finished
process
missing sketch fonts determine from sketch plugin if any fonts used aren t installed notify on render page for author to upload font files figure out how to identify missing fonts in plugin send list of missing font names to db notify user during processing that x fonts are missing upload drop api call to record font w project user download install fonts on plugin machine restart sketch re open previous file reprocess if fonts found otherwise send updated message to db frontend notify uploader when new processing finished
1
16,589
21,638,659,002
IssuesEvent
2022-05-05 16:25:19
shirou/gopsutil
https://api.github.com/repos/shirou/gopsutil
closed
question: `(*process.Process).CPUPercent()` returning values above 100, what does this number represent?
question package:process
when printing `(*process.Process).CPUPercent()` of a process that is pretty heavy i get numbers in the range of 500 or 600 (sometimes more). _(When the tested process finishes the heavy calculations, this number returns to around 2 or 3.)_ Why is that? should't it be a precentage with maximum value of 100?
1.0
question: `(*process.Process).CPUPercent()` returning values above 100, what does this number represent? - when printing `(*process.Process).CPUPercent()` of a process that is pretty heavy i get numbers in the range of 500 or 600 (sometimes more). _(When the tested process finishes the heavy calculations, this number returns to around 2 or 3.)_ Why is that? should't it be a precentage with maximum value of 100?
process
question process process cpupercent returning values above what does this number represent when printing process process cpupercent of a process that is pretty heavy i get numbers in the range of or sometimes more when the tested process finishes the heavy calculations this number returns to around or why is that should t it be a precentage with maximum value of
1
142,360
5,474,514,149
IssuesEvent
2017-03-11 01:18:23
dgore7/knowledge-management
https://api.github.com/repos/dgore7/knowledge-management
closed
Automatic Summarization of Text Documents
enhancement low priority
Provide a system which parses text documents and calls a relevant function which summarizes them. Algorithm commonly used is SMMRY. Information on that/relevant API can be found here: [http://smmry.com/about](http://smmry.com/about). Key points: - Summarizes documents and stores that "value" in the file database in a corresponding field - Create relevant connecting code in searching/viewing of information - Uploading function calls summarization code to ensure a summary is made immediately - Appropriate UI and respective code for the 100 request limit imposed by SMMRY upon free API keys - UI for entry of SMMRY API key along with appropriate link to their sign-up page Further Enhancements: - Automatic tag generation based upon the summary
1.0
Automatic Summarization of Text Documents - Provide a system which parses text documents and calls a relevant function which summarizes them. Algorithm commonly used is SMMRY. Information on that/relevant API can be found here: [http://smmry.com/about](http://smmry.com/about). Key points: - Summarizes documents and stores that "value" in the file database in a corresponding field - Create relevant connecting code in searching/viewing of information - Uploading function calls summarization code to ensure a summary is made immediately - Appropriate UI and respective code for the 100 request limit imposed by SMMRY upon free API keys - UI for entry of SMMRY API key along with appropriate link to their sign-up page Further Enhancements: - Automatic tag generation based upon the summary
non_process
automatic summarization of text documents provide a system which parses text documents and calls a relevant function which summarizes them algorithm commonly used is smmry information on that relevant api can be found here key points summarizes documents and stores that value in the file database in a corresponding field create relevant connecting code in searching viewing of information uploading function calls summarization code to ensure a summary is made immediately appropriate ui and respective code for the request limit imposed by smmry upon free api keys ui for entry of smmry api key along with appropriate link to their sign up page further enhancements automatic tag generation based upon the summary
0
4,178
7,111,637,933
IssuesEvent
2018-01-17 14:48:47
itsyouonline/identityserver
https://api.github.com/repos/itsyouonline/identityserver
closed
Arrived back at step 1 after registration (Safari on iOS 8)
process_wontfix type_bug
I went through steps 1, 2 and 3. Clicked login and ended back at step 1. Happened on iPhone 5S simulator with iOS 8.1 <img width="432" alt="screen shot 2017-10-07 at 17 38 09" src="https://user-images.githubusercontent.com/17762105/31309442-85fd9742-ab86-11e7-9ce0-d7ecc42a2b3d.png">
1.0
Arrived back at step 1 after registration (Safari on iOS 8) - I went through steps 1, 2 and 3. Clicked login and ended back at step 1. Happened on iPhone 5S simulator with iOS 8.1 <img width="432" alt="screen shot 2017-10-07 at 17 38 09" src="https://user-images.githubusercontent.com/17762105/31309442-85fd9742-ab86-11e7-9ce0-d7ecc42a2b3d.png">
process
arrived back at step after registration safari on ios i went through steps and clicked login and ended back at step happened on iphone simulator with ios img width alt screen shot at src
1
212,971
16,505,908,419
IssuesEvent
2021-05-25 19:16:43
giridhar196/pm-notes
https://api.github.com/repos/giridhar196/pm-notes
closed
Write notes for Class 05/13/2021
documentation good first issue
Gather notes for today's class and submit them under the documents after the class.
1.0
Write notes for Class 05/13/2021 - Gather notes for today's class and submit them under the documents after the class.
non_process
write notes for class gather notes for today s class and submit them under the documents after the class
0
11,364
14,175,779,640
IssuesEvent
2020-11-12 22:12:32
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
How to trigger CD after CI
Pri1 devops-cicd-process/tech devops/prod doc-enhancement
I follow the document to create two pipelines, build and release. But whenever I commit change to a branch, it triggers both pipelines. How i can trigger release only after build pipeline completed? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: ee4ec9d0-e0d5-4fb4-7c3e-b84abfa290c2 * Version Independent ID: 3e2b80d9-30e5-0c48-49f0-4fcdfedf5eee * Content: [Resources - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/resources?view=azure-devops&tabs=example#resources-pipelines) * Content Source: [docs/pipelines/process/resources.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/resources.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
How to trigger CD after CI - I follow the document to create two pipelines, build and release. But whenever I commit change to a branch, it triggers both pipelines. How i can trigger release only after build pipeline completed? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: ee4ec9d0-e0d5-4fb4-7c3e-b84abfa290c2 * Version Independent ID: 3e2b80d9-30e5-0c48-49f0-4fcdfedf5eee * Content: [Resources - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/resources?view=azure-devops&tabs=example#resources-pipelines) * Content Source: [docs/pipelines/process/resources.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/resources.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
how to trigger cd after ci i follow the document to create two pipelines build and release but whenever i commit change to a branch it triggers both pipelines how i can trigger release only after build pipeline completed document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
114,312
24,583,615,506
IssuesEvent
2022-10-13 17:39:37
trezor/trezor-suite
https://api.github.com/repos/trezor/trezor-suite
closed
suite-native: Dependencies - upgrade React native
dependencies code mobile-app
According to this [issue](https://github.com/trezor/trezor-suite/issues/6414) (upgrade React in the whole monorepo to version 18) and [PR](https://github.com/trezor/trezor-suite/pull/6437) - we want to upgrade RN to the newest version.
1.0
suite-native: Dependencies - upgrade React native - According to this [issue](https://github.com/trezor/trezor-suite/issues/6414) (upgrade React in the whole monorepo to version 18) and [PR](https://github.com/trezor/trezor-suite/pull/6437) - we want to upgrade RN to the newest version.
non_process
suite native dependencies upgrade react native according to this upgrade react in the whole monorepo to version and we want to upgrade rn to the newest version
0
572
3,037,319,673
IssuesEvent
2015-08-06 16:25:42
pwittchen/prefser
https://api.github.com/repos/pwittchen/prefser
closed
Review & Update README.md
release process
Review & Update README.md after changes in the library functionality and its API.
1.0
Review & Update README.md - Review & Update README.md after changes in the library functionality and its API.
process
review update readme md review update readme md after changes in the library functionality and its api
1
22,028
30,543,376,868
IssuesEvent
2023-07-20 00:20:50
ReMobidyc/ReMobidyc
https://api.github.com/repos/ReMobidyc/ReMobidyc
closed
modelpath should not contain the ":" char
bug processor
The ":" char is used for the drive letter delimiter on the Windows platform. We have to avoid using it.
1.0
modelpath should not contain the ":" char - The ":" char is used for the drive letter delimiter on the Windows platform. We have to avoid using it.
process
modelpath should not contain the char the char is used for the drive letter delimiter on the windows platform we have to avoid using it
1
517,599
15,016,823,132
IssuesEvent
2021-02-01 10:04:09
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
Sending access_token through /oauth2/userinfo API Body in WSO2 5.7.0
Affected/5.7.0 Complexity/Medium Component/Kernel Priority/Highest Severity/Critical WUM bug
In IS-5.3.0 we can send "access_token" through API Body for "/oauth2/userinfo" API. But in IS-5.7.0 if we enabled the correlation value as follows then sending access_token through API body will fail. ``` <Valve className="org.wso2.carbon.tomcat.ext.valves.RequestCorrelationIdValve" headerToCorrelationIdMapping="{'activityid':'Correlation-ID'}" queryToCorrelationIdMapping="{'RelayState':'Correlation-ID'}"/> ``` The reason for this behavior is through correlation valve we are reading the request body\[1] hence when we trying to read the request body again\[2] it is failing because we can only read the request body once. *Steps to reproduce* 1. Add the following property in catalina-server.xml file. 2. Retrieve access token and invoke user info by sending the access token through request body. 3. You won't get any responses. 4. comment the line "Enumeration<String> parameterNames = httpServletRequest.getParameterNames();" in \[1] and rebuild the jar and tried the same flow. 5. You will get the user info Hence we could decide that we need to modify the code \[1] in a way that to only read the query parameters and not to read the request body. For example use *request .getQueryParams*. We need to properly test this as well. \[1] https://github.com/wso2/carbon-kernel/blob/4.4.x/core/org.wso2.carbon.tomcat.ext/src/main/java/org/wso2/carbon/tomcat/ext/valves/RequestCorrelationIdValve.java#L224 \[2] https://github.com/wso2-extensions/identity-inbound-auth-oauth/blob/master/components/org.wso2.carbon.identity.oauth.endpoint/src/main/java/org/wso2/carbon/identity/oauth/endpoint/user/impl/UserInforRequestDefaultValidator.java#L57-L64
1.0
Sending access_token through /oauth2/userinfo API Body in WSO2 5.7.0 - In IS-5.3.0 we can send "access_token" through API Body for "/oauth2/userinfo" API. But in IS-5.7.0 if we enabled the correlation value as follows then sending access_token through API body will fail. ``` <Valve className="org.wso2.carbon.tomcat.ext.valves.RequestCorrelationIdValve" headerToCorrelationIdMapping="{'activityid':'Correlation-ID'}" queryToCorrelationIdMapping="{'RelayState':'Correlation-ID'}"/> ``` The reason for this behavior is through correlation valve we are reading the request body\[1] hence when we trying to read the request body again\[2] it is failing because we can only read the request body once. *Steps to reproduce* 1. Add the following property in catalina-server.xml file. 2. Retrieve access token and invoke user info by sending the access token through request body. 3. You won't get any responses. 4. comment the line "Enumeration<String> parameterNames = httpServletRequest.getParameterNames();" in \[1] and rebuild the jar and tried the same flow. 5. You will get the user info Hence we could decide that we need to modify the code \[1] in a way that to only read the query parameters and not to read the request body. For example use *request .getQueryParams*. We need to properly test this as well. \[1] https://github.com/wso2/carbon-kernel/blob/4.4.x/core/org.wso2.carbon.tomcat.ext/src/main/java/org/wso2/carbon/tomcat/ext/valves/RequestCorrelationIdValve.java#L224 \[2] https://github.com/wso2-extensions/identity-inbound-auth-oauth/blob/master/components/org.wso2.carbon.identity.oauth.endpoint/src/main/java/org/wso2/carbon/identity/oauth/endpoint/user/impl/UserInforRequestDefaultValidator.java#L57-L64
non_process
sending access token through userinfo api body in in is we can send access token through api body for userinfo api but in is if we enabled the correlation value as follows then sending access token through api body will fail valve classname org carbon tomcat ext valves requestcorrelationidvalve headertocorrelationidmapping activityid correlation id querytocorrelationidmapping relaystate correlation id the reason for this behavior is through correlation valve we are reading the request body hence when we trying to read the request body again it is failing because we can only read the request body once steps to reproduce add the following property in catalina server xml file retrieve access token and invoke user info by sending the access token through request body you won t get any responses comment the line enumeration parameternames httpservletrequest getparameternames in and rebuild the jar and tried the same flow you will get the user info hence we could decide that we need to modify the code in a way that to only read the query parameters and not to read the request body for example use request getqueryparams we need to properly test this as well
0
22,702
32,017,867,651
IssuesEvent
2023-09-22 00:13:17
TableRise/tablerise-backend
https://api.github.com/repos/TableRise/tablerise-backend
closed
Montar entidade de usuário
test-process
Agora que já temos acesso as informações do usuário, montaremos a entidade a ser persistida no banco de dados. Objeto de exemplo ( user ) ``` { "email": "some@email.com", "password": "oauth", "nickname": "holder", "tag": "1447", "picture": "http://imgur.com", "createdAt": "timestamp", "updatedAt": "timestamp", } ``` Objeto de exemplo ( user info ) ``` { "userId": "123", "firstName": "John", "lastName": "Doe", "pronoum": "he/his", "secretQuestion": { "question": "is true?", "answer": "yes" }, "gameInfo": { "campaigns": [ '123' ], "characters": [ '321' ], "badges": [ '213' ] }, "biography": "a long long time ago...", "role": "user" } ``` Deve ser criado um método onde esses dois objetos são montados com todos valores null e depois os objetos criados devem ser redefinidos campo por campo em código com as informações vindas do serviço externo, as informações que o serviço não prover devem permancer como null. Deve ser possivel passar uma flag para este método onde caso seja chamado com essa flag ele gerará dois campos com valor 'oauth' ao invés de null, os campos são: - password - secretQuestion - [ ] Desenvolver método que gera uma entidade de usuário - [ ] Redefinir campos null com novas informações do usuário providas pela Google API - [ ] Persistir entidade no banco de dados - [ ] Retornar id do usuário e uma mensagem de sucesso
1.0
Montar entidade de usuário - Agora que já temos acesso as informações do usuário, montaremos a entidade a ser persistida no banco de dados. Objeto de exemplo ( user ) ``` { "email": "some@email.com", "password": "oauth", "nickname": "holder", "tag": "1447", "picture": "http://imgur.com", "createdAt": "timestamp", "updatedAt": "timestamp", } ``` Objeto de exemplo ( user info ) ``` { "userId": "123", "firstName": "John", "lastName": "Doe", "pronoum": "he/his", "secretQuestion": { "question": "is true?", "answer": "yes" }, "gameInfo": { "campaigns": [ '123' ], "characters": [ '321' ], "badges": [ '213' ] }, "biography": "a long long time ago...", "role": "user" } ``` Deve ser criado um método onde esses dois objetos são montados com todos valores null e depois os objetos criados devem ser redefinidos campo por campo em código com as informações vindas do serviço externo, as informações que o serviço não prover devem permancer como null. Deve ser possivel passar uma flag para este método onde caso seja chamado com essa flag ele gerará dois campos com valor 'oauth' ao invés de null, os campos são: - password - secretQuestion - [ ] Desenvolver método que gera uma entidade de usuário - [ ] Redefinir campos null com novas informações do usuário providas pela Google API - [ ] Persistir entidade no banco de dados - [ ] Retornar id do usuário e uma mensagem de sucesso
process
montar entidade de usuário agora que já temos acesso as informações do usuário montaremos a entidade a ser persistida no banco de dados objeto de exemplo user email some email com password oauth nickname holder tag picture createdat timestamp updatedat timestamp objeto de exemplo user info userid firstname john lastname doe pronoum he his secretquestion question is true answer yes gameinfo campaigns characters badges biography a long long time ago role user deve ser criado um método onde esses dois objetos são montados com todos valores null e depois os objetos criados devem ser redefinidos campo por campo em código com as informações vindas do serviço externo as informações que o serviço não prover devem permancer como null deve ser possivel passar uma flag para este método onde caso seja chamado com essa flag ele gerará dois campos com valor oauth ao invés de null os campos são password secretquestion desenvolver método que gera uma entidade de usuário redefinir campos null com novas informações do usuário providas pela google api persistir entidade no banco de dados retornar id do usuário e uma mensagem de sucesso
1
78,324
3,509,571,461
IssuesEvent
2016-01-08 23:32:21
OregonCore/OregonCore
https://api.github.com/repos/OregonCore/OregonCore
opened
[warden] Making character disconnects client (BB #1096)
Category: Client Freeze migrated Priority: Medium Type: Bug
This issue was migrated from bitbucket. **Original Reporter:** Dikkedeur **Original Date:** 14.08.2015 18:13:14 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** new **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/1096 <hr> Rev: c07fe354bafbd8c689bf67de1106c8c31ca1ef93 problem: When somebody makes a new character they get booted/kicked from the server, in the system.err log you see following error: [Char]^[[01;35m[Warden] ^[[00;35mTIMING CHECK FAIL result 0x00^[[0m This is totally random but 1 person claims he has it always. He is running windows 8.1 and think it might be due that. so most of the people dont have this problem..
1.0
[warden] Making character disconnects client (BB #1096) - This issue was migrated from bitbucket. **Original Reporter:** Dikkedeur **Original Date:** 14.08.2015 18:13:14 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** new **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/1096 <hr> Rev: c07fe354bafbd8c689bf67de1106c8c31ca1ef93 problem: When somebody makes a new character they get booted/kicked from the server, in the system.err log you see following error: [Char]^[[01;35m[Warden] ^[[00;35mTIMING CHECK FAIL result 0x00^[[0m This is totally random but 1 person claims he has it always. He is running windows 8.1 and think it might be due that. so most of the people dont have this problem..
non_process
making character disconnects client bb this issue was migrated from bitbucket original reporter dikkedeur original date gmt original priority major original type bug original state new direct link rev problem when somebody makes a new character they get booted kicked from the server in the system err log you see following error check fail result this is totally random but person claims he has it always he is running windows and think it might be due that so most of the people dont have this problem
0
749
3,222,820,084
IssuesEvent
2015-10-09 05:11:08
nodejs/node
https://api.github.com/repos/nodejs/node
closed
The process never ends when piping the stdin to a child process
child_process process
Initially I asked this on [StackOverflow](http://stackoverflow.com/q/31716784/1420197), but looks like a bug (https://github.com/joyent/node/issues/9190). Post follows: --- I'm using `spawn` to create a child process and pipe data: child process | parent process (main) --------------------------------------- stdout -----> process.stdout stderr -----> process.stderr stdin <----- process.stdin The problem is that when piping the `process.stdin` to the child process `stdin`, the main process is not ended when the child process is finished. The code looks like this (not a really good example because `ps` does not use `stdin` data, I guess): ```js var Spawn = require("child_process").spawn; var ps = Spawn("ps"); process.stdin.pipe(ps.stdin); ps.stdout.pipe(process.stdout); ps.stderr.pipe(process.stderr); ``` If I remove the `process.stdin.pipe(ps.stdin)` line, the main process is ended, but the `stdin` data is not piped anymore. Why isn't the main process ended when the `ps` child process is ended? How can I solve this problem? An ugly solution would be: ```js ps.on("close", process.exit.bind(process)); ``` I don't like this, because I don't really want to force the main process to be closed, but I want to be closed *naturally* (e.g. having `setTimeout(function(){}, 1000)` you wait 1000ms and then the process ends). ---- I tried to `ps.stdin.close()` in `ps.on("close", cb)`. It didn't work... :cry:
2.0
The process never ends when piping the stdin to a child process - Initially I asked this on [StackOverflow](http://stackoverflow.com/q/31716784/1420197), but looks like a bug (https://github.com/joyent/node/issues/9190). Post follows: --- I'm using `spawn` to create a child process and pipe data: child process | parent process (main) --------------------------------------- stdout -----> process.stdout stderr -----> process.stderr stdin <----- process.stdin The problem is that when piping the `process.stdin` to the child process `stdin`, the main process is not ended when the child process is finished. The code looks like this (not a really good example because `ps` does not use `stdin` data, I guess): ```js var Spawn = require("child_process").spawn; var ps = Spawn("ps"); process.stdin.pipe(ps.stdin); ps.stdout.pipe(process.stdout); ps.stderr.pipe(process.stderr); ``` If I remove the `process.stdin.pipe(ps.stdin)` line, the main process is ended, but the `stdin` data is not piped anymore. Why isn't the main process ended when the `ps` child process is ended? How can I solve this problem? An ugly solution would be: ```js ps.on("close", process.exit.bind(process)); ``` I don't like this, because I don't really want to force the main process to be closed, but I want to be closed *naturally* (e.g. having `setTimeout(function(){}, 1000)` you wait 1000ms and then the process ends). ---- I tried to `ps.stdin.close()` in `ps.on("close", cb)`. It didn't work... :cry:
process
the process never ends when piping the stdin to a child process initially i asked this on but looks like a bug post follows i m using spawn to create a child process and pipe data child process parent process main stdout process stdout stderr process stderr stdin process stdin the problem is that when piping the process stdin to the child process stdin the main process is not ended when the child process is finished the code looks like this not a really good example because ps does not use stdin data i guess js var spawn require child process spawn var ps spawn ps process stdin pipe ps stdin ps stdout pipe process stdout ps stderr pipe process stderr if i remove the process stdin pipe ps stdin line the main process is ended but the stdin data is not piped anymore why isn t the main process ended when the ps child process is ended how can i solve this problem an ugly solution would be js ps on close process exit bind process i don t like this because i don t really want to force the main process to be closed but i want to be closed naturally e g having settimeout function you wait and then the process ends i tried to ps stdin close in ps on close cb it didn t work cry
1
8,163
11,385,316,181
IssuesEvent
2020-01-29 10:50:52
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
remove comment GO:0080185
multi-species process
GO:0080185 JSON effector-mediated induction of plant hypersensitive response by symbiont A symbiont process whereby a molecule secreted by the symbiont activates plant effector-triggered immunity (ETI)... has the comment: Note that this term should probably be only used to annotate plant pathogens. Evidence of direct recognition of pathogen effectors in animals is currently lacking, see PMID:23411798. this comment is unnecessary because this is names and defined as a pathogen affecting plants specific term.
1.0
remove comment GO:0080185 - GO:0080185 JSON effector-mediated induction of plant hypersensitive response by symbiont A symbiont process whereby a molecule secreted by the symbiont activates plant effector-triggered immunity (ETI)... has the comment: Note that this term should probably be only used to annotate plant pathogens. Evidence of direct recognition of pathogen effectors in animals is currently lacking, see PMID:23411798. this comment is unnecessary because this is names and defined as a pathogen affecting plants specific term.
process
remove comment go go json effector mediated induction of plant hypersensitive response by symbiont a symbiont process whereby a molecule secreted by the symbiont activates plant effector triggered immunity eti has the comment note that this term should probably be only used to annotate plant pathogens evidence of direct recognition of pathogen effectors in animals is currently lacking see pmid this comment is unnecessary because this is names and defined as a pathogen affecting plants specific term
1
417,101
12,155,849,045
IssuesEvent
2020-04-25 14:54:53
wso2/product-is
https://api.github.com/repos/wso2/product-is
opened
[ISSUE] Password grant doesn't return configured sub claim in acces token
Affected/5.11.0-m12 Component/OIDC Priority/Highest Severity/Blocker Type/Bug
**Describe the issue:** With password grant, a self-contained access token doesn't return the confgured custom claim but id token contains the correct custom claim as the sub claim. **Reproduce steps** 1. Add a service provider with OIDC protocol, select JWT token issuer 2. Create a custom claim under local claim, create a oidc claim and mapped with the custom local claim. 3. Add oidc claim to OpenID scope 4. Select the custom claim as the requested claim and sub claim for the created service provider 5. Send an access token request using password grant with the scope openid 6. Decode the retrieved access token, idtoken. You will see sub claim value contains the custom claim value in id token but access token will contain the username as the sub claim. **How to reproduce:** <!-- Steps to reproduce the behavior. --> **Expected behavior:** <!-- A clear and concise description of what you expected to happen. --> **Environment information** (_Please complete the following information; remove any unnecessary fields_) **:** - Product Version: [e.g., IS 5.10.0, IS 5.9.0] - OS: [e.g., Windows, Linux, Mac] - Database: [e.g., MySQL, H2] - Userstore: [e.g., LDAP, JDBC] --- ### Optional Fields **Related issues:** <!-- Any related issues from this/other repositories--> **Suggested labels:** <!-- Only to be used by non-members -->
1.0
[ISSUE] Password grant doesn't return configured sub claim in acces token - **Describe the issue:** With password grant, a self-contained access token doesn't return the confgured custom claim but id token contains the correct custom claim as the sub claim. **Reproduce steps** 1. Add a service provider with OIDC protocol, select JWT token issuer 2. Create a custom claim under local claim, create a oidc claim and mapped with the custom local claim. 3. Add oidc claim to OpenID scope 4. Select the custom claim as the requested claim and sub claim for the created service provider 5. Send an access token request using password grant with the scope openid 6. Decode the retrieved access token, idtoken. You will see sub claim value contains the custom claim value in id token but access token will contain the username as the sub claim. **How to reproduce:** <!-- Steps to reproduce the behavior. --> **Expected behavior:** <!-- A clear and concise description of what you expected to happen. --> **Environment information** (_Please complete the following information; remove any unnecessary fields_) **:** - Product Version: [e.g., IS 5.10.0, IS 5.9.0] - OS: [e.g., Windows, Linux, Mac] - Database: [e.g., MySQL, H2] - Userstore: [e.g., LDAP, JDBC] --- ### Optional Fields **Related issues:** <!-- Any related issues from this/other repositories--> **Suggested labels:** <!-- Only to be used by non-members -->
non_process
password grant doesn t return configured sub claim in acces token describe the issue with password grant a self contained access token doesn t return the confgured custom claim but id token contains the correct custom claim as the sub claim reproduce steps add a service provider with oidc protocol select jwt token issuer create a custom claim under local claim create a oidc claim and mapped with the custom local claim add oidc claim to openid scope select the custom claim as the requested claim and sub claim for the created service provider send an access token request using password grant with the scope openid decode the retrieved access token idtoken you will see sub claim value contains the custom claim value in id token but access token will contain the username as the sub claim how to reproduce expected behavior environment information please complete the following information remove any unnecessary fields product version os database userstore optional fields related issues suggested labels
0
445,904
31,334,944,983
IssuesEvent
2023-08-24 04:52:17
bghira/SimpleTuner
https://api.github.com/repos/bghira/SimpleTuner
closed
Multi consumer GPU
documentation enhancement help wanted question
Have you had results with 2x cards that have 24G of vram, would the batch size be the same as using 48G A6000?
1.0
Multi consumer GPU - Have you had results with 2x cards that have 24G of vram, would the batch size be the same as using 48G A6000?
non_process
multi consumer gpu have you had results with cards that have of vram would the batch size be the same as using
0
20,468
27,129,277,122
IssuesEvent
2023-02-16 08:36:01
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Remove objc_proto_library rule
P3 type: process team-Rules-Server stale
`objc_proto_library` is not usable in Bazel. The code is present because it may be used inside Google. We should remove the rule from Bazel and from its documentation (https://docs.bazel.build/versions/master/be/objective-c.html#objc_proto_library). (tracking bug in Google: b/123888674)
1.0
Remove objc_proto_library rule - `objc_proto_library` is not usable in Bazel. The code is present because it may be used inside Google. We should remove the rule from Bazel and from its documentation (https://docs.bazel.build/versions/master/be/objective-c.html#objc_proto_library). (tracking bug in Google: b/123888674)
process
remove objc proto library rule objc proto library is not usable in bazel the code is present because it may be used inside google we should remove the rule from bazel and from its documentation tracking bug in google b
1
14,026
16,825,724,196
IssuesEvent
2021-06-17 18:17:09
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
Color Calibration Channel Mixer produces black patch
scope: image processing
**Describe the bug/issue** Here is a gradient created and embedded with linear rec 2020 color space, saved as 16 bit tiff. [gradient chart l rec 2020.tif.txt](https://github.com/darktable-org/darktable/files/6554086/gradient.chart.l.rec.2020.tif.txt) I attempt to invert colours using the channel mixer part of color calibration module, with the following settings: CAT: adaptation: none (byass) gamut compression: 0 Clip negative RGB from gamut: unticked R: r-1, g1, b1 G: r1, g-1, b1 B: r1, g1, b-1 As we can see below, the region that was green, and becomes magenta with these color calibration settings, now has a big black patch, instead of smooth gradients. ![gradient chart l rec 2020 col cal](https://user-images.githubusercontent.com/69561443/119829521-ac07c500-bf3e-11eb-9e2b-311709c8de9f.jpg) [gradient chart l rec 2020.tif.xmp.txt](https://github.com/darktable-org/darktable/files/6554117/gradient.chart.l.rec.2020.tif.xmp.txt) Comparably, the same settings used in the old depreciated channel mixer produce smooth gradients with no black patches, as expected: ![gradient chart l rec 2020 ch mix](https://user-images.githubusercontent.com/69561443/119829806-fd17b900-bf3e-11eb-9759-e60ddfe44d4c.jpg) [gradient chart l rec 2020_01.tif .xmp.txt](https://github.com/darktable-org/darktable/files/6554135/gradient.chart.l.rec.2020_01.tif.xmp.txt) Working profile for both is linear rec 2020, output profile is srgb. Perhaps of interest is that changing working profile does alter the shape of the black patch, but doesn't make it disappear. **Platform** _Please fill as much information as possible in the list given below. Please state "unknown" where you do not know the answer and remove any sections that are not applicable _ * darktable version : 3.4.1 * OS : MacOS 10.12.6 * Memory : 4 GB 1600 MHz DDR3 * Graphics card : Intel HD Graphics 4000 1536 MB & NVIDIA GeForce GT 650M * Graphics driver : * OpenCL installed : Y * OpenCL activated : Y - Can you reproduce with a RAW or Jpeg or both? **RAW-file-format/Jpeg/both** I also tried these settings on a RAW file with lots of green, and black patches became noticeable there too. Here is the RAW: https://discuss.pixls.us/t/sapajus-nigritus-helping-me-in-my-first-color-grading-steps/25022 Some screenshots. Pay close attention to the leaf in top left corner. First, with everything in pipe turned off except white balance default: ![col cal off](https://user-images.githubusercontent.com/69561443/119832240-58e34180-bf41-11eb-95b7-e9eabeb726a0.png) Second, with color calibration on, above settings: ![col cal on](https://user-images.githubusercontent.com/69561443/119832343-6ef10200-bf41-11eb-9e36-fbebadb0f2a9.png) Third, with channel mixer on, above settings: ![col cal off, ch mix on](https://user-images.githubusercontent.com/69561443/119832679-c000f600-bf41-11eb-9b2f-0478ab5f7ece.png)
1.0
Color Calibration Channel Mixer produces black patch - **Describe the bug/issue** Here is a gradient created and embedded with linear rec 2020 color space, saved as 16 bit tiff. [gradient chart l rec 2020.tif.txt](https://github.com/darktable-org/darktable/files/6554086/gradient.chart.l.rec.2020.tif.txt) I attempt to invert colours using the channel mixer part of color calibration module, with the following settings: CAT: adaptation: none (byass) gamut compression: 0 Clip negative RGB from gamut: unticked R: r-1, g1, b1 G: r1, g-1, b1 B: r1, g1, b-1 As we can see below, the region that was green, and becomes magenta with these color calibration settings, now has a big black patch, instead of smooth gradients. ![gradient chart l rec 2020 col cal](https://user-images.githubusercontent.com/69561443/119829521-ac07c500-bf3e-11eb-9e2b-311709c8de9f.jpg) [gradient chart l rec 2020.tif.xmp.txt](https://github.com/darktable-org/darktable/files/6554117/gradient.chart.l.rec.2020.tif.xmp.txt) Comparably, the same settings used in the old depreciated channel mixer produce smooth gradients with no black patches, as expected: ![gradient chart l rec 2020 ch mix](https://user-images.githubusercontent.com/69561443/119829806-fd17b900-bf3e-11eb-9759-e60ddfe44d4c.jpg) [gradient chart l rec 2020_01.tif .xmp.txt](https://github.com/darktable-org/darktable/files/6554135/gradient.chart.l.rec.2020_01.tif.xmp.txt) Working profile for both is linear rec 2020, output profile is srgb. Perhaps of interest is that changing working profile does alter the shape of the black patch, but doesn't make it disappear. **Platform** _Please fill as much information as possible in the list given below. Please state "unknown" where you do not know the answer and remove any sections that are not applicable _ * darktable version : 3.4.1 * OS : MacOS 10.12.6 * Memory : 4 GB 1600 MHz DDR3 * Graphics card : Intel HD Graphics 4000 1536 MB & NVIDIA GeForce GT 650M * Graphics driver : * OpenCL installed : Y * OpenCL activated : Y - Can you reproduce with a RAW or Jpeg or both? **RAW-file-format/Jpeg/both** I also tried these settings on a RAW file with lots of green, and black patches became noticeable there too. Here is the RAW: https://discuss.pixls.us/t/sapajus-nigritus-helping-me-in-my-first-color-grading-steps/25022 Some screenshots. Pay close attention to the leaf in top left corner. First, with everything in pipe turned off except white balance default: ![col cal off](https://user-images.githubusercontent.com/69561443/119832240-58e34180-bf41-11eb-95b7-e9eabeb726a0.png) Second, with color calibration on, above settings: ![col cal on](https://user-images.githubusercontent.com/69561443/119832343-6ef10200-bf41-11eb-9e36-fbebadb0f2a9.png) Third, with channel mixer on, above settings: ![col cal off, ch mix on](https://user-images.githubusercontent.com/69561443/119832679-c000f600-bf41-11eb-9b2f-0478ab5f7ece.png)
process
color calibration channel mixer produces black patch describe the bug issue here is a gradient created and embedded with linear rec color space saved as bit tiff i attempt to invert colours using the channel mixer part of color calibration module with the following settings cat adaptation none byass gamut compression clip negative rgb from gamut unticked r r g g b b as we can see below the region that was green and becomes magenta with these color calibration settings now has a big black patch instead of smooth gradients comparably the same settings used in the old depreciated channel mixer produce smooth gradients with no black patches as expected working profile for both is linear rec output profile is srgb perhaps of interest is that changing working profile does alter the shape of the black patch but doesn t make it disappear platform please fill as much information as possible in the list given below please state unknown where you do not know the answer and remove any sections that are not applicable darktable version os macos memory gb mhz graphics card intel hd graphics mb nvidia geforce gt graphics driver opencl installed y opencl activated y can you reproduce with a raw or jpeg or both raw file format jpeg both i also tried these settings on a raw file with lots of green and black patches became noticeable there too here is the raw some screenshots pay close attention to the leaf in top left corner first with everything in pipe turned off except white balance default second with color calibration on above settings third with channel mixer on above settings
1
16,047
20,192,792,408
IssuesEvent
2022-02-11 07:44:42
soederpop/active-mdx-software-project-test-repo
https://api.github.com/repos/soederpop/active-mdx-software-project-test-repo
closed
A customer should be able to pay with paypal
story-created epic-payment-processing
# A customer should be able to pay with paypal As a customer I want to be able to pay with paypal so I can complete my order
1.0
A customer should be able to pay with paypal - # A customer should be able to pay with paypal As a customer I want to be able to pay with paypal so I can complete my order
process
a customer should be able to pay with paypal a customer should be able to pay with paypal as a customer i want to be able to pay with paypal so i can complete my order
1
10,248
13,103,196,438
IssuesEvent
2020-08-04 08:08:59
zammad/zammad
https://api.github.com/repos/zammad/zammad
closed
unprocessible mail "FrozenError: can't modify frozen String"
bug mail processing prioritized by payment verified
<!-- Hi there - thanks for filing an issue. Please ensure the following things before creating an issue - thank you! 🤓 Since november 15th we handle all requests, except real bugs, at our community board. Full explanation: https://community.zammad.org/t/major-change-regarding-github-issues-community-board/21 Please post: - Feature requests - Development questions - Technical questions on the board -> https://community.zammad.org ! If you think you hit a bug, please continue: - Search existing issues and the CHANGELOG.md for your issue - there might be a solution already - Make sure to use the latest version of Zammad if possible - Add the `log/production.log` file from your system. Attention: Make sure no confidential data is in it! - Please write the issue in english - Don't remove the template - otherwise we will close the issue without further comments - Ask questions about Zammad configuration and usage at our mailinglist. See: https://zammad.org/participate Note: We always do our best. Unfortunately, sometimes there are too many requests and we can't handle everything at once. If you want to prioritize/escalate your issue, you can do so by means of a support contract (see https://zammad.com/pricing#selfhosted). * The upper textblock will be removed automatically when you submit your issue * --> ### Infos: * Used Zammad version: 3.4 * Installation method (source, package, ..): any * Operating system: any * Database + version: any * Elasticsearch version: any * Browser + version: any * Ticket-ID: 1078254 ### Expected behavior: Zammad fetches mailer daemon mails without issues. ### Actual behavior: Some mailer daemon mails are so broken that Zammad can't import them. Also trying to manually import them fails. ``` [zammad@lehost zammad]$ rails r 'Channel::EmailParser.process_unprocessable_mails' "ERROR: Can't process email, you will find it for bug reporting under /opt/zammad/tmp/unprocessable_mail/9d140e2bca6927f4e573cad2bccc8cf1.eml, please create an issue at https://github.com/zammad/zammad/issues" "ERROR: #<FrozenError: can't modify frozen String>" Traceback (most recent call last): 35: from bin/rails:9:in `<main>' 34: from /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:291:in `require' 33: from /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:257:in `load_dependency' 32: from /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:291:in `block in require' 31: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:29:in `require' 30: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:20:in `require_with_bootsnap_lfi' 29: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/loaded_features_index.rb:65:in `register' 28: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:21:in `block in require_with_bootsnap_lfi' 27: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:21:in `require' 26: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands.rb:18:in `<main>' 25: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/command.rb:46:in `invoke' 24: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/command/base.rb:69:in `perform' 23: from /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor.rb:392:in `dispatch' 22: from /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor/invocation.rb:127:in `invoke_command' 21: from /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor/command.rb:27:in `run' 20: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `perform' 19: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `eval' 18: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `<main>' 17: from /opt/zammad/app/models/channel/email_parser.rb:481:in `process_unprocessable_mails' 16: from /opt/zammad/app/models/channel/email_parser.rb:481:in `glob' 15: from /opt/zammad/app/models/channel/email_parser.rb:482:in `block in process_unprocessable_mails' 14: from /opt/zammad/app/models/channel/email_parser.rb:117:in `process' 13: from /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:108:in `timeout' 12: from /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:33:in `catch' 11: from /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:33:in `catch' 10: from /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:33:in `block in catch' 9: from /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:93:in `block in timeout' 8: from /opt/zammad/app/models/channel/email_parser.rb:118:in `block in process' 7: from /opt/zammad/app/models/channel/email_parser.rb:149:in `_process' 6: from /opt/zammad/app/models/channel/email_parser.rb:149:in `each' 5: from /opt/zammad/app/models/channel/email_parser.rb:152:in `block in _process' 4: from /opt/zammad/app/models/channel/filter/bounce_follow_up_check.rb:12:in `run' 3: from /opt/zammad/app/models/channel/filter/bounce_follow_up_check.rb:12:in `each' 2: from /opt/zammad/app/models/channel/filter/bounce_follow_up_check.rb:17:in `block in run' 1: from /opt/zammad/app/models/channel/email_parser.rb:75:in `parse' /opt/zammad/app/models/channel/email_parser.rb:75:in `force_encoding': can't modify frozen String (FrozenError) 22: from bin/rails:9:in `<main>' 21: from /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:291:in `require' 20: from /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:257:in `load_dependency' 19: from /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:291:in `block in require' 18: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:29:in `require' 17: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:20:in `require_with_bootsnap_lfi' 16: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/loaded_features_index.rb:65:in `register' 15: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:21:in `block in require_with_bootsnap_lfi' 14: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:21:in `require' 13: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands.rb:18:in `<main>' 12: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/command.rb:46:in `invoke' 11: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/command/base.rb:69:in `perform' 10: from /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor.rb:392:in `dispatch' 9: from /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor/invocation.rb:127:in `invoke_command' 8: from /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor/command.rb:27:in `run' 7: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `perform' 6: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `eval' 5: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `<main>' 4: from /opt/zammad/app/models/channel/email_parser.rb:481:in `process_unprocessable_mails' 3: from /opt/zammad/app/models/channel/email_parser.rb:481:in `glob' 2: from /opt/zammad/app/models/channel/email_parser.rb:482:in `block in process_unprocessable_mails' 1: from /opt/zammad/app/models/channel/email_parser.rb:115:in `process' /opt/zammad/app/models/channel/email_parser.rb:133:in `rescue in process': #<FrozenError: can't modify frozen String> (RuntimeError) /opt/zammad/app/models/channel/email_parser.rb:75:in `force_encoding' /opt/zammad/app/models/channel/email_parser.rb:75:in `parse' /opt/zammad/app/models/channel/filter/bounce_follow_up_check.rb:17:in `block in run' /opt/zammad/app/models/channel/filter/bounce_follow_up_check.rb:12:in `each' /opt/zammad/app/models/channel/filter/bounce_follow_up_check.rb:12:in `run' /opt/zammad/app/models/channel/email_parser.rb:152:in `block in _process' /opt/zammad/app/models/channel/email_parser.rb:149:in `each' /opt/zammad/app/models/channel/email_parser.rb:149:in `_process' /opt/zammad/app/models/channel/email_parser.rb:118:in `block in process' /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:93:in `block in timeout' /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:33:in `block in catch' /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:33:in `catch' /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:33:in `catch' /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:108:in `timeout' /opt/zammad/app/models/channel/email_parser.rb:117:in `process' /opt/zammad/app/models/channel/email_parser.rb:482:in `block in process_unprocessable_mails' /opt/zammad/app/models/channel/email_parser.rb:481:in `glob' /opt/zammad/app/models/channel/email_parser.rb:481:in `process_unprocessable_mails' /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `<main>' /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `eval' /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `perform' /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor/command.rb:27:in `run' /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor/invocation.rb:127:in `invoke_command' /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor.rb:392:in `dispatch' /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/command/base.rb:69:in `perform' /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/command.rb:46:in `invoke' /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands.rb:18:in `<main>' /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:21:in `require' /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:21:in `block in require_with_bootsnap_lfi' /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/loaded_features_index.rb:65:in `register' /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:20:in `require_with_bootsnap_lfi' /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:29:in `require' /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:291:in `block in require' /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:257:in `load_dependency' /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:291:in `require' bin/rails:9:in `<main>' ``` #### Possible help: https://github.com/zammad/zammad/commit/b6f7858dd53836c157f9778a5b7d51a0960569aa ### Steps to reproduce the behavior: * get a mailer-daemon mail that fullfills a specific situation You can find a affected mail in Ticket#1078254 . As it contains sensitive information, I can't post it here. This is to also ensure I don't change encodings that may be responsible for this issue. Yes I'm sure this is a bug and no feature request or a general question.
1.0
unprocessible mail "FrozenError: can't modify frozen String" - <!-- Hi there - thanks for filing an issue. Please ensure the following things before creating an issue - thank you! 🤓 Since november 15th we handle all requests, except real bugs, at our community board. Full explanation: https://community.zammad.org/t/major-change-regarding-github-issues-community-board/21 Please post: - Feature requests - Development questions - Technical questions on the board -> https://community.zammad.org ! If you think you hit a bug, please continue: - Search existing issues and the CHANGELOG.md for your issue - there might be a solution already - Make sure to use the latest version of Zammad if possible - Add the `log/production.log` file from your system. Attention: Make sure no confidential data is in it! - Please write the issue in english - Don't remove the template - otherwise we will close the issue without further comments - Ask questions about Zammad configuration and usage at our mailinglist. See: https://zammad.org/participate Note: We always do our best. Unfortunately, sometimes there are too many requests and we can't handle everything at once. If you want to prioritize/escalate your issue, you can do so by means of a support contract (see https://zammad.com/pricing#selfhosted). * The upper textblock will be removed automatically when you submit your issue * --> ### Infos: * Used Zammad version: 3.4 * Installation method (source, package, ..): any * Operating system: any * Database + version: any * Elasticsearch version: any * Browser + version: any * Ticket-ID: 1078254 ### Expected behavior: Zammad fetches mailer daemon mails without issues. ### Actual behavior: Some mailer daemon mails are so broken that Zammad can't import them. Also trying to manually import them fails. ``` [zammad@lehost zammad]$ rails r 'Channel::EmailParser.process_unprocessable_mails' "ERROR: Can't process email, you will find it for bug reporting under /opt/zammad/tmp/unprocessable_mail/9d140e2bca6927f4e573cad2bccc8cf1.eml, please create an issue at https://github.com/zammad/zammad/issues" "ERROR: #<FrozenError: can't modify frozen String>" Traceback (most recent call last): 35: from bin/rails:9:in `<main>' 34: from /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:291:in `require' 33: from /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:257:in `load_dependency' 32: from /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:291:in `block in require' 31: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:29:in `require' 30: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:20:in `require_with_bootsnap_lfi' 29: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/loaded_features_index.rb:65:in `register' 28: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:21:in `block in require_with_bootsnap_lfi' 27: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:21:in `require' 26: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands.rb:18:in `<main>' 25: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/command.rb:46:in `invoke' 24: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/command/base.rb:69:in `perform' 23: from /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor.rb:392:in `dispatch' 22: from /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor/invocation.rb:127:in `invoke_command' 21: from /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor/command.rb:27:in `run' 20: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `perform' 19: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `eval' 18: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `<main>' 17: from /opt/zammad/app/models/channel/email_parser.rb:481:in `process_unprocessable_mails' 16: from /opt/zammad/app/models/channel/email_parser.rb:481:in `glob' 15: from /opt/zammad/app/models/channel/email_parser.rb:482:in `block in process_unprocessable_mails' 14: from /opt/zammad/app/models/channel/email_parser.rb:117:in `process' 13: from /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:108:in `timeout' 12: from /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:33:in `catch' 11: from /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:33:in `catch' 10: from /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:33:in `block in catch' 9: from /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:93:in `block in timeout' 8: from /opt/zammad/app/models/channel/email_parser.rb:118:in `block in process' 7: from /opt/zammad/app/models/channel/email_parser.rb:149:in `_process' 6: from /opt/zammad/app/models/channel/email_parser.rb:149:in `each' 5: from /opt/zammad/app/models/channel/email_parser.rb:152:in `block in _process' 4: from /opt/zammad/app/models/channel/filter/bounce_follow_up_check.rb:12:in `run' 3: from /opt/zammad/app/models/channel/filter/bounce_follow_up_check.rb:12:in `each' 2: from /opt/zammad/app/models/channel/filter/bounce_follow_up_check.rb:17:in `block in run' 1: from /opt/zammad/app/models/channel/email_parser.rb:75:in `parse' /opt/zammad/app/models/channel/email_parser.rb:75:in `force_encoding': can't modify frozen String (FrozenError) 22: from bin/rails:9:in `<main>' 21: from /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:291:in `require' 20: from /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:257:in `load_dependency' 19: from /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:291:in `block in require' 18: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:29:in `require' 17: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:20:in `require_with_bootsnap_lfi' 16: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/loaded_features_index.rb:65:in `register' 15: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:21:in `block in require_with_bootsnap_lfi' 14: from /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:21:in `require' 13: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands.rb:18:in `<main>' 12: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/command.rb:46:in `invoke' 11: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/command/base.rb:69:in `perform' 10: from /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor.rb:392:in `dispatch' 9: from /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor/invocation.rb:127:in `invoke_command' 8: from /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor/command.rb:27:in `run' 7: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `perform' 6: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `eval' 5: from /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `<main>' 4: from /opt/zammad/app/models/channel/email_parser.rb:481:in `process_unprocessable_mails' 3: from /opt/zammad/app/models/channel/email_parser.rb:481:in `glob' 2: from /opt/zammad/app/models/channel/email_parser.rb:482:in `block in process_unprocessable_mails' 1: from /opt/zammad/app/models/channel/email_parser.rb:115:in `process' /opt/zammad/app/models/channel/email_parser.rb:133:in `rescue in process': #<FrozenError: can't modify frozen String> (RuntimeError) /opt/zammad/app/models/channel/email_parser.rb:75:in `force_encoding' /opt/zammad/app/models/channel/email_parser.rb:75:in `parse' /opt/zammad/app/models/channel/filter/bounce_follow_up_check.rb:17:in `block in run' /opt/zammad/app/models/channel/filter/bounce_follow_up_check.rb:12:in `each' /opt/zammad/app/models/channel/filter/bounce_follow_up_check.rb:12:in `run' /opt/zammad/app/models/channel/email_parser.rb:152:in `block in _process' /opt/zammad/app/models/channel/email_parser.rb:149:in `each' /opt/zammad/app/models/channel/email_parser.rb:149:in `_process' /opt/zammad/app/models/channel/email_parser.rb:118:in `block in process' /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:93:in `block in timeout' /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:33:in `block in catch' /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:33:in `catch' /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:33:in `catch' /usr/local/rvm/rubies/ruby-2.6.5/lib/ruby/2.6.0/timeout.rb:108:in `timeout' /opt/zammad/app/models/channel/email_parser.rb:117:in `process' /opt/zammad/app/models/channel/email_parser.rb:482:in `block in process_unprocessable_mails' /opt/zammad/app/models/channel/email_parser.rb:481:in `glob' /opt/zammad/app/models/channel/email_parser.rb:481:in `process_unprocessable_mails' /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `<main>' /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `eval' /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands/runner/runner_command.rb:41:in `perform' /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor/command.rb:27:in `run' /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor/invocation.rb:127:in `invoke_command' /usr/local/rvm/gems/ruby-2.6.5/gems/thor-1.0.1/lib/thor.rb:392:in `dispatch' /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/command/base.rb:69:in `perform' /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/command.rb:46:in `invoke' /usr/local/rvm/gems/ruby-2.6.5/gems/railties-5.2.4.3/lib/rails/commands.rb:18:in `<main>' /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:21:in `require' /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:21:in `block in require_with_bootsnap_lfi' /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/loaded_features_index.rb:65:in `register' /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:20:in `require_with_bootsnap_lfi' /usr/local/rvm/gems/ruby-2.6.5/gems/bootsnap-1.3.2/lib/bootsnap/load_path_cache/core_ext/kernel_require.rb:29:in `require' /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:291:in `block in require' /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:257:in `load_dependency' /usr/local/rvm/gems/ruby-2.6.5/gems/activesupport-5.2.4.3/lib/active_support/dependencies.rb:291:in `require' bin/rails:9:in `<main>' ``` #### Possible help: https://github.com/zammad/zammad/commit/b6f7858dd53836c157f9778a5b7d51a0960569aa ### Steps to reproduce the behavior: * get a mailer-daemon mail that fullfills a specific situation You can find a affected mail in Ticket#1078254 . As it contains sensitive information, I can't post it here. This is to also ensure I don't change encodings that may be responsible for this issue. Yes I'm sure this is a bug and no feature request or a general question.
process
unprocessible mail frozenerror can t modify frozen string hi there thanks for filing an issue please ensure the following things before creating an issue thank you 🤓 since november we handle all requests except real bugs at our community board full explanation please post feature requests development questions technical questions on the board if you think you hit a bug please continue search existing issues and the changelog md for your issue there might be a solution already make sure to use the latest version of zammad if possible add the log production log file from your system attention make sure no confidential data is in it please write the issue in english don t remove the template otherwise we will close the issue without further comments ask questions about zammad configuration and usage at our mailinglist see note we always do our best unfortunately sometimes there are too many requests and we can t handle everything at once if you want to prioritize escalate your issue you can do so by means of a support contract see the upper textblock will be removed automatically when you submit your issue infos used zammad version installation method source package any operating system any database version any elasticsearch version any browser version any ticket id expected behavior zammad fetches mailer daemon mails without issues actual behavior some mailer daemon mails are so broken that zammad can t import them also trying to manually import them fails rails r channel emailparser process unprocessable mails error can t process email you will find it for bug reporting under opt zammad tmp unprocessable mail eml please create an issue at error traceback most recent call last from bin rails in from usr local rvm gems ruby gems activesupport lib active support dependencies rb in require from usr local rvm gems ruby gems activesupport lib active support dependencies rb in load dependency from usr local rvm gems ruby gems activesupport lib active support dependencies rb in block in require from usr local rvm gems ruby gems bootsnap lib bootsnap load path cache core ext kernel require rb in require from usr local rvm gems ruby gems bootsnap lib bootsnap load path cache core ext kernel require rb in require with bootsnap lfi from usr local rvm gems ruby gems bootsnap lib bootsnap load path cache loaded features index rb in register from usr local rvm gems ruby gems bootsnap lib bootsnap load path cache core ext kernel require rb in block in require with bootsnap lfi from usr local rvm gems ruby gems bootsnap lib bootsnap load path cache core ext kernel require rb in require from usr local rvm gems ruby gems railties lib rails commands rb in from usr local rvm gems ruby gems railties lib rails command rb in invoke from usr local rvm gems ruby gems railties lib rails command base rb in perform from usr local rvm gems ruby gems thor lib thor rb in dispatch from usr local rvm gems ruby gems thor lib thor invocation rb in invoke command from usr local rvm gems ruby gems thor lib thor command rb in run from usr local rvm gems ruby gems railties lib rails commands runner runner command rb in perform from usr local rvm gems ruby gems railties lib rails commands runner runner command rb in eval from usr local rvm gems ruby gems railties lib rails commands runner runner command rb in from opt zammad app models channel email parser rb in process unprocessable mails from opt zammad app models channel email parser rb in glob from opt zammad app models channel email parser rb in block in process unprocessable mails from opt zammad app models channel email parser rb in process from usr local rvm rubies ruby lib ruby timeout rb in timeout from usr local rvm rubies ruby lib ruby timeout rb in catch from usr local rvm rubies ruby lib ruby timeout rb in catch from usr local rvm rubies ruby lib ruby timeout rb in block in catch from usr local rvm rubies ruby lib ruby timeout rb in block in timeout from opt zammad app models channel email parser rb in block in process from opt zammad app models channel email parser rb in process from opt zammad app models channel email parser rb in each from opt zammad app models channel email parser rb in block in process from opt zammad app models channel filter bounce follow up check rb in run from opt zammad app models channel filter bounce follow up check rb in each from opt zammad app models channel filter bounce follow up check rb in block in run from opt zammad app models channel email parser rb in parse opt zammad app models channel email parser rb in force encoding can t modify frozen string frozenerror from bin rails in from usr local rvm gems ruby gems activesupport lib active support dependencies rb in require from usr local rvm gems ruby gems activesupport lib active support dependencies rb in load dependency from usr local rvm gems ruby gems activesupport lib active support dependencies rb in block in require from usr local rvm gems ruby gems bootsnap lib bootsnap load path cache core ext kernel require rb in require from usr local rvm gems ruby gems bootsnap lib bootsnap load path cache core ext kernel require rb in require with bootsnap lfi from usr local rvm gems ruby gems bootsnap lib bootsnap load path cache loaded features index rb in register from usr local rvm gems ruby gems bootsnap lib bootsnap load path cache core ext kernel require rb in block in require with bootsnap lfi from usr local rvm gems ruby gems bootsnap lib bootsnap load path cache core ext kernel require rb in require from usr local rvm gems ruby gems railties lib rails commands rb in from usr local rvm gems ruby gems railties lib rails command rb in invoke from usr local rvm gems ruby gems railties lib rails command base rb in perform from usr local rvm gems ruby gems thor lib thor rb in dispatch from usr local rvm gems ruby gems thor lib thor invocation rb in invoke command from usr local rvm gems ruby gems thor lib thor command rb in run from usr local rvm gems ruby gems railties lib rails commands runner runner command rb in perform from usr local rvm gems ruby gems railties lib rails commands runner runner command rb in eval from usr local rvm gems ruby gems railties lib rails commands runner runner command rb in from opt zammad app models channel email parser rb in process unprocessable mails from opt zammad app models channel email parser rb in glob from opt zammad app models channel email parser rb in block in process unprocessable mails from opt zammad app models channel email parser rb in process opt zammad app models channel email parser rb in rescue in process runtimeerror opt zammad app models channel email parser rb in force encoding opt zammad app models channel email parser rb in parse opt zammad app models channel filter bounce follow up check rb in block in run opt zammad app models channel filter bounce follow up check rb in each opt zammad app models channel filter bounce follow up check rb in run opt zammad app models channel email parser rb in block in process opt zammad app models channel email parser rb in each opt zammad app models channel email parser rb in process opt zammad app models channel email parser rb in block in process usr local rvm rubies ruby lib ruby timeout rb in block in timeout usr local rvm rubies ruby lib ruby timeout rb in block in catch usr local rvm rubies ruby lib ruby timeout rb in catch usr local rvm rubies ruby lib ruby timeout rb in catch usr local rvm rubies ruby lib ruby timeout rb in timeout opt zammad app models channel email parser rb in process opt zammad app models channel email parser rb in block in process unprocessable mails opt zammad app models channel email parser rb in glob opt zammad app models channel email parser rb in process unprocessable mails usr local rvm gems ruby gems railties lib rails commands runner runner command rb in usr local rvm gems ruby gems railties lib rails commands runner runner command rb in eval usr local rvm gems ruby gems railties lib rails commands runner runner command rb in perform usr local rvm gems ruby gems thor lib thor command rb in run usr local rvm gems ruby gems thor lib thor invocation rb in invoke command usr local rvm gems ruby gems thor lib thor rb in dispatch usr local rvm gems ruby gems railties lib rails command base rb in perform usr local rvm gems ruby gems railties lib rails command rb in invoke usr local rvm gems ruby gems railties lib rails commands rb in usr local rvm gems ruby gems bootsnap lib bootsnap load path cache core ext kernel require rb in require usr local rvm gems ruby gems bootsnap lib bootsnap load path cache core ext kernel require rb in block in require with bootsnap lfi usr local rvm gems ruby gems bootsnap lib bootsnap load path cache loaded features index rb in register usr local rvm gems ruby gems bootsnap lib bootsnap load path cache core ext kernel require rb in require with bootsnap lfi usr local rvm gems ruby gems bootsnap lib bootsnap load path cache core ext kernel require rb in require usr local rvm gems ruby gems activesupport lib active support dependencies rb in block in require usr local rvm gems ruby gems activesupport lib active support dependencies rb in load dependency usr local rvm gems ruby gems activesupport lib active support dependencies rb in require bin rails in possible help steps to reproduce the behavior get a mailer daemon mail that fullfills a specific situation you can find a affected mail in ticket as it contains sensitive information i can t post it here this is to also ensure i don t change encodings that may be responsible for this issue yes i m sure this is a bug and no feature request or a general question
1
21,531
29,820,135,675
IssuesEvent
2023-06-17 00:58:33
bitfocus/companion-module-requests
https://api.github.com/repos/bitfocus/companion-module-requests
opened
Module for Shure MV7
NOT YET PROCESSED
- [ ] **I have researched the list of existing Companion modules and requests and have determined this has not yet been requested** The name of the device, hardware, or software you would like to control: Shure MV7 What you would like to be able to make it do from Companion: Change mic gain, headphone volume, eq, compression, etc. Direct links or attachments to the ethernet control protocol or API: https://www.shure.com/en-US/products/software/shure_plus_motiv_desktop
1.0
Module for Shure MV7 - - [ ] **I have researched the list of existing Companion modules and requests and have determined this has not yet been requested** The name of the device, hardware, or software you would like to control: Shure MV7 What you would like to be able to make it do from Companion: Change mic gain, headphone volume, eq, compression, etc. Direct links or attachments to the ethernet control protocol or API: https://www.shure.com/en-US/products/software/shure_plus_motiv_desktop
process
module for shure i have researched the list of existing companion modules and requests and have determined this has not yet been requested the name of the device hardware or software you would like to control shure what you would like to be able to make it do from companion change mic gain headphone volume eq compression etc direct links or attachments to the ethernet control protocol or api
1
16,673
21,776,852,381
IssuesEvent
2022-05-13 14:34:36
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
opened
Adding more operators to `depth_integration`
enhancement preprocessor
**Is your feature request related to a problem? Please describe.** A while back, Javi worked on generalising the `depth_integration` preprocessor to accept other operators. It would be useful to have this changes officially in the core. Additionally, it could be an opportunity to improve some lines to make the preprocessor lazy, namely: ```python if depth.ndim == 1: slices = [None for i in cube.shape] coord_dim = cube.coord_dims(cube.coord(axis='z'))[0] slices[coord_dim] = slice(None) thickness = np.abs(thickness[tuple(slices)]) ones = np.ones_like(cube.data) ``` **Would you be able to help out?** yes
1.0
Adding more operators to `depth_integration` - **Is your feature request related to a problem? Please describe.** A while back, Javi worked on generalising the `depth_integration` preprocessor to accept other operators. It would be useful to have this changes officially in the core. Additionally, it could be an opportunity to improve some lines to make the preprocessor lazy, namely: ```python if depth.ndim == 1: slices = [None for i in cube.shape] coord_dim = cube.coord_dims(cube.coord(axis='z'))[0] slices[coord_dim] = slice(None) thickness = np.abs(thickness[tuple(slices)]) ones = np.ones_like(cube.data) ``` **Would you be able to help out?** yes
process
adding more operators to depth integration is your feature request related to a problem please describe a while back javi worked on generalising the depth integration preprocessor to accept other operators it would be useful to have this changes officially in the core additionally it could be an opportunity to improve some lines to make the preprocessor lazy namely python if depth ndim slices coord dim cube coord dims cube coord axis z slices slice none thickness np abs thickness ones np ones like cube data would you be able to help out yes
1
20,258
26,875,838,154
IssuesEvent
2023-02-05 02:00:09
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Fri, 3 Feb 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### Cooperative Saliency-based Obstacle Detection and AR Rendering for Increased Situational Awareness - **Authors:** Gerasimos Arvanitis, Nikolaos Stagakis, Evangelia I. Zacharaki, Konstantinos Moustakas - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2302.00916 - **Pdf link:** https://arxiv.org/pdf/2302.00916 - **Abstract** Autonomous vehicles are expected to operate safely in real-life road conditions in the next years. Nevertheless, unanticipated events such as the existence of unexpected objects in the range of the road, can put safety at risk. The advancement of sensing and communication technologies and Internet of Things may facilitate the recognition of hazardous situations and information exchange in a cooperative driving scheme, providing new opportunities for the increase of collaborative situational awareness. Safe and unobtrusive visualization of the obtained information may nowadays be enabled through the adoption of novel Augmented Reality (AR) interfaces in the form of windshields. Motivated by these technological opportunities, we propose in this work a saliency-based distributed, cooperative obstacle detection and rendering scheme for increasing the driver's situational awareness through (i) automated obstacle detection, (ii) AR visualization and (iii) information sharing (upcoming potential dangers) with other connected vehicles or road infrastructure. An extensive evaluation study using a variety of real datasets for pothole detection showed that the proposed method provides favorable results and features compared to other recent and relevant approaches. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Online Continual Learning via the Knowledge Invariant and Spread-out Properties - **Authors:** Ya-nan Han, Jian-wei Liu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.00858 - **Pdf link:** https://arxiv.org/pdf/2302.00858 - **Abstract** The goal of continual learning is to provide intelligent agents that are capable of learning continually a sequence of tasks using the knowledge obtained from previous tasks while performing well on prior tasks. However, a key challenge in this continual learning paradigm is catastrophic forgetting, namely adapting a model to new tasks often leads to severe performance degradation on prior tasks. Current memory-based approaches show their success in alleviating the catastrophic forgetting problem by replaying examples from past tasks when new tasks are learned. However, these methods are infeasible to transfer the structural knowledge from previous tasks i.e., similarities or dissimilarities between different instances. Furthermore, the learning bias between the current and prior tasks is also an urgent problem that should be solved. In this work, we propose a new method, named Online Continual Learning via the Knowledge Invariant and Spread-out Properties (OCLKISP), in which we constrain the evolution of the embedding features via Knowledge Invariant and Spread-out Properties (KISP). Thus, we can further transfer the inter-instance structural knowledge of previous tasks while alleviating the forgetting due to the learning bias. We empirically evaluate our proposed method on four popular benchmarks for continual learning: Split CIFAR 100, Split SVHN, Split CUB200 and Split Tiny-Image-Net. The experimental results show the efficacy of our proposed method compared to the state-of-the-art continual learning algorithms. ### Advances and Challenges in Multimodal Remote Sensing Image Registration - **Authors:** Bai Zhu, Liang Zhou, Simiao Pu, Jianwei Fan, Yuanxin Ye - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.00912 - **Pdf link:** https://arxiv.org/pdf/2302.00912 - **Abstract** Over the past few decades, with the rapid development of global aerospace and aerial remote sensing technology, the types of sensors have evolved from the traditional monomodal sensors (e.g., optical sensors) to the new generation of multimodal sensors [e.g., multispectral, hyperspectral, light detection and ranging (LiDAR) and synthetic aperture radar (SAR) sensors]. These advanced devices can dynamically provide various and abundant multimodal remote sensing images with different spatial, temporal, and spectral resolutions according to different application requirements. Since then, it is of great scientific significance to carry out the research of multimodal remote sensing image registration, which is a crucial step for integrating the complementary information among multimodal data and making comprehensive observations and analysis of the Earths surface. In this work, we will present our own contributions to the field of multimodal image registration, summarize the advantages and limitations of existing multimodal image registration methods, and then discuss the remaining challenges and make a forward-looking prospect for the future development of the field. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### STEPS: Joint Self-supervised Nighttime Image Enhancement and Depth Estimation - **Authors:** Yupeng Zheng, Chengliang Zhong, Pengfei Li, Huan-ang Gao, Yuhang Zheng, Bu Jin, Ling Wang, Hao Zhao, Guyue Zhou, Qichao Zhang, Dongbin Zhao - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.01334 - **Pdf link:** https://arxiv.org/pdf/2302.01334 - **Abstract** Self-supervised depth estimation draws a lot of attention recently as it can promote the 3D sensing capabilities of self-driving vehicles. However, it intrinsically relies upon the photometric consistency assumption, which hardly holds during nighttime. Although various supervised nighttime image enhancement methods have been proposed, their generalization performance in challenging driving scenarios is not satisfactory. To this end, we propose the first method that jointly learns a nighttime image enhancer and a depth estimator, without using ground truth for either task. Our method tightly entangles two self-supervised tasks using a newly proposed uncertain pixel masking strategy. This strategy originates from the observation that nighttime images not only suffer from underexposed regions but also from overexposed regions. By fitting a bridge-shaped curve to the illumination map distribution, both regions are suppressed and two tasks are bridged naturally. We benchmark the method on two established datasets: nuScenes and RobotCar and demonstrate state-of-the-art performance on both of them. Detailed ablations also reveal the mechanism of our proposal. Last but not least, to mitigate the problem of sparse ground truth of existing datasets, we provide a new photo-realistically enhanced nighttime dataset based upon CARLA. It brings meaningful new challenges to the community. Codes, data, and models are available at https://github.com/ucaszyp/STEPS. ## Keyword: raw image There is no result
2.0
New submissions for Fri, 3 Feb 23 - ## Keyword: events ### Cooperative Saliency-based Obstacle Detection and AR Rendering for Increased Situational Awareness - **Authors:** Gerasimos Arvanitis, Nikolaos Stagakis, Evangelia I. Zacharaki, Konstantinos Moustakas - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2302.00916 - **Pdf link:** https://arxiv.org/pdf/2302.00916 - **Abstract** Autonomous vehicles are expected to operate safely in real-life road conditions in the next years. Nevertheless, unanticipated events such as the existence of unexpected objects in the range of the road, can put safety at risk. The advancement of sensing and communication technologies and Internet of Things may facilitate the recognition of hazardous situations and information exchange in a cooperative driving scheme, providing new opportunities for the increase of collaborative situational awareness. Safe and unobtrusive visualization of the obtained information may nowadays be enabled through the adoption of novel Augmented Reality (AR) interfaces in the form of windshields. Motivated by these technological opportunities, we propose in this work a saliency-based distributed, cooperative obstacle detection and rendering scheme for increasing the driver's situational awareness through (i) automated obstacle detection, (ii) AR visualization and (iii) information sharing (upcoming potential dangers) with other connected vehicles or road infrastructure. An extensive evaluation study using a variety of real datasets for pothole detection showed that the proposed method provides favorable results and features compared to other recent and relevant approaches. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Online Continual Learning via the Knowledge Invariant and Spread-out Properties - **Authors:** Ya-nan Han, Jian-wei Liu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.00858 - **Pdf link:** https://arxiv.org/pdf/2302.00858 - **Abstract** The goal of continual learning is to provide intelligent agents that are capable of learning continually a sequence of tasks using the knowledge obtained from previous tasks while performing well on prior tasks. However, a key challenge in this continual learning paradigm is catastrophic forgetting, namely adapting a model to new tasks often leads to severe performance degradation on prior tasks. Current memory-based approaches show their success in alleviating the catastrophic forgetting problem by replaying examples from past tasks when new tasks are learned. However, these methods are infeasible to transfer the structural knowledge from previous tasks i.e., similarities or dissimilarities between different instances. Furthermore, the learning bias between the current and prior tasks is also an urgent problem that should be solved. In this work, we propose a new method, named Online Continual Learning via the Knowledge Invariant and Spread-out Properties (OCLKISP), in which we constrain the evolution of the embedding features via Knowledge Invariant and Spread-out Properties (KISP). Thus, we can further transfer the inter-instance structural knowledge of previous tasks while alleviating the forgetting due to the learning bias. We empirically evaluate our proposed method on four popular benchmarks for continual learning: Split CIFAR 100, Split SVHN, Split CUB200 and Split Tiny-Image-Net. The experimental results show the efficacy of our proposed method compared to the state-of-the-art continual learning algorithms. ### Advances and Challenges in Multimodal Remote Sensing Image Registration - **Authors:** Bai Zhu, Liang Zhou, Simiao Pu, Jianwei Fan, Yuanxin Ye - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.00912 - **Pdf link:** https://arxiv.org/pdf/2302.00912 - **Abstract** Over the past few decades, with the rapid development of global aerospace and aerial remote sensing technology, the types of sensors have evolved from the traditional monomodal sensors (e.g., optical sensors) to the new generation of multimodal sensors [e.g., multispectral, hyperspectral, light detection and ranging (LiDAR) and synthetic aperture radar (SAR) sensors]. These advanced devices can dynamically provide various and abundant multimodal remote sensing images with different spatial, temporal, and spectral resolutions according to different application requirements. Since then, it is of great scientific significance to carry out the research of multimodal remote sensing image registration, which is a crucial step for integrating the complementary information among multimodal data and making comprehensive observations and analysis of the Earths surface. In this work, we will present our own contributions to the field of multimodal image registration, summarize the advantages and limitations of existing multimodal image registration methods, and then discuss the remaining challenges and make a forward-looking prospect for the future development of the field. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### STEPS: Joint Self-supervised Nighttime Image Enhancement and Depth Estimation - **Authors:** Yupeng Zheng, Chengliang Zhong, Pengfei Li, Huan-ang Gao, Yuhang Zheng, Bu Jin, Ling Wang, Hao Zhao, Guyue Zhou, Qichao Zhang, Dongbin Zhao - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.01334 - **Pdf link:** https://arxiv.org/pdf/2302.01334 - **Abstract** Self-supervised depth estimation draws a lot of attention recently as it can promote the 3D sensing capabilities of self-driving vehicles. However, it intrinsically relies upon the photometric consistency assumption, which hardly holds during nighttime. Although various supervised nighttime image enhancement methods have been proposed, their generalization performance in challenging driving scenarios is not satisfactory. To this end, we propose the first method that jointly learns a nighttime image enhancer and a depth estimator, without using ground truth for either task. Our method tightly entangles two self-supervised tasks using a newly proposed uncertain pixel masking strategy. This strategy originates from the observation that nighttime images not only suffer from underexposed regions but also from overexposed regions. By fitting a bridge-shaped curve to the illumination map distribution, both regions are suppressed and two tasks are bridged naturally. We benchmark the method on two established datasets: nuScenes and RobotCar and demonstrate state-of-the-art performance on both of them. Detailed ablations also reveal the mechanism of our proposal. Last but not least, to mitigate the problem of sparse ground truth of existing datasets, we provide a new photo-realistically enhanced nighttime dataset based upon CARLA. It brings meaningful new challenges to the community. Codes, data, and models are available at https://github.com/ucaszyp/STEPS. ## Keyword: raw image There is no result
process
new submissions for fri feb keyword events cooperative saliency based obstacle detection and ar rendering for increased situational awareness authors gerasimos arvanitis nikolaos stagakis evangelia i zacharaki konstantinos moustakas subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract autonomous vehicles are expected to operate safely in real life road conditions in the next years nevertheless unanticipated events such as the existence of unexpected objects in the range of the road can put safety at risk the advancement of sensing and communication technologies and internet of things may facilitate the recognition of hazardous situations and information exchange in a cooperative driving scheme providing new opportunities for the increase of collaborative situational awareness safe and unobtrusive visualization of the obtained information may nowadays be enabled through the adoption of novel augmented reality ar interfaces in the form of windshields motivated by these technological opportunities we propose in this work a saliency based distributed cooperative obstacle detection and rendering scheme for increasing the driver s situational awareness through i automated obstacle detection ii ar visualization and iii information sharing upcoming potential dangers with other connected vehicles or road infrastructure an extensive evaluation study using a variety of real datasets for pothole detection showed that the proposed method provides favorable results and features compared to other recent and relevant approaches keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp online continual learning via the knowledge invariant and spread out properties authors ya nan han jian wei liu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract the goal of continual learning is to provide intelligent agents that are capable of learning continually a sequence of tasks using the knowledge obtained from previous tasks while performing well on prior tasks however a key challenge in this continual learning paradigm is catastrophic forgetting namely adapting a model to new tasks often leads to severe performance degradation on prior tasks current memory based approaches show their success in alleviating the catastrophic forgetting problem by replaying examples from past tasks when new tasks are learned however these methods are infeasible to transfer the structural knowledge from previous tasks i e similarities or dissimilarities between different instances furthermore the learning bias between the current and prior tasks is also an urgent problem that should be solved in this work we propose a new method named online continual learning via the knowledge invariant and spread out properties oclkisp in which we constrain the evolution of the embedding features via knowledge invariant and spread out properties kisp thus we can further transfer the inter instance structural knowledge of previous tasks while alleviating the forgetting due to the learning bias we empirically evaluate our proposed method on four popular benchmarks for continual learning split cifar split svhn split and split tiny image net the experimental results show the efficacy of our proposed method compared to the state of the art continual learning algorithms advances and challenges in multimodal remote sensing image registration authors bai zhu liang zhou simiao pu jianwei fan yuanxin ye subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract over the past few decades with the rapid development of global aerospace and aerial remote sensing technology the types of sensors have evolved from the traditional monomodal sensors e g optical sensors to the new generation of multimodal sensors these advanced devices can dynamically provide various and abundant multimodal remote sensing images with different spatial temporal and spectral resolutions according to different application requirements since then it is of great scientific significance to carry out the research of multimodal remote sensing image registration which is a crucial step for integrating the complementary information among multimodal data and making comprehensive observations and analysis of the earths surface in this work we will present our own contributions to the field of multimodal image registration summarize the advantages and limitations of existing multimodal image registration methods and then discuss the remaining challenges and make a forward looking prospect for the future development of the field keyword image signal processing there is no result keyword image signal process there is no result keyword compression there is no result keyword raw steps joint self supervised nighttime image enhancement and depth estimation authors yupeng zheng chengliang zhong pengfei li huan ang gao yuhang zheng bu jin ling wang hao zhao guyue zhou qichao zhang dongbin zhao subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract self supervised depth estimation draws a lot of attention recently as it can promote the sensing capabilities of self driving vehicles however it intrinsically relies upon the photometric consistency assumption which hardly holds during nighttime although various supervised nighttime image enhancement methods have been proposed their generalization performance in challenging driving scenarios is not satisfactory to this end we propose the first method that jointly learns a nighttime image enhancer and a depth estimator without using ground truth for either task our method tightly entangles two self supervised tasks using a newly proposed uncertain pixel masking strategy this strategy originates from the observation that nighttime images not only suffer from underexposed regions but also from overexposed regions by fitting a bridge shaped curve to the illumination map distribution both regions are suppressed and two tasks are bridged naturally we benchmark the method on two established datasets nuscenes and robotcar and demonstrate state of the art performance on both of them detailed ablations also reveal the mechanism of our proposal last but not least to mitigate the problem of sparse ground truth of existing datasets we provide a new photo realistically enhanced nighttime dataset based upon carla it brings meaningful new challenges to the community codes data and models are available at keyword raw image there is no result
1
3,559
6,593,461,498
IssuesEvent
2017-09-15 01:20:53
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Check if we can open the directory on Linux when UseShellExecute is set
area-System.Diagnostics.Process enhancement os-linux up-for-grabs
I am creating this issue to track if we can open the directory on Linux when having ShellExecute is set. - Ideally we would want to throw exception only when UseShellExecute option was false.  - In OSX, the program "open" successfully opens folder using "finder". - But Linux needs extra effort to be able to open a folder here. In Linux, xdg-open cannot open folder, so I decided to throw when startInfo.FileName is a directory, regardless of UseShellExecute option being true or false.  Probably the failure happens when the OS doesnt have the installed libraries to open. The current behavior in Linux is captured [here](https://github.com/dotnet/corefx/blob/master/src/System.Diagnostics.Process/tests/ProcessTests.Unix.cs#L67L71) Part of this goal is to keep a consistent behavior with Windows. - When UseShellExecute is set on windows desktop, Process Start opens directory using File Explorer. - Also, the current behavior in Windows running on .NET Core is captured [here](https://github.com/dotnet/corefx/blob/e670fd396d1c2a6ae99f2bb0ec24a81e30c668e7/src/System.Diagnostics.Process/tests/ProcessTests.cs#L134-L149 ) Code Sample: ``` var startInfo = new ProcessStartInfo { UseShellExecute = true, FileName = Environment.CurrentDirectory }; using (var px = Process.Start(startInfo)) { ... } ``` @tarekgh
1.0
Check if we can open the directory on Linux when UseShellExecute is set - I am creating this issue to track if we can open the directory on Linux when having ShellExecute is set. - Ideally we would want to throw exception only when UseShellExecute option was false.  - In OSX, the program "open" successfully opens folder using "finder". - But Linux needs extra effort to be able to open a folder here. In Linux, xdg-open cannot open folder, so I decided to throw when startInfo.FileName is a directory, regardless of UseShellExecute option being true or false.  Probably the failure happens when the OS doesnt have the installed libraries to open. The current behavior in Linux is captured [here](https://github.com/dotnet/corefx/blob/master/src/System.Diagnostics.Process/tests/ProcessTests.Unix.cs#L67L71) Part of this goal is to keep a consistent behavior with Windows. - When UseShellExecute is set on windows desktop, Process Start opens directory using File Explorer. - Also, the current behavior in Windows running on .NET Core is captured [here](https://github.com/dotnet/corefx/blob/e670fd396d1c2a6ae99f2bb0ec24a81e30c668e7/src/System.Diagnostics.Process/tests/ProcessTests.cs#L134-L149 ) Code Sample: ``` var startInfo = new ProcessStartInfo { UseShellExecute = true, FileName = Environment.CurrentDirectory }; using (var px = Process.Start(startInfo)) { ... } ``` @tarekgh
process
check if we can open the directory on linux when useshellexecute is set i am creating this issue to track if we can open the directory on linux when having shellexecute is set ideally we would want to throw exception only when useshellexecute option was false   in osx the program open successfully opens folder using finder but linux needs extra effort to be able to open a folder here in linux xdg open cannot open folder so i decided to throw when startinfo filename is a directory regardless of useshellexecute option being true or false   probably the failure happens when the os doesnt have the installed libraries to open the current behavior in linux is captured part of this goal is to keep a consistent behavior with windows when useshellexecute is set on windows desktop process start opens directory using file explorer also the current behavior in windows running on net core is captured code sample var startinfo new processstartinfo useshellexecute true filename environment currentdirectory using var px process start startinfo tarekgh
1
291,563
25,155,732,802
IssuesEvent
2022-11-10 13:25:32
Uuvana-Studios/longvinter-windows-client
https://api.github.com/repos/Uuvana-Studios/longvinter-windows-client
closed
Can't place items on sand
Bug Not Tested
A path of sand appeared through my land after the new update. I cannot place anything on these parts leaving me unable to close off my fence. It says it is overlapping with something. ![7c09649a96cafb1c66280287734928e8](https://user-images.githubusercontent.com/103220758/190828068-ddd17de8-d2f3-4fd0-8372-ae2d83a5ab13.png)
1.0
Can't place items on sand - A path of sand appeared through my land after the new update. I cannot place anything on these parts leaving me unable to close off my fence. It says it is overlapping with something. ![7c09649a96cafb1c66280287734928e8](https://user-images.githubusercontent.com/103220758/190828068-ddd17de8-d2f3-4fd0-8372-ae2d83a5ab13.png)
non_process
can t place items on sand a path of sand appeared through my land after the new update i cannot place anything on these parts leaving me unable to close off my fence it says it is overlapping with something
0
14,048
4,807,807,376
IssuesEvent
2016-11-02 22:40:00
BowlerHatLLC/vscode-nextgenas
https://api.github.com/repos/BowlerHatLLC/vscode-nextgenas
closed
Not stopping at breakpoints in main class constructor
bug vscode-dependency
Ok, so I managed to get the compiling working on my mac, it now produces the bin folder and the output js files. The debugger executes the compile task then fires the chrome browser and the script runs but it doesn't stop at the breakpoints. I'm trying to debug using the local file option as per instructions.
1.0
Not stopping at breakpoints in main class constructor - Ok, so I managed to get the compiling working on my mac, it now produces the bin folder and the output js files. The debugger executes the compile task then fires the chrome browser and the script runs but it doesn't stop at the breakpoints. I'm trying to debug using the local file option as per instructions.
non_process
not stopping at breakpoints in main class constructor ok so i managed to get the compiling working on my mac it now produces the bin folder and the output js files the debugger executes the compile task then fires the chrome browser and the script runs but it doesn t stop at the breakpoints i m trying to debug using the local file option as per instructions
0
18,680
24,594,536,848
IssuesEvent
2022-10-14 07:07:00
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[FHIR] Questionnaire resource > JSON > Branching logic data is not getting mapped into the JSON file
Bug Blocker P0 Response datastore Process: Fixed Process: Tested QA Process: Tested dev
AR: Questionnaire resource > JSON > Branching logic data is not mapped into the JSON file ER: Branching logic data should get mapped into the JSON file as configured in the SB
3.0
[FHIR] Questionnaire resource > JSON > Branching logic data is not getting mapped into the JSON file - AR: Questionnaire resource > JSON > Branching logic data is not mapped into the JSON file ER: Branching logic data should get mapped into the JSON file as configured in the SB
process
questionnaire resource json branching logic data is not getting mapped into the json file ar questionnaire resource json branching logic data is not mapped into the json file er branching logic data should get mapped into the json file as configured in the sb
1
46,815
19,517,794,086
IssuesEvent
2021-12-29 13:16:23
hashicorp/terraform-provider-azurerm
https://api.github.com/repos/hashicorp/terraform-provider-azurerm
closed
Support for creating child DNS-zone
enhancement service/dns sdk/not-yet-supported
<!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description <!--- Please leave a helpful description of the feature request here. ---> Kindly ask you to support creating a child DNS-zone for existing parent Azure DNS-zone. ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * azurerm_dns_zone ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "azurerm_dns_zone" "parent_zone" { name = "parent.com" resource_group_name = azurerm_resource_group.resource_group.name } resource "azurerm_dns_zone" "child_zone" { name = "child.parent.com" resource_group_name = azurerm_resource_group.resource_group.name parent_zone_name = azure_dns_zone.parent_zone.name } ``` ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example: * https://azure.microsoft.com/en-us/roadmap/virtual-network-service-endpoint-for-azure-cosmos-db/ ---> * [Tutorial: Creating a new Child DNS zone](https://docs.microsoft.com/en-us/azure/dns/tutorial-public-dns-zones-child)
1.0
Support for creating child DNS-zone - <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description <!--- Please leave a helpful description of the feature request here. ---> Kindly ask you to support creating a child DNS-zone for existing parent Azure DNS-zone. ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * azurerm_dns_zone ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "azurerm_dns_zone" "parent_zone" { name = "parent.com" resource_group_name = azurerm_resource_group.resource_group.name } resource "azurerm_dns_zone" "child_zone" { name = "child.parent.com" resource_group_name = azurerm_resource_group.resource_group.name parent_zone_name = azure_dns_zone.parent_zone.name } ``` ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example: * https://azure.microsoft.com/en-us/roadmap/virtual-network-service-endpoint-for-azure-cosmos-db/ ---> * [Tutorial: Creating a new Child DNS zone](https://docs.microsoft.com/en-us/azure/dns/tutorial-public-dns-zones-child)
non_process
support for creating child dns zone community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description kindly ask you to support creating a child dns zone for existing parent azure dns zone new or affected resource s azurerm dns zone potential terraform configuration hcl resource azurerm dns zone parent zone name parent com resource group name azurerm resource group resource group name resource azurerm dns zone child zone name child parent com resource group name azurerm resource group resource group name parent zone name azure dns zone parent zone name references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation for example
0
6,750
9,879,973,225
IssuesEvent
2019-06-24 11:25:39
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Specifying cell size instead of size when reprojecting and resampling a raster
Feature Request Processing
Author Name: **Mattias Lindman** (Mattias Lindman) Original Redmine Issue: [14510](https://issues.qgis.org/issues/14510) Redmine category:processing/gdal --- When using Raster/Projections/Warp(Reproject) in order to reproject a raster to a different projection it would be convenient to specify the desired cell size in the resampled raster rather than the dimensions of the raster (through the size option in the GUI). By searching the Internet I learned that it was possible to specify the desired cell size by adding -tr and the desired cell size in the command window at the bottom of the GUI window. To find this out was however not straight forward and it would be more convenient if the GUI provided an option to specify the cell size of the reprojected and resampled raster.
1.0
Specifying cell size instead of size when reprojecting and resampling a raster - Author Name: **Mattias Lindman** (Mattias Lindman) Original Redmine Issue: [14510](https://issues.qgis.org/issues/14510) Redmine category:processing/gdal --- When using Raster/Projections/Warp(Reproject) in order to reproject a raster to a different projection it would be convenient to specify the desired cell size in the resampled raster rather than the dimensions of the raster (through the size option in the GUI). By searching the Internet I learned that it was possible to specify the desired cell size by adding -tr and the desired cell size in the command window at the bottom of the GUI window. To find this out was however not straight forward and it would be more convenient if the GUI provided an option to specify the cell size of the reprojected and resampled raster.
process
specifying cell size instead of size when reprojecting and resampling a raster author name mattias lindman mattias lindman original redmine issue redmine category processing gdal when using raster projections warp reproject in order to reproject a raster to a different projection it would be convenient to specify the desired cell size in the resampled raster rather than the dimensions of the raster through the size option in the gui by searching the internet i learned that it was possible to specify the desired cell size by adding tr and the desired cell size in the command window at the bottom of the gui window to find this out was however not straight forward and it would be more convenient if the gui provided an option to specify the cell size of the reprojected and resampled raster
1
142,902
19,122,596,882
IssuesEvent
2021-12-01 01:17:21
Tim-sandbox/webgoat-trng
https://api.github.com/repos/Tim-sandbox/webgoat-trng
opened
CVE-2021-22096 (Medium) detected in spring-web-5.2.2.RELEASE.jar
security vulnerability
## CVE-2021-22096 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-5.2.2.RELEASE.jar</b></p></summary> <p>Spring Web</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: webgoat-trng/webgoat-integration-tests/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-web/5.2.2.RELEASE/spring-web-5.2.2.RELEASE.jar</p> <p> Dependency Hierarchy: - webwolf-v8.1.0.jar (Root Library) - spring-boot-starter-web-2.2.2.RELEASE.jar - :x: **spring-web-5.2.2.RELEASE.jar** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework versions 5.3.0 - 5.3.10, 5.2.0 - 5.2.17, and older unsupported versions, it is possible for a user to provide malicious input to cause the insertion of additional log entries. <p>Publish Date: 2021-10-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22096>CVE-2021-22096</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22096">https://tanzu.vmware.com/security/cve-2021-22096</a></p> <p>Release Date: 2021-10-28</p> <p>Fix Resolution: org.springframework:spring:5.2.18.RELEASE,5.3.12</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-web","packageVersion":"5.2.2.RELEASE","packageFilePaths":["/webgoat-integration-tests/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.owasp.webgoat:webwolf:v8.1.0;org.springframework.boot:spring-boot-starter-web:2.2.2.RELEASE;org.springframework:spring-web:5.2.2.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework:spring:5.2.18.RELEASE,5.3.12","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-22096","vulnerabilityDetails":"In Spring Framework versions 5.3.0 - 5.3.10, 5.2.0 - 5.2.17, and older unsupported versions, it is possible for a user to provide malicious input to cause the insertion of additional log entries.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22096","cvss3Severity":"medium","cvss3Score":"4.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-22096 (Medium) detected in spring-web-5.2.2.RELEASE.jar - ## CVE-2021-22096 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-5.2.2.RELEASE.jar</b></p></summary> <p>Spring Web</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: webgoat-trng/webgoat-integration-tests/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-web/5.2.2.RELEASE/spring-web-5.2.2.RELEASE.jar</p> <p> Dependency Hierarchy: - webwolf-v8.1.0.jar (Root Library) - spring-boot-starter-web-2.2.2.RELEASE.jar - :x: **spring-web-5.2.2.RELEASE.jar** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework versions 5.3.0 - 5.3.10, 5.2.0 - 5.2.17, and older unsupported versions, it is possible for a user to provide malicious input to cause the insertion of additional log entries. <p>Publish Date: 2021-10-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22096>CVE-2021-22096</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22096">https://tanzu.vmware.com/security/cve-2021-22096</a></p> <p>Release Date: 2021-10-28</p> <p>Fix Resolution: org.springframework:spring:5.2.18.RELEASE,5.3.12</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-web","packageVersion":"5.2.2.RELEASE","packageFilePaths":["/webgoat-integration-tests/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.owasp.webgoat:webwolf:v8.1.0;org.springframework.boot:spring-boot-starter-web:2.2.2.RELEASE;org.springframework:spring-web:5.2.2.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework:spring:5.2.18.RELEASE,5.3.12","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-22096","vulnerabilityDetails":"In Spring Framework versions 5.3.0 - 5.3.10, 5.2.0 - 5.2.17, and older unsupported versions, it is possible for a user to provide malicious input to cause the insertion of additional log entries.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22096","cvss3Severity":"medium","cvss3Score":"4.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in spring web release jar cve medium severity vulnerability vulnerable library spring web release jar spring web library home page a href path to dependency file webgoat trng webgoat integration tests pom xml path to vulnerable library home wss scanner repository org springframework spring web release spring web release jar dependency hierarchy webwolf jar root library spring boot starter web release jar x spring web release jar vulnerable library found in base branch main vulnerability details in spring framework versions and older unsupported versions it is possible for a user to provide malicious input to cause the insertion of additional log entries publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring release isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org owasp webgoat webwolf org springframework boot spring boot starter web release org springframework spring web release isminimumfixversionavailable true minimumfixversion org springframework spring release isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails in spring framework versions and older unsupported versions it is possible for a user to provide malicious input to cause the insertion of additional log entries vulnerabilityurl
0
41,290
2,868,994,330
IssuesEvent
2015-06-05 22:26:43
dart-lang/pub-dartlang
https://api.github.com/repos/dart-lang/pub-dartlang
closed
Supply stats for pub authors
enhancement MovedToGithub Priority-Medium
<a href="https://github.com/sethladd"><img src="https://avatars.githubusercontent.com/u/5479?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sethladd](https://github.com/sethladd)** _Originally opened as dart-lang/sdk#5424_ ---- A pub author will be interested in their own stats for their pub packages. Some stats would include: \* Page views \* Downloads (per version) \* Updates (if we track those differently from initial downloads) Ideally, pub can allow a pub author to supply their own Google Analytics code for their specific page. Note on implementation: if we stored all the binaries in google cloud storage, it'd be easy to move the logs into bigquery for analysis. Bonus: we could also use prediction api to offer recommendations for users (users who installed library A also installed library B)
1.0
Supply stats for pub authors - <a href="https://github.com/sethladd"><img src="https://avatars.githubusercontent.com/u/5479?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sethladd](https://github.com/sethladd)** _Originally opened as dart-lang/sdk#5424_ ---- A pub author will be interested in their own stats for their pub packages. Some stats would include: \* Page views \* Downloads (per version) \* Updates (if we track those differently from initial downloads) Ideally, pub can allow a pub author to supply their own Google Analytics code for their specific page. Note on implementation: if we stored all the binaries in google cloud storage, it'd be easy to move the logs into bigquery for analysis. Bonus: we could also use prediction api to offer recommendations for users (users who installed library A also installed library B)
non_process
supply stats for pub authors issue by originally opened as dart lang sdk a pub author will be interested in their own stats for their pub packages some stats would include page views downloads per version updates if we track those differently from initial downloads ideally pub can allow a pub author to supply their own google analytics code for their specific page note on implementation if we stored all the binaries in google cloud storage it d be easy to move the logs into bigquery for analysis bonus we could also use prediction api to offer recommendations for users users who installed library a also installed library b
0
12,388
14,908,341,709
IssuesEvent
2021-01-22 05:39:19
ncbo/bioportal-project
https://api.github.com/repos/ncbo/bioportal-project
opened
MedDRA concept tree doesn't display
bug ontology processing problem
Since UMLS 2020AB upload, the MedDRA ontology isn't displaying its concept tree, the issue being that no root concepts were identified. Misha and I have done some early investigation, and learned the following. No other UMLS concept trees in production have this issue; all the other ontologies not display a tree are "flat". In staging (version 2020AA) MedDRA displays correctly, with many roots. [1,2] The first root has 17 concepts under it. The first root concept ID is http://purl.bioontology.org/ontology/MEDDRA/10005329, the first ID below it is http://purl.bioontology.org/ontology/MEDDRA/10002086 and has an annotation as a subClassOf the first root concept ID (Blood and lymphatic system disorders, 10005329). "Inverse of SIB" relations, which identifies all the siblings of this class, is populated. The first root concept (10005329) shows it is a subclass of Owl#Thing. In production, while the top tree can not be viewed, the concepts are visible through searching. The 10005329 concept shows no children or parent, does not show it is a subclass of Owl#Thing, and does not show SIB relations. [3] The first concept under it in the staging case—10002086—does show a small tree under it. [4] It does not show itself as a subclass of the first root concept ID. A download of the 2020AA and 2020AB .ttl files were taken from production logs (respectively before and after the adjustment Alex made to reprocess everything). These were compared; there are over 37000 changes, but at least some of these are ordering changes. Both show approximately the same number of internal subClass relations (~ 200 more in 2020AB). [5] However, there are zero declarations that have the first root class ID above (10005329) as the object—the declaration of this for 10002086 is missing that subclass declaration. [6,7] Opening the 2020AB .ttl file in Protege confirms this subclass relationship is missing for 10002086. [8] Misha opened the 2020AB .xrdf file in Protege, and found the hierarchies are maintained in that document. (I don't have a screenshot of that and it may be worth confirming with a fresh download; I can't explain this inconsistency with the .ttl file.) Looking at the [UMLS Statistics page for MedDRA](https://www.nlm.nih.gov/research/umls/sourcereleasedocs/current/MDR/stats.html), we can see there are 37543 CHD (child) relations, which is within 300 of the total number of subClassOf relations, including those to terms outside MedDRA; it makes sense that we'd convert CHD relations to subClassOf, so that seems consistent. What is not clear is (1) what is keeping many subClassOf relations from showing up in the .ttl. (Especially if they are found in the .xrdf.), and (2) why SIB relations are not being captured in BioPortal for this ontology's terms. [1] Staging root class (10005329) ![Screen Shot 2021-01-21 at 8.40.45 PM.png](https://images.zenhubusercontent.com/5cc9bdc26c1fd80bacaab82a/8b933b98-fa23-407c-9d0d-1be6e6f76280) [2] Staging subclass (10002086) ![Screen Shot 2021-01-21 at 8.40.57 PM.png](https://images.zenhubusercontent.com/5cc9bdc26c1fd80bacaab82a/d672ae1d-91bf-411c-88a7-bb4fad46fd4c) [3] Production 'root class' (10005329) ![Screen Shot 2021-01-21 at 8.50.55 PM.png](https://images.zenhubusercontent.com/5cc9bdc26c1fd80bacaab82a/c0893a4d-e084-4643-b1ba-5fe43bca2878) [4] Production 'subclass' (10002086) ![Screen Shot 2021-01-21 at 8.39.59 PM.png](https://images.zenhubusercontent.com/5cc9bdc26c1fd80bacaab82a/b6013853-bc47-4297-b3f9-ef339b824b91) [5] Comparison of internal subclass counts ![Screen Shot 2021-01-21 at 9.04.30 PM.png](https://images.zenhubusercontent.com/5cc9bdc26c1fd80bacaab82a/96c212b0-102a-433c-9ec4-743d2ebc169a) [6] 2020AA concept 10005329 with subclass declaration ![Screen Shot 2021-01-21 at 9.19.33 PM.png](https://images.zenhubusercontent.com/5cc9bdc26c1fd80bacaab82a/c8273ffd-8f47-43b1-b0af-bbe8ac0e9bb1) [7] 2020AA concept 10005329 without subclass declaration ![Screen Shot 2021-01-21 at 9.20.54 PM.png](https://images.zenhubusercontent.com/5cc9bdc26c1fd80bacaab82a/aa81b46d-b85a-4f98-99c9-e20670555fc6) [8] Protege doesn't see subclass declaration either ![Screen Shot 2021-01-21 at 9.10.41 PM.png](https://images.zenhubusercontent.com/5cc9bdc26c1fd80bacaab82a/320ffae0-5354-4b28-8d99-ace505e8c5eb)
1.0
MedDRA concept tree doesn't display - Since UMLS 2020AB upload, the MedDRA ontology isn't displaying its concept tree, the issue being that no root concepts were identified. Misha and I have done some early investigation, and learned the following. No other UMLS concept trees in production have this issue; all the other ontologies not display a tree are "flat". In staging (version 2020AA) MedDRA displays correctly, with many roots. [1,2] The first root has 17 concepts under it. The first root concept ID is http://purl.bioontology.org/ontology/MEDDRA/10005329, the first ID below it is http://purl.bioontology.org/ontology/MEDDRA/10002086 and has an annotation as a subClassOf the first root concept ID (Blood and lymphatic system disorders, 10005329). "Inverse of SIB" relations, which identifies all the siblings of this class, is populated. The first root concept (10005329) shows it is a subclass of Owl#Thing. In production, while the top tree can not be viewed, the concepts are visible through searching. The 10005329 concept shows no children or parent, does not show it is a subclass of Owl#Thing, and does not show SIB relations. [3] The first concept under it in the staging case—10002086—does show a small tree under it. [4] It does not show itself as a subclass of the first root concept ID. A download of the 2020AA and 2020AB .ttl files were taken from production logs (respectively before and after the adjustment Alex made to reprocess everything). These were compared; there are over 37000 changes, but at least some of these are ordering changes. Both show approximately the same number of internal subClass relations (~ 200 more in 2020AB). [5] However, there are zero declarations that have the first root class ID above (10005329) as the object—the declaration of this for 10002086 is missing that subclass declaration. [6,7] Opening the 2020AB .ttl file in Protege confirms this subclass relationship is missing for 10002086. [8] Misha opened the 2020AB .xrdf file in Protege, and found the hierarchies are maintained in that document. (I don't have a screenshot of that and it may be worth confirming with a fresh download; I can't explain this inconsistency with the .ttl file.) Looking at the [UMLS Statistics page for MedDRA](https://www.nlm.nih.gov/research/umls/sourcereleasedocs/current/MDR/stats.html), we can see there are 37543 CHD (child) relations, which is within 300 of the total number of subClassOf relations, including those to terms outside MedDRA; it makes sense that we'd convert CHD relations to subClassOf, so that seems consistent. What is not clear is (1) what is keeping many subClassOf relations from showing up in the .ttl. (Especially if they are found in the .xrdf.), and (2) why SIB relations are not being captured in BioPortal for this ontology's terms. [1] Staging root class (10005329) ![Screen Shot 2021-01-21 at 8.40.45 PM.png](https://images.zenhubusercontent.com/5cc9bdc26c1fd80bacaab82a/8b933b98-fa23-407c-9d0d-1be6e6f76280) [2] Staging subclass (10002086) ![Screen Shot 2021-01-21 at 8.40.57 PM.png](https://images.zenhubusercontent.com/5cc9bdc26c1fd80bacaab82a/d672ae1d-91bf-411c-88a7-bb4fad46fd4c) [3] Production 'root class' (10005329) ![Screen Shot 2021-01-21 at 8.50.55 PM.png](https://images.zenhubusercontent.com/5cc9bdc26c1fd80bacaab82a/c0893a4d-e084-4643-b1ba-5fe43bca2878) [4] Production 'subclass' (10002086) ![Screen Shot 2021-01-21 at 8.39.59 PM.png](https://images.zenhubusercontent.com/5cc9bdc26c1fd80bacaab82a/b6013853-bc47-4297-b3f9-ef339b824b91) [5] Comparison of internal subclass counts ![Screen Shot 2021-01-21 at 9.04.30 PM.png](https://images.zenhubusercontent.com/5cc9bdc26c1fd80bacaab82a/96c212b0-102a-433c-9ec4-743d2ebc169a) [6] 2020AA concept 10005329 with subclass declaration ![Screen Shot 2021-01-21 at 9.19.33 PM.png](https://images.zenhubusercontent.com/5cc9bdc26c1fd80bacaab82a/c8273ffd-8f47-43b1-b0af-bbe8ac0e9bb1) [7] 2020AA concept 10005329 without subclass declaration ![Screen Shot 2021-01-21 at 9.20.54 PM.png](https://images.zenhubusercontent.com/5cc9bdc26c1fd80bacaab82a/aa81b46d-b85a-4f98-99c9-e20670555fc6) [8] Protege doesn't see subclass declaration either ![Screen Shot 2021-01-21 at 9.10.41 PM.png](https://images.zenhubusercontent.com/5cc9bdc26c1fd80bacaab82a/320ffae0-5354-4b28-8d99-ace505e8c5eb)
process
meddra concept tree doesn t display since umls upload the meddra ontology isn t displaying its concept tree the issue being that no root concepts were identified misha and i have done some early investigation and learned the following no other umls concept trees in production have this issue all the other ontologies not display a tree are flat in staging version meddra displays correctly with many roots the first root has concepts under it the first root concept id is the first id below it is and has an annotation as a subclassof the first root concept id blood and lymphatic system disorders inverse of sib relations which identifies all the siblings of this class is populated the first root concept shows it is a subclass of owl thing in production while the top tree can not be viewed the concepts are visible through searching the concept shows no children or parent does not show it is a subclass of owl thing and does not show sib relations the first concept under it in the staging case— —does show a small tree under it it does not show itself as a subclass of the first root concept id a download of the and ttl files were taken from production logs respectively before and after the adjustment alex made to reprocess everything these were compared there are over changes but at least some of these are ordering changes both show approximately the same number of internal subclass relations more in however there are zero declarations that have the first root class id above as the object—the declaration of this for is missing that subclass declaration opening the ttl file in protege confirms this subclass relationship is missing for misha opened the xrdf file in protege and found the hierarchies are maintained in that document i don t have a screenshot of that and it may be worth confirming with a fresh download i can t explain this inconsistency with the ttl file looking at the we can see there are chd child relations which is within of the total number of subclassof relations including those to terms outside meddra it makes sense that we d convert chd relations to subclassof so that seems consistent what is not clear is what is keeping many subclassof relations from showing up in the ttl especially if they are found in the xrdf and why sib relations are not being captured in bioportal for this ontology s terms staging root class staging subclass production root class production subclass comparison of internal subclass counts concept with subclass declaration concept without subclass declaration protege doesn t see subclass declaration either
1
59,214
14,539,801,766
IssuesEvent
2020-12-15 12:25:17
angular/angular-cli
https://api.github.com/repos/angular/angular-cli
closed
`--hmr` with `enableProdMode()` breaks HMR
comp: devkit/build-angular devkit/build-angular: dev-server freq1: low severity1: confusing type: bug/fix
<!--🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 Oh hi there! 😄 To expedite issue processing please search open and closed issues before submitting a new one. Existing issues often contain information about workarounds, resolution, or progress updates. 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅--> # 🐞 Bug report ### Command (mark with an `x`) <!-- Can you pin-point the command or commands that are effected by this bug? --> <!-- ✍️edit: --> - [ ] new - [ ] build - [x] serve - [ ] test - [ ] e2e - [ ] generate - [ ] add - [ ] update - [ ] lint - [ ] extract-i18n - [ ] run - [ ] config - [ ] help - [ ] version - [ ] doc ### Is this a regression? Not sure. ### Description Using the new `ng serve --hmr` functionality in Angular 11 while the app was also enabling production mode results in an error while swapping in the modules, and the page falling back to a normal reload with minimal indication that this was happening. **I was only able to see what's going on by enabling "Preserve Log" in the Chrome console.** A warning in the console if `enableProdMode()` is called while HMR is active might be a better indication about this. ## 🔬 Minimal Reproduction ``` ng serve --hmr ``` ```js enableProdMode() ``` ## 🔥 Exception or Error <pre><code> [HMR] Cannot apply update. Need to do a full reload! push.dZZH.module.exports @ polyfills.js:6327 (anonymous) @ polyfills.js:1434 invoke @ polyfills.js:13461 run @ polyfills.js:13220 (anonymous) @ polyfills.js:13954 invokeTask @ polyfills.js:13496 runTask @ polyfills.js:13264 drainMicroTaskQueue @ polyfills.js:13666 Promise.then (async) scheduleMicroTask @ polyfills.js:13649 scheduleTask @ polyfills.js:13485 scheduleTask @ polyfills.js:13307 scheduleMicroTask @ polyfills.js:13327 scheduleResolveOrReject @ polyfills.js:13944 then @ polyfills.js:14076 hotUpdateDownloaded @ runtime.js:386 hotAddUpdateChunk @ runtime.js:362 webpackHotUpdateCallback @ runtime.js:58 (anonymous) @ main.4b7dbf18777e0ec0b808.hot-update.js:1 log.js:26 [HMR] Error: Cannot enable prod mode after platform setup. at enableProdMode (http://localhost:3000/vendor.js:129787:15) [<root>] </code></pre> <img width="853" alt="Screenshot 2020-12-11 at 10 24 39" src="https://user-images.githubusercontent.com/697707/101880138-18625800-3b9b-11eb-87ab-e559a28bd7f1.png"> ## 🌍 Your Environment <pre><code> ng version _ _ ____ _ ___ / \ _ __ __ _ _ _| | __ _ _ __ / ___| | |_ _| / △ \ | '_ \ / _` | | | | |/ _` | '__| | | | | | | / ___ \| | | | (_| | |_| | | (_| | | | |___| |___ | | /_/ \_\_| |_|\__, |\__,_|_|\__,_|_| \____|_____|___| |___/ Angular CLI: 11.1.0-next.2 Node: 15.4.0 OS: darwin arm64 Angular: 11.0.4 ... animations, common, compiler, compiler-cli, core, forms ... language-service, platform-browser, platform-browser-dynamic ... router, service-worker Ivy Workspace: Yes Package Version ------------------------------------------------------------ @angular-devkit/architect 0.1101.0-next.2 @angular-devkit/build-angular 0.1101.0-next.2 @angular-devkit/core 11.1.0-next.2 @angular-devkit/schematics 11.1.0-next.2 @angular/cdk 11.0.2 @angular/cli 11.1.0-next.2 @angular/flex-layout 9.0.0-beta.31 @angular/material 11.0.2 @angular/material-moment-adapter 11.0.2 @schematics/angular 11.1.0-next.2 @schematics/update 0.1101.0-next.2 rxjs 6.6.3 typescript 4.0.5 </code></pre>
2.0
`--hmr` with `enableProdMode()` breaks HMR - <!--🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 Oh hi there! 😄 To expedite issue processing please search open and closed issues before submitting a new one. Existing issues often contain information about workarounds, resolution, or progress updates. 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅--> # 🐞 Bug report ### Command (mark with an `x`) <!-- Can you pin-point the command or commands that are effected by this bug? --> <!-- ✍️edit: --> - [ ] new - [ ] build - [x] serve - [ ] test - [ ] e2e - [ ] generate - [ ] add - [ ] update - [ ] lint - [ ] extract-i18n - [ ] run - [ ] config - [ ] help - [ ] version - [ ] doc ### Is this a regression? Not sure. ### Description Using the new `ng serve --hmr` functionality in Angular 11 while the app was also enabling production mode results in an error while swapping in the modules, and the page falling back to a normal reload with minimal indication that this was happening. **I was only able to see what's going on by enabling "Preserve Log" in the Chrome console.** A warning in the console if `enableProdMode()` is called while HMR is active might be a better indication about this. ## 🔬 Minimal Reproduction ``` ng serve --hmr ``` ```js enableProdMode() ``` ## 🔥 Exception or Error <pre><code> [HMR] Cannot apply update. Need to do a full reload! push.dZZH.module.exports @ polyfills.js:6327 (anonymous) @ polyfills.js:1434 invoke @ polyfills.js:13461 run @ polyfills.js:13220 (anonymous) @ polyfills.js:13954 invokeTask @ polyfills.js:13496 runTask @ polyfills.js:13264 drainMicroTaskQueue @ polyfills.js:13666 Promise.then (async) scheduleMicroTask @ polyfills.js:13649 scheduleTask @ polyfills.js:13485 scheduleTask @ polyfills.js:13307 scheduleMicroTask @ polyfills.js:13327 scheduleResolveOrReject @ polyfills.js:13944 then @ polyfills.js:14076 hotUpdateDownloaded @ runtime.js:386 hotAddUpdateChunk @ runtime.js:362 webpackHotUpdateCallback @ runtime.js:58 (anonymous) @ main.4b7dbf18777e0ec0b808.hot-update.js:1 log.js:26 [HMR] Error: Cannot enable prod mode after platform setup. at enableProdMode (http://localhost:3000/vendor.js:129787:15) [<root>] </code></pre> <img width="853" alt="Screenshot 2020-12-11 at 10 24 39" src="https://user-images.githubusercontent.com/697707/101880138-18625800-3b9b-11eb-87ab-e559a28bd7f1.png"> ## 🌍 Your Environment <pre><code> ng version _ _ ____ _ ___ / \ _ __ __ _ _ _| | __ _ _ __ / ___| | |_ _| / △ \ | '_ \ / _` | | | | |/ _` | '__| | | | | | | / ___ \| | | | (_| | |_| | | (_| | | | |___| |___ | | /_/ \_\_| |_|\__, |\__,_|_|\__,_|_| \____|_____|___| |___/ Angular CLI: 11.1.0-next.2 Node: 15.4.0 OS: darwin arm64 Angular: 11.0.4 ... animations, common, compiler, compiler-cli, core, forms ... language-service, platform-browser, platform-browser-dynamic ... router, service-worker Ivy Workspace: Yes Package Version ------------------------------------------------------------ @angular-devkit/architect 0.1101.0-next.2 @angular-devkit/build-angular 0.1101.0-next.2 @angular-devkit/core 11.1.0-next.2 @angular-devkit/schematics 11.1.0-next.2 @angular/cdk 11.0.2 @angular/cli 11.1.0-next.2 @angular/flex-layout 9.0.0-beta.31 @angular/material 11.0.2 @angular/material-moment-adapter 11.0.2 @schematics/angular 11.1.0-next.2 @schematics/update 0.1101.0-next.2 rxjs 6.6.3 typescript 4.0.5 </code></pre>
non_process
hmr with enableprodmode breaks hmr 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 oh hi there 😄 to expedite issue processing please search open and closed issues before submitting a new one existing issues often contain information about workarounds resolution or progress updates 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 🐞 bug report command mark with an x new build serve test generate add update lint extract run config help version doc is this a regression not sure description using the new ng serve hmr functionality in angular while the app was also enabling production mode results in an error while swapping in the modules and the page falling back to a normal reload with minimal indication that this was happening i was only able to see what s going on by enabling preserve log in the chrome console a warning in the console if enableprodmode is called while hmr is active might be a better indication about this 🔬 minimal reproduction ng serve hmr js enableprodmode 🔥 exception or error cannot apply update need to do a full reload push dzzh module exports polyfills js anonymous polyfills js invoke polyfills js run polyfills js anonymous polyfills js invoketask polyfills js runtask polyfills js drainmicrotaskqueue polyfills js promise then async schedulemicrotask polyfills js scheduletask polyfills js scheduletask polyfills js schedulemicrotask polyfills js scheduleresolveorreject polyfills js then polyfills js hotupdatedownloaded runtime js hotaddupdatechunk runtime js webpackhotupdatecallback runtime js anonymous main hot update js log js error cannot enable prod mode after platform setup at enableprodmode img width alt screenshot at src 🌍 your environment ng version △ angular cli next node os darwin angular animations common compiler compiler cli core forms language service platform browser platform browser dynamic router service worker ivy workspace yes package version angular devkit architect next angular devkit build angular next angular devkit core next angular devkit schematics next angular cdk angular cli next angular flex layout beta angular material angular material moment adapter schematics angular next schematics update next rxjs typescript
0
16,816
22,060,927,478
IssuesEvent
2022-05-30 17:42:39
bitPogo/kmock
https://api.github.com/repos/bitPogo/kmock
closed
Move UnitRelaxer Logic into ProxyFactory
enhancement kmock kmock-processor
## Description <!--- Provide a detailed introduction to the issue itself, and why you consider it to be a bug --> Currently the Processor hardwires all the Relaxers into the Mocks. At least the UnitRelaxer can be invoked with the ProxyFactory. Acceptance Criteria: At least `relaxVoidFunction` is wired by the ProxyFactory. Optional: Find a way to make that for all relaxers and Spies, in order to improve testability.
1.0
Move UnitRelaxer Logic into ProxyFactory - ## Description <!--- Provide a detailed introduction to the issue itself, and why you consider it to be a bug --> Currently the Processor hardwires all the Relaxers into the Mocks. At least the UnitRelaxer can be invoked with the ProxyFactory. Acceptance Criteria: At least `relaxVoidFunction` is wired by the ProxyFactory. Optional: Find a way to make that for all relaxers and Spies, in order to improve testability.
process
move unitrelaxer logic into proxyfactory description currently the processor hardwires all the relaxers into the mocks at least the unitrelaxer can be invoked with the proxyfactory acceptance criteria at least relaxvoidfunction is wired by the proxyfactory optional find a way to make that for all relaxers and spies in order to improve testability
1
235,344
19,338,989,467
IssuesEvent
2021-12-15 00:42:47
wuespace/telestion-client
https://api.github.com/repos/wuespace/telestion-client
closed
Update Storybook packages and configuration
:hammer: enhancement :link: dependencies :eyeglasses: tests
Please update the Storybook packages/components and their configuration. Some parts of storybook have changed drastically, e.g. Create React App is now in a separate thread and the default PostCSS plugin is deprecated.
1.0
Update Storybook packages and configuration - Please update the Storybook packages/components and their configuration. Some parts of storybook have changed drastically, e.g. Create React App is now in a separate thread and the default PostCSS plugin is deprecated.
non_process
update storybook packages and configuration please update the storybook packages components and their configuration some parts of storybook have changed drastically e g create react app is now in a separate thread and the default postcss plugin is deprecated
0
9,356
12,366,368,880
IssuesEvent
2020-05-18 10:19:39
DiSSCo/user-stories
https://api.github.com/repos/DiSSCo/user-stories
opened
an automated identification tool
1. NH museum 4. Data processing ICEDIG-SURVEY Research Specimen level
As a Citizen Scientist I want to identify a specimen so that I can know which organism I have encountered for this I need an automated identification tool
1.0
an automated identification tool - As a Citizen Scientist I want to identify a specimen so that I can know which organism I have encountered for this I need an automated identification tool
process
an automated identification tool as a citizen scientist i want to identify a specimen so that i can know which organism i have encountered for this i need an automated identification tool
1
22,190
30,744,080,720
IssuesEvent
2023-07-28 13:45:53
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Status of Bazel 7.0.0-pre.20230710.5
P1 type: process release team-OSS
- Expected release date: 2023-07-27 Task list: - [x] Pick release baseline: [7845acae](https://github.com/bazelbuild/bazel/commit/7845acae9769a72dc507dc2f57c4e032ebf429d3) with cherrypicks [d9e2f918](https://github.com/bazelbuild/bazel/commit/d9e2f9181f8fa283e3986ee3b261e610c41cf61b) [da23370d](https://github.com/bazelbuild/bazel/commit/da23370dcdf6ea19545002fb86bd5d3e6519cdf6) [4714cd70](https://github.com/bazelbuild/bazel/commit/4714cd70c83de43ba7ab10c29e63af716216770a) [0a1d5053](https://github.com/bazelbuild/bazel/commit/0a1d50533078283fab93024681c917303a91c46c) - [x] Create release candidate: https://releases.bazel.build/7.0.0/rolling/7.0.0-pre.20230710.5rc1/index.html - [x] Post-submit: https://buildkite.com/bazel/bazel-bazel/builds?branch=release-7.0.0-pre.20230710.5rc1 - [x] Push the release: https://releases.bazel.build/7.0.0/rolling/7.0.0-pre.20230710.5/index.html - [x] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
1.0
Status of Bazel 7.0.0-pre.20230710.5 - - Expected release date: 2023-07-27 Task list: - [x] Pick release baseline: [7845acae](https://github.com/bazelbuild/bazel/commit/7845acae9769a72dc507dc2f57c4e032ebf429d3) with cherrypicks [d9e2f918](https://github.com/bazelbuild/bazel/commit/d9e2f9181f8fa283e3986ee3b261e610c41cf61b) [da23370d](https://github.com/bazelbuild/bazel/commit/da23370dcdf6ea19545002fb86bd5d3e6519cdf6) [4714cd70](https://github.com/bazelbuild/bazel/commit/4714cd70c83de43ba7ab10c29e63af716216770a) [0a1d5053](https://github.com/bazelbuild/bazel/commit/0a1d50533078283fab93024681c917303a91c46c) - [x] Create release candidate: https://releases.bazel.build/7.0.0/rolling/7.0.0-pre.20230710.5rc1/index.html - [x] Post-submit: https://buildkite.com/bazel/bazel-bazel/builds?branch=release-7.0.0-pre.20230710.5rc1 - [x] Push the release: https://releases.bazel.build/7.0.0/rolling/7.0.0-pre.20230710.5/index.html - [x] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
process
status of bazel pre expected release date task list pick release baseline with cherrypicks create release candidate post submit push the release update the
1
163,897
12,749,751,187
IssuesEvent
2020-06-27 00:08:57
mozilla-mobile/fenix
https://api.github.com/repos/mozilla-mobile/fenix
opened
Intermittent test failure - FenixSearchEngineProviderTest - sharedprefs contains installed engines when installedSearchEngineIdentifiers
intermittent-test
### Test Run: https://firefoxci.taskcluster-artifacts.net/eh0VCU3UQLaXUooAI5QEuw/0/public/reports/test/testGeckoNightlyDebugUnitTest/classes/org.mozilla.fenix.components.searchengine.FenixSearchEngineProviderTest.html#GIVEN%20sharedprefs%20contains%20installed%20engines%20WHEN%20installedSearchEngineIdentifiers%20THEN%20defaultEngines%20+%20customEngines%20ids%20are%20returned ### Stacktrace: ``` java.lang.IllegalStateException: This job has not completed yet at kotlinx.coroutines.JobSupport.getCompletionExceptionOrNull(JobSupport.kt:1189) at kotlinx.coroutines.test.TestBuildersKt.runBlockingTest(TestBuilders.kt:53) at kotlinx.coroutines.test.TestBuildersKt.runBlockingTest$default(TestBuilders.kt:45) at org.mozilla.fenix.components.searchengine.FenixSearchEngineProviderTest.GIVEN sharedprefs contains installed engines WHEN installedSearchEngineIdentifiers THEN defaultEngines + customEngines ids are returned(FenixSearchEngineProviderTest.kt:84) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.robolectric.RobolectricTestRunner$HelperTestRunner$1.evaluate(RobolectricTestRunner.java:546) at org.robolectric.internal.SandboxTestRunner$2.lambda$evaluate$0(SandboxTestRunner.java:252) at org.robolectric.internal.bytecode.Sandbox.lambda$runOnMainThread$0(Sandbox.java:89) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) ``` ### Build: https://github.com/mozilla-mobile/fenix/pull/12041#issuecomment-650457568
1.0
Intermittent test failure - FenixSearchEngineProviderTest - sharedprefs contains installed engines when installedSearchEngineIdentifiers - ### Test Run: https://firefoxci.taskcluster-artifacts.net/eh0VCU3UQLaXUooAI5QEuw/0/public/reports/test/testGeckoNightlyDebugUnitTest/classes/org.mozilla.fenix.components.searchengine.FenixSearchEngineProviderTest.html#GIVEN%20sharedprefs%20contains%20installed%20engines%20WHEN%20installedSearchEngineIdentifiers%20THEN%20defaultEngines%20+%20customEngines%20ids%20are%20returned ### Stacktrace: ``` java.lang.IllegalStateException: This job has not completed yet at kotlinx.coroutines.JobSupport.getCompletionExceptionOrNull(JobSupport.kt:1189) at kotlinx.coroutines.test.TestBuildersKt.runBlockingTest(TestBuilders.kt:53) at kotlinx.coroutines.test.TestBuildersKt.runBlockingTest$default(TestBuilders.kt:45) at org.mozilla.fenix.components.searchengine.FenixSearchEngineProviderTest.GIVEN sharedprefs contains installed engines WHEN installedSearchEngineIdentifiers THEN defaultEngines + customEngines ids are returned(FenixSearchEngineProviderTest.kt:84) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.robolectric.RobolectricTestRunner$HelperTestRunner$1.evaluate(RobolectricTestRunner.java:546) at org.robolectric.internal.SandboxTestRunner$2.lambda$evaluate$0(SandboxTestRunner.java:252) at org.robolectric.internal.bytecode.Sandbox.lambda$runOnMainThread$0(Sandbox.java:89) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) ``` ### Build: https://github.com/mozilla-mobile/fenix/pull/12041#issuecomment-650457568
non_process
intermittent test failure fenixsearchengineprovidertest sharedprefs contains installed engines when installedsearchengineidentifiers test run stacktrace java lang illegalstateexception this job has not completed yet at kotlinx coroutines jobsupport getcompletionexceptionornull jobsupport kt at kotlinx coroutines test testbuilderskt runblockingtest testbuilders kt at kotlinx coroutines test testbuilderskt runblockingtest default testbuilders kt at org mozilla fenix components searchengine fenixsearchengineprovidertest given sharedprefs contains installed engines when installedsearchengineidentifiers then defaultengines customengines ids are returned fenixsearchengineprovidertest kt at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org robolectric robolectrictestrunner helpertestrunner evaluate robolectrictestrunner java at org robolectric internal sandboxtestrunner lambda evaluate sandboxtestrunner java at org robolectric internal bytecode sandbox lambda runonmainthread sandbox java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java build
0
79,491
7,717,717,395
IssuesEvent
2018-05-23 14:26:08
openshift/origin
https://api.github.com/repos/openshift/origin
opened
TestOAuthServiceAccountClientEvent is timing sensitive
kind/test-flake
in reference to https://github.com/openshift/origin/pull/16850, https://github.com/openshift/origin/issues/16940, and https://github.com/openshift/origin/pull/19790 Is one of the integration tests that still appears to be flaky https://openshift-gce-devel.appspot.com/build/origin-ci-test/pr-logs/pull/19759/test_pull_request_origin_integration/18011/ https://openshift-gce-devel.appspot.com/build/origin-ci-test/pr-logs/pull/19759/test_pull_request_origin_integration/18027/ https://openshift-gce-devel.appspot.com/build/origin-ci-test/pr-logs/pull/19759/test_pull_request_origin_integration/18066/ https://openshift-gce-devel.appspot.com/build/origin-ci-test/pr-logs/pull/19759/test_pull_request_origin_integration/18075/ Assigning @mrogers950 as he appears to be the last one touching the test. /cc: @enj , @bparees
1.0
TestOAuthServiceAccountClientEvent is timing sensitive - in reference to https://github.com/openshift/origin/pull/16850, https://github.com/openshift/origin/issues/16940, and https://github.com/openshift/origin/pull/19790 Is one of the integration tests that still appears to be flaky https://openshift-gce-devel.appspot.com/build/origin-ci-test/pr-logs/pull/19759/test_pull_request_origin_integration/18011/ https://openshift-gce-devel.appspot.com/build/origin-ci-test/pr-logs/pull/19759/test_pull_request_origin_integration/18027/ https://openshift-gce-devel.appspot.com/build/origin-ci-test/pr-logs/pull/19759/test_pull_request_origin_integration/18066/ https://openshift-gce-devel.appspot.com/build/origin-ci-test/pr-logs/pull/19759/test_pull_request_origin_integration/18075/ Assigning @mrogers950 as he appears to be the last one touching the test. /cc: @enj , @bparees
non_process
testoauthserviceaccountclientevent is timing sensitive in reference to and is one of the integration tests that still appears to be flaky assigning as he appears to be the last one touching the test cc enj bparees
0
17,638
23,460,975,111
IssuesEvent
2022-08-16 13:06:05
NationalSecurityAgency/ghidra
https://api.github.com/repos/NationalSecurityAgency/ghidra
closed
Context Dependent Decoding of Jump Instructions
Feature: Processor/x86 Status: Prioritize
**Is your feature request related to a problem? Please describe.** *Ghidra* decodes `0F 84` as `JZ` following a comparison with a non-zero value. ![image](https://user-images.githubusercontent.com/10724464/114182542-058e4300-9943-11eb-97e6-7f6718812537.png) **Describe the solution you'd like** Wouldn't it be nicer to decode as `JE`? **Environment** OS: Windows 10 10.0 amd64 (VM) Java Version: AdoptOpenJDK 11.0.10 Ghidra Version: 9.3 DEV, commit b69c3d6 Ghidra Origin: locally built
1.0
Context Dependent Decoding of Jump Instructions - **Is your feature request related to a problem? Please describe.** *Ghidra* decodes `0F 84` as `JZ` following a comparison with a non-zero value. ![image](https://user-images.githubusercontent.com/10724464/114182542-058e4300-9943-11eb-97e6-7f6718812537.png) **Describe the solution you'd like** Wouldn't it be nicer to decode as `JE`? **Environment** OS: Windows 10 10.0 amd64 (VM) Java Version: AdoptOpenJDK 11.0.10 Ghidra Version: 9.3 DEV, commit b69c3d6 Ghidra Origin: locally built
process
context dependent decoding of jump instructions is your feature request related to a problem please describe ghidra decodes as jz following a comparison with a non zero value describe the solution you d like wouldn t it be nicer to decode as je environment os windows vm java version adoptopenjdk ghidra version dev commit ghidra origin locally built
1
94,993
10,863,361,223
IssuesEvent
2019-11-14 15:00:00
quarkusio/quarkus
https://api.github.com/repos/quarkusio/quarkus
closed
Make an explicit section on Hibernate ORM in production with the different defaults
component:documentation enhancement
The default import.sql file is not executed when running in Native mode. According to the docs, Hibernate ORM, upon boot, will read and execute the SQL statements in the /import.sql file (if present). That works in JVM Mode. In native mode the script is not executed. As workaround, you have to set explicitly the property: `quarkus.hibernate-orm.sql-load-script=import.sql` I have used this project to reproduce the issue: https://bit.ly/32Iw7Zh Environment: Quarkus 1.0.0.CR1 $ uname -a Linux fedora 4.18.16-300.fc29.x86_64 #1 SMP Sat Oct 20 23:24:08 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux GraalVM: graalvm-ce-19.2.0 $GRAALVM_HOME/bin/java -version openjdk version "1.8.0_222" OpenJDK Runtime Environment (build 1.8.0_222-20190711120915.buildslave.jdk8u-src-tar--b08) OpenJDK 64-Bit GraalVM CE 19.2.0 (build 25.222-b08-jvmci-19.2-b02, mixed mode)
1.0
Make an explicit section on Hibernate ORM in production with the different defaults - The default import.sql file is not executed when running in Native mode. According to the docs, Hibernate ORM, upon boot, will read and execute the SQL statements in the /import.sql file (if present). That works in JVM Mode. In native mode the script is not executed. As workaround, you have to set explicitly the property: `quarkus.hibernate-orm.sql-load-script=import.sql` I have used this project to reproduce the issue: https://bit.ly/32Iw7Zh Environment: Quarkus 1.0.0.CR1 $ uname -a Linux fedora 4.18.16-300.fc29.x86_64 #1 SMP Sat Oct 20 23:24:08 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux GraalVM: graalvm-ce-19.2.0 $GRAALVM_HOME/bin/java -version openjdk version "1.8.0_222" OpenJDK Runtime Environment (build 1.8.0_222-20190711120915.buildslave.jdk8u-src-tar--b08) OpenJDK 64-Bit GraalVM CE 19.2.0 (build 25.222-b08-jvmci-19.2-b02, mixed mode)
non_process
make an explicit section on hibernate orm in production with the different defaults the default import sql file is not executed when running in native mode according to the docs hibernate orm upon boot will read and execute the sql statements in the import sql file if present that works in jvm mode in native mode the script is not executed as workaround you have to set explicitly the property quarkus hibernate orm sql load script import sql i have used this project to reproduce the issue environment quarkus uname a linux fedora smp sat oct utc gnu linux graalvm graalvm ce graalvm home bin java version openjdk version openjdk runtime environment build buildslave src tar openjdk bit graalvm ce build jvmci mixed mode
0
9,785
12,801,170,542
IssuesEvent
2020-07-02 18:35:12
solid/process
https://api.github.com/repos/solid/process
closed
Test Suite not really a Panel
process proposal
The test suite is in a separate repository from the specification so it could be interpreted as falling outside of the editorial process. Although there is a test suite panel, there is no corresponding editorial assignment so it is unclear who would review proposals from the test suite panel. One solution could be to have an independent group from the editorial process working on the test suite which is based on the specification produced by the editorial process.
1.0
Test Suite not really a Panel - The test suite is in a separate repository from the specification so it could be interpreted as falling outside of the editorial process. Although there is a test suite panel, there is no corresponding editorial assignment so it is unclear who would review proposals from the test suite panel. One solution could be to have an independent group from the editorial process working on the test suite which is based on the specification produced by the editorial process.
process
test suite not really a panel the test suite is in a separate repository from the specification so it could be interpreted as falling outside of the editorial process although there is a test suite panel there is no corresponding editorial assignment so it is unclear who would review proposals from the test suite panel one solution could be to have an independent group from the editorial process working on the test suite which is based on the specification produced by the editorial process
1
345,690
24,870,690,426
IssuesEvent
2022-10-27 15:01:26
kubernetes-client/python
https://api.github.com/repos/kubernetes-client/python
opened
Body is not acccepted as None for patch_certificate_signing_request_approval
kind/documentation
**Link to the issue (please include a link to the specific documentation or example)**: https://github.com/kubernetes-client/python/blob/master/kubernetes/docs/CertificatesV1Api.md#patch_certificate_signing_request_approval **Description of the issue (please include outputs or screenshots if possible)**: ```/home/cloud-user/cnSBC/lib64/python3.6/site-packages/requests/__init__.py:104: RequestsDependencyWarning: urllib3 (1.26.12) or chardet (5.0.0 )/charset_normalizer (2.0.12) doesn't match a supported version! RequestsDependencyWarning) Traceback (most recent call last): File "approve_crt.py", line 16, in <module> api_response = v1.patch_certificate_signing_request_approval(name, body, pretty=True) File "/home/cloud-user/cnSBC/lib64/python3.6/site-packages/kubernetes/client/api/certificates_v1_api.py", line 928, in patch_certificate_si gning_request_approval return self.patch_certificate_signing_request_approval_with_http_info(name, body, **kwargs) # noqa: E501 File "/home/cloud-user/cnSBC/lib64/python3.6/site-packages/kubernetes/client/api/certificates_v1_api.py", line 996, in patch_certificate_si gning_request_approval_with_http_info raise ApiValueError("Missing the required parameter `body` when calling `patch_certificate_signing_request_approval`") # noqa: E501 kubernetes.client.exceptions.ApiValueError: Missing the required parameter `body` when calling `patch_certificate_signing_request_approval` ``` As a workaround -> I tried by adding `body = {"metadata": {"namespace": "cnsbcapp-operator-system"}}` but even then, the approval didn't worked. Output: ``` /home/cloud-user/cnSBC/lib64/python3.6/site-packages/requests/__init__.py:104: RequestsDependencyWarning: u rllib3 (1.26.12) or chardet (5.0.0)/charset_normalizer (2.0.12) doesn't match a supported version! RequestsDependencyWarning) {'api_version': 'certificates.k8s.io/v1', 'kind': 'CertificateSigningRequest', 'metadata': {'annotations': None, 'cluster_name': None, 'creation_timestamp': datetime.datetime(2022, 10, 27, 10, 16, 31, tzinfo=tzutc()), 'deletion_grace_period_seconds': None, 'deletion_timestamp': None, 'finalizers': None, 'generate_name': None, 'generation': None, 'labels': None, 'managed_fields': None, 'name': 'configservice.cnsbcapp-operator-system', 'namespace': None, 'owner_references': None, 'resource_version': '1715067', 'self_link': None, 'uid': '72b8e205-8730-47f4-b82f-264173082e54'}, 'spec': {'expiration_seconds': None, 'extra': None, 'groups': ['system:masters', 'system:authenticated'], 'request': 'LS0tLS1CRUdJTiBDRVJUSUZJQ0FURSBSRVFVRVNULS0tLS0KTUlJQlZUQ0IvUUlCQURBL01UMHdPd1lEVlFRR EV6UmpiMjVtYVdkelpYSjJhV05sTG1OdWMySmpZWEJ3TFc5dwpaWEpoZEc5eUxYTjVjM1JsYlM1amJIVnpkR1Z5TG14dlkyRnNNRmt3RXdZSEtvWkl6ajBDQVFZSUtvWkl6ajBECkFRY0RRZ0FFOFJ5aTlRQWk0ZmxpNEY4aThuVnJKdE02aU1jMStzOUxwQVBjbC9NZGREMjA0cHVhSmtYUVJ1TVQKaW9PUzRhUHNJemh1Zmk5anRmUHVrN0ltZWtSRGphQmNNRm9HQ1NxR1NJYjNEUUVKRGpGTk1Fc3dTUVlEVlIwUgpCRUl3UUlJNFkyOXVabWxuYzJWeWRtbGpaUzVqYm5OaVkyRndjQzF2Y0dWeVlYUnZjaTF6ZVhOMFpXMHVjM1pqCkxtTnNkWE4wWlhJdWJHOWpZV3lIQkFwcXNNY3dDZ1lJS29aSXpqMEVBd0lEUndBd1JBSWdSU3BpdHRDOE0ycE4KZHU3RXdFY3RuRzc4eTdpNHI4N2ZVYkUwaXBydkx0SUNJR2JjUFdNTmpiOXpFUWF0SHJsQjFDUWZDM2ViVk93WAo0dmtWVkFLWmF6T3oKLS0tLS1FTkQgQ0VSVElGSUNBVEUgUkVRVUVTVC0tLS0tCg==', 'signer_name': 'cnsbc.cgbu.oracle.com/configservice', 'uid': None, 'usages': ['digital signature', 'key encipherment', 'server auth'], 'username': 'kubernetes-admin'}, 'status': {'certificate': 'LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUMyakNDQWNLZ0F3SUJBZ0lVZkZFeTBBMmtBdUZ qdkd4Szd3TG5PS0JMNzg4d0RRWUpLb1pJaHZjTkFRRUwKQlFBd0ZERVNNQkFHQTFVRUF4TUpZMjR0ZEdWemRDMHhNQ0FYRFRJeU1UQXlOekV4TURZd01Gb1lEekl4TWpJeApNREF6TVRFd05qQXdXakEvTVQwd093WURWUVFERXpSamIyNW1hV2R6WlhKMmFXTmxMbU51YzJKallYQndMVzl3ClpYSmhkRzl5TFhONWMzUmxiUzVqYkhWemRHVnlMbXh2WTJGc01Ga3dFd1lIS29aSXpqMENBUVlJS29aSXpqMEQKQVFjRFFnQUU4UnlpOVFBaTRmbGk0RjhpOG5Wckp0TTZpTWMxK3M5THBBUGNsL01kZEQyMDRwdWFKa1hRUnVNVAppb09TNGFQc0l6aHVmaTlqdGZQdWs3SW1la1JEamFPQndUQ0J2akFPQmdOVkhROEJBZjhFQkFNQ0JhQXdFd1lEClZSMGxCQXd3Q2dZSUt3WUJCUVVIQXdFd0RBWURWUjBUQVFIL0JBSXdBREFkQmdOVkhRNEVGZ1FVaDhpSVZJSkIKNzBLb1QvbklVaU5lMlUwZWhBQXdId1lEVlIwakJCZ3dGb0FVOWE1MXZ0Vk9rV09nUkU5OFJIUmR3Q0pWYXFNdwpTUVlEVlIwUkJFSXdRSUk0WTI5dVptbG5jMlZ5ZG1salpTNWpibk5pWTJGd2NDMXZjR1Z5WVhSdmNpMXplWE4wClpXMHVjM1pqTG1Oc2RYTjBaWEl1Ykc5allXeUhCQXBxc01jd0RRWUpLb1pJaHZjTkFRRUxCUUFEZ2dFQkFIdkIKM2VYN0d5Z1l6WXJUdEFiR3h3Y2JvcXpHN0NlTlFFN3A0MGxQRFcwTXFSQ21QaFFTai8yZUFWL3E1N2lEVURaZAphc0xHMU9xbUJwYkRUSVgzNitXUm9rTDRKN2V6VXZBVTFmNWF4cUFrTXJVcE1WeFNBWVFMc3pKRkRHTVRjOS9QCjd4OFdaL0k2WUdFUFpXUVJzaGh4RU5LVmhYNThHZGU4dUZVSWpsSU5SVDFiTVFHRlV2N1BLZXErRitDNmpTZm8Kcml5L0w2U0ZqMTNWU0FvOHdXbnU5Rk0vWk55bTh5VytuL0pzM0sxbzU4ckQrK0l3UkFGcitnbEFWYTlDZnNHawpqSDZvQUxuUThSZlJPSFZ5SVNNTG1LYTVmSUtTTTNsT2lJVnRqQ3MzY0l1NGNVL0M4b2VBTGtyZ0ZlYVM1bVh0Cm0waG5mSWZ0MlNZUlIrU2M1VDg9Ci0tLS0tRU5EIENFUlRJRklDQVRFLS0tLS0K', 'conditions': None}} ```
1.0
Body is not acccepted as None for patch_certificate_signing_request_approval - **Link to the issue (please include a link to the specific documentation or example)**: https://github.com/kubernetes-client/python/blob/master/kubernetes/docs/CertificatesV1Api.md#patch_certificate_signing_request_approval **Description of the issue (please include outputs or screenshots if possible)**: ```/home/cloud-user/cnSBC/lib64/python3.6/site-packages/requests/__init__.py:104: RequestsDependencyWarning: urllib3 (1.26.12) or chardet (5.0.0 )/charset_normalizer (2.0.12) doesn't match a supported version! RequestsDependencyWarning) Traceback (most recent call last): File "approve_crt.py", line 16, in <module> api_response = v1.patch_certificate_signing_request_approval(name, body, pretty=True) File "/home/cloud-user/cnSBC/lib64/python3.6/site-packages/kubernetes/client/api/certificates_v1_api.py", line 928, in patch_certificate_si gning_request_approval return self.patch_certificate_signing_request_approval_with_http_info(name, body, **kwargs) # noqa: E501 File "/home/cloud-user/cnSBC/lib64/python3.6/site-packages/kubernetes/client/api/certificates_v1_api.py", line 996, in patch_certificate_si gning_request_approval_with_http_info raise ApiValueError("Missing the required parameter `body` when calling `patch_certificate_signing_request_approval`") # noqa: E501 kubernetes.client.exceptions.ApiValueError: Missing the required parameter `body` when calling `patch_certificate_signing_request_approval` ``` As a workaround -> I tried by adding `body = {"metadata": {"namespace": "cnsbcapp-operator-system"}}` but even then, the approval didn't worked. Output: ``` /home/cloud-user/cnSBC/lib64/python3.6/site-packages/requests/__init__.py:104: RequestsDependencyWarning: u rllib3 (1.26.12) or chardet (5.0.0)/charset_normalizer (2.0.12) doesn't match a supported version! RequestsDependencyWarning) {'api_version': 'certificates.k8s.io/v1', 'kind': 'CertificateSigningRequest', 'metadata': {'annotations': None, 'cluster_name': None, 'creation_timestamp': datetime.datetime(2022, 10, 27, 10, 16, 31, tzinfo=tzutc()), 'deletion_grace_period_seconds': None, 'deletion_timestamp': None, 'finalizers': None, 'generate_name': None, 'generation': None, 'labels': None, 'managed_fields': None, 'name': 'configservice.cnsbcapp-operator-system', 'namespace': None, 'owner_references': None, 'resource_version': '1715067', 'self_link': None, 'uid': '72b8e205-8730-47f4-b82f-264173082e54'}, 'spec': {'expiration_seconds': None, 'extra': None, 'groups': ['system:masters', 'system:authenticated'], 'request': 'LS0tLS1CRUdJTiBDRVJUSUZJQ0FURSBSRVFVRVNULS0tLS0KTUlJQlZUQ0IvUUlCQURBL01UMHdPd1lEVlFRR EV6UmpiMjVtYVdkelpYSjJhV05sTG1OdWMySmpZWEJ3TFc5dwpaWEpoZEc5eUxYTjVjM1JsYlM1amJIVnpkR1Z5TG14dlkyRnNNRmt3RXdZSEtvWkl6ajBDQVFZSUtvWkl6ajBECkFRY0RRZ0FFOFJ5aTlRQWk0ZmxpNEY4aThuVnJKdE02aU1jMStzOUxwQVBjbC9NZGREMjA0cHVhSmtYUVJ1TVQKaW9PUzRhUHNJemh1Zmk5anRmUHVrN0ltZWtSRGphQmNNRm9HQ1NxR1NJYjNEUUVKRGpGTk1Fc3dTUVlEVlIwUgpCRUl3UUlJNFkyOXVabWxuYzJWeWRtbGpaUzVqYm5OaVkyRndjQzF2Y0dWeVlYUnZjaTF6ZVhOMFpXMHVjM1pqCkxtTnNkWE4wWlhJdWJHOWpZV3lIQkFwcXNNY3dDZ1lJS29aSXpqMEVBd0lEUndBd1JBSWdSU3BpdHRDOE0ycE4KZHU3RXdFY3RuRzc4eTdpNHI4N2ZVYkUwaXBydkx0SUNJR2JjUFdNTmpiOXpFUWF0SHJsQjFDUWZDM2ViVk93WAo0dmtWVkFLWmF6T3oKLS0tLS1FTkQgQ0VSVElGSUNBVEUgUkVRVUVTVC0tLS0tCg==', 'signer_name': 'cnsbc.cgbu.oracle.com/configservice', 'uid': None, 'usages': ['digital signature', 'key encipherment', 'server auth'], 'username': 'kubernetes-admin'}, 'status': {'certificate': 'LS0tLS1CRUdJTiBDRVJUSUZJQ0FURS0tLS0tCk1JSUMyakNDQWNLZ0F3SUJBZ0lVZkZFeTBBMmtBdUZ qdkd4Szd3TG5PS0JMNzg4d0RRWUpLb1pJaHZjTkFRRUwKQlFBd0ZERVNNQkFHQTFVRUF4TUpZMjR0ZEdWemRDMHhNQ0FYRFRJeU1UQXlOekV4TURZd01Gb1lEekl4TWpJeApNREF6TVRFd05qQXdXakEvTVQwd093WURWUVFERXpSamIyNW1hV2R6WlhKMmFXTmxMbU51YzJKallYQndMVzl3ClpYSmhkRzl5TFhONWMzUmxiUzVqYkhWemRHVnlMbXh2WTJGc01Ga3dFd1lIS29aSXpqMENBUVlJS29aSXpqMEQKQVFjRFFnQUU4UnlpOVFBaTRmbGk0RjhpOG5Wckp0TTZpTWMxK3M5THBBUGNsL01kZEQyMDRwdWFKa1hRUnVNVAppb09TNGFQc0l6aHVmaTlqdGZQdWs3SW1la1JEamFPQndUQ0J2akFPQmdOVkhROEJBZjhFQkFNQ0JhQXdFd1lEClZSMGxCQXd3Q2dZSUt3WUJCUVVIQXdFd0RBWURWUjBUQVFIL0JBSXdBREFkQmdOVkhRNEVGZ1FVaDhpSVZJSkIKNzBLb1QvbklVaU5lMlUwZWhBQXdId1lEVlIwakJCZ3dGb0FVOWE1MXZ0Vk9rV09nUkU5OFJIUmR3Q0pWYXFNdwpTUVlEVlIwUkJFSXdRSUk0WTI5dVptbG5jMlZ5ZG1salpTNWpibk5pWTJGd2NDMXZjR1Z5WVhSdmNpMXplWE4wClpXMHVjM1pqTG1Oc2RYTjBaWEl1Ykc5allXeUhCQXBxc01jd0RRWUpLb1pJaHZjTkFRRUxCUUFEZ2dFQkFIdkIKM2VYN0d5Z1l6WXJUdEFiR3h3Y2JvcXpHN0NlTlFFN3A0MGxQRFcwTXFSQ21QaFFTai8yZUFWL3E1N2lEVURaZAphc0xHMU9xbUJwYkRUSVgzNitXUm9rTDRKN2V6VXZBVTFmNWF4cUFrTXJVcE1WeFNBWVFMc3pKRkRHTVRjOS9QCjd4OFdaL0k2WUdFUFpXUVJzaGh4RU5LVmhYNThHZGU4dUZVSWpsSU5SVDFiTVFHRlV2N1BLZXErRitDNmpTZm8Kcml5L0w2U0ZqMTNWU0FvOHdXbnU5Rk0vWk55bTh5VytuL0pzM0sxbzU4ckQrK0l3UkFGcitnbEFWYTlDZnNHawpqSDZvQUxuUThSZlJPSFZ5SVNNTG1LYTVmSUtTTTNsT2lJVnRqQ3MzY0l1NGNVL0M4b2VBTGtyZ0ZlYVM1bVh0Cm0waG5mSWZ0MlNZUlIrU2M1VDg9Ci0tLS0tRU5EIENFUlRJRklDQVRFLS0tLS0K', 'conditions': None}} ```
non_process
body is not acccepted as none for patch certificate signing request approval link to the issue please include a link to the specific documentation or example description of the issue please include outputs or screenshots if possible home cloud user cnsbc site packages requests init py requestsdependencywarning or chardet charset normalizer doesn t match a supported version requestsdependencywarning traceback most recent call last file approve crt py line in api response patch certificate signing request approval name body pretty true file home cloud user cnsbc site packages kubernetes client api certificates api py line in patch certificate si gning request approval return self patch certificate signing request approval with http info name body kwargs noqa file home cloud user cnsbc site packages kubernetes client api certificates api py line in patch certificate si gning request approval with http info raise apivalueerror missing the required parameter body when calling patch certificate signing request approval noqa kubernetes client exceptions apivalueerror missing the required parameter body when calling patch certificate signing request approval as a workaround i tried by adding body metadata namespace cnsbcapp operator system but even then the approval didn t worked output home cloud user cnsbc site packages requests init py requestsdependencywarning u or chardet charset normalizer doesn t match a supported version requestsdependencywarning api version certificates io kind certificatesigningrequest metadata annotations none cluster name none creation timestamp datetime datetime tzinfo tzutc deletion grace period seconds none deletion timestamp none finalizers none generate name none generation none labels none managed fields none name configservice cnsbcapp operator system namespace none owner references none resource version self link none uid spec expiration seconds none extra none groups request signer name cnsbc cgbu oracle com configservice uid none usages username kubernetes admin status certificate conditions none
0
7,973
11,161,944,967
IssuesEvent
2019-12-26 15:50:00
pytorch/xla
https://api.github.com/repos/pytorch/xla
closed
Training is sooooooooo slow
multiprocessing needs reproduction perf stale transformers
Hello, I'm trying to train the transformer model based on PyTorch's nn.Transformer using multiprocessing with 8 TPUs. However, the forward is super slow (even the first iteration, with 10 batch size). Each iteration takes a few minutes. here are logs: =============================================== 2019-10-30 07:44:07.324500: I 29379 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:07.324563: I 29379 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.314788: I 29430 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:09.314850: I 29430 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.317559: I 29427 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:09.317618: I 29427 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.336030: I 29431 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:09.336093: I 29431 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.338445: I 29433 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:09.338496: I 29433 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.343114: I 29430 tensorflow/compiler/xla/xla_client/mesh_service.cc:168] Waiting to connect to client mesh master (300 seconds) localhost:43255 2019-10-30 07:44:09.347678: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) CPU:0 -> /job:tpu_worker/replica:0/task:0/device:XLA_CPU:0 2019-10-30 07:44:09.347753: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) TPU:0 -> /job:tpu_worker/replica:0/task:0/device:TPU:0 2019-10-30 07:44:09.347762: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:1 -> /job:tpu_worker/replica:0/task:0/device:TPU:1 2019-10-30 07:44:09.347769: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:2 -> /job:tpu_worker/replica:0/task:0/device:TPU:2 2019-10-30 07:44:09.347776: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:3 -> /job:tpu_worker/replica:0/task:0/device:TPU:3 2019-10-30 07:44:09.347783: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:4 -> /job:tpu_worker/replica:0/task:0/device:TPU:4 2019-10-30 07:44:09.347789: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:5 -> /job:tpu_worker/replica:0/task:0/device:TPU:5 2019-10-30 07:44:09.347797: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:6 -> /job:tpu_worker/replica:0/task:0/device:TPU:6 2019-10-30 07:44:09.347814: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:7 -> /job:tpu_worker/replica:0/task:0/device:TPU:7 2019-10-30 07:44:09.348037: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:221] Worker grpc://10.0.101.2:8470 for /job:tpu_worker/replica:0/task:0 2019-10-30 07:44:09.348103: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:225] XRT default device: TPU:0 2019-10-30 07:44:09.348133: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:1114] Configuring TPU for master worker tpu_worker:0 at grpc://10.0.101.2:8470 2019-10-30 07:44:09.348722: I 29434 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:09.348768: I 29434 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.361452: I 29428 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:09.361517: I 29428 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.367103: I 29431 tensorflow/compiler/xla/xla_client/mesh_service.cc:168] Waiting to connect to client mesh master (300 seconds) localhost:43255 2019-10-30 07:44:09.371809: I 29433 tensorflow/compiler/xla/xla_client/mesh_service.cc:168] Waiting to connect to client mesh master (300 seconds) localhost:43255 2019-10-30 07:44:09.381630: I 29434 tensorflow/compiler/xla/xla_client/mesh_service.cc:168] Waiting to connect to client mesh master (300 seconds) localhost:43255 2019-10-30 07:44:09.383934: I 29429 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:09.383985: I 29429 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.392137: I 29428 tensorflow/compiler/xla/xla_client/mesh_service.cc:168] Waiting to connect to client mesh master (300 seconds) localhost:43255 2019-10-30 07:44:09.412761: I 29429 tensorflow/compiler/xla/xla_client/mesh_service.cc:168] Waiting to connect to client mesh master (300 seconds) localhost:43255 2019-10-30 07:44:09.430583: I 29432 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:09.430648: I 29432 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.459201: I 29432 tensorflow/compiler/xla/xla_client/mesh_service.cc:168] Waiting to connect to client mesh master (300 seconds) localhost:43255 2019-10-30 07:44:12.129402: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:1125] TPU topology: mesh_shape: 2 mesh_shape: 2 mesh_shape: 2 num_tasks: 1 num_tpu_devices_per_task: 8 device_coordinates: 0 device_coordinates: 0 device_coordinates: 0 device_coordinates: 0 device_coordinates: 0 device_coordinates: 1 device_coordinates: 0 device_coordinates: 1 device_coordinates: 0 device_coordinates: 0 device_coordinates: 1 device_coordinates: 1 device_coordinates: 1 device_coordinates: 0 device_coordinates: 0 device_coordinates: 1 device_coordinates: 0 device_coordinates: 1 device_coordinates: 1 device_coordinates: 1 device_coordinates: 0 device_coordinates: 1 device_coordinates: 1 device_coordinates: 1 2019-10-30 07:44:12.129501: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:1194] Creating mesh service bound to localhost:43255 Rank: 0 Load data 2019-10-30 07:44:12.182163: I 29430 tensorflow/compiler/xla/xla_client/computation_client.cc:195] Fetching mesh configuration for worker tpu_worker:0 from mesh service at localhost:43255 2019-10-30 07:44:12.182913: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) CPU:0 -> /job:tpu_worker/replica:0/task:0/device:XLA_CPU:0 2019-10-30 07:44:12.182948: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:0 -> /job:tpu_worker/replica:0/task:0/device:TPU:0 2019-10-30 07:44:12.182956: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:1 -> /job:tpu_worker/replica:0/task:0/device:TPU:1 2019-10-30 07:44:12.182962: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:2 -> /job:tpu_worker/replica:0/task:0/device:TPU:2 2019-10-30 07:44:12.182969: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) TPU:3 -> /job:tpu_worker/replica:0/task:0/device:TPU:3 2019-10-30 07:44:12.182974: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:4 -> /job:tpu_worker/replica:0/task:0/device:TPU:4 2019-10-30 07:44:12.182981: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:5 -> /job:tpu_worker/replica:0/task:0/device:TPU:5 2019-10-30 07:44:12.182987: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:6 -> /job:tpu_worker/replica:0/task:0/device:TPU:6 2019-10-30 07:44:12.182993: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:7 -> /job:tpu_worker/replica:0/task:0/device:TPU:7 2019-10-30 07:44:12.183000: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:221] Worker grpc://10.0.101.2:8470 for /job:tpu_worker/replica:0/task:0 2019-10-30 07:44:12.183007: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:225] XRT default device: TPU:3 Rank: 3 Load data 2019-10-30 07:44:12.223982: I 29431 tensorflow/compiler/xla/xla_client/computation_client.cc:195] Fetching mesh configuration for worker tpu_worker:0 from mesh service at localhost:43255 2019-10-30 07:44:12.224456: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) CPU:0 -> /job:tpu_worker/replica:0/task:0/device:XLA_CPU:0 2019-10-30 07:44:12.224487: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:0 -> /job:tpu_worker/replica:0/task:0/device:TPU:0 2019-10-30 07:44:12.224496: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:1 -> /job:tpu_worker/replica:0/task:0/device:TPU:1 2019-10-30 07:44:12.224516: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:2 -> /job:tpu_worker/replica:0/task:0/device:TPU:2 2019-10-30 07:44:12.224530: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:3 -> /job:tpu_worker/replica:0/task:0/device:TPU:3 2019-10-30 07:44:12.224536: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) TPU:4 -> /job:tpu_worker/replica:0/task:0/device:TPU:4 2019-10-30 07:44:12.224707: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:5 -> /job:tpu_worker/replica:0/task:0/device:TPU:5 2019-10-30 07:44:12.224801: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:6 -> /job:tpu_worker/replica:0/task:0/device:TPU:6 2019-10-30 07:44:12.224836: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:7 -> /job:tpu_worker/replica:0/task:0/device:TPU:7 2019-10-30 07:44:12.224846: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:221] Worker grpc://10.0.101.2:8470 for /job:tpu_worker/replica:0/task:0 2019-10-30 07:44:12.224856: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:225] XRT default device: TPU:4 Rank: 4 Load data 2019-10-30 07:44:12.268021: I 29429 tensorflow/compiler/xla/xla_client/computation_client.cc:195] Fetching mesh configuration for worker tpu_worker:0 from mesh service at localhost:43255 2019-10-30 07:44:12.268547: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) CPU:0 -> /job:tpu_worker/replica:0/task:0/device:XLA_CPU:0 2019-10-30 07:44:12.268584: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:0 -> /job:tpu_worker/replica:0/task:0/device:TPU:0 2019-10-30 07:44:12.268594: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:1 -> /job:tpu_worker/replica:0/task:0/device:TPU:1 2019-10-30 07:44:12.268600: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) TPU:2 -> /job:tpu_worker/replica:0/task:0/device:TPU:2 2019-10-30 07:44:12.268607: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:3 -> /job:tpu_worker/replica:0/task:0/device:TPU:3 2019-10-30 07:44:12.268623: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:4 -> /job:tpu_worker/replica:0/task:0/device:TPU:4 2019-10-30 07:44:12.268631: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:5 -> /job:tpu_worker/replica:0/task:0/device:TPU:5 2019-10-30 07:44:12.268661: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:6 -> /job:tpu_worker/replica:0/task:0/device:TPU:6 2019-10-30 07:44:12.268669: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:7 -> /job:tpu_worker/replica:0/task:0/device:TPU:7 2019-10-30 07:44:12.268680: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:221] Worker grpc://10.0.101.2:8470 for /job:tpu_worker/replica:0/task:0 2019-10-30 07:44:12.268690: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:225] XRT default device: TPU:2 Rank: 2 Load data 2019-10-30 07:44:14.294320: I 29432 tensorflow/compiler/xla/xla_client/computation_client.cc:195] Fetching mesh configuration for worker tpu_worker:0 from mesh service at localhost:43255 2019-10-30 07:44:14.294793: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) CPU:0 -> /job:tpu_worker/replica:0/task:0/device:XLA_CPU:0 2019-10-30 07:44:14.294829: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:0 -> /job:tpu_worker/replica:0/task:0/device:TPU:0 2019-10-30 07:44:14.294839: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:1 -> /job:tpu_worker/replica:0/task:0/device:TPU:1 2019-10-30 07:44:14.294855: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:2 -> /job:tpu_worker/replica:0/task:0/device:TPU:2 2019-10-30 07:44:14.294867: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:3 -> /job:tpu_worker/replica:0/task:0/device:TPU:3 2019-10-30 07:44:14.294874: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:4 -> /job:tpu_worker/replica:0/task:0/device:TPU:4 2019-10-30 07:44:14.294881: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) TPU:5 -> /job:tpu_worker/replica:0/task:0/device:TPU:5 2019-10-30 07:44:14.294891: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:6 -> /job:tpu_worker/replica:0/task:0/device:TPU:6 2019-10-30 07:44:14.294901: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:7 -> /job:tpu_worker/replica:0/task:0/device:TPU:7 2019-10-30 07:44:14.294912: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:221] Worker grpc://10.0.101.2:8470 for /job:tpu_worker/replica:0/task:0 2019-10-30 07:44:14.294921: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:225] XRT default device: TPU:5 Rank: 5 Load data 2019-10-30 07:44:14.478214: I 29428 tensorflow/compiler/xla/xla_client/computation_client.cc:195] Fetching mesh configuration for worker tpu_worker:0 from mesh service at localhost:43255 2019-10-30 07:44:14.478892: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) CPU:0 -> /job:tpu_worker/replica:0/task:0/device:XLA_CPU:0 2019-10-30 07:44:14.478929: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:0 -> /job:tpu_worker/replica:0/task:0/device:TPU:0 2019-10-30 07:44:14.478937: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) TPU:1 -> /job:tpu_worker/replica:0/task:0/device:TPU:1 2019-10-30 07:44:14.478944: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:2 -> /job:tpu_worker/replica:0/task:0/device:TPU:2 2019-10-30 07:44:14.478950: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:3 -> /job:tpu_worker/replica:0/task:0/device:TPU:3 2019-10-30 07:44:14.478956: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:4 -> /job:tpu_worker/replica:0/task:0/device:TPU:4 2019-10-30 07:44:14.478972: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:5 -> /job:tpu_worker/replica:0/task:0/device:TPU:5 2019-10-30 07:44:14.478985: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:6 -> /job:tpu_worker/replica:0/task:0/device:TPU:6 2019-10-30 07:44:14.478993: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:7 -> /job:tpu_worker/replica:0/task:0/device:TPU:7 2019-10-30 07:44:14.479008: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:221] Worker grpc://10.0.101.2:8470 for /job:tpu_worker/replica:0/task:0 2019-10-30 07:44:14.479014: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:225] XRT default device: TPU:1 Rank: 1 Load data 2019-10-30 07:44:14.601763: I 29434 tensorflow/compiler/xla/xla_client/computation_client.cc:195] Fetching mesh configuration for worker tpu_worker:0 from mesh service at localhost:43255 2019-10-30 07:44:14.602318: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) CPU:0 -> /job:tpu_worker/replica:0/task:0/device:XLA_CPU:0 2019-10-30 07:44:14.602360: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:0 -> /job:tpu_worker/replica:0/task:0/device:TPU:0 2019-10-30 07:44:14.602368: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:1 -> /job:tpu_worker/replica:0/task:0/device:TPU:1 2019-10-30 07:44:14.602375: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:2 -> /job:tpu_worker/replica:0/task:0/device:TPU:2 2019-10-30 07:44:14.602381: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:3 -> /job:tpu_worker/replica:0/task:0/device:TPU:3 2019-10-30 07:44:14.602397: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:4 -> /job:tpu_worker/replica:0/task:0/device:TPU:4 2019-10-30 07:44:14.602417: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:5 -> /job:tpu_worker/replica:0/task:0/device:TPU:5 2019-10-30 07:44:14.602424: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:6 -> /job:tpu_worker/replica:0/task:0/device:TPU:6 2019-10-30 07:44:14.602430: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) TPU:7 -> /job:tpu_worker/replica:0/task:0/device:TPU:7 2019-10-30 07:44:14.602438: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:221] Worker grpc://10.0.101.2:8470 for /job:tpu_worker/replica:0/task:0 2019-10-30 07:44:14.602447: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:225] XRT default device: TPU:7 Rank: 7 Load data 2019-10-30 07:44:14.850202: I 29433 tensorflow/compiler/xla/xla_client/computation_client.cc:195] Fetching mesh configuration for worker tpu_worker:0 from mesh service at localhost:43255 2019-10-30 07:44:14.850964: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) CPU:0 -> /job:tpu_worker/replica:0/task:0/device:XLA_CPU:0 2019-10-30 07:44:14.851015: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:0 -> /job:tpu_worker/replica:0/task:0/device:TPU:0 2019-10-30 07:44:14.851025: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:1 -> /job:tpu_worker/replica:0/task:0/device:TPU:1 2019-10-30 07:44:14.851033: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:2 -> /job:tpu_worker/replica:0/task:0/device:TPU:2 2019-10-30 07:44:14.851042: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:3 -> /job:tpu_worker/replica:0/task:0/device:TPU:3 2019-10-30 07:44:14.851049: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:4 -> /job:tpu_worker/replica:0/task:0/device:TPU:4 2019-10-30 07:44:14.851058: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:5 -> /job:tpu_worker/replica:0/task:0/device:TPU:5 2019-10-30 07:44:14.851065: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) TPU:6 -> /job:tpu_worker/replica:0/task:0/device:TPU:6 2019-10-30 07:44:14.851074: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:7 -> /job:tpu_worker/replica:0/task:0/device:TPU:7 2019-10-30 07:44:14.851094: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:221] Worker grpc://10.0.101.2:8470 for /job:tpu_worker/replica:0/task:0 2019-10-30 07:44:14.851106: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:225] XRT default device: TPU:6 Rank: 6 Load data Build model Build model Total Parameters: 86205440 Total Parameters: 86205440 Training start - Total iter: 3506 Build model Training start - Total iter: 3506 Build model Total Parameters: 86205440 Training start - Total iter: 3506 Total Parameters: 86205440 Training start - Total iter: 3506 Build model Build model Build model Build model Total Parameters: 86205440 Training start - Total iter: 3506 Total Parameters: 86205440 Total Parameters: 86205440 Training start - Total iter: 3506 Total Parameters: 86205440 Training start - Total iter: 3506 Training start - Total iter: 3506 2019-10-30 07:45:10.059324: I 30786 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:45:10.059376: I 30786 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:45:10.232121: I 30787 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:45:10.232176: I 30787 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:45:11.010605: I 30950 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:45:11.010664: I 30950 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:45:11.117959: I 30951 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:45:11.118026: I 30951 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:45:13.167248: I 31034 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:45:13.167308: I 31034 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:45:13.341512: I 31035 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:45:13.341574: I 31035 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:45:13.549634: I 31037 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:45:13.549699: I 31037 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:45:13.578042: I 31070 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:45:13.578114: I 31070 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 =============================================== There is no errors except these logs. Do you have any idea to solve this slow training?
1.0
Training is sooooooooo slow - Hello, I'm trying to train the transformer model based on PyTorch's nn.Transformer using multiprocessing with 8 TPUs. However, the forward is super slow (even the first iteration, with 10 batch size). Each iteration takes a few minutes. here are logs: =============================================== 2019-10-30 07:44:07.324500: I 29379 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:07.324563: I 29379 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.314788: I 29430 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:09.314850: I 29430 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.317559: I 29427 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:09.317618: I 29427 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.336030: I 29431 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:09.336093: I 29431 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.338445: I 29433 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:09.338496: I 29433 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.343114: I 29430 tensorflow/compiler/xla/xla_client/mesh_service.cc:168] Waiting to connect to client mesh master (300 seconds) localhost:43255 2019-10-30 07:44:09.347678: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) CPU:0 -> /job:tpu_worker/replica:0/task:0/device:XLA_CPU:0 2019-10-30 07:44:09.347753: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) TPU:0 -> /job:tpu_worker/replica:0/task:0/device:TPU:0 2019-10-30 07:44:09.347762: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:1 -> /job:tpu_worker/replica:0/task:0/device:TPU:1 2019-10-30 07:44:09.347769: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:2 -> /job:tpu_worker/replica:0/task:0/device:TPU:2 2019-10-30 07:44:09.347776: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:3 -> /job:tpu_worker/replica:0/task:0/device:TPU:3 2019-10-30 07:44:09.347783: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:4 -> /job:tpu_worker/replica:0/task:0/device:TPU:4 2019-10-30 07:44:09.347789: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:5 -> /job:tpu_worker/replica:0/task:0/device:TPU:5 2019-10-30 07:44:09.347797: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:6 -> /job:tpu_worker/replica:0/task:0/device:TPU:6 2019-10-30 07:44:09.347814: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:7 -> /job:tpu_worker/replica:0/task:0/device:TPU:7 2019-10-30 07:44:09.348037: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:221] Worker grpc://10.0.101.2:8470 for /job:tpu_worker/replica:0/task:0 2019-10-30 07:44:09.348103: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:225] XRT default device: TPU:0 2019-10-30 07:44:09.348133: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:1114] Configuring TPU for master worker tpu_worker:0 at grpc://10.0.101.2:8470 2019-10-30 07:44:09.348722: I 29434 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:09.348768: I 29434 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.361452: I 29428 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:09.361517: I 29428 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.367103: I 29431 tensorflow/compiler/xla/xla_client/mesh_service.cc:168] Waiting to connect to client mesh master (300 seconds) localhost:43255 2019-10-30 07:44:09.371809: I 29433 tensorflow/compiler/xla/xla_client/mesh_service.cc:168] Waiting to connect to client mesh master (300 seconds) localhost:43255 2019-10-30 07:44:09.381630: I 29434 tensorflow/compiler/xla/xla_client/mesh_service.cc:168] Waiting to connect to client mesh master (300 seconds) localhost:43255 2019-10-30 07:44:09.383934: I 29429 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:09.383985: I 29429 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.392137: I 29428 tensorflow/compiler/xla/xla_client/mesh_service.cc:168] Waiting to connect to client mesh master (300 seconds) localhost:43255 2019-10-30 07:44:09.412761: I 29429 tensorflow/compiler/xla/xla_client/mesh_service.cc:168] Waiting to connect to client mesh master (300 seconds) localhost:43255 2019-10-30 07:44:09.430583: I 29432 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:44:09.430648: I 29432 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:44:09.459201: I 29432 tensorflow/compiler/xla/xla_client/mesh_service.cc:168] Waiting to connect to client mesh master (300 seconds) localhost:43255 2019-10-30 07:44:12.129402: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:1125] TPU topology: mesh_shape: 2 mesh_shape: 2 mesh_shape: 2 num_tasks: 1 num_tpu_devices_per_task: 8 device_coordinates: 0 device_coordinates: 0 device_coordinates: 0 device_coordinates: 0 device_coordinates: 0 device_coordinates: 1 device_coordinates: 0 device_coordinates: 1 device_coordinates: 0 device_coordinates: 0 device_coordinates: 1 device_coordinates: 1 device_coordinates: 1 device_coordinates: 0 device_coordinates: 0 device_coordinates: 1 device_coordinates: 0 device_coordinates: 1 device_coordinates: 1 device_coordinates: 1 device_coordinates: 0 device_coordinates: 1 device_coordinates: 1 device_coordinates: 1 2019-10-30 07:44:12.129501: I 29427 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:1194] Creating mesh service bound to localhost:43255 Rank: 0 Load data 2019-10-30 07:44:12.182163: I 29430 tensorflow/compiler/xla/xla_client/computation_client.cc:195] Fetching mesh configuration for worker tpu_worker:0 from mesh service at localhost:43255 2019-10-30 07:44:12.182913: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) CPU:0 -> /job:tpu_worker/replica:0/task:0/device:XLA_CPU:0 2019-10-30 07:44:12.182948: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:0 -> /job:tpu_worker/replica:0/task:0/device:TPU:0 2019-10-30 07:44:12.182956: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:1 -> /job:tpu_worker/replica:0/task:0/device:TPU:1 2019-10-30 07:44:12.182962: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:2 -> /job:tpu_worker/replica:0/task:0/device:TPU:2 2019-10-30 07:44:12.182969: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) TPU:3 -> /job:tpu_worker/replica:0/task:0/device:TPU:3 2019-10-30 07:44:12.182974: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:4 -> /job:tpu_worker/replica:0/task:0/device:TPU:4 2019-10-30 07:44:12.182981: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:5 -> /job:tpu_worker/replica:0/task:0/device:TPU:5 2019-10-30 07:44:12.182987: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:6 -> /job:tpu_worker/replica:0/task:0/device:TPU:6 2019-10-30 07:44:12.182993: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:7 -> /job:tpu_worker/replica:0/task:0/device:TPU:7 2019-10-30 07:44:12.183000: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:221] Worker grpc://10.0.101.2:8470 for /job:tpu_worker/replica:0/task:0 2019-10-30 07:44:12.183007: I 29430 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:225] XRT default device: TPU:3 Rank: 3 Load data 2019-10-30 07:44:12.223982: I 29431 tensorflow/compiler/xla/xla_client/computation_client.cc:195] Fetching mesh configuration for worker tpu_worker:0 from mesh service at localhost:43255 2019-10-30 07:44:12.224456: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) CPU:0 -> /job:tpu_worker/replica:0/task:0/device:XLA_CPU:0 2019-10-30 07:44:12.224487: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:0 -> /job:tpu_worker/replica:0/task:0/device:TPU:0 2019-10-30 07:44:12.224496: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:1 -> /job:tpu_worker/replica:0/task:0/device:TPU:1 2019-10-30 07:44:12.224516: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:2 -> /job:tpu_worker/replica:0/task:0/device:TPU:2 2019-10-30 07:44:12.224530: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:3 -> /job:tpu_worker/replica:0/task:0/device:TPU:3 2019-10-30 07:44:12.224536: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) TPU:4 -> /job:tpu_worker/replica:0/task:0/device:TPU:4 2019-10-30 07:44:12.224707: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:5 -> /job:tpu_worker/replica:0/task:0/device:TPU:5 2019-10-30 07:44:12.224801: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:6 -> /job:tpu_worker/replica:0/task:0/device:TPU:6 2019-10-30 07:44:12.224836: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:7 -> /job:tpu_worker/replica:0/task:0/device:TPU:7 2019-10-30 07:44:12.224846: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:221] Worker grpc://10.0.101.2:8470 for /job:tpu_worker/replica:0/task:0 2019-10-30 07:44:12.224856: I 29431 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:225] XRT default device: TPU:4 Rank: 4 Load data 2019-10-30 07:44:12.268021: I 29429 tensorflow/compiler/xla/xla_client/computation_client.cc:195] Fetching mesh configuration for worker tpu_worker:0 from mesh service at localhost:43255 2019-10-30 07:44:12.268547: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) CPU:0 -> /job:tpu_worker/replica:0/task:0/device:XLA_CPU:0 2019-10-30 07:44:12.268584: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:0 -> /job:tpu_worker/replica:0/task:0/device:TPU:0 2019-10-30 07:44:12.268594: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:1 -> /job:tpu_worker/replica:0/task:0/device:TPU:1 2019-10-30 07:44:12.268600: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) TPU:2 -> /job:tpu_worker/replica:0/task:0/device:TPU:2 2019-10-30 07:44:12.268607: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:3 -> /job:tpu_worker/replica:0/task:0/device:TPU:3 2019-10-30 07:44:12.268623: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:4 -> /job:tpu_worker/replica:0/task:0/device:TPU:4 2019-10-30 07:44:12.268631: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:5 -> /job:tpu_worker/replica:0/task:0/device:TPU:5 2019-10-30 07:44:12.268661: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:6 -> /job:tpu_worker/replica:0/task:0/device:TPU:6 2019-10-30 07:44:12.268669: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:7 -> /job:tpu_worker/replica:0/task:0/device:TPU:7 2019-10-30 07:44:12.268680: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:221] Worker grpc://10.0.101.2:8470 for /job:tpu_worker/replica:0/task:0 2019-10-30 07:44:12.268690: I 29429 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:225] XRT default device: TPU:2 Rank: 2 Load data 2019-10-30 07:44:14.294320: I 29432 tensorflow/compiler/xla/xla_client/computation_client.cc:195] Fetching mesh configuration for worker tpu_worker:0 from mesh service at localhost:43255 2019-10-30 07:44:14.294793: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) CPU:0 -> /job:tpu_worker/replica:0/task:0/device:XLA_CPU:0 2019-10-30 07:44:14.294829: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:0 -> /job:tpu_worker/replica:0/task:0/device:TPU:0 2019-10-30 07:44:14.294839: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:1 -> /job:tpu_worker/replica:0/task:0/device:TPU:1 2019-10-30 07:44:14.294855: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:2 -> /job:tpu_worker/replica:0/task:0/device:TPU:2 2019-10-30 07:44:14.294867: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:3 -> /job:tpu_worker/replica:0/task:0/device:TPU:3 2019-10-30 07:44:14.294874: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:4 -> /job:tpu_worker/replica:0/task:0/device:TPU:4 2019-10-30 07:44:14.294881: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) TPU:5 -> /job:tpu_worker/replica:0/task:0/device:TPU:5 2019-10-30 07:44:14.294891: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:6 -> /job:tpu_worker/replica:0/task:0/device:TPU:6 2019-10-30 07:44:14.294901: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:7 -> /job:tpu_worker/replica:0/task:0/device:TPU:7 2019-10-30 07:44:14.294912: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:221] Worker grpc://10.0.101.2:8470 for /job:tpu_worker/replica:0/task:0 2019-10-30 07:44:14.294921: I 29432 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:225] XRT default device: TPU:5 Rank: 5 Load data 2019-10-30 07:44:14.478214: I 29428 tensorflow/compiler/xla/xla_client/computation_client.cc:195] Fetching mesh configuration for worker tpu_worker:0 from mesh service at localhost:43255 2019-10-30 07:44:14.478892: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) CPU:0 -> /job:tpu_worker/replica:0/task:0/device:XLA_CPU:0 2019-10-30 07:44:14.478929: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:0 -> /job:tpu_worker/replica:0/task:0/device:TPU:0 2019-10-30 07:44:14.478937: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) TPU:1 -> /job:tpu_worker/replica:0/task:0/device:TPU:1 2019-10-30 07:44:14.478944: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:2 -> /job:tpu_worker/replica:0/task:0/device:TPU:2 2019-10-30 07:44:14.478950: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:3 -> /job:tpu_worker/replica:0/task:0/device:TPU:3 2019-10-30 07:44:14.478956: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:4 -> /job:tpu_worker/replica:0/task:0/device:TPU:4 2019-10-30 07:44:14.478972: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:5 -> /job:tpu_worker/replica:0/task:0/device:TPU:5 2019-10-30 07:44:14.478985: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:6 -> /job:tpu_worker/replica:0/task:0/device:TPU:6 2019-10-30 07:44:14.478993: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:7 -> /job:tpu_worker/replica:0/task:0/device:TPU:7 2019-10-30 07:44:14.479008: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:221] Worker grpc://10.0.101.2:8470 for /job:tpu_worker/replica:0/task:0 2019-10-30 07:44:14.479014: I 29428 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:225] XRT default device: TPU:1 Rank: 1 Load data 2019-10-30 07:44:14.601763: I 29434 tensorflow/compiler/xla/xla_client/computation_client.cc:195] Fetching mesh configuration for worker tpu_worker:0 from mesh service at localhost:43255 2019-10-30 07:44:14.602318: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) CPU:0 -> /job:tpu_worker/replica:0/task:0/device:XLA_CPU:0 2019-10-30 07:44:14.602360: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:0 -> /job:tpu_worker/replica:0/task:0/device:TPU:0 2019-10-30 07:44:14.602368: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:1 -> /job:tpu_worker/replica:0/task:0/device:TPU:1 2019-10-30 07:44:14.602375: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:2 -> /job:tpu_worker/replica:0/task:0/device:TPU:2 2019-10-30 07:44:14.602381: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:3 -> /job:tpu_worker/replica:0/task:0/device:TPU:3 2019-10-30 07:44:14.602397: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:4 -> /job:tpu_worker/replica:0/task:0/device:TPU:4 2019-10-30 07:44:14.602417: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:5 -> /job:tpu_worker/replica:0/task:0/device:TPU:5 2019-10-30 07:44:14.602424: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:6 -> /job:tpu_worker/replica:0/task:0/device:TPU:6 2019-10-30 07:44:14.602430: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) TPU:7 -> /job:tpu_worker/replica:0/task:0/device:TPU:7 2019-10-30 07:44:14.602438: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:221] Worker grpc://10.0.101.2:8470 for /job:tpu_worker/replica:0/task:0 2019-10-30 07:44:14.602447: I 29434 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:225] XRT default device: TPU:7 Rank: 7 Load data 2019-10-30 07:44:14.850202: I 29433 tensorflow/compiler/xla/xla_client/computation_client.cc:195] Fetching mesh configuration for worker tpu_worker:0 from mesh service at localhost:43255 2019-10-30 07:44:14.850964: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) CPU:0 -> /job:tpu_worker/replica:0/task:0/device:XLA_CPU:0 2019-10-30 07:44:14.851015: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:0 -> /job:tpu_worker/replica:0/task:0/device:TPU:0 2019-10-30 07:44:14.851025: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:1 -> /job:tpu_worker/replica:0/task:0/device:TPU:1 2019-10-30 07:44:14.851033: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:2 -> /job:tpu_worker/replica:0/task:0/device:TPU:2 2019-10-30 07:44:14.851042: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:3 -> /job:tpu_worker/replica:0/task:0/device:TPU:3 2019-10-30 07:44:14.851049: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:4 -> /job:tpu_worker/replica:0/task:0/device:TPU:4 2019-10-30 07:44:14.851058: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:5 -> /job:tpu_worker/replica:0/task:0/device:TPU:5 2019-10-30 07:44:14.851065: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (LOCAL) TPU:6 -> /job:tpu_worker/replica:0/task:0/device:TPU:6 2019-10-30 07:44:14.851074: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:217] XRT device (REMOTE) TPU:7 -> /job:tpu_worker/replica:0/task:0/device:TPU:7 2019-10-30 07:44:14.851094: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:221] Worker grpc://10.0.101.2:8470 for /job:tpu_worker/replica:0/task:0 2019-10-30 07:44:14.851106: I 29433 tensorflow/compiler/xla/xla_client/xrt_computation_client.cc:225] XRT default device: TPU:6 Rank: 6 Load data Build model Build model Total Parameters: 86205440 Total Parameters: 86205440 Training start - Total iter: 3506 Build model Training start - Total iter: 3506 Build model Total Parameters: 86205440 Training start - Total iter: 3506 Total Parameters: 86205440 Training start - Total iter: 3506 Build model Build model Build model Build model Total Parameters: 86205440 Training start - Total iter: 3506 Total Parameters: 86205440 Total Parameters: 86205440 Training start - Total iter: 3506 Total Parameters: 86205440 Training start - Total iter: 3506 Training start - Total iter: 3506 2019-10-30 07:45:10.059324: I 30786 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:45:10.059376: I 30786 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:45:10.232121: I 30787 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:45:10.232176: I 30787 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:45:11.010605: I 30950 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:45:11.010664: I 30950 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:45:11.117959: I 30951 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:45:11.118026: I 30951 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:45:13.167248: I 31034 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:45:13.167308: I 31034 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:45:13.341512: I 31035 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:45:13.341574: I 31035 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:45:13.549634: I 31037 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:45:13.549699: I 31037 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 2019-10-30 07:45:13.578042: I 31070 torch_xla/csrc/aten_xla_type.cpp:86] PyTorch GIT revision: c89340f06877024a5a81393db35637a10b10568f 2019-10-30 07:45:13.578114: I 31070 torch_xla/csrc/aten_xla_type.cpp:87] XLA GIT revision: 2ddfcffd3cf91e7105bae61f7defa4591cdcb900 =============================================== There is no errors except these logs. Do you have any idea to solve this slow training?
process
training is sooooooooo slow hello i m trying to train the transformer model based on pytorch s nn transformer using multiprocessing with tpus however the forward is super slow even the first iteration with batch size each iteration takes a few minutes here are logs i torch xla csrc aten xla type cpp pytorch git revision i torch xla csrc aten xla type cpp xla git revision i torch xla csrc aten xla type cpp pytorch git revision i torch xla csrc aten xla type cpp xla git revision i torch xla csrc aten xla type cpp pytorch git revision i torch xla csrc aten xla type cpp xla git revision i torch xla csrc aten xla type cpp pytorch git revision i torch xla csrc aten xla type cpp xla git revision i torch xla csrc aten xla type cpp pytorch git revision i torch xla csrc aten xla type cpp xla git revision i tensorflow compiler xla xla client mesh service cc waiting to connect to client mesh master seconds localhost i tensorflow compiler xla xla client xrt computation client cc xrt device local cpu job tpu worker replica task device xla cpu i tensorflow compiler xla xla client xrt computation client cc xrt device local tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc worker grpc for job tpu worker replica task i tensorflow compiler xla xla client xrt computation client cc xrt default device tpu i tensorflow compiler xla xla client xrt computation client cc configuring tpu for master worker tpu worker at grpc i torch xla csrc aten xla type cpp pytorch git revision i torch xla csrc aten xla type cpp xla git revision i torch xla csrc aten xla type cpp pytorch git revision i torch xla csrc aten xla type cpp xla git revision i tensorflow compiler xla xla client mesh service cc waiting to connect to client mesh master seconds localhost i tensorflow compiler xla xla client mesh service cc waiting to connect to client mesh master seconds localhost i tensorflow compiler xla xla client mesh service cc waiting to connect to client mesh master seconds localhost i torch xla csrc aten xla type cpp pytorch git revision i torch xla csrc aten xla type cpp xla git revision i tensorflow compiler xla xla client mesh service cc waiting to connect to client mesh master seconds localhost i tensorflow compiler xla xla client mesh service cc waiting to connect to client mesh master seconds localhost i torch xla csrc aten xla type cpp pytorch git revision i torch xla csrc aten xla type cpp xla git revision i tensorflow compiler xla xla client mesh service cc waiting to connect to client mesh master seconds localhost i tensorflow compiler xla xla client xrt computation client cc tpu topology mesh shape mesh shape mesh shape num tasks num tpu devices per task device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates device coordinates i tensorflow compiler xla xla client xrt computation client cc creating mesh service bound to localhost rank load data i tensorflow compiler xla xla client computation client cc fetching mesh configuration for worker tpu worker from mesh service at localhost i tensorflow compiler xla xla client xrt computation client cc xrt device remote cpu job tpu worker replica task device xla cpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device local tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc worker grpc for job tpu worker replica task i tensorflow compiler xla xla client xrt computation client cc xrt default device tpu rank load data i tensorflow compiler xla xla client computation client cc fetching mesh configuration for worker tpu worker from mesh service at localhost i tensorflow compiler xla xla client xrt computation client cc xrt device remote cpu job tpu worker replica task device xla cpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device local tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc worker grpc for job tpu worker replica task i tensorflow compiler xla xla client xrt computation client cc xrt default device tpu rank load data i tensorflow compiler xla xla client computation client cc fetching mesh configuration for worker tpu worker from mesh service at localhost i tensorflow compiler xla xla client xrt computation client cc xrt device remote cpu job tpu worker replica task device xla cpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device local tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc worker grpc for job tpu worker replica task i tensorflow compiler xla xla client xrt computation client cc xrt default device tpu rank load data i tensorflow compiler xla xla client computation client cc fetching mesh configuration for worker tpu worker from mesh service at localhost i tensorflow compiler xla xla client xrt computation client cc xrt device remote cpu job tpu worker replica task device xla cpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device local tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc worker grpc for job tpu worker replica task i tensorflow compiler xla xla client xrt computation client cc xrt default device tpu rank load data i tensorflow compiler xla xla client computation client cc fetching mesh configuration for worker tpu worker from mesh service at localhost i tensorflow compiler xla xla client xrt computation client cc xrt device remote cpu job tpu worker replica task device xla cpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device local tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc worker grpc for job tpu worker replica task i tensorflow compiler xla xla client xrt computation client cc xrt default device tpu rank load data i tensorflow compiler xla xla client computation client cc fetching mesh configuration for worker tpu worker from mesh service at localhost i tensorflow compiler xla xla client xrt computation client cc xrt device remote cpu job tpu worker replica task device xla cpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device local tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc worker grpc for job tpu worker replica task i tensorflow compiler xla xla client xrt computation client cc xrt default device tpu rank load data i tensorflow compiler xla xla client computation client cc fetching mesh configuration for worker tpu worker from mesh service at localhost i tensorflow compiler xla xla client xrt computation client cc xrt device remote cpu job tpu worker replica task device xla cpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device local tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc xrt device remote tpu job tpu worker replica task device tpu i tensorflow compiler xla xla client xrt computation client cc worker grpc for job tpu worker replica task i tensorflow compiler xla xla client xrt computation client cc xrt default device tpu rank load data build model build model total parameters total parameters training start total iter build model training start total iter build model total parameters training start total iter total parameters training start total iter build model build model build model build model total parameters training start total iter total parameters total parameters training start total iter total parameters training start total iter training start total iter i torch xla csrc aten xla type cpp pytorch git revision i torch xla csrc aten xla type cpp xla git revision i torch xla csrc aten xla type cpp pytorch git revision i torch xla csrc aten xla type cpp xla git revision i torch xla csrc aten xla type cpp pytorch git revision i torch xla csrc aten xla type cpp xla git revision i torch xla csrc aten xla type cpp pytorch git revision i torch xla csrc aten xla type cpp xla git revision i torch xla csrc aten xla type cpp pytorch git revision i torch xla csrc aten xla type cpp xla git revision i torch xla csrc aten xla type cpp pytorch git revision i torch xla csrc aten xla type cpp xla git revision i torch xla csrc aten xla type cpp pytorch git revision i torch xla csrc aten xla type cpp xla git revision i torch xla csrc aten xla type cpp pytorch git revision i torch xla csrc aten xla type cpp xla git revision there is no errors except these logs do you have any idea to solve this slow training
1
3,193
3,833,126,424
IssuesEvent
2016-04-01 01:00:36
anholt/linux
https://api.github.com/repos/anholt/linux
closed
vc4: Queue V3D bin/render jobs in parallel
performance
Varad has been working on taking my anholt/vc4-kms-v3d-rpi2-binrendersubmit-2 branch and fixing it. Hopefully done soon.
True
vc4: Queue V3D bin/render jobs in parallel - Varad has been working on taking my anholt/vc4-kms-v3d-rpi2-binrendersubmit-2 branch and fixing it. Hopefully done soon.
non_process
queue bin render jobs in parallel varad has been working on taking my anholt kms binrendersubmit branch and fixing it hopefully done soon
0
38,374
8,468,424,006
IssuesEvent
2018-10-23 19:42:24
pnp/pnpjs
https://api.github.com/repos/pnp/pnpjs
closed
Question: Setting property bag values (e.g. on folders) - possible at all via REST?
area: code status: answered type: question
Is it at all possible to update the properties (in the property bag, not object properties or item fields) of folders/items/webs via REST? A MERGE fails.
1.0
Question: Setting property bag values (e.g. on folders) - possible at all via REST? - Is it at all possible to update the properties (in the property bag, not object properties or item fields) of folders/items/webs via REST? A MERGE fails.
non_process
question setting property bag values e g on folders possible at all via rest is it at all possible to update the properties in the property bag not object properties or item fields of folders items webs via rest a merge fails
0
16,535
21,563,635,512
IssuesEvent
2022-05-01 14:40:41
MartinBruun/P6
https://api.github.com/repos/MartinBruun/P6
opened
(CI) Add the user that creates a Pull Request as the Assignee
3: Should have Need grooming Process
Make an addition to labels-on-pr.yml (or a new) workflow in the .github/workflows folder, which adds the current user as the assignee when the person creates a pull request. Makes it easier to see who created what.
1.0
(CI) Add the user that creates a Pull Request as the Assignee - Make an addition to labels-on-pr.yml (or a new) workflow in the .github/workflows folder, which adds the current user as the assignee when the person creates a pull request. Makes it easier to see who created what.
process
ci add the user that creates a pull request as the assignee make an addition to labels on pr yml or a new workflow in the github workflows folder which adds the current user as the assignee when the person creates a pull request makes it easier to see who created what
1
722,722
24,872,669,547
IssuesEvent
2022-10-27 16:21:38
AY2223S1-CS2103T-W16-3/tp
https://api.github.com/repos/AY2223S1-CS2103T-W16-3/tp
closed
Editing upcoming appointment deletes all past appointments
priority.High type.Bug severity.High
This is due to a bug with the `EditPersonDescriptor` class within the `EditCommand` class, where past appointments are taken from `editPersonDescriptor` instead of `personToEdit`.
1.0
Editing upcoming appointment deletes all past appointments - This is due to a bug with the `EditPersonDescriptor` class within the `EditCommand` class, where past appointments are taken from `editPersonDescriptor` instead of `personToEdit`.
non_process
editing upcoming appointment deletes all past appointments this is due to a bug with the editpersondescriptor class within the editcommand class where past appointments are taken from editpersondescriptor instead of persontoedit
0
11,771
14,601,071,344
IssuesEvent
2020-12-21 08:03:38
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] [Dev] Multiple entries are found in sites tab for the same study
Bug P1 Participant manager Process: Fixed Process: Tested dev
Steps: 1. Login to PM as superadmin 2. Navigate to sites tab 3. Observe multiple enteries record A/R: Multiple entries are found in sites tab for the same study E/R: Only single record should be available ![Screenshot_1](https://user-images.githubusercontent.com/60386291/99935687-48ca8980-2d87-11eb-8ad1-8a2744bf515c.png)
2.0
[PM] [Dev] Multiple entries are found in sites tab for the same study - Steps: 1. Login to PM as superadmin 2. Navigate to sites tab 3. Observe multiple enteries record A/R: Multiple entries are found in sites tab for the same study E/R: Only single record should be available ![Screenshot_1](https://user-images.githubusercontent.com/60386291/99935687-48ca8980-2d87-11eb-8ad1-8a2744bf515c.png)
process
multiple entries are found in sites tab for the same study steps login to pm as superadmin navigate to sites tab observe multiple enteries record a r multiple entries are found in sites tab for the same study e r only single record should be available
1
8,990
6,085,262,491
IssuesEvent
2017-06-17 13:08:51
MISP/MISP
https://api.github.com/repos/MISP/MISP
closed
Do not redirect to event page after publishing from events list
usability
Usually when a user chooses to publish from the events list, he wants things to get done quickly by avoiding to go on the event page. Not redirecting to the event page after publishing from the events list, helps this goal. Related to #276
True
Do not redirect to event page after publishing from events list - Usually when a user chooses to publish from the events list, he wants things to get done quickly by avoiding to go on the event page. Not redirecting to the event page after publishing from the events list, helps this goal. Related to #276
non_process
do not redirect to event page after publishing from events list usually when a user chooses to publish from the events list he wants things to get done quickly by avoiding to go on the event page not redirecting to the event page after publishing from the events list helps this goal related to
0
8,956
12,061,468,239
IssuesEvent
2020-04-15 23:53:55
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Disabling concurrent scheduled runs
Pri1 devops-cicd-process/tech devops/prod
This may be a limitation of the product and not this documentation but I was wondering if there is a feature that would stop concurrent builds, like setting `trigger.batch: true` for the pipeline triggers (see [these docs](https://docs.microsoft.com/en-us/azure/devops/pipelines/yaml-schema?view=azure-devops&tabs=schema%2Cparameter-schema#push-trigger). To give you more information; I have a pipeline scheduled to run every fifteen minutes but which occasionally takes longer than fifteen minutes. If the previous pipeline is still running I'd like the next one to wait so they're not running at the same time. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 2ea2c851-bd1e-cddc-b4d0-e9f4112b8565 * Version Independent ID: 07c23fdd-14b5-985b-1c63-3f26f3a216ad * Content: [Configure schedules to run pipelines - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/scheduled-triggers?view=azure-devops&tabs=yaml#feedback) * Content Source: [docs/pipelines/process/scheduled-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/scheduled-triggers.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @steved0x * Microsoft Alias: **sdanie**
1.0
Disabling concurrent scheduled runs - This may be a limitation of the product and not this documentation but I was wondering if there is a feature that would stop concurrent builds, like setting `trigger.batch: true` for the pipeline triggers (see [these docs](https://docs.microsoft.com/en-us/azure/devops/pipelines/yaml-schema?view=azure-devops&tabs=schema%2Cparameter-schema#push-trigger). To give you more information; I have a pipeline scheduled to run every fifteen minutes but which occasionally takes longer than fifteen minutes. If the previous pipeline is still running I'd like the next one to wait so they're not running at the same time. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 2ea2c851-bd1e-cddc-b4d0-e9f4112b8565 * Version Independent ID: 07c23fdd-14b5-985b-1c63-3f26f3a216ad * Content: [Configure schedules to run pipelines - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/scheduled-triggers?view=azure-devops&tabs=yaml#feedback) * Content Source: [docs/pipelines/process/scheduled-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/scheduled-triggers.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @steved0x * Microsoft Alias: **sdanie**
process
disabling concurrent scheduled runs this may be a limitation of the product and not this documentation but i was wondering if there is a feature that would stop concurrent builds like setting trigger batch true for the pipeline triggers see to give you more information i have a pipeline scheduled to run every fifteen minutes but which occasionally takes longer than fifteen minutes if the previous pipeline is still running i d like the next one to wait so they re not running at the same time document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id cddc version independent id content content source product devops technology devops cicd process github login microsoft alias sdanie
1
18,968
24,938,628,964
IssuesEvent
2022-10-31 17:01:22
googleapis/google-cloud-go
https://api.github.com/repos/googleapis/google-cloud-go
closed
Warning: a recent release failed
type: process
The following release PRs may have failed: * #6934 - The release job is 'autorelease: pending', but expected 'autorelease: tagged'.
1.0
Warning: a recent release failed - The following release PRs may have failed: * #6934 - The release job is 'autorelease: pending', but expected 'autorelease: tagged'.
process
warning a recent release failed the following release prs may have failed the release job is autorelease pending but expected autorelease tagged
1
14,466
17,571,095,398
IssuesEvent
2021-08-14 18:05:57
flancast90/Speech-To-Text-in-TW5
https://api.github.com/repos/flancast90/Speech-To-Text-in-TW5
closed
How do we want to start?
process-description
Hi @flancast90 , I'm writing this first issue to get an overview how we want to start this little project :)
1.0
How do we want to start? - Hi @flancast90 , I'm writing this first issue to get an overview how we want to start this little project :)
process
how do we want to start hi i m writing this first issue to get an overview how we want to start this little project
1
19,482
25,792,866,014
IssuesEvent
2022-12-10 08:29:26
COPIM/open-book-collective
https://api.github.com/repos/COPIM/open-book-collective
closed
Allow initiatives to access subscribers' data
userstory development membership management (pillar 4) organisational process
As an open access book publisher (OABP) or an open access infrastructure provider (OAIP) ... I want a mechanism for receiving and storing member library data (eg IP addresses, contact points) ... ... so that I am able to provide membership services.
1.0
Allow initiatives to access subscribers' data - As an open access book publisher (OABP) or an open access infrastructure provider (OAIP) ... I want a mechanism for receiving and storing member library data (eg IP addresses, contact points) ... ... so that I am able to provide membership services.
process
allow initiatives to access subscribers data as an open access book publisher oabp or an open access infrastructure provider oaip i want a mechanism for receiving and storing member library data eg ip addresses contact points so that i am able to provide membership services
1
13,107
15,497,098,241
IssuesEvent
2021-03-11 03:57:03
2i2c-org/team-compass
https://api.github.com/repos/2i2c-org/team-compass
closed
Tech Team Update: March 2021
team-process
Hey @2i2c-org/tech-team - time to fill in some updates about what you've been up to the last couple of weeks! Can folks fill out the [HackMD](https://hackmd.io/i2Siurp1TkmPYgn3ZgxFQw) with their own updates? ✨ - **Updates HackMD**: https://hackmd.io/i2Siurp1TkmPYgn3ZgxFQw - **Team Sync history**: https://2i2c.org/team-compass/team/tech/sync/ # ToDo - [x] Clean up the [HackMD](https://hackmd.io/i2Siurp1TkmPYgn3ZgxFQw) for this update - [x] Ping the team members in [`#tech-updates`](https://2i2c.slack.com/archives/C01GLCC1VCN) - [x] Wait 2-3 days - [x] Copy/paste into the `team-compass` repository - [x] Clean up the HackMD - [x] Link to new updates in `team-compass/` in [`#tech-updates`](https://2i2c.slack.com/archives/C01GLCC1VCN)
1.0
Tech Team Update: March 2021 - Hey @2i2c-org/tech-team - time to fill in some updates about what you've been up to the last couple of weeks! Can folks fill out the [HackMD](https://hackmd.io/i2Siurp1TkmPYgn3ZgxFQw) with their own updates? ✨ - **Updates HackMD**: https://hackmd.io/i2Siurp1TkmPYgn3ZgxFQw - **Team Sync history**: https://2i2c.org/team-compass/team/tech/sync/ # ToDo - [x] Clean up the [HackMD](https://hackmd.io/i2Siurp1TkmPYgn3ZgxFQw) for this update - [x] Ping the team members in [`#tech-updates`](https://2i2c.slack.com/archives/C01GLCC1VCN) - [x] Wait 2-3 days - [x] Copy/paste into the `team-compass` repository - [x] Clean up the HackMD - [x] Link to new updates in `team-compass/` in [`#tech-updates`](https://2i2c.slack.com/archives/C01GLCC1VCN)
process
tech team update march hey org tech team time to fill in some updates about what you ve been up to the last couple of weeks can folks fill out the with their own updates ✨ updates hackmd team sync history todo clean up the for this update ping the team members in wait days copy paste into the team compass repository clean up the hackmd link to new updates in team compass in
1
19,008
25,007,167,784
IssuesEvent
2022-11-03 12:47:31
googleapis/gapic-generator-csharp
https://api.github.com/repos/googleapis/gapic-generator-csharp
closed
Consider removing unit test generation
priority: p2 type: process
This should probably only happen after we've got integration tests working with show case, but our unit test generation is frustrating: - Generating test projects means keeping more dependencies up to date - Adding a generator feature often means modifying test generation too - These tests don't actually add value, or at least that value hasn't been noticeable - Running generated tests of generated code takes significant CI time for no tangible benefit
1.0
Consider removing unit test generation - This should probably only happen after we've got integration tests working with show case, but our unit test generation is frustrating: - Generating test projects means keeping more dependencies up to date - Adding a generator feature often means modifying test generation too - These tests don't actually add value, or at least that value hasn't been noticeable - Running generated tests of generated code takes significant CI time for no tangible benefit
process
consider removing unit test generation this should probably only happen after we ve got integration tests working with show case but our unit test generation is frustrating generating test projects means keeping more dependencies up to date adding a generator feature often means modifying test generation too these tests don t actually add value or at least that value hasn t been noticeable running generated tests of generated code takes significant ci time for no tangible benefit
1
145,037
13,132,761,061
IssuesEvent
2020-08-06 19:32:51
Qiskit/qiskit-tutorials
https://api.github.com/repos/Qiskit/qiskit-tutorials
closed
1_getting_started_with_qiskit.ipynb should improve markdown for rst conversion
documentation
## Target file tutorials/circuits/1_getting_started_with_qiskit.ipynb. ## Improvement points ### 1. rst constraints Markdown can use emphasis and link at the same time, but rst cannot. It should be modified to a notation that can be converted to rst. The following is the converted statement. > The fundamental unit of Qiskit is the \*\*quantum circuit\*\*. In this case, it is better to remove the emphasis. ### 2. Markdown notation when using List When using List, it is necessary to insert a line break before and after each line. Otherwise, the symbol "*" will be treated as a string during rst conversion and merged into a single line. The following is the converted statement. > To create such a state, we start with a three-qubit quantum register. By default, each qubit in the register is initialized to |0⟩. To make the GHZ state, we apply the following gates: \* A Hadamard gate H on qubit 0, which puts it into the superposition state (|0⟩+|1⟩)/2–√. \* A controlled-Not operation (CX) between qubit 0 and qubit 1. \* A controlled-Not operation between qubit 0 and qubit 2. ## Note Some other files have similar issues. Therefore, other files also need to be checked and modified.
1.0
1_getting_started_with_qiskit.ipynb should improve markdown for rst conversion - ## Target file tutorials/circuits/1_getting_started_with_qiskit.ipynb. ## Improvement points ### 1. rst constraints Markdown can use emphasis and link at the same time, but rst cannot. It should be modified to a notation that can be converted to rst. The following is the converted statement. > The fundamental unit of Qiskit is the \*\*quantum circuit\*\*. In this case, it is better to remove the emphasis. ### 2. Markdown notation when using List When using List, it is necessary to insert a line break before and after each line. Otherwise, the symbol "*" will be treated as a string during rst conversion and merged into a single line. The following is the converted statement. > To create such a state, we start with a three-qubit quantum register. By default, each qubit in the register is initialized to |0⟩. To make the GHZ state, we apply the following gates: \* A Hadamard gate H on qubit 0, which puts it into the superposition state (|0⟩+|1⟩)/2–√. \* A controlled-Not operation (CX) between qubit 0 and qubit 1. \* A controlled-Not operation between qubit 0 and qubit 2. ## Note Some other files have similar issues. Therefore, other files also need to be checked and modified.
non_process
getting started with qiskit ipynb should improve markdown for rst conversion target file tutorials circuits getting started with qiskit ipynb improvement points rst constraints markdown can use emphasis and link at the same time but rst cannot it should be modified to a notation that can be converted to rst the following is the converted statement the fundamental unit of qiskit is the quantum circuit in this case it is better to remove the emphasis markdown notation when using list when using list it is necessary to insert a line break before and after each line otherwise the symbol will be treated as a string during rst conversion and merged into a single line the following is the converted statement to create such a state we start with a three qubit quantum register by default each qubit in the register is initialized to ⟩ to make the ghz state we apply the following gates a hadamard gate h on qubit which puts it into the superposition state ⟩ ⟩ –√ a controlled not operation cx between qubit and qubit a controlled not operation between qubit and qubit note some other files have similar issues therefore other files also need to be checked and modified
0
767,751
26,938,836,831
IssuesEvent
2023-02-07 23:27:02
phetsims/chipper
https://api.github.com/repos/phetsims/chipper
closed
Locale info is still a "ROUGH DRAFT". Make it production ready.
priority:2-high
While working on https://github.com/phetsims/chipper/issues/1374, I was examining commits for localeInfoModule.js, the primary source for PhET locale information that is displayed in Preferences and elsewhere. There is a single comment by @jonathanolson about this file in https://github.com/phetsims/chipper/issues/1302#issuecomment-1213644765: > Added a VERY rough-draft locale switcher in the navbar for demo-ing, see https://bayes.colorado.edu/dev/olsonjb/locale-test.html This is a single commit by @jonathanolson in GitHub for this file, with this message: > Adding ROUGH DRAFT locale switcher for development purposes, see https://github.com/phetsims/chipper/issues/1302 In both cases the use of cap ("VERY rough-draft" and "ROUGH DRAFT") seems significant. And there's even an acknowledge that this isn't ready for production ("for development purposes") - which I can vouch for given issues like https://github.com/phetsims/chipper/issues/1374. So given that a huge amount of work has been done since August to support dynamic locale, and that this is the foundation that's now "included in all sims"... Do you think we should move this from the "ROUGH DRAFT" stage to production-ready? @jonathanolson for comment. @kathy-phet for prioritization.
1.0
Locale info is still a "ROUGH DRAFT". Make it production ready. - While working on https://github.com/phetsims/chipper/issues/1374, I was examining commits for localeInfoModule.js, the primary source for PhET locale information that is displayed in Preferences and elsewhere. There is a single comment by @jonathanolson about this file in https://github.com/phetsims/chipper/issues/1302#issuecomment-1213644765: > Added a VERY rough-draft locale switcher in the navbar for demo-ing, see https://bayes.colorado.edu/dev/olsonjb/locale-test.html This is a single commit by @jonathanolson in GitHub for this file, with this message: > Adding ROUGH DRAFT locale switcher for development purposes, see https://github.com/phetsims/chipper/issues/1302 In both cases the use of cap ("VERY rough-draft" and "ROUGH DRAFT") seems significant. And there's even an acknowledge that this isn't ready for production ("for development purposes") - which I can vouch for given issues like https://github.com/phetsims/chipper/issues/1374. So given that a huge amount of work has been done since August to support dynamic locale, and that this is the foundation that's now "included in all sims"... Do you think we should move this from the "ROUGH DRAFT" stage to production-ready? @jonathanolson for comment. @kathy-phet for prioritization.
non_process
locale info is still a rough draft make it production ready while working on i was examining commits for localeinfomodule js the primary source for phet locale information that is displayed in preferences and elsewhere there is a single comment by jonathanolson about this file in added a very rough draft locale switcher in the navbar for demo ing see this is a single commit by jonathanolson in github for this file with this message adding rough draft locale switcher for development purposes see in both cases the use of cap very rough draft and rough draft seems significant and there s even an acknowledge that this isn t ready for production for development purposes which i can vouch for given issues like so given that a huge amount of work has been done since august to support dynamic locale and that this is the foundation that s now included in all sims do you think we should move this from the rough draft stage to production ready jonathanolson for comment kathy phet for prioritization
0
682,837
23,359,190,035
IssuesEvent
2022-08-10 10:06:45
exeGesIS-SDM/NetworkDesignTools_Feb21
https://api.github.com/repos/exeGesIS-SDM/NetworkDesignTools_Feb21
closed
Create property count layer for properties in a polygon
Priority 1 New feature
Request user selects a PN (primary node) polygon (which covers multiple buildings) Select all addresses in the polygon Add a single point for each building on a property summary layer Set the property ref Count the number of properties in the building(ie with the same property ID or at the same point) and save in a field
1.0
Create property count layer for properties in a polygon - Request user selects a PN (primary node) polygon (which covers multiple buildings) Select all addresses in the polygon Add a single point for each building on a property summary layer Set the property ref Count the number of properties in the building(ie with the same property ID or at the same point) and save in a field
non_process
create property count layer for properties in a polygon request user selects a pn primary node polygon which covers multiple buildings select all addresses in the polygon add a single point for each building on a property summary layer set the property ref count the number of properties in the building ie with the same property id or at the same point and save in a field
0
325,061
27,845,629,376
IssuesEvent
2023-03-20 15:19:47
QubesOS/updates-status
https://api.github.com/repos/QubesOS/updates-status
closed
video-companion v1.0.1-1 (r4.1)
r4.1-buster-cur-test r4.1-dom0-cur-test r4.1-bullseye-cur-test r4.1-centos-stream8-cur-test r4.1-bookworm-cur-test r4.1-fc36-cur-test r4.1-fc37-cur-test r4.1-fc38-cur-test
Update of video-companion to v1.0.1-1 for Qubes r4.1, see comments below for details. Built from: https://github.com/QubesOS/qubes-video-companion/commit/7b568f15cffe8ec0fde5aab37f5afe3d62ce6885 [Changes since previous version](https://github.com/QubesOS/qubes-video-companion/compare/v2.0.0...v1.0.1-1): QubesOS/qubes-video-companion@7b568f1 version 1.0.1-1 QubesOS/qubes-video-companion@313ce9b pylint: disable consider-using-f-string QubesOS/qubes-video-companion@038b4f8 tests: webcam test is supposed to work now QubesOS/qubes-video-companion@ce725b8 webcam: workaround gstreamer bug for video/x-raw source QubesOS/qubes-video-companion@741a487 Drop --buffer-size=0, it doesn't do anything QubesOS/qubes-video-companion@dd8e6be tests: log more details on failure QubesOS/qubes-video-companion@9bb6481 sender/webcam: support raw source stream too QubesOS/qubes-video-companion@9dafa4e Remove 'colorimetry' pipeline constraint QubesOS/qubes-video-companion@b3706f8 ci: limit pylint to actual package QubesOS/qubes-video-companion@cb91413 tests: add skeleton for webcam test too QubesOS/qubes-video-companion@14b44ce tests: add basic integration test for screen share QubesOS/qubes-video-companion@f710822 Add v4l-utils dependency QubesOS/qubes-video-companion@82c1038 Make pylint happy again QubesOS/qubes-video-companion@683e401 Reformat code according to python-black QubesOS/qubes-video-companion@240c140 Make pylint happy QubesOS/qubes-video-companion@7800118 Add .pylintrc from core-admin and fix deps QubesOS/qubes-video-companion@58cf260 Make shellcheck happy QubesOS/qubes-video-companion@a790ea9 Qubes Builder integration QubesOS/qubes-video-companion@51f9c35 Rework .gitlab-ci.yml QubesOS/qubes-video-companion@c7215ba Allow @default in the qrexec policies QubesOS/qubes-video-companion@e67ad19 Use @default as the destination if none is passed QubesOS/qubes-video-companion@4939934 Fix debian/rules QubesOS/qubes-video-companion@418a7c1 Convert icon.png QubesOS/qubes-video-companion@695043a Make install a bit more robust QubesOS/qubes-video-companion@3c7d3cd Check for /etc/qubes-release in dom0 QubesOS/qubes-video-companion@7af0651 Avoid overbroad except QubesOS/qubes-video-companion@302352f Minor Python improvements QubesOS/qubes-video-companion@0b8b901 Add basic comments to the qrexec policies QubesOS/qubes-video-companion@fb6db6f Don't hardcode dom0 as a destination QubesOS/qubes-video-companion@ca39b24 Avoid tripping -o pipefail QubesOS/qubes-video-companion@43343bc Improve documentation QubesOS/qubes-video-companion@ae28a82 Merge branch 'packaging-improvements' from DemiMarie QubesOS/qubes-video-companion@72cd1c6 Small fix ups QubesOS/qubes-video-companion@d7804d3 Improved packaging script QubesOS/qubes-video-companion@ed259e9 Split the RPM package into separate packages QubesOS/qubes-video-companion@ce98c5d Mark various targets as .PHONY QubesOS/qubes-video-companion@e81f522 Avoid installing v4l2loopback scripts by default Referenced issues: If you're release manager, you can issue GPG-inline signed command: * `Upload video-companion 7b568f15cffe8ec0fde5aab37f5afe3d62ce6885 r4.1 current repo` (available 7 days from now) * `Upload video-companion 7b568f15cffe8ec0fde5aab37f5afe3d62ce6885 r4.1 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now) * `Upload video-companion 7b568f15cffe8ec0fde5aab37f5afe3d62ce6885 r4.1 security-testing repo` Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
8.0
video-companion v1.0.1-1 (r4.1) - Update of video-companion to v1.0.1-1 for Qubes r4.1, see comments below for details. Built from: https://github.com/QubesOS/qubes-video-companion/commit/7b568f15cffe8ec0fde5aab37f5afe3d62ce6885 [Changes since previous version](https://github.com/QubesOS/qubes-video-companion/compare/v2.0.0...v1.0.1-1): QubesOS/qubes-video-companion@7b568f1 version 1.0.1-1 QubesOS/qubes-video-companion@313ce9b pylint: disable consider-using-f-string QubesOS/qubes-video-companion@038b4f8 tests: webcam test is supposed to work now QubesOS/qubes-video-companion@ce725b8 webcam: workaround gstreamer bug for video/x-raw source QubesOS/qubes-video-companion@741a487 Drop --buffer-size=0, it doesn't do anything QubesOS/qubes-video-companion@dd8e6be tests: log more details on failure QubesOS/qubes-video-companion@9bb6481 sender/webcam: support raw source stream too QubesOS/qubes-video-companion@9dafa4e Remove 'colorimetry' pipeline constraint QubesOS/qubes-video-companion@b3706f8 ci: limit pylint to actual package QubesOS/qubes-video-companion@cb91413 tests: add skeleton for webcam test too QubesOS/qubes-video-companion@14b44ce tests: add basic integration test for screen share QubesOS/qubes-video-companion@f710822 Add v4l-utils dependency QubesOS/qubes-video-companion@82c1038 Make pylint happy again QubesOS/qubes-video-companion@683e401 Reformat code according to python-black QubesOS/qubes-video-companion@240c140 Make pylint happy QubesOS/qubes-video-companion@7800118 Add .pylintrc from core-admin and fix deps QubesOS/qubes-video-companion@58cf260 Make shellcheck happy QubesOS/qubes-video-companion@a790ea9 Qubes Builder integration QubesOS/qubes-video-companion@51f9c35 Rework .gitlab-ci.yml QubesOS/qubes-video-companion@c7215ba Allow @default in the qrexec policies QubesOS/qubes-video-companion@e67ad19 Use @default as the destination if none is passed QubesOS/qubes-video-companion@4939934 Fix debian/rules QubesOS/qubes-video-companion@418a7c1 Convert icon.png QubesOS/qubes-video-companion@695043a Make install a bit more robust QubesOS/qubes-video-companion@3c7d3cd Check for /etc/qubes-release in dom0 QubesOS/qubes-video-companion@7af0651 Avoid overbroad except QubesOS/qubes-video-companion@302352f Minor Python improvements QubesOS/qubes-video-companion@0b8b901 Add basic comments to the qrexec policies QubesOS/qubes-video-companion@fb6db6f Don't hardcode dom0 as a destination QubesOS/qubes-video-companion@ca39b24 Avoid tripping -o pipefail QubesOS/qubes-video-companion@43343bc Improve documentation QubesOS/qubes-video-companion@ae28a82 Merge branch 'packaging-improvements' from DemiMarie QubesOS/qubes-video-companion@72cd1c6 Small fix ups QubesOS/qubes-video-companion@d7804d3 Improved packaging script QubesOS/qubes-video-companion@ed259e9 Split the RPM package into separate packages QubesOS/qubes-video-companion@ce98c5d Mark various targets as .PHONY QubesOS/qubes-video-companion@e81f522 Avoid installing v4l2loopback scripts by default Referenced issues: If you're release manager, you can issue GPG-inline signed command: * `Upload video-companion 7b568f15cffe8ec0fde5aab37f5afe3d62ce6885 r4.1 current repo` (available 7 days from now) * `Upload video-companion 7b568f15cffe8ec0fde5aab37f5afe3d62ce6885 r4.1 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now) * `Upload video-companion 7b568f15cffe8ec0fde5aab37f5afe3d62ce6885 r4.1 security-testing repo` Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
non_process
video companion update of video companion to for qubes see comments below for details built from qubesos qubes video companion version qubesos qubes video companion pylint disable consider using f string qubesos qubes video companion tests webcam test is supposed to work now qubesos qubes video companion webcam workaround gstreamer bug for video x raw source qubesos qubes video companion drop buffer size it doesn t do anything qubesos qubes video companion tests log more details on failure qubesos qubes video companion sender webcam support raw source stream too qubesos qubes video companion remove colorimetry pipeline constraint qubesos qubes video companion ci limit pylint to actual package qubesos qubes video companion tests add skeleton for webcam test too qubesos qubes video companion tests add basic integration test for screen share qubesos qubes video companion add utils dependency qubesos qubes video companion make pylint happy again qubesos qubes video companion reformat code according to python black qubesos qubes video companion make pylint happy qubesos qubes video companion add pylintrc from core admin and fix deps qubesos qubes video companion make shellcheck happy qubesos qubes video companion qubes builder integration qubesos qubes video companion rework gitlab ci yml qubesos qubes video companion allow default in the qrexec policies qubesos qubes video companion use default as the destination if none is passed qubesos qubes video companion fix debian rules qubesos qubes video companion convert icon png qubesos qubes video companion make install a bit more robust qubesos qubes video companion check for etc qubes release in qubesos qubes video companion avoid overbroad except qubesos qubes video companion minor python improvements qubesos qubes video companion add basic comments to the qrexec policies qubesos qubes video companion don t hardcode as a destination qubesos qubes video companion avoid tripping o pipefail qubesos qubes video companion improve documentation qubesos qubes video companion merge branch packaging improvements from demimarie qubesos qubes video companion small fix ups qubesos qubes video companion improved packaging script qubesos qubes video companion split the rpm package into separate packages qubesos qubes video companion mark various targets as phony qubesos qubes video companion avoid installing scripts by default referenced issues if you re release manager you can issue gpg inline signed command upload video companion current repo available days from now upload video companion current dists repo you can choose subset of distributions like vm vm available days from now upload video companion security testing repo above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it
0
388,465
26,767,299,128
IssuesEvent
2023-01-31 11:35:57
nuxt/nuxt
https://api.github.com/repos/nuxt/nuxt
closed
docs: markdown badly rendered
documentation 3.x upstream
### Environment all ### Reproduction https://nuxt.com/docs/api/composables/use-fetch https://nuxt.com/docs/api/composables/use-async-data ### Describe the bug A lot of characters are not rendered in the code block for useFetch and useAsyncData type ### Additional context \o/ nuxt 3 finally released \o/ thank you for all your hard work ! \o/ ### Logs _No response_
1.0
docs: markdown badly rendered - ### Environment all ### Reproduction https://nuxt.com/docs/api/composables/use-fetch https://nuxt.com/docs/api/composables/use-async-data ### Describe the bug A lot of characters are not rendered in the code block for useFetch and useAsyncData type ### Additional context \o/ nuxt 3 finally released \o/ thank you for all your hard work ! \o/ ### Logs _No response_
non_process
docs markdown badly rendered environment all reproduction describe the bug a lot of characters are not rendered in the code block for usefetch and useasyncdata type additional context o nuxt finally released o thank you for all your hard work o logs no response
0
393,463
11,616,176,440
IssuesEvent
2020-02-26 15:19:18
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.theguardian.com - see bug description
browser-fenix engine-gecko priority-important
<!-- @browser: Firefox Mobile 75.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:75.0) Gecko/75.0 Firefox/75.0 --> <!-- @reported_with: --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/49016 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.theguardian.com/politics/2020/feb/25/keep-chlorinated-chicken-ban-to-win-trade-deal-eu-tells-uk **Browser / Version**: Firefox Mobile 75.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Something else **Description**: every page loads somewhere in the middle, never at the top/beginning of an article. was the same on BBC.com and sz.de **Steps to Reproduce**: See above, loading newspaper sites gets me landing on the middle of the starting page or an article. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.theguardian.com - see bug description - <!-- @browser: Firefox Mobile 75.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:75.0) Gecko/75.0 Firefox/75.0 --> <!-- @reported_with: --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/49016 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.theguardian.com/politics/2020/feb/25/keep-chlorinated-chicken-ban-to-win-trade-deal-eu-tells-uk **Browser / Version**: Firefox Mobile 75.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Something else **Description**: every page loads somewhere in the middle, never at the top/beginning of an article. was the same on BBC.com and sz.de **Steps to Reproduce**: See above, loading newspaper sites gets me landing on the middle of the starting page or an article. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
see bug description url browser version firefox mobile operating system android tested another browser no problem type something else description every page loads somewhere in the middle never at the top beginning of an article was the same on bbc com and sz de steps to reproduce see above loading newspaper sites gets me landing on the middle of the starting page or an article browser configuration none from with ❤️
0
84,743
24,404,916,181
IssuesEvent
2022-10-05 07:00:35
godotengine/godot-proposals
https://api.github.com/repos/godotengine/godot-proposals
closed
Musl libc support, scons detection for libexecinfo
topic:buildsystem topic:porting platform:linuxbsd
<!-- Please fill in *all* the questions below and don't remove any of them. Proposals not following the template below will be closed immediately. --> ### Describe the project you are working on Musl support for Godot On distributions like Alpine Linux or Gentoo Linux musl support seems to be already feature-complete. With the use of glibc macros in the Godot source code (for debugging / stack traces) Godot can't be compiled on those systems by default. This proposal is about modifying the linuxbsd platform implementation to apply glibc macros only if glibc is available on the system. As an example the `basename` function which is included in the glibc headers differ from the musl implementation so it might be needed to "reimplement" it in the godot source code itself (see patch below). Of course the glibc implementation is more widespread on linux desktops but with the support of other libc implementations for linuxbsd (at least at source level) godot would be POSIX-compliant and therefore open for other linux platforms like servers or embedded systems without glibc. ### Describe the problem or limitation you are having in your project Godot currently only runs on Linux distributions with glibc as libc implementation. On distros with musl libc, it won't compile. ### Describe the feature / enhancement and how it helps to overcome the problem or limitation Patch the Godot source code so glibc specific features will only be enabled if glibc is available on the system. ### Describe how your proposal will work, with code, pseudo-code, mock-ups, and/or diagrams Example patch for musl ( all deleted lines of code should contain `#ifdef __GLIBC__` instead so it doesn't break compatibility with existing glibc systems running godot): ```diff diff --git a/platform/linuxbsd/crash_handler_linuxbsd.cpp b/platform/linuxbsd/crash_handler_linuxbsd.cpp index ea0222cb19..845a555706 100644 --- a/platform/linuxbsd/crash_handler_linuxbsd.cpp +++ b/platform/linuxbsd/crash_handler_linuxbsd.cpp @@ -38,89 +38,6 @@ #define CRASH_HANDLER_ENABLED 1 #endif -#ifdef CRASH_HANDLER_ENABLED -#include <cxxabi.h> -#include <dlfcn.h> -#include <execinfo.h> -#include <signal.h> -#include <stdlib.h> - -static void handle_crash(int sig) { - if (OS::get_singleton() == nullptr) { - abort(); - } - - void *bt_buffer[256]; - size_t size = backtrace(bt_buffer, 256); - String _execpath = OS::get_singleton()->get_executable_path(); - - String msg; - const ProjectSettings *proj_settings = ProjectSettings::get_singleton(); - if (proj_settings) { - msg = proj_settings->get("debug/settings/crash_handler/message"); - } - - // Dump the backtrace to stderr with a message to the user - fprintf(stderr, "%s: Program crashed with signal %d\n", __FUNCTION__, sig); - - if (OS::get_singleton()->get_main_loop()) { - OS::get_singleton()->get_main_loop()->notification(MainLoop::NOTIFICATION_CRASH); - } - - fprintf(stderr, "Dumping the backtrace. %s\n", msg.utf8().get_data()); - char **strings = backtrace_symbols(bt_buffer, size); - if (strings) { - for (size_t i = 1; i < size; i++) { - char fname[1024]; - Dl_info info; - - snprintf(fname, 1024, "%s", strings[i]); - - // Try to demangle the function name to provide a more readable one - if (dladdr(bt_buffer[i], &info) && info.dli_sname) { - if (info.dli_sname[0] == '_') { - int status; - char *demangled = abi::__cxa_demangle(info.dli_sname, nullptr, nullptr, &status); - - if (status == 0 && demangled) { - snprintf(fname, 1024, "%s", demangled); - } - - if (demangled) { - free(demangled); - } - } - } - - List<String> args; - - char str[1024]; - snprintf(str, 1024, "%p", bt_buffer[i]); - args.push_back(str); - args.push_back("-e"); - args.push_back(_execpath); - - String output = ""; - - // Try to get the file/line number using addr2line - int ret; - Error err = OS::get_singleton()->execute(String("addr2line"), args, &output, &ret); - if (err == OK) { - output.erase(output.length() - 1, 1); - } - - fprintf(stderr, "[%ld] %s (%s)\n", (long int)i, fname, output.utf8().get_data()); - } - - free(strings); - } - fprintf(stderr, "-- END OF BACKTRACE --\n"); - - // Abort to pass the error to the OS - abort(); -} -#endif - CrashHandler::CrashHandler() { disabled = false; } @@ -134,19 +51,8 @@ void CrashHandler::disable() { return; } -#ifdef CRASH_HANDLER_ENABLED - signal(SIGSEGV, nullptr); - signal(SIGFPE, nullptr); - signal(SIGILL, nullptr); -#endif - disabled = true; } void CrashHandler::initialize() { -#ifdef CRASH_HANDLER_ENABLED - signal(SIGSEGV, handle_crash); - signal(SIGFPE, handle_crash); - signal(SIGILL, handle_crash); -#endif } diff --git a/platform/linuxbsd/os_linuxbsd.cpp b/platform/linuxbsd/os_linuxbsd.cpp index 09e1f9461c..08bfead7f2 100644 --- a/platform/linuxbsd/os_linuxbsd.cpp +++ b/platform/linuxbsd/os_linuxbsd.cpp @@ -44,7 +44,7 @@ #include <stdio.h> #include <stdlib.h> #include <string.h> - +#include <libgen.h> #include <dlfcn.h> #include <fcntl.h> #include <sys/stat.h> @@ -302,6 +302,8 @@ static String get_mountpoint(const String &p_path) { return ""; } +const char *gnu_basename(const char *path) {const char *base = strrchr(path, '/'); return base ? base+1: path;} + Error OS_LinuxBSD::move_to_trash(const String &p_path) { int err_code; List<String> args; @@ -382,7 +384,8 @@ Error OS_LinuxBSD::move_to_trash(const String &p_path) { // The trash can is successfully created, now we check that we don't exceed our file name length limit. // If the file name is too long trim it so we can add the identifying number and ".trashinfo". // Assumes that the file name length limit is 255 characters. - String file_name = basename(p_path.utf8().get_data()); + String file_name = gnu_basename(p_path.utf8().get_data()); + if (file_name.length() > 240) { file_name = file_name.substr(0, file_name.length() - 15); } ``` ### If this enhancement will not be used often, can it be worked around with a few lines of script? Providing the patch in a separate "gd_tweaks" repository or fork would be possible but because this has to be applied to the source code, an upstream pull request for the main godot repository would be better. ### Is there a reason why this should be core and not an add-on in the asset library? It can't fix the buildsystem.
1.0
Musl libc support, scons detection for libexecinfo - <!-- Please fill in *all* the questions below and don't remove any of them. Proposals not following the template below will be closed immediately. --> ### Describe the project you are working on Musl support for Godot On distributions like Alpine Linux or Gentoo Linux musl support seems to be already feature-complete. With the use of glibc macros in the Godot source code (for debugging / stack traces) Godot can't be compiled on those systems by default. This proposal is about modifying the linuxbsd platform implementation to apply glibc macros only if glibc is available on the system. As an example the `basename` function which is included in the glibc headers differ from the musl implementation so it might be needed to "reimplement" it in the godot source code itself (see patch below). Of course the glibc implementation is more widespread on linux desktops but with the support of other libc implementations for linuxbsd (at least at source level) godot would be POSIX-compliant and therefore open for other linux platforms like servers or embedded systems without glibc. ### Describe the problem or limitation you are having in your project Godot currently only runs on Linux distributions with glibc as libc implementation. On distros with musl libc, it won't compile. ### Describe the feature / enhancement and how it helps to overcome the problem or limitation Patch the Godot source code so glibc specific features will only be enabled if glibc is available on the system. ### Describe how your proposal will work, with code, pseudo-code, mock-ups, and/or diagrams Example patch for musl ( all deleted lines of code should contain `#ifdef __GLIBC__` instead so it doesn't break compatibility with existing glibc systems running godot): ```diff diff --git a/platform/linuxbsd/crash_handler_linuxbsd.cpp b/platform/linuxbsd/crash_handler_linuxbsd.cpp index ea0222cb19..845a555706 100644 --- a/platform/linuxbsd/crash_handler_linuxbsd.cpp +++ b/platform/linuxbsd/crash_handler_linuxbsd.cpp @@ -38,89 +38,6 @@ #define CRASH_HANDLER_ENABLED 1 #endif -#ifdef CRASH_HANDLER_ENABLED -#include <cxxabi.h> -#include <dlfcn.h> -#include <execinfo.h> -#include <signal.h> -#include <stdlib.h> - -static void handle_crash(int sig) { - if (OS::get_singleton() == nullptr) { - abort(); - } - - void *bt_buffer[256]; - size_t size = backtrace(bt_buffer, 256); - String _execpath = OS::get_singleton()->get_executable_path(); - - String msg; - const ProjectSettings *proj_settings = ProjectSettings::get_singleton(); - if (proj_settings) { - msg = proj_settings->get("debug/settings/crash_handler/message"); - } - - // Dump the backtrace to stderr with a message to the user - fprintf(stderr, "%s: Program crashed with signal %d\n", __FUNCTION__, sig); - - if (OS::get_singleton()->get_main_loop()) { - OS::get_singleton()->get_main_loop()->notification(MainLoop::NOTIFICATION_CRASH); - } - - fprintf(stderr, "Dumping the backtrace. %s\n", msg.utf8().get_data()); - char **strings = backtrace_symbols(bt_buffer, size); - if (strings) { - for (size_t i = 1; i < size; i++) { - char fname[1024]; - Dl_info info; - - snprintf(fname, 1024, "%s", strings[i]); - - // Try to demangle the function name to provide a more readable one - if (dladdr(bt_buffer[i], &info) && info.dli_sname) { - if (info.dli_sname[0] == '_') { - int status; - char *demangled = abi::__cxa_demangle(info.dli_sname, nullptr, nullptr, &status); - - if (status == 0 && demangled) { - snprintf(fname, 1024, "%s", demangled); - } - - if (demangled) { - free(demangled); - } - } - } - - List<String> args; - - char str[1024]; - snprintf(str, 1024, "%p", bt_buffer[i]); - args.push_back(str); - args.push_back("-e"); - args.push_back(_execpath); - - String output = ""; - - // Try to get the file/line number using addr2line - int ret; - Error err = OS::get_singleton()->execute(String("addr2line"), args, &output, &ret); - if (err == OK) { - output.erase(output.length() - 1, 1); - } - - fprintf(stderr, "[%ld] %s (%s)\n", (long int)i, fname, output.utf8().get_data()); - } - - free(strings); - } - fprintf(stderr, "-- END OF BACKTRACE --\n"); - - // Abort to pass the error to the OS - abort(); -} -#endif - CrashHandler::CrashHandler() { disabled = false; } @@ -134,19 +51,8 @@ void CrashHandler::disable() { return; } -#ifdef CRASH_HANDLER_ENABLED - signal(SIGSEGV, nullptr); - signal(SIGFPE, nullptr); - signal(SIGILL, nullptr); -#endif - disabled = true; } void CrashHandler::initialize() { -#ifdef CRASH_HANDLER_ENABLED - signal(SIGSEGV, handle_crash); - signal(SIGFPE, handle_crash); - signal(SIGILL, handle_crash); -#endif } diff --git a/platform/linuxbsd/os_linuxbsd.cpp b/platform/linuxbsd/os_linuxbsd.cpp index 09e1f9461c..08bfead7f2 100644 --- a/platform/linuxbsd/os_linuxbsd.cpp +++ b/platform/linuxbsd/os_linuxbsd.cpp @@ -44,7 +44,7 @@ #include <stdio.h> #include <stdlib.h> #include <string.h> - +#include <libgen.h> #include <dlfcn.h> #include <fcntl.h> #include <sys/stat.h> @@ -302,6 +302,8 @@ static String get_mountpoint(const String &p_path) { return ""; } +const char *gnu_basename(const char *path) {const char *base = strrchr(path, '/'); return base ? base+1: path;} + Error OS_LinuxBSD::move_to_trash(const String &p_path) { int err_code; List<String> args; @@ -382,7 +384,8 @@ Error OS_LinuxBSD::move_to_trash(const String &p_path) { // The trash can is successfully created, now we check that we don't exceed our file name length limit. // If the file name is too long trim it so we can add the identifying number and ".trashinfo". // Assumes that the file name length limit is 255 characters. - String file_name = basename(p_path.utf8().get_data()); + String file_name = gnu_basename(p_path.utf8().get_data()); + if (file_name.length() > 240) { file_name = file_name.substr(0, file_name.length() - 15); } ``` ### If this enhancement will not be used often, can it be worked around with a few lines of script? Providing the patch in a separate "gd_tweaks" repository or fork would be possible but because this has to be applied to the source code, an upstream pull request for the main godot repository would be better. ### Is there a reason why this should be core and not an add-on in the asset library? It can't fix the buildsystem.
non_process
musl libc support scons detection for libexecinfo please fill in all the questions below and don t remove any of them proposals not following the template below will be closed immediately describe the project you are working on musl support for godot on distributions like alpine linux or gentoo linux musl support seems to be already feature complete with the use of glibc macros in the godot source code for debugging stack traces godot can t be compiled on those systems by default this proposal is about modifying the linuxbsd platform implementation to apply glibc macros only if glibc is available on the system as an example the basename function which is included in the glibc headers differ from the musl implementation so it might be needed to reimplement it in the godot source code itself see patch below of course the glibc implementation is more widespread on linux desktops but with the support of other libc implementations for linuxbsd at least at source level godot would be posix compliant and therefore open for other linux platforms like servers or embedded systems without glibc describe the problem or limitation you are having in your project godot currently only runs on linux distributions with glibc as libc implementation on distros with musl libc it won t compile describe the feature enhancement and how it helps to overcome the problem or limitation patch the godot source code so glibc specific features will only be enabled if glibc is available on the system describe how your proposal will work with code pseudo code mock ups and or diagrams example patch for musl all deleted lines of code should contain ifdef glibc instead so it doesn t break compatibility with existing glibc systems running godot diff diff git a platform linuxbsd crash handler linuxbsd cpp b platform linuxbsd crash handler linuxbsd cpp index a platform linuxbsd crash handler linuxbsd cpp b platform linuxbsd crash handler linuxbsd cpp define crash handler enabled endif ifdef crash handler enabled include include include include include static void handle crash int sig if os get singleton nullptr abort void bt buffer size t size backtrace bt buffer string execpath os get singleton get executable path string msg const projectsettings proj settings projectsettings get singleton if proj settings msg proj settings get debug settings crash handler message dump the backtrace to stderr with a message to the user fprintf stderr s program crashed with signal d n function sig if os get singleton get main loop os get singleton get main loop notification mainloop notification crash fprintf stderr dumping the backtrace s n msg get data char strings backtrace symbols bt buffer size if strings for size t i i size i char fname dl info info snprintf fname s strings try to demangle the function name to provide a more readable one if dladdr bt buffer info info dli sname if info dli sname int status char demangled abi cxa demangle info dli sname nullptr nullptr status if status demangled snprintf fname s demangled if demangled free demangled list args char str snprintf str p bt buffer args push back str args push back e args push back execpath string output try to get the file line number using int ret error err os get singleton execute string args output ret if err ok output erase output length fprintf stderr s s n long int i fname output get data free strings fprintf stderr end of backtrace n abort to pass the error to the os abort endif crashhandler crashhandler disabled false void crashhandler disable return ifdef crash handler enabled signal sigsegv nullptr signal sigfpe nullptr signal sigill nullptr endif disabled true void crashhandler initialize ifdef crash handler enabled signal sigsegv handle crash signal sigfpe handle crash signal sigill handle crash endif diff git a platform linuxbsd os linuxbsd cpp b platform linuxbsd os linuxbsd cpp index a platform linuxbsd os linuxbsd cpp b platform linuxbsd os linuxbsd cpp include include include include include include include static string get mountpoint const string p path return const char gnu basename const char path const char base strrchr path return base base path error os linuxbsd move to trash const string p path int err code list args error os linuxbsd move to trash const string p path the trash can is successfully created now we check that we don t exceed our file name length limit if the file name is too long trim it so we can add the identifying number and trashinfo assumes that the file name length limit is characters string file name basename p path get data string file name gnu basename p path get data if file name length file name file name substr file name length if this enhancement will not be used often can it be worked around with a few lines of script providing the patch in a separate gd tweaks repository or fork would be possible but because this has to be applied to the source code an upstream pull request for the main godot repository would be better is there a reason why this should be core and not an add on in the asset library it can t fix the buildsystem
0
8,136
11,339,552,995
IssuesEvent
2020-01-23 02:28:45
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
reopened
[1.0.0] TestUserCredentialsPropertiesOnWindows failing on Windows (v1.0 only)
area-System.Diagnostics.Process blocked bug disabled-test
``` System.Diagnostics.Tests.ProcessStartInfoTests.TestUserCredentialsPropertiesOnWindows [FAIL] System.InvalidOperationException : No process is associated with this object. Stack Trace: D:\j\workspace\outerloop_win---2155886e\src\System.Diagnostics.Process\src\System\Diagnostics\Process.cs(779,0): at System.Diagnostics.Process.EnsureState(State state) D:\j\workspace\outerloop_win---2155886e\src\System.Diagnostics.Process\src\System\Diagnostics\Process.cs(189,0): at System.Diagnostics.Process.get_HasExited() D:\j\workspace\outerloop_win---2155886e\src\System.Diagnostics.Process\tests\ProcessStartInfoTests.cs(393,0): at System.Diagnostics.Tests.ProcessStartInfoTests.TestUserCredentialsPropertiesOnWindows() ``` Examples: https://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_windows_nt_release/13/ https://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_windows_nt_debug/12/ https://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_win10_release/12/ https://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_win10_debug/12/ https://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_windows_nt_release/12/ https://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_windows_nt_debug/11/ https://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_win10_release/11/ https://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_win10_debug/11/ http://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_windows_nt_release/11/ ...
1.0
[1.0.0] TestUserCredentialsPropertiesOnWindows failing on Windows (v1.0 only) - ``` System.Diagnostics.Tests.ProcessStartInfoTests.TestUserCredentialsPropertiesOnWindows [FAIL] System.InvalidOperationException : No process is associated with this object. Stack Trace: D:\j\workspace\outerloop_win---2155886e\src\System.Diagnostics.Process\src\System\Diagnostics\Process.cs(779,0): at System.Diagnostics.Process.EnsureState(State state) D:\j\workspace\outerloop_win---2155886e\src\System.Diagnostics.Process\src\System\Diagnostics\Process.cs(189,0): at System.Diagnostics.Process.get_HasExited() D:\j\workspace\outerloop_win---2155886e\src\System.Diagnostics.Process\tests\ProcessStartInfoTests.cs(393,0): at System.Diagnostics.Tests.ProcessStartInfoTests.TestUserCredentialsPropertiesOnWindows() ``` Examples: https://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_windows_nt_release/13/ https://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_windows_nt_debug/12/ https://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_win10_release/12/ https://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_win10_debug/12/ https://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_windows_nt_release/12/ https://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_windows_nt_debug/11/ https://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_win10_release/11/ https://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_win10_debug/11/ http://ci.dot.net/job/dotnet_corefx/job/release_1.0.0/job/outerloop_windows_nt_release/11/ ...
process
testusercredentialspropertiesonwindows failing on windows only system diagnostics tests processstartinfotests testusercredentialspropertiesonwindows system invalidoperationexception no process is associated with this object stack trace d j workspace outerloop win src system diagnostics process src system diagnostics process cs at system diagnostics process ensurestate state state d j workspace outerloop win src system diagnostics process src system diagnostics process cs at system diagnostics process get hasexited d j workspace outerloop win src system diagnostics process tests processstartinfotests cs at system diagnostics tests processstartinfotests testusercredentialspropertiesonwindows examples
1
792,251
27,952,616,819
IssuesEvent
2023-03-24 10:00:27
nimblehq/android-templates
https://api.github.com/repos/nimblehq/android-templates
opened
Add new sample yaml file for setting up CI/CD with CodeMagic
type : feature status : approved priority : normal
## Why [Original RFC](https://github.com/nimblehq/android-templates/discussions/414) As of now, we are using [CodeMagic](https://codemagic.io/start/) to execute CI/CD pipelines on some projects, there is a lack of a sample yaml file for quick setup when initializing a new project ([.cicdtemplate](https://github.com/nimblehq/android-templates/tree/develop/.cicdtemplate)) ## Solution - Create a new `.codemagic` folder - Add new sample yaml file `codemagic.yml` and `README.md` for instruction ## Who Benefits? Developers ## Acceptance Criteria - [ ] Create a new `.codemagic` folder - [ ] Add new sample yaml file `codemagic.yml` and `README.md` for instruction - Sample should include running unit tests, detekt and deploying to Firebase App Distribution (similar to `.github` folder)
1.0
Add new sample yaml file for setting up CI/CD with CodeMagic - ## Why [Original RFC](https://github.com/nimblehq/android-templates/discussions/414) As of now, we are using [CodeMagic](https://codemagic.io/start/) to execute CI/CD pipelines on some projects, there is a lack of a sample yaml file for quick setup when initializing a new project ([.cicdtemplate](https://github.com/nimblehq/android-templates/tree/develop/.cicdtemplate)) ## Solution - Create a new `.codemagic` folder - Add new sample yaml file `codemagic.yml` and `README.md` for instruction ## Who Benefits? Developers ## Acceptance Criteria - [ ] Create a new `.codemagic` folder - [ ] Add new sample yaml file `codemagic.yml` and `README.md` for instruction - Sample should include running unit tests, detekt and deploying to Firebase App Distribution (similar to `.github` folder)
non_process
add new sample yaml file for setting up ci cd with codemagic why as of now we are using to execute ci cd pipelines on some projects there is a lack of a sample yaml file for quick setup when initializing a new project solution create a new codemagic folder add new sample yaml file codemagic yml and readme md for instruction who benefits developers acceptance criteria create a new codemagic folder add new sample yaml file codemagic yml and readme md for instruction sample should include running unit tests detekt and deploying to firebase app distribution similar to github folder
0
82,953
15,682,637,227
IssuesEvent
2021-03-25 07:38:03
soumya102/test_java
https://api.github.com/repos/soumya102/test_java
opened
CVE-2020-25638 (High) detected in hibernate-core-4.1.3.Final.jar
security vulnerability
## CVE-2020-25638 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-core-4.1.3.Final.jar</b></p></summary> <p>A module of the Hibernate Core project</p> <p>Library home page: <a href="http://hibernate.org">http://hibernate.org</a></p> <p>Path to dependency file: test_java/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/org/hibernate/hibernate-core/4.1.3.Final/hibernate-core-4.1.3.Final.jar</p> <p> Dependency Hierarchy: - :x: **hibernate-core-4.1.3.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/soumya102/test_java/commit/eeca4b314074f32b1a1c3d2fd7240f23ff45e18a">eeca4b314074f32b1a1c3d2fd7240f23ff45e18a</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in hibernate-core in versions prior to and including 5.4.23.Final. A SQL injection in the implementation of the JPA Criteria API can permit unsanitized literals when a literal is used in the SQL comments of the query. This flaw could allow an attacker to access unauthorized information or possibly conduct further attacks. The highest threat from this vulnerability is to data confidentiality and integrity. <p>Publish Date: 2020-12-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25638>CVE-2020-25638</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://in.relation.to/2020/11/19/hibernate-orm-5424-final-release/">https://in.relation.to/2020/11/19/hibernate-orm-5424-final-release/</a></p> <p>Release Date: 2020-12-02</p> <p>Fix Resolution: org.hibernate:hibernate-core:5.3.20.Final,5.4.24.Final</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-25638 (High) detected in hibernate-core-4.1.3.Final.jar - ## CVE-2020-25638 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-core-4.1.3.Final.jar</b></p></summary> <p>A module of the Hibernate Core project</p> <p>Library home page: <a href="http://hibernate.org">http://hibernate.org</a></p> <p>Path to dependency file: test_java/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/org/hibernate/hibernate-core/4.1.3.Final/hibernate-core-4.1.3.Final.jar</p> <p> Dependency Hierarchy: - :x: **hibernate-core-4.1.3.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/soumya102/test_java/commit/eeca4b314074f32b1a1c3d2fd7240f23ff45e18a">eeca4b314074f32b1a1c3d2fd7240f23ff45e18a</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in hibernate-core in versions prior to and including 5.4.23.Final. A SQL injection in the implementation of the JPA Criteria API can permit unsanitized literals when a literal is used in the SQL comments of the query. This flaw could allow an attacker to access unauthorized information or possibly conduct further attacks. The highest threat from this vulnerability is to data confidentiality and integrity. <p>Publish Date: 2020-12-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25638>CVE-2020-25638</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://in.relation.to/2020/11/19/hibernate-orm-5424-final-release/">https://in.relation.to/2020/11/19/hibernate-orm-5424-final-release/</a></p> <p>Release Date: 2020-12-02</p> <p>Fix Resolution: org.hibernate:hibernate-core:5.3.20.Final,5.4.24.Final</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in hibernate core final jar cve high severity vulnerability vulnerable library hibernate core final jar a module of the hibernate core project library home page a href path to dependency file test java pom xml path to vulnerable library canner repository org hibernate hibernate core final hibernate core final jar dependency hierarchy x hibernate core final jar vulnerable library found in head commit a href found in base branch main vulnerability details a flaw was found in hibernate core in versions prior to and including final a sql injection in the implementation of the jpa criteria api can permit unsanitized literals when a literal is used in the sql comments of the query this flaw could allow an attacker to access unauthorized information or possibly conduct further attacks the highest threat from this vulnerability is to data confidentiality and integrity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org hibernate hibernate core final final step up your open source security game with whitesource
0
156,930
5,995,069,348
IssuesEvent
2017-06-03 00:17:07
MTS11648/Don-NonAutomated-Bot
https://api.github.com/repos/MTS11648/Don-NonAutomated-Bot
closed
"Thank You" clip is awful.
Priority: 4-Minor Task: Needs Audio File Type: Bug
Don cuts himself off as he's saying it; it's not good and needs to be replaced.
1.0
"Thank You" clip is awful. - Don cuts himself off as he's saying it; it's not good and needs to be replaced.
non_process
thank you clip is awful don cuts himself off as he s saying it it s not good and needs to be replaced
0
8,548
11,723,934,715
IssuesEvent
2020-03-10 10:01:30
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
Basic adjustments compatibility broken in 3.0.1
bug: pending difficulty: trivial priority: high reproduce: confirmed scope: image processing understood: clear
af34b44459 introduced a vibrance setting in the params of `basicadj.c` and increased `DT_MODULE_INTROSPECTION(2, dt_iop_basicadj_params_t)`. Unfortunately, the import function `legacy_params()` to migrate the params structure from v1 to v2 has been forgotten, and users see the error `module basicadj version mismatch: 2 != 1` and lost their edits. See https://www.reddit.com/r/DarkTable/comments/fg2c7f/basic_adjustments_module_version_mismatch_after/
1.0
Basic adjustments compatibility broken in 3.0.1 - af34b44459 introduced a vibrance setting in the params of `basicadj.c` and increased `DT_MODULE_INTROSPECTION(2, dt_iop_basicadj_params_t)`. Unfortunately, the import function `legacy_params()` to migrate the params structure from v1 to v2 has been forgotten, and users see the error `module basicadj version mismatch: 2 != 1` and lost their edits. See https://www.reddit.com/r/DarkTable/comments/fg2c7f/basic_adjustments_module_version_mismatch_after/
process
basic adjustments compatibility broken in introduced a vibrance setting in the params of basicadj c and increased dt module introspection dt iop basicadj params t unfortunately the import function legacy params to migrate the params structure from to has been forgotten and users see the error module basicadj version mismatch and lost their edits see
1
13,352
15,816,685,330
IssuesEvent
2021-04-05 13:30:30
NationalSecurityAgency/ghidra
https://api.github.com/repos/NationalSecurityAgency/ghidra
closed
TBLPAG and PSVPAG addresses off by one for some of the PIC24/dsPIC30F/dsPIC33 family
Feature: Processor/PIC Type: Bug
**Describe the bug** The addresses for the TBLPAG and PSVPAG registers are not on an even word boundary. The problem is `pic24.sinc` at this block of text: ``` @if defined(PIC24E) || defined(dsPIC33E) || defined(dsPIC33C) define ram offset=0x32 size=2 [ DSRPAG ]; # 9bit Data Space Read Page Address define ram offset=0x34 size=2 [ DSWPAG ]; # 8bit Data Space Write Page Address define ram offset=0x36 size=2 [ RCOUNT ]; # Repeat counter # TODO: Re-implement with shadow stack # define ram offset=0x38 size=2 [ DCOUNT ]; # 13 bits long DO Loop counter define ram offset=0x54 size=1 [ TBLPAG ]; # 7bit Data Table Page Address @else define ram offset=0x31 size=1 [ TBLPAG ]; # 8bit Data Table Page Address define ram offset=0x33 size=1 [ PSVPAG ]; # Program Memory Visibility Page Address Pointer define ram offset=0x36 size=2 [ RCOUNT ]; # Repeat counter define ram offset=0x38 size=2 [ DCOUNT ]; # 13 bits long DO Loop counter define ram offset=0x3A size=3 [ DOSTART ]; define ram offset=0x3C size=3 [ DOEND ]; @endif ``` **Fix** After the `@else`, change the address of `TBLPAG` from `0x31` to `0x32`, and change the address of `PSVPAG` from `0x33` to `0x34`.
1.0
TBLPAG and PSVPAG addresses off by one for some of the PIC24/dsPIC30F/dsPIC33 family - **Describe the bug** The addresses for the TBLPAG and PSVPAG registers are not on an even word boundary. The problem is `pic24.sinc` at this block of text: ``` @if defined(PIC24E) || defined(dsPIC33E) || defined(dsPIC33C) define ram offset=0x32 size=2 [ DSRPAG ]; # 9bit Data Space Read Page Address define ram offset=0x34 size=2 [ DSWPAG ]; # 8bit Data Space Write Page Address define ram offset=0x36 size=2 [ RCOUNT ]; # Repeat counter # TODO: Re-implement with shadow stack # define ram offset=0x38 size=2 [ DCOUNT ]; # 13 bits long DO Loop counter define ram offset=0x54 size=1 [ TBLPAG ]; # 7bit Data Table Page Address @else define ram offset=0x31 size=1 [ TBLPAG ]; # 8bit Data Table Page Address define ram offset=0x33 size=1 [ PSVPAG ]; # Program Memory Visibility Page Address Pointer define ram offset=0x36 size=2 [ RCOUNT ]; # Repeat counter define ram offset=0x38 size=2 [ DCOUNT ]; # 13 bits long DO Loop counter define ram offset=0x3A size=3 [ DOSTART ]; define ram offset=0x3C size=3 [ DOEND ]; @endif ``` **Fix** After the `@else`, change the address of `TBLPAG` from `0x31` to `0x32`, and change the address of `PSVPAG` from `0x33` to `0x34`.
process
tblpag and psvpag addresses off by one for some of the family describe the bug the addresses for the tblpag and psvpag registers are not on an even word boundary the problem is sinc at this block of text if defined defined defined define ram offset size data space read page address define ram offset size data space write page address define ram offset size repeat counter todo re implement with shadow stack define ram offset size bits long do loop counter define ram offset size data table page address else define ram offset size data table page address define ram offset size program memory visibility page address pointer define ram offset size repeat counter define ram offset size bits long do loop counter define ram offset size define ram offset size endif fix after the else change the address of tblpag from to and change the address of psvpag from to
1
19,466
25,762,712,580
IssuesEvent
2022-12-08 22:03:37
brucemiller/LaTeXML
https://api.github.com/repos/brucemiller/LaTeXML
closed
Option for the magnification of images
enhancement postprocessing
Images produces by LaTeXML are fine if you leave the CSS untouched. However, if you want to - say - increase the body text size at 150%, then the images look way too small. My current "dirty hack", since I define the width of all my images wrt `\textwidth`, is to put a `\addtolength{\textwidth}{0.5\textwidth}` after `\begin{document}` when I'm converting with LaTeXML; however, it's definitely an unorthodox solution which could lead to unexpected results sooner or later. It would be nice, instead, to have an option which specifies the magnification factor for the images - much like you do with `--mathimagemagnification` for the math. The same goes for the picture environments.
1.0
Option for the magnification of images - Images produces by LaTeXML are fine if you leave the CSS untouched. However, if you want to - say - increase the body text size at 150%, then the images look way too small. My current "dirty hack", since I define the width of all my images wrt `\textwidth`, is to put a `\addtolength{\textwidth}{0.5\textwidth}` after `\begin{document}` when I'm converting with LaTeXML; however, it's definitely an unorthodox solution which could lead to unexpected results sooner or later. It would be nice, instead, to have an option which specifies the magnification factor for the images - much like you do with `--mathimagemagnification` for the math. The same goes for the picture environments.
process
option for the magnification of images images produces by latexml are fine if you leave the css untouched however if you want to say increase the body text size at then the images look way too small my current dirty hack since i define the width of all my images wrt textwidth is to put a addtolength textwidth textwidth after begin document when i m converting with latexml however it s definitely an unorthodox solution which could lead to unexpected results sooner or later it would be nice instead to have an option which specifies the magnification factor for the images much like you do with mathimagemagnification for the math the same goes for the picture environments
1
110,051
9,428,620,866
IssuesEvent
2019-04-12 02:02:39
bwsw/cloudstack-ui
https://api.github.com/repos/bwsw/cloudstack-ui
closed
[e2e tests] Set xit to failed tests add TODO for failed asserts for known issues
e2e tests
### Description Need to make running suite CI stable and set xit to failed tests add TODO for failed asserts for known issues. ### Acceptance criteria: There are no failed tests for executed CI suite.
1.0
[e2e tests] Set xit to failed tests add TODO for failed asserts for known issues - ### Description Need to make running suite CI stable and set xit to failed tests add TODO for failed asserts for known issues. ### Acceptance criteria: There are no failed tests for executed CI suite.
non_process
set xit to failed tests add todo for failed asserts for known issues description need to make running suite ci stable and set xit to failed tests add todo for failed asserts for known issues acceptance criteria there are no failed tests for executed ci suite
0
1,360
3,918,683,899
IssuesEvent
2016-04-21 13:26:30
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
April release
category: release / binary P1 type: process
I'll try to create the candidate today, from the release candidate in Google: mainline: 759bbfedbd8acd1324211d68b69e302478428e32 cherry-picks: - 1250fdac4c7769cfa200af8b4f9b061024356fea - ba8700ee63efe26c1a09d288129ced18a265ff89 - Rollback of https://bazel-review.googlesource.com/#/c/3220/
1.0
April release - I'll try to create the candidate today, from the release candidate in Google: mainline: 759bbfedbd8acd1324211d68b69e302478428e32 cherry-picks: - 1250fdac4c7769cfa200af8b4f9b061024356fea - ba8700ee63efe26c1a09d288129ced18a265ff89 - Rollback of https://bazel-review.googlesource.com/#/c/3220/
process
april release i ll try to create the candidate today from the release candidate in google mainline cherry picks rollback of
1
11,372
14,214,433,982
IssuesEvent
2020-11-17 05:14:08
kubeflow/pipelines
https://api.github.com/repos/kubeflow/pipelines
closed
[Process] Update backend development README
area/engprod help wanted kind/process
We have https://github.com/kubeflow/pipelines/tree/master/backend, but it is highly outdated because we have moved away from bazel. It'll be great if we can update it to reflect current best practices.
1.0
[Process] Update backend development README - We have https://github.com/kubeflow/pipelines/tree/master/backend, but it is highly outdated because we have moved away from bazel. It'll be great if we can update it to reflect current best practices.
process
update backend development readme we have but it is highly outdated because we have moved away from bazel it ll be great if we can update it to reflect current best practices
1
22,580
31,805,991,883
IssuesEvent
2023-09-13 13:59:25
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
reopened
Obsoletion notice: GO:0044278 disruption of cell wall in another organism
obsoletion multi-species process
Unnecessary grouping term; plant cell walls are only disrupted by from fungal and bacterial symbionts; bacterial cell walls are disrupted by phages, but there are no known non-symbiont predator toxins poisons that act on cell walls.
1.0
Obsoletion notice: GO:0044278 disruption of cell wall in another organism - Unnecessary grouping term; plant cell walls are only disrupted by from fungal and bacterial symbionts; bacterial cell walls are disrupted by phages, but there are no known non-symbiont predator toxins poisons that act on cell walls.
process
obsoletion notice go disruption of cell wall in another organism unnecessary grouping term plant cell walls are only disrupted by from fungal and bacterial symbionts bacterial cell walls are disrupted by phages but there are no known non symbiont predator toxins poisons that act on cell walls
1
197,710
14,940,310,448
IssuesEvent
2021-01-25 18:06:03
Carlosma7/MedAuth
https://api.github.com/repos/Carlosma7/MedAuth
closed
Realización de tests unitarios
testing
Se necesita realizar test para la consecución de un mvp de calidad, que compruebe que se realizan las operaciones correctamente.
1.0
Realización de tests unitarios - Se necesita realizar test para la consecución de un mvp de calidad, que compruebe que se realizan las operaciones correctamente.
non_process
realización de tests unitarios se necesita realizar test para la consecución de un mvp de calidad que compruebe que se realizan las operaciones correctamente
0
11,760
14,592,574,703
IssuesEvent
2020-12-19 18:16:27
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Translate raster in .asc format (AAIgrid) with a specific value NODATA_value=-9999 in the header is replaced by nan in the main text...
Bug Feedback Processing
<!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue --> **Describe the bug** <!-- A clear and concise description of what the bug is. --> When you translate a raster (geotiff) file with a specified nodata value (-9999 to replace nan) to an asc file, you get an .asc file with the good header (NODATA_value -9999) but with bad main text (NODATA_value nan) **Example : ** [Myfile_export_asc_1.zip](https://github.com/qgis/QGIS/files/5559530/Myfile_export_asc_1.zip) ncols 10 nrows 9 xllcorner 698362.000000000000 yllcorner 6946119.000000000000 cellsize 1.000000000000 **NODATA_value -9999** **nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan** 55.299999237060546875 55.299999237060546875 55.200000762939453125 55.200000762939453125 55.200000762939453125 **nan nan nan nan** 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.200000762939453125 55.200000762939453125 **nan nan nan** 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.200000762939453125 **nan nan** 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.200000762939453125 **nan** 55.5 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.5 55.5 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.5 55.5 55.5 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 **How to Reproduce** Use a <!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome --> 1. Load Myfile.tif (a geotiff with NODATA_Value=nan) [Myfile.zip](https://github.com/qgis/QGIS/files/5559454/Myfile.zip) 2. Go to 'Menu Raster > Conversion > Translate (Convert Format)...' ![1_Translate_tif_with_nan_to_asc_with_specified_nodata_value_to_asc_export_1](https://user-images.githubusercontent.com/7483093/99521216-703ee200-2994-11eb-9574-a7e85a88dfa6.jpg) 3. Assign a specified nodata value : nodata_value=-9999 4. Convert it to .asc file : C:/Temp/Myfile_export_asc_1.asc If you edit the .asc file, you get a good header but with bad main text [Myfile_export_asc_1.zip](https://github.com/qgis/QGIS/files/5559530/Myfile_export_asc_1.zip)... like described above. **To get the good value : -9999 in the main text, you must do these processings :** 1. Load Myfile.tif (a geotiff with NODATA_Value=nan) [Myfile.zip](https://github.com/qgis/QGIS/files/5559454/Myfile.zip) 2. Go to 'Menu Raster > Conversion > Translate (Convert Format)...' ![1_Translate_tif_with_nan_to_asc_with_specified_nodata_value_to_asc_export_1](https://user-images.githubusercontent.com/7483093/99521216-703ee200-2994-11eb-9574-a7e85a88dfa6.jpg) 3. Assign a specified nodata value : nodata_value=-9999 4. Convert it to .asc file : C:/Temp/Myfile_export_asc_1.asc (you get a good header with bad main text : [Myfile_export_asc_1.zip](https://github.com/qgis/QGIS/files/5559530/Myfile_export_asc_1.zip)) The GDAL/OGR console call gives: gdal_translate -a_srs EPSG:2154 -a_nodata -9999.0 -of AAIGrid C:/Temp/Myfile.tif C:/Temp/Myfile_export_asc_1.asc 5. Load .asc file (good header with bad main text) : Myfile_export_asc_1.asc 6. Click right mouse and the file : Save as... ![2a_saveas_export_1_asc_to_geotiff](https://user-images.githubusercontent.com/7483093/99521530-cf045b80-2994-11eb-985d-42940be74af8.jpg) Save this first .asc file to GeoTiff... ![2b_saveas_(Myfile_export_asc_1)_to_geotiff_(My_export_asc_1_saveas_to_geotiff)](https://user-images.githubusercontent.com/7483093/99521554-d6c40000-2994-11eb-9318-9f3ff4ec18dd.jpg) 7. open the new GeoTiff : [Myfile_export_asc_1_saveas_to_geotiff.zip](https://github.com/qgis/QGIS/files/5559541/Myfile_export_asc_1_saveas_to_geotiff.zip) 8. Translate the new GeoTiff to .asc file again. Go to 'Menu Raster > Conversion > Translate (Convert Format)...' ![3_Translate_(My_export_asc_1_saveas_to_geotiff)_to_asc_export_2](https://user-images.githubusercontent.com/7483093/99521584-e3e0ef00-2994-11eb-872e-d2f162370871.jpg) The GDAL/OGR console call gives: gdal_translate -a_srs EPSG:2154 -of AAIGrid C:\Temp\Myfile_export_asc_1_saveas_to_geotiff.tif C:/Temp/Myfile_export_asc_2.asc 9. Then, you get an .asc file with the good header and good main text like this : [Myfile_export_asc_2.zip](https://github.com/qgis/QGIS/files/5559537/Myfile_export_asc_2.zip) ncols 10 nrows 9 xllcorner 698362.000000000000 yllcorner 6946119.000000000000 cellsize 1.000000000000 **NODATA_value -9999 -9999.0 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999** 55.299999237060546875 55.299999237060546875 55.200000762939453125 55.200000762939453125 55.200000762939453125 **-9999 -9999 -9999 -9999** 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.200000762939453125 55.200000762939453125 **-9999 -9999 -9999** 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.200000762939453125 **-9999 -9999** 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.200000762939453125 **-9999** 55.5 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.5 55.5 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.5 55.5 55.5 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 **QGIS and OS versions** <!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here --> QGIS version 3.16.0-Hannover QGIS code revision 43b64b13f3 Compiled against Qt 5.11.2 Running against Qt 5.11.2 Compiled against GDAL/OGR 3.1.4 Running against GDAL/OGR 3.1.4 Compiled against GEOS 3.8.1-CAPI-1.13.3 Running against GEOS 3.8.1-CAPI-1.13.3 Compiled against SQLite 3.29.0 Running against SQLite 3.29.0 PostgreSQL Client Version 11.5 SpatiaLite Version 4.3.0 QWT Version 6.1.3 QScintilla2 Version 2.10.8 Compiled against PROJ 6.3.2 Running against PROJ Rel. 6.3.2, May 1st, 2020 OS Version Windows 10 (10.0) **Additional context** <!-- Add any other context about the problem here. -->
1.0
Translate raster in .asc format (AAIgrid) with a specific value NODATA_value=-9999 in the header is replaced by nan in the main text... - <!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue --> **Describe the bug** <!-- A clear and concise description of what the bug is. --> When you translate a raster (geotiff) file with a specified nodata value (-9999 to replace nan) to an asc file, you get an .asc file with the good header (NODATA_value -9999) but with bad main text (NODATA_value nan) **Example : ** [Myfile_export_asc_1.zip](https://github.com/qgis/QGIS/files/5559530/Myfile_export_asc_1.zip) ncols 10 nrows 9 xllcorner 698362.000000000000 yllcorner 6946119.000000000000 cellsize 1.000000000000 **NODATA_value -9999** **nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan** 55.299999237060546875 55.299999237060546875 55.200000762939453125 55.200000762939453125 55.200000762939453125 **nan nan nan nan** 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.200000762939453125 55.200000762939453125 **nan nan nan** 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.200000762939453125 **nan nan** 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.200000762939453125 **nan** 55.5 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.5 55.5 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.5 55.5 55.5 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 **How to Reproduce** Use a <!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome --> 1. Load Myfile.tif (a geotiff with NODATA_Value=nan) [Myfile.zip](https://github.com/qgis/QGIS/files/5559454/Myfile.zip) 2. Go to 'Menu Raster > Conversion > Translate (Convert Format)...' ![1_Translate_tif_with_nan_to_asc_with_specified_nodata_value_to_asc_export_1](https://user-images.githubusercontent.com/7483093/99521216-703ee200-2994-11eb-9574-a7e85a88dfa6.jpg) 3. Assign a specified nodata value : nodata_value=-9999 4. Convert it to .asc file : C:/Temp/Myfile_export_asc_1.asc If you edit the .asc file, you get a good header but with bad main text [Myfile_export_asc_1.zip](https://github.com/qgis/QGIS/files/5559530/Myfile_export_asc_1.zip)... like described above. **To get the good value : -9999 in the main text, you must do these processings :** 1. Load Myfile.tif (a geotiff with NODATA_Value=nan) [Myfile.zip](https://github.com/qgis/QGIS/files/5559454/Myfile.zip) 2. Go to 'Menu Raster > Conversion > Translate (Convert Format)...' ![1_Translate_tif_with_nan_to_asc_with_specified_nodata_value_to_asc_export_1](https://user-images.githubusercontent.com/7483093/99521216-703ee200-2994-11eb-9574-a7e85a88dfa6.jpg) 3. Assign a specified nodata value : nodata_value=-9999 4. Convert it to .asc file : C:/Temp/Myfile_export_asc_1.asc (you get a good header with bad main text : [Myfile_export_asc_1.zip](https://github.com/qgis/QGIS/files/5559530/Myfile_export_asc_1.zip)) The GDAL/OGR console call gives: gdal_translate -a_srs EPSG:2154 -a_nodata -9999.0 -of AAIGrid C:/Temp/Myfile.tif C:/Temp/Myfile_export_asc_1.asc 5. Load .asc file (good header with bad main text) : Myfile_export_asc_1.asc 6. Click right mouse and the file : Save as... ![2a_saveas_export_1_asc_to_geotiff](https://user-images.githubusercontent.com/7483093/99521530-cf045b80-2994-11eb-985d-42940be74af8.jpg) Save this first .asc file to GeoTiff... ![2b_saveas_(Myfile_export_asc_1)_to_geotiff_(My_export_asc_1_saveas_to_geotiff)](https://user-images.githubusercontent.com/7483093/99521554-d6c40000-2994-11eb-9318-9f3ff4ec18dd.jpg) 7. open the new GeoTiff : [Myfile_export_asc_1_saveas_to_geotiff.zip](https://github.com/qgis/QGIS/files/5559541/Myfile_export_asc_1_saveas_to_geotiff.zip) 8. Translate the new GeoTiff to .asc file again. Go to 'Menu Raster > Conversion > Translate (Convert Format)...' ![3_Translate_(My_export_asc_1_saveas_to_geotiff)_to_asc_export_2](https://user-images.githubusercontent.com/7483093/99521584-e3e0ef00-2994-11eb-872e-d2f162370871.jpg) The GDAL/OGR console call gives: gdal_translate -a_srs EPSG:2154 -of AAIGrid C:\Temp\Myfile_export_asc_1_saveas_to_geotiff.tif C:/Temp/Myfile_export_asc_2.asc 9. Then, you get an .asc file with the good header and good main text like this : [Myfile_export_asc_2.zip](https://github.com/qgis/QGIS/files/5559537/Myfile_export_asc_2.zip) ncols 10 nrows 9 xllcorner 698362.000000000000 yllcorner 6946119.000000000000 cellsize 1.000000000000 **NODATA_value -9999 -9999.0 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999 -9999** 55.299999237060546875 55.299999237060546875 55.200000762939453125 55.200000762939453125 55.200000762939453125 **-9999 -9999 -9999 -9999** 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.200000762939453125 55.200000762939453125 **-9999 -9999 -9999** 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.200000762939453125 **-9999 -9999** 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.200000762939453125 **-9999** 55.5 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.5 55.5 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 55.5 55.5 55.5 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.40000152587890625 55.299999237060546875 55.299999237060546875 55.299999237060546875 **QGIS and OS versions** <!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here --> QGIS version 3.16.0-Hannover QGIS code revision 43b64b13f3 Compiled against Qt 5.11.2 Running against Qt 5.11.2 Compiled against GDAL/OGR 3.1.4 Running against GDAL/OGR 3.1.4 Compiled against GEOS 3.8.1-CAPI-1.13.3 Running against GEOS 3.8.1-CAPI-1.13.3 Compiled against SQLite 3.29.0 Running against SQLite 3.29.0 PostgreSQL Client Version 11.5 SpatiaLite Version 4.3.0 QWT Version 6.1.3 QScintilla2 Version 2.10.8 Compiled against PROJ 6.3.2 Running against PROJ Rel. 6.3.2, May 1st, 2020 OS Version Windows 10 (10.0) **Additional context** <!-- Add any other context about the problem here. -->
process
translate raster in asc format aaigrid with a specific value nodata value in the header is replaced by nan in the main text bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue describe the bug when you translate a raster geotiff file with a specified nodata value to replace nan to an asc file you get an asc file with the good header nodata value but with bad main text nodata value nan example ncols nrows xllcorner yllcorner cellsize nodata value nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan how to reproduce use a load myfile tif a geotiff with nodata value nan go to menu raster conversion translate convert format assign a specified nodata value nodata value convert it to asc file c temp myfile export asc asc if you edit the asc file you get a good header but with bad main text like described above to get the good value in the main text you must do these processings load myfile tif a geotiff with nodata value nan go to menu raster conversion translate convert format assign a specified nodata value nodata value convert it to asc file c temp myfile export asc asc you get a good header with bad main text the gdal ogr console call gives gdal translate a srs epsg a nodata of aaigrid c temp myfile tif c temp myfile export asc asc load asc file good header with bad main text myfile export asc asc click right mouse and the file save as save this first asc file to geotiff open the new geotiff translate the new geotiff to asc file again go to menu raster conversion translate convert format the gdal ogr console call gives gdal translate a srs epsg of aaigrid c temp myfile export asc saveas to geotiff tif c temp myfile export asc asc then you get an asc file with the good header and good main text like this ncols nrows xllcorner yllcorner cellsize nodata value qgis and os versions about click in the table ctrl a and then ctrl c finally paste here qgis version hannover qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel may os version windows additional context
1
2,093
4,928,975,219
IssuesEvent
2016-11-27 16:12:53
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
closed
"Lock" rows from being edited ?
help wanted inprocess
Hi Allen, I am hoping there is a way to do this. Based on certain criteria, I'd like to make certain entire rows uneditable. Currently from what I have seen, columns can be either editable or not; I don't know if I can put logic somewhere that will control whether or not to "lock" certain rows, so no cells can be edited within those rows. Possible? Thanks, Eric.
1.0
"Lock" rows from being edited ? - Hi Allen, I am hoping there is a way to do this. Based on certain criteria, I'd like to make certain entire rows uneditable. Currently from what I have seen, columns can be either editable or not; I don't know if I can put logic somewhere that will control whether or not to "lock" certain rows, so no cells can be edited within those rows. Possible? Thanks, Eric.
process
lock rows from being edited hi allen i am hoping there is a way to do this based on certain criteria i d like to make certain entire rows uneditable currently from what i have seen columns can be either editable or not i don t know if i can put logic somewhere that will control whether or not to lock certain rows so no cells can be edited within those rows possible thanks eric
1
760,887
26,659,492,730
IssuesEvent
2023-01-25 19:44:25
IBMa/equal-access
https://api.github.com/repos/IBMa/equal-access
closed
[ACT - new rule] 78fd32 - Line height in style attributes is not !important
engine SME priority-2 (med) development ACT
Related to #1124 Word spacing #1125 Letter spacing No existing rules map to IBM Accessibility requirement [1.4.12 Text Spacing (Level AA)](https://www.ibm.com/able/requirements/requirements/#1_4_12) - No loss of content or functionality occurs when users change letter, word and paragraph spacing, as well as line height. No Checker rules listed under 1.4.12 in https://unpkg.com/accessibility-checker-engine@3.1.40/help/rules.html - [x] Create rules to cover [ACT 78fd32 - Line height in style attributes is not !important](https://www.w3.org/WAI/standards-guidelines/act/rules/78fd32/) ### Background See new rule discussion on box See **_Details by test case_** tab in the spreadsheet [ACT vs Engine inconsistencies.xlsx](https://ibm.ent.box.com/file/1040535127435?s=9o6m83gzzc9sp8pu8lwtyesh0b8wwxln) on box.
1.0
[ACT - new rule] 78fd32 - Line height in style attributes is not !important - Related to #1124 Word spacing #1125 Letter spacing No existing rules map to IBM Accessibility requirement [1.4.12 Text Spacing (Level AA)](https://www.ibm.com/able/requirements/requirements/#1_4_12) - No loss of content or functionality occurs when users change letter, word and paragraph spacing, as well as line height. No Checker rules listed under 1.4.12 in https://unpkg.com/accessibility-checker-engine@3.1.40/help/rules.html - [x] Create rules to cover [ACT 78fd32 - Line height in style attributes is not !important](https://www.w3.org/WAI/standards-guidelines/act/rules/78fd32/) ### Background See new rule discussion on box See **_Details by test case_** tab in the spreadsheet [ACT vs Engine inconsistencies.xlsx](https://ibm.ent.box.com/file/1040535127435?s=9o6m83gzzc9sp8pu8lwtyesh0b8wwxln) on box.
non_process
line height in style attributes is not important related to word spacing letter spacing no existing rules map to ibm accessibility requirement no loss of content or functionality occurs when users change letter word and paragraph spacing as well as line height no checker rules listed under in create rules to cover background see new rule discussion on box see details by test case tab in the spreadsheet on box
0
19,441
25,708,739,807
IssuesEvent
2022-12-07 04:02:10
AusClimateService/axiom
https://api.github.com/repos/AusClimateService/axiom
closed
Process ERA5 at aus10i
processing v1.0.0
As per discussion with @tha051 05/09/2022 we need to process CCAM ERA5 @ aus10i. All payloads should be ready.
1.0
Process ERA5 at aus10i - As per discussion with @tha051 05/09/2022 we need to process CCAM ERA5 @ aus10i. All payloads should be ready.
process
process at as per discussion with we need to process ccam all payloads should be ready
1
12,305
14,858,945,722
IssuesEvent
2021-01-18 17:35:57
yuta252/startlens_web_backend
https://api.github.com/repos/yuta252/startlens_web_backend
closed
観光地訪問時のログ収集モデルの作成
dev process
## 概要 ユーザーが観光地をオンラインで訪問した際に(具体的にはフロントエンドで観光地ページをレンダリングした際のuseEffect関数が呼ばれた時に)ログ情報をデータベースで管理する。 ## 変更点 --- - [x] userStatisticモデルの作成 - 同日に複数回訪問したユーザーは全て1回の訪問とカウントするため、同日に複数回ログを保存しないようにバリデーション追加 - [x] userStatisticContorllerの作成 - [x] Rspecテストの作成 ## 課題 --- - Exhibitごとのログ収集 ## 参照 --- - [カスタムバリデーションの追加](https://teratail.com/questions/50423) - [RubyとRailsにおけるTime, Date, DateTime, TimeWithZoneの違い](https://qiita.com/jnchito/items/cae89ee43c30f5d6fa2c) - [ActiveSupportの期間指定メソッド](https://qiita.com/terufumi1122/items/3aa21c20eeacbce33b93) - [Railsでの日付操作](https://qiita.com/mmmm/items/efda48f1ac0267c95c29) ## 備考 ---
1.0
観光地訪問時のログ収集モデルの作成 - ## 概要 ユーザーが観光地をオンラインで訪問した際に(具体的にはフロントエンドで観光地ページをレンダリングした際のuseEffect関数が呼ばれた時に)ログ情報をデータベースで管理する。 ## 変更点 --- - [x] userStatisticモデルの作成 - 同日に複数回訪問したユーザーは全て1回の訪問とカウントするため、同日に複数回ログを保存しないようにバリデーション追加 - [x] userStatisticContorllerの作成 - [x] Rspecテストの作成 ## 課題 --- - Exhibitごとのログ収集 ## 参照 --- - [カスタムバリデーションの追加](https://teratail.com/questions/50423) - [RubyとRailsにおけるTime, Date, DateTime, TimeWithZoneの違い](https://qiita.com/jnchito/items/cae89ee43c30f5d6fa2c) - [ActiveSupportの期間指定メソッド](https://qiita.com/terufumi1122/items/3aa21c20eeacbce33b93) - [Railsでの日付操作](https://qiita.com/mmmm/items/efda48f1ac0267c95c29) ## 備考 ---
process
観光地訪問時のログ収集モデルの作成 概要 ユーザーが観光地をオンラインで訪問した際に(具体的にはフロントエンドで観光地ページをレンダリングした際のuseeffect関数が呼ばれた時に)ログ情報をデータベースで管理する。 変更点 userstatisticモデルの作成 、同日に複数回ログを保存しないようにバリデーション追加 userstatisticcontorllerの作成 rspecテストの作成 課題 exhibitごとのログ収集 参照 備考
1
4,989
7,822,174,389
IssuesEvent
2018-06-14 00:48:58
StrikeNP/trac_test
https://api.github.com/repos/StrikeNP/trac_test
closed
Link Trac site to internal group page (Trac #3)
Migrated from Trac post_processing senkbeil@uwm.edu task
In order for this CLUBB Trac site to be useful to the group, it must be readily accessible. A link should be placed on the internal group page. Also there is no component for internal site issues. Perhaps this means I am using this trac site wrong. Attachments: Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/3 ```json { "status": "closed", "changetime": "2009-05-13T18:09:47", "description": "In order for this CLUBB Trac site to be useful to the group, it must be readily accessible. A link should be placed on the internal group page.\n\nAlso there is no component for internal site issues. Perhaps this means I am using this trac site wrong.", "reporter": "fasching@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1242238187000000", "component": "post_processing", "summary": "Link Trac site to internal group page", "priority": "minor", "keywords": "", "time": "2009-05-01T21:17:44", "milestone": "", "owner": "senkbeil@uwm.edu", "type": "task" } ```
1.0
Link Trac site to internal group page (Trac #3) - In order for this CLUBB Trac site to be useful to the group, it must be readily accessible. A link should be placed on the internal group page. Also there is no component for internal site issues. Perhaps this means I am using this trac site wrong. Attachments: Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/3 ```json { "status": "closed", "changetime": "2009-05-13T18:09:47", "description": "In order for this CLUBB Trac site to be useful to the group, it must be readily accessible. A link should be placed on the internal group page.\n\nAlso there is no component for internal site issues. Perhaps this means I am using this trac site wrong.", "reporter": "fasching@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1242238187000000", "component": "post_processing", "summary": "Link Trac site to internal group page", "priority": "minor", "keywords": "", "time": "2009-05-01T21:17:44", "milestone": "", "owner": "senkbeil@uwm.edu", "type": "task" } ```
process
link trac site to internal group page trac in order for this clubb trac site to be useful to the group it must be readily accessible a link should be placed on the internal group page also there is no component for internal site issues perhaps this means i am using this trac site wrong attachments migrated from json status closed changetime description in order for this clubb trac site to be useful to the group it must be readily accessible a link should be placed on the internal group page n nalso there is no component for internal site issues perhaps this means i am using this trac site wrong reporter fasching uwm edu cc resolution verified by v larson ts component post processing summary link trac site to internal group page priority minor keywords time milestone owner senkbeil uwm edu type task
1
684,845
23,435,107,985
IssuesEvent
2022-08-15 08:57:13
duzda/deezer-enhanced
https://api.github.com/repos/duzda/deezer-enhanced
opened
Deemix notifications for tracks
enhancement low priority
With the 0.3.0 update it's possible to use deemix-integration, commit https://github.com/duzda/deezer-enhanced/commit/b09202256408b71e2baf3016dc13abbee4bd96f8 adds notification when album, playlist or artist's discography gets downloaded. Tracks are not supported, because the id can't be easily obtained from the url, instead the id has to obtained any other way.
1.0
Deemix notifications for tracks - With the 0.3.0 update it's possible to use deemix-integration, commit https://github.com/duzda/deezer-enhanced/commit/b09202256408b71e2baf3016dc13abbee4bd96f8 adds notification when album, playlist or artist's discography gets downloaded. Tracks are not supported, because the id can't be easily obtained from the url, instead the id has to obtained any other way.
non_process
deemix notifications for tracks with the update it s possible to use deemix integration commit adds notification when album playlist or artist s discography gets downloaded tracks are not supported because the id can t be easily obtained from the url instead the id has to obtained any other way
0
286,985
31,805,760,219
IssuesEvent
2023-09-13 13:52:13
hinoshiba/news
https://api.github.com/repos/hinoshiba/news
opened
[SecurityWeek] Airbus Launches Investigation After Hacker Leaks Data
SecurityWeek
Airbus has launched an investigation after a hacker claimed to have breached the company’s systems and leaked some business documents. The post [Airbus Launches Investigation After Hacker Leaks Data](https://www.securityweek.com/airbus-launches-investigation-after-hacker-leaks-data/) appeared first on [SecurityWeek](https://www.securityweek.com). <https://www.securityweek.com/airbus-launches-investigation-after-hacker-leaks-data/>
True
[SecurityWeek] Airbus Launches Investigation After Hacker Leaks Data - Airbus has launched an investigation after a hacker claimed to have breached the company’s systems and leaked some business documents. The post [Airbus Launches Investigation After Hacker Leaks Data](https://www.securityweek.com/airbus-launches-investigation-after-hacker-leaks-data/) appeared first on [SecurityWeek](https://www.securityweek.com). <https://www.securityweek.com/airbus-launches-investigation-after-hacker-leaks-data/>
non_process
airbus launches investigation after hacker leaks data airbus has launched an investigation after a hacker claimed to have breached the company’s systems and leaked some business documents the post appeared first on
0
102,266
4,153,004,938
IssuesEvent
2016-06-16 05:11:29
MobleyLab/alchemical-analysis
https://api.github.com/repos/MobleyLab/alchemical-analysis
closed
Consider implementing extrapolation scheme for MBAR to handle AMBER's lack of endpoint data
enhancement Medium priority
I wonder if it would make sense to introduce an extrapolation scheme for "incomplete" MBAR data data. The principal problem for me is sander which cannot sample the end-points. I also heard from some other person that he doesn't sample the end-points and relies on extrapolation apparently. Is there any good reason why that should not be done? I certainly understand that this shouldn't be used to bridge larger gaps.
1.0
Consider implementing extrapolation scheme for MBAR to handle AMBER's lack of endpoint data - I wonder if it would make sense to introduce an extrapolation scheme for "incomplete" MBAR data data. The principal problem for me is sander which cannot sample the end-points. I also heard from some other person that he doesn't sample the end-points and relies on extrapolation apparently. Is there any good reason why that should not be done? I certainly understand that this shouldn't be used to bridge larger gaps.
non_process
consider implementing extrapolation scheme for mbar to handle amber s lack of endpoint data i wonder if it would make sense to introduce an extrapolation scheme for incomplete mbar data data the principal problem for me is sander which cannot sample the end points i also heard from some other person that he doesn t sample the end points and relies on extrapolation apparently is there any good reason why that should not be done i certainly understand that this shouldn t be used to bridge larger gaps
0
164,274
12,795,344,164
IssuesEvent
2020-07-02 08:36:03
Rocologo/MobHunting
https://api.github.com/repos/Rocologo/MobHunting
closed
NPE on server start
Fixed - To be tested
Spigot 1.16.1 MH 7.5.0-SNAPSHOT-B1054 updated from 7.5.0-SNAPSHOT-B1051 On server start: ``` [05:11:19] [Server thread/ERROR]: Error occurred while enabling MobHunting v7.5.0-SNAPSHOT-B1054 (Is it up to date?) java.lang.NullPointerException: null at one.lindegaard.MobHunting.MobHunting.onEnable(MobHunting.java:142) ~[?:?] at org.bukkit.plugin.java.JavaPlugin.setEnabled(JavaPlugin.java:263) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at org.bukkit.plugin.java.JavaPluginLoader.enablePlugin(JavaPluginLoader.java:351) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at org.bukkit.plugin.SimplePluginManager.enablePlugin(SimplePluginManager.java:480) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at org.bukkit.craftbukkit.v1_16_R1.CraftServer.enablePlugin(CraftServer.java:491) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at org.bukkit.craftbukkit.v1_16_R1.CraftServer.enablePlugins(CraftServer.java:405) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at net.minecraft.server.v1_16_R1.MinecraftServer.loadWorld(MinecraftServer.java:438) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at net.minecraft.server.v1_16_R1.DedicatedServer.init(DedicatedServer.java:219) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at net.minecraft.server.v1_16_R1.MinecraftServer.v(MinecraftServer.java:810) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at net.minecraft.server.v1_16_R1.MinecraftServer.lambda$0(MinecraftServer.java:164) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at java.lang.Thread.run(Thread.java:834) [?:?] ```
1.0
NPE on server start - Spigot 1.16.1 MH 7.5.0-SNAPSHOT-B1054 updated from 7.5.0-SNAPSHOT-B1051 On server start: ``` [05:11:19] [Server thread/ERROR]: Error occurred while enabling MobHunting v7.5.0-SNAPSHOT-B1054 (Is it up to date?) java.lang.NullPointerException: null at one.lindegaard.MobHunting.MobHunting.onEnable(MobHunting.java:142) ~[?:?] at org.bukkit.plugin.java.JavaPlugin.setEnabled(JavaPlugin.java:263) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at org.bukkit.plugin.java.JavaPluginLoader.enablePlugin(JavaPluginLoader.java:351) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at org.bukkit.plugin.SimplePluginManager.enablePlugin(SimplePluginManager.java:480) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at org.bukkit.craftbukkit.v1_16_R1.CraftServer.enablePlugin(CraftServer.java:491) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at org.bukkit.craftbukkit.v1_16_R1.CraftServer.enablePlugins(CraftServer.java:405) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at net.minecraft.server.v1_16_R1.MinecraftServer.loadWorld(MinecraftServer.java:438) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at net.minecraft.server.v1_16_R1.DedicatedServer.init(DedicatedServer.java:219) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at net.minecraft.server.v1_16_R1.MinecraftServer.v(MinecraftServer.java:810) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at net.minecraft.server.v1_16_R1.MinecraftServer.lambda$0(MinecraftServer.java:164) ~[spigot-1.16.1.jar-2020-06-30-2152:git-Spigot-4d9262c-3b9cf0f] at java.lang.Thread.run(Thread.java:834) [?:?] ```
non_process
npe on server start spigot mh snapshot updated from snapshot on server start error occurred while enabling mobhunting snapshot is it up to date java lang nullpointerexception null at one lindegaard mobhunting mobhunting onenable mobhunting java at org bukkit plugin java javaplugin setenabled javaplugin java at org bukkit plugin java javapluginloader enableplugin javapluginloader java at org bukkit plugin simplepluginmanager enableplugin simplepluginmanager java at org bukkit craftbukkit craftserver enableplugin craftserver java at org bukkit craftbukkit craftserver enableplugins craftserver java at net minecraft server minecraftserver loadworld minecraftserver java at net minecraft server dedicatedserver init dedicatedserver java at net minecraft server minecraftserver v minecraftserver java at net minecraft server minecraftserver lambda minecraftserver java at java lang thread run thread java
0
445,737
12,835,649,573
IssuesEvent
2020-07-07 13:13:28
cds-snc/covid-shield-mobile
https://api.github.com/repos/cds-snc/covid-shield-mobile
opened
Back link changed from blue to grey
a11y high priority v1launch
The ‘Back’ link in onboarding and in ‘How it works’ are all faded grey rather than blue like other hyperlinks. This should be changed to the blue found on the cancel links. ![Image from iOS (12)](https://user-images.githubusercontent.com/35537308/86786481-0f8cb980-c032-11ea-8212-4b233c86bd48.png)
1.0
Back link changed from blue to grey - The ‘Back’ link in onboarding and in ‘How it works’ are all faded grey rather than blue like other hyperlinks. This should be changed to the blue found on the cancel links. ![Image from iOS (12)](https://user-images.githubusercontent.com/35537308/86786481-0f8cb980-c032-11ea-8212-4b233c86bd48.png)
non_process
back link changed from blue to grey the ‘back’ link in onboarding and in ‘how it works’ are all faded grey rather than blue like other hyperlinks this should be changed to the blue found on the cancel links
0
21,343
29,116,663,499
IssuesEvent
2023-05-17 02:00:09
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Wed, 17 May 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### Common Diffusion Noise Schedules and Sample Steps are Flawed - **Authors:** Shanchuan Lin, Bingchen Liu, Jiashi Li, Xiao Yang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.08891 - **Pdf link:** https://arxiv.org/pdf/2305.08891 - **Abstract** We discover that common diffusion noise schedules do not enforce the last timestep to have zero signal-to-noise ratio (SNR), and some implementations of diffusion samplers do not start from the last timestep. Such designs are flawed and do not reflect the fact that the model is given pure Gaussian noise at inference, creating a discrepancy between training and inference. We show that the flawed design causes real problems in existing implementations. In Stable Diffusion, it severely limits the model to only generate images with medium brightness and prevents it from generating very bright and dark samples. We propose a few simple fixes: (1) rescale the noise schedule to enforce zero terminal SNR; (2) train the model with v prediction; (3) change the sampler to always start from the last timestep; (4) rescale classifier-free guidance to prevent over-exposure. These simple changes ensure the diffusion process is congruent between training and inference and allow the model to generate samples more faithful to the original data distribution. ### Motion Question Answering via Modular Motion Programs - **Authors:** Mark Endo, Joy Hsu, Jiaman Li, Jiajun Wu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2305.08953 - **Pdf link:** https://arxiv.org/pdf/2305.08953 - **Abstract** In order to build artificial intelligence systems that can perceive and reason with human behavior in the real world, we must first design models that conduct complex spatio-temporal reasoning over motion sequences. Moving towards this goal, we propose the HumanMotionQA task to evaluate complex, multi-step reasoning abilities of models on long-form human motion sequences. We generate a dataset of question-answer pairs that require detecting motor cues in small portions of motion sequences, reasoning temporally about when events occur, and querying specific motion attributes. In addition, we propose NSPose, a neuro-symbolic method for this task that uses symbolic reasoning and a modular design to ground motion through learning motion concepts, attribute neural operators, and temporal relations. We demonstrate the suitability of NSPose for the HumanMotionQA task, outperforming all baseline methods. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Skin Deep: Investigating Subjectivity in Skin Tone Annotations for Computer Vision Benchmark Datasets - **Authors:** Teanna Barrett, Quan Ze Chen, Amy X. Zhang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.09072 - **Pdf link:** https://arxiv.org/pdf/2305.09072 - **Abstract** To investigate the well-observed racial disparities in computer vision systems that analyze images of humans, researchers have turned to skin tone as more objective annotation than race metadata for fairness performance evaluations. However, the current state of skin tone annotation procedures is highly varied. For instance, researchers use a range of untested scales and skin tone categories, have unclear annotation procedures, and provide inadequate analyses of uncertainty. In addition, little attention is paid to the positionality of the humans involved in the annotation process--both designers and annotators alike--and the historical and sociological context of skin tone in the United States. Our work is the first to investigate the skin tone annotation process as a sociotechnical project. We surveyed recent skin tone annotation procedures and conducted annotation experiments to examine how subjective understandings of skin tone are embedded in skin tone annotation procedures. Our systematic literature review revealed the uninterrogated association between skin tone and race and the limited effort to analyze annotator uncertainty in current procedures for skin tone annotation in computer vision evaluation. Our experiments demonstrated that design decisions in the annotation procedure such as the order in which the skin tone scale is presented or additional context in the image (i.e., presence of a face) significantly affected the resulting inter-annotator agreement and individual uncertainty of skin tone annotations. We call for greater reflexivity in the design, analysis, and documentation of procedures for evaluation using skin tone. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### PIQI: Perceptual Image Quality Index based on Ensemble of Gaussian Process Regression - **Authors:** Nisar Ahmed, Hafiz Muhammad Shahzad Asif, Hassan Khalid - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2305.09214 - **Pdf link:** https://arxiv.org/pdf/2305.09214 - **Abstract** Digital images contain a lot of redundancies, therefore, compression techniques are applied to reduce the image size without loss of reasonable image quality. Same become more prominent in the case of videos which contains image sequences and higher compression ratios are achieved in low throughput networks. Assessment of quality of images in such scenarios has become of particular interest. Subjective evaluation in most of the scenarios is infeasible so objective evaluation is preferred. Among the three objective quality measures, full-reference and reduced-reference methods require an original image in some form to calculate the image quality which is unfeasible in scenarios such as broadcasting, acquisition or enhancement. Therefore, a no-reference Perceptual Image Quality Index (PIQI) is proposed in this paper to assess the quality of digital images which calculates luminance and gradient statistics along with mean subtracted contrast normalized products in multiple scales and color spaces. These extracted features are provided to a stacked ensemble of Gaussian Process Regression (GPR) to perform the perceptual quality evaluation. The performance of the PIQI is checked on six benchmark databases and compared with twelve state-of-the-art methods and competitive results are achieved. The comparison is made based on RMSE, Pearson and Spearman correlation coefficients between ground truth and predicted quality scores. The scores of 0.0552, 0.9802 and 0.9776 are achieved respectively for these metrics on CSIQ database. Two cross-dataset evaluation experiments are performed to check the generalization of PIQI. ## Keyword: RAW ### Make-An-Animation: Large-Scale Text-conditional 3D Human Motion Generation - **Authors:** Samaneh Azadi, Akbar Shah, Thomas Hayes, Devi Parikh, Sonal Gupta - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2305.09662 - **Pdf link:** https://arxiv.org/pdf/2305.09662 - **Abstract** Text-guided human motion generation has drawn significant interest because of its impactful applications spanning animation and robotics. Recently, application of diffusion models for motion generation has enabled improvements in the quality of generated motions. However, existing approaches are limited by their reliance on relatively small-scale motion capture data, leading to poor performance on more diverse, in-the-wild prompts. In this paper, we introduce Make-An-Animation, a text-conditioned human motion generation model which learns more diverse poses and prompts from large-scale image-text datasets, enabling significant improvement in performance over prior works. Make-An-Animation is trained in two stages. First, we train on a curated large-scale dataset of (text, static pseudo-pose) pairs extracted from image-text datasets. Second, we fine-tune on motion capture data, adding additional layers to model the temporal dimension. Unlike prior diffusion models for motion generation, Make-An-Animation uses a U-Net architecture similar to recent text-to-video generation models. Human evaluation of motion realism and alignment with input text shows that our model reaches state-of-the-art performance on text-to-motion generation. ## Keyword: raw image There is no result
2.0
New submissions for Wed, 17 May 23 - ## Keyword: events ### Common Diffusion Noise Schedules and Sample Steps are Flawed - **Authors:** Shanchuan Lin, Bingchen Liu, Jiashi Li, Xiao Yang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.08891 - **Pdf link:** https://arxiv.org/pdf/2305.08891 - **Abstract** We discover that common diffusion noise schedules do not enforce the last timestep to have zero signal-to-noise ratio (SNR), and some implementations of diffusion samplers do not start from the last timestep. Such designs are flawed and do not reflect the fact that the model is given pure Gaussian noise at inference, creating a discrepancy between training and inference. We show that the flawed design causes real problems in existing implementations. In Stable Diffusion, it severely limits the model to only generate images with medium brightness and prevents it from generating very bright and dark samples. We propose a few simple fixes: (1) rescale the noise schedule to enforce zero terminal SNR; (2) train the model with v prediction; (3) change the sampler to always start from the last timestep; (4) rescale classifier-free guidance to prevent over-exposure. These simple changes ensure the diffusion process is congruent between training and inference and allow the model to generate samples more faithful to the original data distribution. ### Motion Question Answering via Modular Motion Programs - **Authors:** Mark Endo, Joy Hsu, Jiaman Li, Jiajun Wu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2305.08953 - **Pdf link:** https://arxiv.org/pdf/2305.08953 - **Abstract** In order to build artificial intelligence systems that can perceive and reason with human behavior in the real world, we must first design models that conduct complex spatio-temporal reasoning over motion sequences. Moving towards this goal, we propose the HumanMotionQA task to evaluate complex, multi-step reasoning abilities of models on long-form human motion sequences. We generate a dataset of question-answer pairs that require detecting motor cues in small portions of motion sequences, reasoning temporally about when events occur, and querying specific motion attributes. In addition, we propose NSPose, a neuro-symbolic method for this task that uses symbolic reasoning and a modular design to ground motion through learning motion concepts, attribute neural operators, and temporal relations. We demonstrate the suitability of NSPose for the HumanMotionQA task, outperforming all baseline methods. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Skin Deep: Investigating Subjectivity in Skin Tone Annotations for Computer Vision Benchmark Datasets - **Authors:** Teanna Barrett, Quan Ze Chen, Amy X. Zhang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.09072 - **Pdf link:** https://arxiv.org/pdf/2305.09072 - **Abstract** To investigate the well-observed racial disparities in computer vision systems that analyze images of humans, researchers have turned to skin tone as more objective annotation than race metadata for fairness performance evaluations. However, the current state of skin tone annotation procedures is highly varied. For instance, researchers use a range of untested scales and skin tone categories, have unclear annotation procedures, and provide inadequate analyses of uncertainty. In addition, little attention is paid to the positionality of the humans involved in the annotation process--both designers and annotators alike--and the historical and sociological context of skin tone in the United States. Our work is the first to investigate the skin tone annotation process as a sociotechnical project. We surveyed recent skin tone annotation procedures and conducted annotation experiments to examine how subjective understandings of skin tone are embedded in skin tone annotation procedures. Our systematic literature review revealed the uninterrogated association between skin tone and race and the limited effort to analyze annotator uncertainty in current procedures for skin tone annotation in computer vision evaluation. Our experiments demonstrated that design decisions in the annotation procedure such as the order in which the skin tone scale is presented or additional context in the image (i.e., presence of a face) significantly affected the resulting inter-annotator agreement and individual uncertainty of skin tone annotations. We call for greater reflexivity in the design, analysis, and documentation of procedures for evaluation using skin tone. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### PIQI: Perceptual Image Quality Index based on Ensemble of Gaussian Process Regression - **Authors:** Nisar Ahmed, Hafiz Muhammad Shahzad Asif, Hassan Khalid - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2305.09214 - **Pdf link:** https://arxiv.org/pdf/2305.09214 - **Abstract** Digital images contain a lot of redundancies, therefore, compression techniques are applied to reduce the image size without loss of reasonable image quality. Same become more prominent in the case of videos which contains image sequences and higher compression ratios are achieved in low throughput networks. Assessment of quality of images in such scenarios has become of particular interest. Subjective evaluation in most of the scenarios is infeasible so objective evaluation is preferred. Among the three objective quality measures, full-reference and reduced-reference methods require an original image in some form to calculate the image quality which is unfeasible in scenarios such as broadcasting, acquisition or enhancement. Therefore, a no-reference Perceptual Image Quality Index (PIQI) is proposed in this paper to assess the quality of digital images which calculates luminance and gradient statistics along with mean subtracted contrast normalized products in multiple scales and color spaces. These extracted features are provided to a stacked ensemble of Gaussian Process Regression (GPR) to perform the perceptual quality evaluation. The performance of the PIQI is checked on six benchmark databases and compared with twelve state-of-the-art methods and competitive results are achieved. The comparison is made based on RMSE, Pearson and Spearman correlation coefficients between ground truth and predicted quality scores. The scores of 0.0552, 0.9802 and 0.9776 are achieved respectively for these metrics on CSIQ database. Two cross-dataset evaluation experiments are performed to check the generalization of PIQI. ## Keyword: RAW ### Make-An-Animation: Large-Scale Text-conditional 3D Human Motion Generation - **Authors:** Samaneh Azadi, Akbar Shah, Thomas Hayes, Devi Parikh, Sonal Gupta - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2305.09662 - **Pdf link:** https://arxiv.org/pdf/2305.09662 - **Abstract** Text-guided human motion generation has drawn significant interest because of its impactful applications spanning animation and robotics. Recently, application of diffusion models for motion generation has enabled improvements in the quality of generated motions. However, existing approaches are limited by their reliance on relatively small-scale motion capture data, leading to poor performance on more diverse, in-the-wild prompts. In this paper, we introduce Make-An-Animation, a text-conditioned human motion generation model which learns more diverse poses and prompts from large-scale image-text datasets, enabling significant improvement in performance over prior works. Make-An-Animation is trained in two stages. First, we train on a curated large-scale dataset of (text, static pseudo-pose) pairs extracted from image-text datasets. Second, we fine-tune on motion capture data, adding additional layers to model the temporal dimension. Unlike prior diffusion models for motion generation, Make-An-Animation uses a U-Net architecture similar to recent text-to-video generation models. Human evaluation of motion realism and alignment with input text shows that our model reaches state-of-the-art performance on text-to-motion generation. ## Keyword: raw image There is no result
process
new submissions for wed may keyword events common diffusion noise schedules and sample steps are flawed authors shanchuan lin bingchen liu jiashi li xiao yang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract we discover that common diffusion noise schedules do not enforce the last timestep to have zero signal to noise ratio snr and some implementations of diffusion samplers do not start from the last timestep such designs are flawed and do not reflect the fact that the model is given pure gaussian noise at inference creating a discrepancy between training and inference we show that the flawed design causes real problems in existing implementations in stable diffusion it severely limits the model to only generate images with medium brightness and prevents it from generating very bright and dark samples we propose a few simple fixes rescale the noise schedule to enforce zero terminal snr train the model with v prediction change the sampler to always start from the last timestep rescale classifier free guidance to prevent over exposure these simple changes ensure the diffusion process is congruent between training and inference and allow the model to generate samples more faithful to the original data distribution motion question answering via modular motion programs authors mark endo joy hsu jiaman li jiajun wu subjects computer vision and pattern recognition cs cv artificial intelligence cs ai machine learning cs lg arxiv link pdf link abstract in order to build artificial intelligence systems that can perceive and reason with human behavior in the real world we must first design models that conduct complex spatio temporal reasoning over motion sequences moving towards this goal we propose the humanmotionqa task to evaluate complex multi step reasoning abilities of models on long form human motion sequences we generate a dataset of question answer pairs that require detecting motor cues in small portions of motion sequences reasoning temporally about when events occur and querying specific motion attributes in addition we propose nspose a neuro symbolic method for this task that uses symbolic reasoning and a modular design to ground motion through learning motion concepts attribute neural operators and temporal relations we demonstrate the suitability of nspose for the humanmotionqa task outperforming all baseline methods keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp skin deep investigating subjectivity in skin tone annotations for computer vision benchmark datasets authors teanna barrett quan ze chen amy x zhang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract to investigate the well observed racial disparities in computer vision systems that analyze images of humans researchers have turned to skin tone as more objective annotation than race metadata for fairness performance evaluations however the current state of skin tone annotation procedures is highly varied for instance researchers use a range of untested scales and skin tone categories have unclear annotation procedures and provide inadequate analyses of uncertainty in addition little attention is paid to the positionality of the humans involved in the annotation process both designers and annotators alike and the historical and sociological context of skin tone in the united states our work is the first to investigate the skin tone annotation process as a sociotechnical project we surveyed recent skin tone annotation procedures and conducted annotation experiments to examine how subjective understandings of skin tone are embedded in skin tone annotation procedures our systematic literature review revealed the uninterrogated association between skin tone and race and the limited effort to analyze annotator uncertainty in current procedures for skin tone annotation in computer vision evaluation our experiments demonstrated that design decisions in the annotation procedure such as the order in which the skin tone scale is presented or additional context in the image i e presence of a face significantly affected the resulting inter annotator agreement and individual uncertainty of skin tone annotations we call for greater reflexivity in the design analysis and documentation of procedures for evaluation using skin tone keyword image signal processing there is no result keyword image signal process there is no result keyword compression piqi perceptual image quality index based on ensemble of gaussian process regression authors nisar ahmed hafiz muhammad shahzad asif hassan khalid subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract digital images contain a lot of redundancies therefore compression techniques are applied to reduce the image size without loss of reasonable image quality same become more prominent in the case of videos which contains image sequences and higher compression ratios are achieved in low throughput networks assessment of quality of images in such scenarios has become of particular interest subjective evaluation in most of the scenarios is infeasible so objective evaluation is preferred among the three objective quality measures full reference and reduced reference methods require an original image in some form to calculate the image quality which is unfeasible in scenarios such as broadcasting acquisition or enhancement therefore a no reference perceptual image quality index piqi is proposed in this paper to assess the quality of digital images which calculates luminance and gradient statistics along with mean subtracted contrast normalized products in multiple scales and color spaces these extracted features are provided to a stacked ensemble of gaussian process regression gpr to perform the perceptual quality evaluation the performance of the piqi is checked on six benchmark databases and compared with twelve state of the art methods and competitive results are achieved the comparison is made based on rmse pearson and spearman correlation coefficients between ground truth and predicted quality scores the scores of and are achieved respectively for these metrics on csiq database two cross dataset evaluation experiments are performed to check the generalization of piqi keyword raw make an animation large scale text conditional human motion generation authors samaneh azadi akbar shah thomas hayes devi parikh sonal gupta subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract text guided human motion generation has drawn significant interest because of its impactful applications spanning animation and robotics recently application of diffusion models for motion generation has enabled improvements in the quality of generated motions however existing approaches are limited by their reliance on relatively small scale motion capture data leading to poor performance on more diverse in the wild prompts in this paper we introduce make an animation a text conditioned human motion generation model which learns more diverse poses and prompts from large scale image text datasets enabling significant improvement in performance over prior works make an animation is trained in two stages first we train on a curated large scale dataset of text static pseudo pose pairs extracted from image text datasets second we fine tune on motion capture data adding additional layers to model the temporal dimension unlike prior diffusion models for motion generation make an animation uses a u net architecture similar to recent text to video generation models human evaluation of motion realism and alignment with input text shows that our model reaches state of the art performance on text to motion generation keyword raw image there is no result
1